LLM safety
Explainable LLM Unlearning: Making AI Forget With Reasoning
New research introduces explainable approaches to LLM unlearning, enabling models to selectively forget information while providing transparent reasoning for the process.
LLM safety
New research introduces explainable approaches to LLM unlearning, enabling models to selectively forget information while providing transparent reasoning for the process.
Machine Unlearning
New research explores machine unlearning for LLM agents, addressing how autonomous AI systems can selectively forget data while maintaining tool-use and reasoning capabilities.
Machine Unlearning
New research introduces a principled approach to removing harmful concepts from generative AI models using tempering and classifier guidance, with major implications for synthetic media safety.