AI security
Open Framework Detects Attack Patterns in Multi-Agent AI Systems
New research introduces an open framework for training security models that detect temporal attack patterns in multi-agent AI workflows through trace-based analysis.
AI security
New research introduces an open framework for training security models that detect temporal attack patterns in multi-agent AI workflows through trace-based analysis.
neural-networks
New neural architecture creates mathematically guaranteed decision regions using hyperspheres, enabling AI systems to know when they're uncertain rather than making unreliable predictions.
AI Detection
New research tackles the challenge of attributing AI-generated content to specific models while handling unknown generators—critical for deepfake detection and digital authenticity verification.
LLM Security
Researchers reveal how malicious actors can embed hidden backdoors in LLMs through vocabulary manipulation, enabling stealthy sabotage that evades detection methods.
LLM
New research introduces HaluNet, a framework using multi-granular uncertainty modeling to efficiently detect hallucinations in LLM question answering systems.
AI Safety
New research proposes integrating actions, compositional structure, and episodic memory from neuroscience to build safer, more interpretable AI systems that could transform how we approach AI trustworthiness.
AI Safety
Researchers introduce DarkPatterns-LLM, a multi-layer benchmark designed to identify and evaluate manipulative behaviors in large language models, advancing AI safety and authenticity research.
AI Detection
Research reveals significant limitations in human ability to detect AI-generated images, raising critical questions about synthetic media verification and the future of visual authenticity.
synthetic data
New research explores how reinforcement learning can optimize synthetic data generation, with implications for training more capable AI video and media generation models.
AI agents
New research proposes combining blockchain monitoring with agentic AI to create verifiable perception-reasoning-action pipelines, addressing critical trust and authenticity challenges in autonomous AI systems.
Generative AI
Researchers propose a Taylor-based approach that outperforms the classic Paterson-Stockmeyer method for computing matrix exponentials in flow-based generative AI models, offering efficiency gains for video and image synthesis.
AI Safety
New research bridges efficiency and safety by developing formal verification methods for neural networks with early exits, enabling mathematically proven safety guarantees for adaptive AI systems.