LLM optimization
LLM Agent Automates Hardware-Aware Model Quantization
New research introduces an LLM-based agent that automatically selects optimal quantization strategies for deploying large language models across diverse hardware platforms.
LLM optimization
New research introduces an LLM-based agent that automatically selects optimal quantization strategies for deploying large language models across diverse hardware platforms.
LLM fine-tuning
New research introduces Ratio-Variance Regularized Policy Optimization (RVPO), a method that stabilizes reinforcement learning from human feedback by controlling importance sampling variance in LLM training.
LLM Training
New research introduces SIGMA, a scalable spectral method using eigenvalue analysis to detect model collapse during LLM training before performance degrades catastrophically.
AI Agents
Beyond prompt engineering, context engineering is emerging as the critical discipline for building reliable AI agents—managing what information models see, when, and how.
deepfakes
Fabricated AI-generated images depicting Venezuelan President Maduro's arrest highlight growing concerns about synthetic media's role in political disinformation and public trust.
Agentic AI
As AI agents tackle complex multi-step tasks, traditional memory systems are hitting fundamental scaling limits. New architectural approaches are emerging to handle persistent context across extended workflows.
deepfake detection
Gen Digital partners with Intel to deploy hardware-accelerated deepfake detection directly on consumer devices, enabling real-time protection against AI-generated scam calls and video fraud.
LLM Agents
New research presents SimpleMem, an efficient memory architecture enabling LLM agents to maintain persistent context across extended interactions without traditional retrieval overhead.
AI Agents
New research introduces Orchestral AI, a framework for coordinating multiple AI agents in complex workflows, addressing key challenges in task distribution and agent communication.
Explainable AI
Researchers introduce prompt-counterfactual explanations, a new method for understanding generative AI behavior by identifying minimal prompt changes that alter outputs.
LLM Infrastructure
New research proposes joint encoding of KV-cache blocks to improve memory efficiency in large language model inference, addressing a key bottleneck in scalable AI deployment.
neural architecture
New research explores whether large language models can creatively design novel neural network architectures rather than simply recombining existing patterns from training data.