LoRA
LoRA Fine-Tuning: Run Massive AI Models on Consumer Hardware
Learn how Low-Rank Adaptation lets you customize billion-parameter AI models on standard laptops—the same technique powering custom deepfakes and AI video generation.
LoRA
Learn how Low-Rank Adaptation lets you customize billion-parameter AI models on standard laptops—the same technique powering custom deepfakes and AI video generation.
LLM Reasoning
New research reveals that even frontier AI models like GPT-4 and Claude struggle with basic reasoning puzzles, exposing fundamental limitations in how large language models process logic.
AI benchmarks
New benchmark evaluates whether frontier AI models can perform PhD-level scientific research tasks, revealing significant gaps between current capabilities and expert human performance.
AI Safety
New arXiv research investigates how varying levels of information access affect LLM monitors' ability to detect sabotage, with implications for AI safety and oversight systems.
LLM research
New research introduces test-time policy evolution to scale LLM reasoning without additional training, enabling models to dynamically improve their problem-solving strategies during inference.
AI Systems
Researchers introduce SETA, a statistical method for identifying which component in complex AI pipelines causes failures—critical for debugging multi-stage systems like video generation workflows.
LLM research
New research introduces a method to preserve correct reasoning steps while penalizing errors, improving LLM performance through more nuanced reinforcement learning credit assignment.
LLM research
New research identifies specific neurons responsible for reasoning in LLMs and demonstrates how transferring their activation patterns can significantly improve inference reliability across models.
LLM research
New research introduces embedded reasoning to improve how LLMs handle function parameters, addressing a critical bottleneck in AI agent reliability for tool-using applications.
LLM Agents
New research explores how reinforcement learning training affects LLM agent generalization across domains, introducing the concept of 'generalization tax' and strategies to minimize performance degradation.
multimodal AI
Researchers introduce MMR-Bench, a comprehensive benchmark evaluating how well routing systems direct queries to optimal multimodal LLMs across diverse visual reasoning tasks.
LLM research
New research combines graph-based local reasoning with belief propagation to help LLMs tackle complex investigative tasks, enabling more reliable multi-step analysis in AI systems.