LLM research
Study Reveals How Users Build Trust When LLMs Hallucinate
New research examines how users develop calibrated trust strategies when interacting with hallucination-prone LLMs, offering frameworks for safer human-AI collaboration.
LLM research
New research examines how users develop calibrated trust strategies when interacting with hallucination-prone LLMs, offering frameworks for safer human-AI collaboration.
DeepMind
Google DeepMind deepens collaboration with UK AI Security Institute on frontier AI safety evaluation, establishing frameworks that could shape how synthetic media and generative models are assessed globally.
AI Regulation
A coalition of state attorneys general has issued formal warnings to Microsoft, OpenAI, Google, and other major AI companies demanding fixes to AI systems that generate false or misleading information.
AI video generation
AI video startup Tavus reports users spending hours conversing with its synthetic Santa avatar, revealing unexpected engagement patterns in real-time AI video generation technology.
Intel
Intel is reportedly nearing a deal to acquire SambaNova Systems, an AI chip startup that could strengthen Intel's position in the competitive AI accelerator market against Nvidia.
LLM Deployment
A technical deep-dive into deploying quantized large language models using AWQ compression, vLLM inference engine, and FastAPI for production-ready AI applications.
LLM Compression
Learn how to reduce a 7 billion parameter language model from ~14GB to 4.5GB using quantization, pruning, and knowledge distillation while maintaining accuracy.
LLM Agents
New research introduces a co-adaptive dual-strategy framework combining fast intuitive reasoning with slow deliberative thinking to improve LLM-based agent performance.
AI Evaluation
New research explores using generative AI agents as reliable proxies for human evaluation of AI-generated content, potentially transforming how we assess synthetic media quality at scale.
LLM Agents
New research introduces SABER, a safeguarding framework that identifies how small errors in LLM agent actions can cascade into significant failures, proposing intervention mechanisms.
AI Agents
New arXiv research explores whether AI agents can autonomously build, operate, and utilize complete data infrastructure, examining the boundaries of agentic AI capabilities.
AI Agents
Learn to build AI agents that learn, store, and reuse skills as modular neural components. This technical guide covers procedural memory architecture for persistent skill acquisition.