AI Hardware
Taalas Hardwired AI Chips Hit 17K Tokens Per Second
Startup Taalas is challenging GPU dominance with hardwired AI chips designed specifically for inference, claiming 17,000 tokens per second throughput for ubiquitous AI deployment.
AI Hardware
Startup Taalas is challenging GPU dominance with hardwired AI chips designed specifically for inference, claiming 17,000 tokens per second throughput for ubiquitous AI deployment.
Meta
Meta has struck a major deal with Nvidia to acquire millions of next-generation AI chips, dramatically expanding its compute infrastructure for AI model training and deployment.
AI Hardware
Startup Positron secures $230M Series B to develop alternative AI chips, potentially reshaping the hardware landscape that powers video generation and synthetic media systems.
AI Hardware
Researchers introduce DABench-LLM, a standardized framework for evaluating dataflow AI accelerators designed for large language model inference in the post-Moore era.
Microsoft
Microsoft announces its Maia 200 custom AI accelerator, entering direct competition with Amazon and Google in the race to build proprietary silicon for AI workloads.
Apple
Apple is developing an AI-powered wearable device, responding to OpenAI's hardware ambitions. The move signals intensifying competition in AI interfaces beyond smartphones.
AI Hardware
Micron declares AI-driven memory shortage 'unprecedented,' predicting supply constraints will persist beyond 2026 as demand for high-bandwidth memory outpaces production capacity.
AI Hardware
Groq's Language Processing Unit takes a radically different approach to AI inference, replacing GPU parallelism with deterministic compute for predictable, ultra-fast performance.
AMD
AMD researchers unveil AIE4ML, an end-to-end compiler framework that maps neural networks to next-gen AI Engines, achieving significant speedups over CPU implementations for ML workloads.
LLM Infrastructure
New research proposes CXL-SpecKV, a disaggregated FPGA architecture using CXL memory pooling and speculative prefetching to overcome memory bottlenecks in large language model inference at datacenter scale.
AI Hardware
New benchmarks of Apple's MLX framework demonstrate the growing capability of on-device AI processing, with major implications for local synthetic media generation.