ElastixAI Inc. today emerged from stealth to tackle the systemic inefficiencies and high costs of generative AI (GenAI) inference. Founded by former Apple and Meta machine learning (ML) researchers, ...
Inception, the company behind the first commercial diffusion large language models (dLLMs), today announced the launch of Mercury 2, the fastest reasoning LLM and first reasoning dLLM. Mercury 2 ...
If mHC scales the way early benchmarks suggest, it could reshape how we think about model capacity, compute budgets and the ...
Researchers at the Tokyo-based startup Sakana AI have developed a new technique that enables language models to use memory more efficiently, helping enterprises cut the costs of building applications ...
With reported 3x speed gains and limited degradation in output quality, the method targets one of the biggest pain points in production AI systems: latency at scale.
A new technical paper titled “Combating the Memory Walls: Optimization Pathways for Long-Context Agentic LLM Inference” was published by researchers at University of Cambridge, Imperial College London ...
AWS Premier Tier Partner leverages its AI Services Competency and expertise to help founders cut LLM costs using ...
BELLEVUE, Wash.--(BUSINESS WIRE)--MangoBoost, a provider of cutting-edge system solutions designed to maximize AI data center efficiency, is announcing the launch of Mango LLMBoost™, system ...
Marketing, technology, and business leaders today are asking an important question: how do you optimize for large language models (LLMs) like ChatGPT, Gemini, and Claude? LLM optimization is taking ...
New deployment data from four inference providers shows where the savings actually come from — and what teams should evaluate ...