Inception, the company behind the first commercial diffusion large language models (dLLMs), today announced the launch of Mercury 2, the fastest reasoning LLM and first reasoning dLLM. Mercury 2 ...
The new Mercury 2 AI model uses diffusion reasoning to generate 1,000 tokens per second; it runs about 5x faster than Haiku, speed limits are ...
Mercury 2 introduces diffusion LLMs to text, delivering 10x faster speeds for AI agents and production workflows without sacrificing reasoning power.
The development of large language models (LLMs) is entering a pivotal phase with the emergence of diffusion-based architectures. These models, spearheaded by Inception Labs through its new Mercury ...