Metilience unveils a hybrid AI reasoning engine for high-stakes exams, leveraging structured cognitive error analysis ...
Anthropic has unveiled Claude 3.7 Sonnet, a notable addition to its lineup of large language models (LLMs), building on the foundation of Claude 3.5 Sonnet. Marketed as the first hybrid reasoning ...
The new Mercury 2 AI model uses diffusion reasoning to generate 1,000 tokens per second; it runs about 5x faster than Haiku, speed limits are ...
Reasoning large language models (LLMs) are designed to solve complex problems by breaking them down into a series of smaller ...
It's cheap to copy already built models from their outputs, but likely still expensive to train new models that push the boundaries. Reading time 4 minutes It is becoming increasingly clear that AI ...
Microsoft has released its Phi-4-mini-flash-reasoning small language model for on-device AI. With this, the Redmond giant promises a much more efficient Phi model strong in math and logic. Microsoft ...
Unlock the full InfoQ experience by logging in! Stay updated with your favorite authors and topics, engage with content, and download exclusive resources. Cory Benfield discusses the evolution of ...
Nvidia researchers developed dynamic memory sparsification (DMS), a technique that compresses the KV cache in large language models by up to 8x while maintaining reasoning accuracy — and it can be ...
Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Very small language models (SLMs) can ...
The global large language model market size was estimated at USD 7.77 billion in 2025 and is projected to reach around USD ...