The company said that the model was trained on 15 trillion mixed visual and text tokens.
Moonshot AI’s Kimi K2.5 Reddit AMA revealed why the powerful open-weight model is hard to run, plus new details on agent ...
Abu Dhabi-based Mohamed bin Zayed University of Artificial Intelligence’s (MBZUAI) Institute of Foundation Models has ...
30-person startup Arcee AI has released a 400B model called Trinity, which it says is one of the biggest open source foundation models from a US company.
B, an open-source AI coding model trained in four days on Nvidia B200 GPUs, publishing its full reinforcement-learning stack ...
Kimi has a standard mode and a Thinking mode that offers higher output quality. Additionally, a capability called K2.5 Agent ...
According to the Allen Institute for AI, coding agents suffer from a fundamental problem: Most are closed, expensive to train ...
Kimi K2.5 adds Agent Swarm with up to 100 parallel helpers and a 256k window, so teams solve complex work faster.
The American startup is pitching investors on a $1 billion+ valuation to train a model over a trillion parameters, aiming to reclaim the open-weight lead from Chinese labs like Moonshot and DeepSeek.
The non-profit Allen Institute for AI (AI2) has launched a family of open-source coding models targeting independent developers and SMEs.
HONG KONG, CHINA - JANUARY 28: In this photo illustration, the DeepSeek logo is seen next to the Chat GPT logo on a phone on January 28, 2025 in Hong Kong, China. (Photo illustration by Anthony ...