Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now
Deep Cogito, a lesser-known AI research startup based in San Francisco, founded by ex-Googlers, today released four new open-ish large language models (LLMs) that attempt something few others do: learn how to reason more effectively over time — and get better at it on their own.
The models, released as part of Cogito’s v2 family, range from 70 billion to 671 billion parameters and are available for AI developers and enterprises to use under a mix of limited and fully open licensing terms. They include:
Cogito v2-70B (Dense)
Cogito v2-109B (Mixture-of-Experts)
Cogito v2-405B (Dense)
Cogito v2-671B (Mixture-of-Experts)
The Cogito v2 series includes both dense and Mixture-of-Experts (MoE) models, each suited to different needs. Dense models, like the 70B and 405B variants, activate all parameters on every forward pass, making them more predictable and easier to deploy across a wide range of hardware.
They’re ideal for low-latency applications, fine-tuning, and environments with limited GPU capacity. MoE models, such as the 109B and 671B versions, use a sparse routing mechanism to activate only a few specialized “expert” subnetworks at a time, allowing for much larger total model sizes without proportional increases in compute cost.
The AI Impact Series Returns to San Francisco - August 5 The next phase of AI is here - are you ready? Join leaders from Block, GSK, and SAP for an exclusive look at how autonomous agents are reshaping enterprise workflows - from real-time decision-making to end-to-end automation. Secure your spot now - space is limited: https://bit.ly/3GuuPLF
... continue reading