Deep Cogito

15 posts

Deep Cogito

Deep Cogito

@DeepCogito

San Francisco Katılım Nisan 2025
2 Takip Edilen3.1K Takipçiler
Deep Cogito retweetledi
Sriram Krishnan
Sriram Krishnan@sriramk·
Great to see more US open weight models.
Drishan Arora@drishanarora

Today, we are releasing the best open-weight LLM by a US company: Cogito v2.1 671B. On most industry benchmarks and our internal evals, the model performs competitively with frontier closed and open models, while being ahead of any US open model (such as the best versions of OpenAI’s GPT-OSS, Nvidia’s Nemotron and Meta’s Llama). We also built an interface where you can try the model (it’s free and we don’t store any chats): chat.deepcogito.com Additionally, you can download the model on @huggingface, or try it out on @openrouter, @togethercompute, @FireworksAI_HQ , @ollama cloud, @runpod, @baseten, or run it locally using @ollama or @UnslothAI. This model uses significantly fewer tokens amongst any similar capability models, because it has better reasoning capabilities. You will also notice improvements across instruction following, coding, longer queries, multi-turn and creativity. 📌 Model Weights: huggingface.co/collections/de… 📌Openrouter: openrouter.ai/deepcogito/cog… 📌 HF Blog: huggingface.co/blog/deepcogit… Some notes on our approach + design choices below 👇

English
4
12
128
46.8K
Deep Cogito retweetledi
Gokul Rajaram
Gokul Rajaram@gokulr·
Most tech startups in the US and elsewhere use Chinese open source models. Today, tech companies have a best-in-class US alternative. Congrats to @drishanarora Arora and DeepCogito team for releasing the best open-weight LLM by a US company: Cogito v2.1 671B. The model performs better than every US open model (GPT-OSS, Llama, Nemotron) on industry benchmarks. Importantly, it uses significantly fewer tokens because it has better reasoning capabilities, AND delivers improvements across instruction following, coding, longer queries, multi-turn and creativity. Try it out in one of the following ways: 1. check it out at the free interface below 2. download it on Hugging Face 3. try it out on OpenRouter, Inc, Together AI, Fireworks AI, Ollama cloud, Runpod, Baseten 4. Run it locally using Ollama or Unsloth AI. This is the shot in the arm that US open source needs! LFG.
Drishan Arora@drishanarora

Today, we are releasing the best open-weight LLM by a US company: Cogito v2.1 671B. On most industry benchmarks and our internal evals, the model performs competitively with frontier closed and open models, while being ahead of any US open model (such as the best versions of OpenAI’s GPT-OSS, Nvidia’s Nemotron and Meta’s Llama). We also built an interface where you can try the model (it’s free and we don’t store any chats): chat.deepcogito.com Additionally, you can download the model on @huggingface, or try it out on @openrouter, @togethercompute, @FireworksAI_HQ , @ollama cloud, @runpod, @baseten, or run it locally using @ollama or @UnslothAI. This model uses significantly fewer tokens amongst any similar capability models, because it has better reasoning capabilities. You will also notice improvements across instruction following, coding, longer queries, multi-turn and creativity. 📌 Model Weights: huggingface.co/collections/de… 📌Openrouter: openrouter.ai/deepcogito/cog… 📌 HF Blog: huggingface.co/blog/deepcogit… Some notes on our approach + design choices below 👇

English
10
5
25
12K
Deep Cogito retweetledi
Anastasios Nikolas Angelopoulos
Anastasios Nikolas Angelopoulos@ml_angelopoulos·
@drishanarora Congrats team! Great results. x.com/arena/status/1…
Arena.ai@arena

🚨Leaderboard Update New model provider in the Arena: @DeepCogito has released Cogito v2.1 (MIT licensed) 🔹Top 10 Open Source Model for WebDev, rank #10 🔹Tie ranks #18 overall for WebDev This puts Cogito v2.1 on par with community favorites like Qwen 3 Coder Plus & Kimi K2 Instruct. Congrats to the @DeepCogito team for this achievement! 👏

English
1
1
14
9.4K
Deep Cogito retweetledi
Vasudev Gupta
Vasudev Gupta@thevasudevgupta·
Check it out: chat.deepcogito.com !!!
Drishan Arora@drishanarora

Today, we are releasing the best open-weight LLM by a US company: Cogito v2.1 671B. On most industry benchmarks and our internal evals, the model performs competitively with frontier closed and open models, while being ahead of any US open model (such as the best versions of OpenAI’s GPT-OSS, Nvidia’s Nemotron and Meta’s Llama). We also built an interface where you can try the model (it’s free and we don’t store any chats): chat.deepcogito.com Additionally, you can download the model on @huggingface, or try it out on @openrouter, @togethercompute, @FireworksAI_HQ , @ollama cloud, @runpod, @baseten, or run it locally using @ollama or @UnslothAI. This model uses significantly fewer tokens amongst any similar capability models, because it has better reasoning capabilities. You will also notice improvements across instruction following, coding, longer queries, multi-turn and creativity. 📌 Model Weights: huggingface.co/collections/de… 📌Openrouter: openrouter.ai/deepcogito/cog… 📌 HF Blog: huggingface.co/blog/deepcogit… Some notes on our approach + design choices below 👇

English
0
3
3
2.5K
Deep Cogito retweetledi
ollama
ollama@ollama·
New open-weight LLM by Deep Cogito! Locally (671B): ollama run cogito-2.1 Ollama's Cloud: ollama run cogito-2.1:671b-cloud
Drishan Arora@drishanarora

Today, we are releasing the best open-weight LLM by a US company: Cogito v2.1 671B. On most industry benchmarks and our internal evals, the model performs competitively with frontier closed and open models, while being ahead of any US open model (such as the best versions of OpenAI’s GPT-OSS, Nvidia’s Nemotron and Meta’s Llama). We also built an interface where you can try the model (it’s free and we don’t store any chats): chat.deepcogito.com Additionally, you can download the model on @huggingface, or try it out on @openrouter, @togethercompute, @FireworksAI_HQ , @ollama cloud, @runpod, @baseten, or run it locally using @ollama or @UnslothAI. This model uses significantly fewer tokens amongst any similar capability models, because it has better reasoning capabilities. You will also notice improvements across instruction following, coding, longer queries, multi-turn and creativity. 📌 Model Weights: huggingface.co/collections/de… 📌Openrouter: openrouter.ai/deepcogito/cog… 📌 HF Blog: huggingface.co/blog/deepcogit… Some notes on our approach + design choices below 👇

English
5
30
178
32.1K
Deep Cogito retweetledi
Runpod
Runpod@runpod·
The best open-weight LLM by a US company, Cogito v2.1 671B, was trained on Runpod as part of their frontier stack. 🇺🇸🦅🚀 Try it now with no setup: console.runpod.io/hub/playground…
Drishan Arora@drishanarora

Today, we are releasing the best open-weight LLM by a US company: Cogito v2.1 671B. On most industry benchmarks and our internal evals, the model performs competitively with frontier closed and open models, while being ahead of any US open model (such as the best versions of OpenAI’s GPT-OSS, Nvidia’s Nemotron and Meta’s Llama). We also built an interface where you can try the model (it’s free and we don’t store any chats): chat.deepcogito.com Additionally, you can download the model on @huggingface, or try it out on @openrouter, @togethercompute, @FireworksAI_HQ , @ollama cloud, @runpod, @baseten, or run it locally using @ollama or @UnslothAI. This model uses significantly fewer tokens amongst any similar capability models, because it has better reasoning capabilities. You will also notice improvements across instruction following, coding, longer queries, multi-turn and creativity. 📌 Model Weights: huggingface.co/collections/de… 📌Openrouter: openrouter.ai/deepcogito/cog… 📌 HF Blog: huggingface.co/blog/deepcogit… Some notes on our approach + design choices below 👇

English
0
4
12
2.9K
Deep Cogito retweetledi
Ali Elqursh
Ali Elqursh@alielqursh·
Extremely proud to be part of this launch! 🚀 Cogito v2.1 is officially here, and it’s the best open-weight LLM released by a US company to date. 🇺🇸 The performance on this is incredible—go try it out. 👇
Drishan Arora@drishanarora

Today, we are releasing the best open-weight LLM by a US company: Cogito v2.1 671B. On most industry benchmarks and our internal evals, the model performs competitively with frontier closed and open models, while being ahead of any US open model (such as the best versions of OpenAI’s GPT-OSS, Nvidia’s Nemotron and Meta’s Llama). We also built an interface where you can try the model (it’s free and we don’t store any chats): chat.deepcogito.com Additionally, you can download the model on @huggingface, or try it out on @openrouter, @togethercompute, @FireworksAI_HQ , @ollama cloud, @runpod, @baseten, or run it locally using @ollama or @UnslothAI. This model uses significantly fewer tokens amongst any similar capability models, because it has better reasoning capabilities. You will also notice improvements across instruction following, coding, longer queries, multi-turn and creativity. 📌 Model Weights: huggingface.co/collections/de… 📌Openrouter: openrouter.ai/deepcogito/cog… 📌 HF Blog: huggingface.co/blog/deepcogit… Some notes on our approach + design choices below 👇

English
1
1
2
597
Deep Cogito retweetledi
Arena.ai
Arena.ai@arena·
🚨Leaderboard Update New model provider in the Arena: @DeepCogito has released Cogito v2.1 (MIT licensed) 🔹Top 10 Open Source Model for WebDev, rank #10 🔹Tie ranks #18 overall for WebDev This puts Cogito v2.1 on par with community favorites like Qwen 3 Coder Plus & Kimi K2 Instruct. Congrats to the @DeepCogito team for this achievement! 👏
Arena.ai tweet media
Drishan Arora@drishanarora

Today, we are releasing the best open-weight LLM by a US company: Cogito v2.1 671B. On most industry benchmarks and our internal evals, the model performs competitively with frontier closed and open models, while being ahead of any US open model (such as the best versions of OpenAI’s GPT-OSS, Nvidia’s Nemotron and Meta’s Llama). We also built an interface where you can try the model (it’s free and we don’t store any chats): chat.deepcogito.com Additionally, you can download the model on @huggingface, or try it out on @openrouter, @togethercompute, @FireworksAI_HQ , @ollama cloud, @runpod, @baseten, or run it locally using @ollama or @UnslothAI. This model uses significantly fewer tokens amongst any similar capability models, because it has better reasoning capabilities. You will also notice improvements across instruction following, coding, longer queries, multi-turn and creativity. 📌 Model Weights: huggingface.co/collections/de… 📌Openrouter: openrouter.ai/deepcogito/cog… 📌 HF Blog: huggingface.co/blog/deepcogit… Some notes on our approach + design choices below 👇

English
5
13
115
24.7K
Deep Cogito retweetledi
Drishan Arora
Drishan Arora@drishanarora·
Today, we are releasing the best open-weight LLM by a US company: Cogito v2.1 671B. On most industry benchmarks and our internal evals, the model performs competitively with frontier closed and open models, while being ahead of any US open model (such as the best versions of OpenAI’s GPT-OSS, Nvidia’s Nemotron and Meta’s Llama). We also built an interface where you can try the model (it’s free and we don’t store any chats): chat.deepcogito.com Additionally, you can download the model on @huggingface, or try it out on @openrouter, @togethercompute, @FireworksAI_HQ , @ollama cloud, @runpod, @baseten, or run it locally using @ollama or @UnslothAI. This model uses significantly fewer tokens amongst any similar capability models, because it has better reasoning capabilities. You will also notice improvements across instruction following, coding, longer queries, multi-turn and creativity. 📌 Model Weights: huggingface.co/collections/de… 📌Openrouter: openrouter.ai/deepcogito/cog… 📌 HF Blog: huggingface.co/blog/deepcogit… Some notes on our approach + design choices below 👇
Drishan Arora tweet mediaDrishan Arora tweet mediaDrishan Arora tweet media
English
86
111
730
870.1K
Deep Cogito retweetledi
Gokul Rajaram
Gokul Rajaram@gokulr·
Love how @DeepCogito is building its models using self-play - turns out this makes models better both in reasoning and non-thinking modes. Kudos @drishanarora and @DeepCogito team!
Drishan Arora@drishanarora

It is intuitively easy to understand why self play *can* work for LLMs, if we are able to provide a value function at intermediate steps (although not as clearly guaranteed as in two-player zero-sum games). In chess / go / poker, we have a reward associated with every next move, but as Noam points out, natural language is messy. It is hard to define a value function at intermediate steps like tokens. As a result, in usual reinforcement learning (like RLVR), LLMs get a reward at the end. They end up learning to 'meander' more for hard problems. In a way, we reward brute forcing with more tokens to end up at the right answer as the right approach. However, at @DeepCogito, we provide a signal for the thinking process itself. Conceptually, you can imagine this as post-hoc assigning a reward to better search trajectories. This teaches the model to develop a stronger intuition for 'how to search' while reasoning. In practice, the model ends up with significantly shorter reasoning chains for harder problems in a reasoning mode. Somewhat surprisingly, it also ends up being better in a non-thinking mode. One way to think about it is that since the model knows how to search better, it 'picks' the most likely trajectory better in the non-thinking mode.

English
4
1
11
6.5K
Deep Cogito retweetledi
Drishan Arora
Drishan Arora@drishanarora·
It is intuitively easy to understand why self play *can* work for LLMs, if we are able to provide a value function at intermediate steps (although not as clearly guaranteed as in two-player zero-sum games). In chess / go / poker, we have a reward associated with every next move, but as Noam points out, natural language is messy. It is hard to define a value function at intermediate steps like tokens. As a result, in usual reinforcement learning (like RLVR), LLMs get a reward at the end. They end up learning to 'meander' more for hard problems. In a way, we reward brute forcing with more tokens to end up at the right answer as the right approach. However, at @DeepCogito, we provide a signal for the thinking process itself. Conceptually, you can imagine this as post-hoc assigning a reward to better search trajectories. This teaches the model to develop a stronger intuition for 'how to search' while reasoning. In practice, the model ends up with significantly shorter reasoning chains for harder problems in a reasoning mode. Somewhat surprisingly, it also ends up being better in a non-thinking mode. One way to think about it is that since the model knows how to search better, it 'picks' the most likely trajectory better in the non-thinking mode.
Noam Brown@polynoamial

Self play works so well in chess, go, and poker because those games are two-player zero-sum. That simplifies a lot of problems. The real world is messier, which is why we haven’t seen many successes from self play in LLMs yet. Btw @karpathy did great and I mostly agree with him!

English
13
11
37
16.3K
Deep Cogito retweetledi
Vipul Ved Prakash
Vipul Ved Prakash@vipulved·
Cogito 671B is an impressive model, a material improvement over DSV3… often better than Sonnet & 4o. Nicely done @drishanarora & @DeepCogito team!
Drishan Arora@drishanarora

A small update - we had more traffic than anticipated. However, the endpoints are now scalable on Together AI for all models, including the 671B MoE. Test out the model here: together.ai/models/cogito-… (A huge thanks to the folks at @togethercompute for making this happen so quickly.)

English
1
3
23
7K
Deep Cogito retweetledi
Dhruv Malrana
Dhruv Malrana@drvdhruv·
We are putting out four frontier intelligence models today, each hitting heavy in its weight class. The large ones are among the smartest open models in the world right now. It has become increasingly clear that superintelligence is within reach. As we continue to iterate and refine self improvement, the models will inevitably get smarter. Hence it is all the more important to be able to open source our work. Technology this powerful should not be built in isolation. It has been an honor to learn from the best while tinkering with and ultimately pushing the ability of machines to think over the past decade. We continue to do that at Deep Cogito.
Drishan Arora@drishanarora

Today, we are releasing 4 hybrid reasoning models of sizes 70B, 109B MoE, 405B, 671B MoE under open license. These are some of the strongest LLMs in the world, and serve as a proof of concept for a novel AI paradigm - iterative self-improvement (AI systems improving themselves). The largest 671B MoE model is amongst the strongest open models in the world. It matches/exceeds the performance of the latest DeepSeek v3 and DeepSeek R1 models both, and approaches closed frontier models like o3 and Claude 4 Opus.

English
3
4
44
8.1K
Deep Cogito retweetledi
Drishan Arora
Drishan Arora@drishanarora·
Today, we are releasing 4 hybrid reasoning models of sizes 70B, 109B MoE, 405B, 671B MoE under open license. These are some of the strongest LLMs in the world, and serve as a proof of concept for a novel AI paradigm - iterative self-improvement (AI systems improving themselves). The largest 671B MoE model is amongst the strongest open models in the world. It matches/exceeds the performance of the latest DeepSeek v3 and DeepSeek R1 models both, and approaches closed frontier models like o3 and Claude 4 Opus.
Drishan Arora tweet media
English
43
257
2K
451.6K