Darshan Tank

639 posts

Darshan Tank banner
Darshan Tank

Darshan Tank

@TankDarshan7

AI + Data guy @SentientAGI | prev-@GnaniAi, @deloitte

India Katılım Haziran 2019
368 Takip Edilen44 Takipçiler
Darshan Tank
Darshan Tank@TankDarshan7·
@hwchase17 Research direction: > Extracting useful signals from that data for eval/environment generation, harness engineering, and post-training
English
0
0
1
31
Darshan Tank
Darshan Tank@TankDarshan7·
@trq212 This is really good approach to understand anything new. HTML gives better visualisation!
English
0
0
0
4
Darshan Tank
Darshan Tank@TankDarshan7·
30$/M input tokens and $180/M output tokens!! On what basis? on the name of benchmark numbers! Many questions: - Need of further challenging benchmarks? - Dependence shift to OS models? - Normalisation of using model locally?
OpenAI@OpenAI

Introducing GPT-5.5 A new class of intelligence for real work and powering agents, built to understand complex goals, use tools, check its work, and carry more tasks through to completion. It marks a new way of getting computer work done. Now available in ChatGPT and Codex.

English
0
0
0
22
Darshan Tank
Darshan Tank@TankDarshan7·
@theo Price is a relative parameter! The day Anthropic makes it cheaper, it will also go down, but this will never happen and it will rise from here onwards on the name of model hype or benchmark numbers.
English
0
0
1
2.2K
Theo - t3.gg
Theo - t3.gg@theo·
$5 per mil in, $30 per mil out. GPT-5.5 is smart. I've been using it for a bit. It's also weird, hard to wrangle, and too expensive IMO. Double the price of GPT-5.4. 20% more expensive than Opus 4.7.
OpenAI@OpenAI

Introducing GPT-5.5 A new class of intelligence for real work and powering agents, built to understand complex goals, use tools, check its work, and carry more tasks through to completion. It marks a new way of getting computer work done. Now available in ChatGPT and Codex.

English
170
54
2.7K
399.2K
OpenAI
OpenAI@OpenAI·
Introducing GPT-5.5 A new class of intelligence for real work and powering agents, built to understand complex goals, use tools, check its work, and carry more tasks through to completion. It marks a new way of getting computer work done. Now available in ChatGPT and Codex.
English
2.5K
7K
51.8K
13M
Darshan Tank
Darshan Tank@TankDarshan7·
Coverage, Complexity and Quality!
English
0
0
0
14
SpaceX
SpaceX@SpaceX·
SpaceXAI and @cursor_ai are now working closely together to create the world’s best coding and knowledge work AI. The combination of Cursor’s leading product and distribution to expert software engineers with SpaceX’s million H100 equivalent Colossus training supercomputer will allow us to build the world’s most useful models. Cursor has also given SpaceX the right to acquire Cursor later this year for $60 billion or pay $10 billion for our work together.
English
2.4K
5K
38.4K
20.8M
Darshan Tank
Darshan Tank@TankDarshan7·
@bcherny @stolinski It would be interesting to know, if you guys analyse each session by clustering users- you can kill any growing startup/company 👀
English
0
0
0
714
Boris Cherny
Boris Cherny@bcherny·
@stolinski That's weird. Could you run /feedback and share the id here? Would love to debug
English
8
1
153
46.5K
Darshan Tank
Darshan Tank@TankDarshan7·
@alexalbert__ It’s strange! You are showcasing what is realistically possible with Mythos and selling something low!
English
0
0
0
92
Alex Albert
Alex Albert@alexalbert__·
Some of my favorite things in Opus 4.7: - Very good at async work and following instructions - Effort levels are far more predictable for token control (+ new xhigh level) - No more downscaling of high-res images - Noticeably more taste in UIs, slides, docs
Claude@claudeai

Introducing Claude Opus 4.7, our most capable Opus model yet. It handles long-running tasks with more rigor, follows instructions more precisely, and verifies its own outputs before reporting back. You can hand off your hardest work with less supervision.

English
107
32
885
88.7K
Darshan Tank
Darshan Tank@TankDarshan7·
@trq212 Context rot is a real problem it hallucinates- miss or hit problem
English
0
0
0
19
Darshan Tank
Darshan Tank@TankDarshan7·
@trq212 rewind and subagent are something that I've not used that much will try to adapt it!
English
0
0
0
20
Darshan Tank
Darshan Tank@TankDarshan7·
@aakashgupta If they think of building one time infrastructure and use OS models that would be game changer! 1. It will save cost! 2. Your data stay within your system only!
English
0
0
0
572
Aakash Gupta
Aakash Gupta@aakashgupta·
Uber gave 5,000 engineers access to Claude Code in December. By February, usage had nearly doubled. By April, the CTO told the company they'd burned through the entire annual AI budget. The adoption curve tells you everything about what happened. In December 2024, 32% of Uber's engineers were using Claude Code. By February 2026, that number was 63%. That's not a gradual rollout. That's a product so useful that engineers pulled it into their workflow faster than finance could model the spend. Uber has about 34,000 employees. Engineering is roughly 15% of that headcount, somewhere around 5,100 people. At enterprise API pricing, Claude Code runs $100 to $200 per developer per month on Sonnet alone. But that's the subscription math. The real number is token consumption, and Uber's engineers aren't building hello-world apps. They're building rider-driver matching algorithms, dynamic pricing engines, and real-time logistics across 70+ countries. Every one of those tasks eats context windows for breakfast. The scale of what these engineers are actually doing with AI is wild. 92% of Uber's developers use AI agents monthly. 65 to 72% of code written inside IDEs is now AI-generated. 11% of all pull requests are opened by agents, not humans. The company's AI code review system, uReview, analyzes over 90% of the 65,000 diffs Uber ships per week. AI-related costs at Uber are up 6x since 2024. CTO Praveen Neppalli Naga's quote was "I'm back to the drawing board." That's the CTO of a $144 billion company admitting that the tools work so well his team can't afford to keep using them at this rate. Here's the part nobody is pricing in. Anthropic's Claude Code hit $2.5 billion in annualized revenue by February 2026. That's up from $1 billion in November 2025. The fastest enterprise software ramp in history, and a huge portion of that growth is coming from exactly this pattern: companies deploy Claude Code, engineers love it, usage explodes, budgets evaporate. Uber won't be the last company to have this conversation. The average Claude Code developer burns about $6 per day. Multiply that across thousands of engineers running complex agentic workflows, spawning sub-agents that each maintain their own context windows, and the math compounds fast. One engineering team running Claude Code in automated CI/CD loops can drain a monthly budget in days. The CFO problem is now the bottleneck for AI adoption at the enterprise level. The technology works. The productivity gains are real. Uber's own data says 75% of AI code review comments are marked helpful by engineers. The constraint is that traditional annual budgeting was designed for tools with predictable per-seat costs, and AI coding agents have usage curves that look like cloud compute bills from 2015: exponential until someone notices. Every enterprise CTO is about to have the same meeting Praveen just had. The tools are too good to pull back. The costs are too unpredictable to ignore. And the companies that figure out token cost optimization first will have a structural advantage over every competitor still running annual budget cycles against exponential adoption curves.
Aakash Gupta tweet mediaAakash Gupta tweet media
English
170
241
1.9K
393.9K
Darshan Tank
Darshan Tank@TankDarshan7·
@aakashgupta We have best open source model like MiniMax M2.7, GLM 5.1 with fraction of cost to opus4.6 but still people are falling for claude code! It is not about LLM anymore it is about agent and it's harness
English
1
0
3
568
Darshan Tank
Darshan Tank@TankDarshan7·
> use claude code for heavy lifting > cursor as reviewer and to make minor changes observation: cursor is quicker as they are have done indexing
English
0
0
0
32