Eric Vishria

2.9K posts

Eric Vishria banner
Eric Vishria

Eric Vishria

@ericvishria

GP @Benchmark Director @confluentinc @Cerebras @Contentful @Benchling @CommerceLayer @acuitymd @FireworksAI_HQ @quilterai @pomerium_io @greptile @sundayrobotics

Katılım Haziran 2008
848 Takip Edilen19.7K Takipçiler
Sabitlenmiş Tweet
Eric Vishria
Eric Vishria@ericvishria·
.@benchmark co-lead the initial round for Cerebras 10 years ago. Over the following 5 years, the team amazed, delivering the technological marvel of a wafer-scale chip, the system to heat and cool it, and more recently the software layer that allows giant fleets of Cerebras systems to work together for very large MoE models. But even more impressive, is they just never fucking quit, despite kissing death like 3 times, getting made fun of for unusual early customers, and getting passed over by virtually every respected semi investor (who have all converted now). The team knew, IF they could stay alive, it was just a matter of time…. In tech, speed ultimately wins, and nothing is close to as fast as Cerebras.
Andrew Feldman@andrewdfeldman

@OpenAI and @Cerebras have signed a multi-year agreement to deploy 750 megawatts of Cerebras wafer-scale systems to serve OpenAI customers. This has been a decade in the making. Deployment begins in early 2026, and when fully rolled out, it will be the largest high-speed AI inference deployment in the world. OpenAI and Cerebras were both founded in 2015 with radically ambitious goals. OpenAI set out to build the software that would push AI toward general intelligence. Cerebras set out to rethink computing hardware from first principles. Our teams met as far back as 2017. We shared ideas, early work, and a common belief: there would come a point when model scale and hardware architecture would have to converge. That point has arrived. ChatGPT set the direction for the entire industry. It showed the world what AI could be. Now we’re in the next phase - not proving capability, but delivering it at global scale. The history of technology is clear on one thing: speed drives adoption. The PC industry didn’t operate at kilohertz. The internet didn’t change the world on dial-up. AI is no different. As models grow more capable, speed becomes the bottleneck. Slow systems limit what users can do, how often they engage, and whether AI becomes infrastructure or remains a novelty. Cerebras was built for this moment. By keeping computation and memory on a single wafer-scale processor, we eliminate the data-movement penalties that dominate GPU systems. The result is up to 15× faster inference, without sacrificing model size or accuracy. That speed changes product design, user behavior, and ultimately productivity. For consumers, it means AI that feels instantaneous. For the economy, it means agents that can finally drive serious productivity growth. For Cerebras, 2026 will be a defining year. With this collaboration with OpenAI, Cerebras’ wafer-scale technology will reach hundreds of millions - and eventually billions - of users. We’re proud to work alongside OpenAI to bring fast, frontier AI to people around the world. This is what a decade of long-term thinking looks like.

English
13
37
305
121.2K
Eric Vishria retweetledi
Dmytro Dzhulgakov
Dmytro Dzhulgakov@dzhulgakov·
Composer 2 beats Opus on TerminalBench at a fraction of the cost. The ingredients: coding focus only, data flywheel, cracked RL team, and infrastructure that can keep up. @FireworksAI_HQ powered the inference and RL scaling behind Composer 2. Scaling RL is still genuinely hard, and we're proud we could help make it less so. Congrats to @cursor_ai on shipping a great model!
Cursor@cursor_ai

Composer 2 is now available in Cursor.

English
2
8
48
11.7K
Eric Vishria retweetledi
Lin Qiao
Lin Qiao@lqiao·
🔥 Cursor Composer2 launched on Fireworks 🔥 This time it's not just inference but also RL powered by @FireworksAI_HQ. So much hard work and sleepless nights to get this gift out. Congrats @cursor_ai team on launching this SOTA model beating Opus 4.6 on terminal bench! 🚀 x.com/cursor_ai/stat…
English
5
15
103
18.1K
Eric Vishria retweetledi
Mike Freedman
Mike Freedman@michaelfreedman·
Introducing TigerFS - a filesystem backed by PostgreSQL, and a filesystem interface to PostgreSQL. Idea is simple: Agents don't need fancy APIs or SDKs, they love the file system. ls, cat, find, grep. Pipelined UNIX tools. So let’s make files transactional and concurrent by backing them with a real database. There are two ways to use it: File-first: Write markdown, organize into directories. Writes are atomic, everything is auto-versioned. Any tool that works with files -- Claude Code, Cursor, grep, emacs -- just works. Multi-agent task coordination is just mv'ing files between todo/doing/done directories. Data-first: Mount any Postgres database and explore it with Unix tools. For large databases, chain filters into paths that push down to SQL: .by/customer_id/123/.order/created_at/.last/10/.export/json. Bulk import/export, no SQL needed, and ships with Claude Code skills. Every file is a real PostgreSQL row. Multiple agents and humans read and write concurrently with full ACID guarantees. The filesystem /is/ the API. Mounts via FUSE on Linux and NFS on macOS, no extra dependencies. Point it at an existing Postgres database, or spin up a free one on Tiger Cloud or Ghost. I built this mostly for agent workflows, but curious what else people would use it for. It's early but the core is solid. Feedback welcome. tigerfs.io
English
77
99
1.1K
116.9K
davidlee
davidlee@davidlee·
Pride only hurts, it never helps. -Marcellus Wallace "We make them in charge. My teams have to work for them, which makes them really unhappy. And not many of them like it. But I'm like, look, these guys went out there, raised money, kicked your ass in your category, and you want them to work for you? That makes no sense to me. You're going to work for them. Learn from them."
TBPN@tbpn

"I don't think many tech companies execute M&A well." Palo Alto Networks CEO @nikesharora breaks down his strategy for successful M&A: "Purchase price is an irrelevant artifact. If it's going to work, it's going to work phenomenally well, or you're going to screw it up. It's not what you paid, it's what you're able to do with it." "You could say that Instagram was expensive, or YouTube was expensive, or DoubleClick was expensive. They all worked perfectly. AOL Time Warner is a different story. So it boils down to how you execute past the price you pay for it." "In tech, when you buy a company, you buy a team, you buy an existing product, and you buy a roadmap for the future. The question is: can you deliver on that roadmap? Can you accelerate that roadmap? Does it work?" "We sign a term sheet, and we ask the founders to sit with our team and redesign the product roadmap so we like it and they like it. And if they don't agree with our expectations and we don't agree with theirs, we don't buy the company." "We make them in charge. My teams have to work for them, which makes them really unhappy. And not many of them like it. But I'm like, look, these guys went out there, raised money, kicked your ass in your category, and you want them to work for you? That makes no sense to me. You're going to work for them. Learn from them." "So our job is to enable these people. We look at them and say, whatever your business plan was when you were a small private company, find me a business plan that's twice as assertive and bold as the one you had then." "We've built a phenomenal system to take them to market. I have 3,000 people in the field... 3,000 people go out there and see 10,000 customers. So that's where the secret sauce kicks in." "We've bought 34 companies so far. I think our hit rate on things that have worked is over 70%."

English
2
1
17
4.5K
Eric Vishria
Eric Vishria@ericvishria·
With the IBM’s $11B acquisition of Confluent officially closing, I want to say congrats to the entire team for navigating the tricky course from open source Kafka support to open core to single-tenant cloud to multi-tenant cloud and all the ensuing GTM changes along the way and 5 (chaotic) years as a public co. The company relentlessly built >$1B business (actual revenue not last wk x 52 :-)) under the incredible leadership of @jaykreps and fantastic execution from the team. On a personal note I've been lucky enough to serve on the board from Day 1 to Day End through our investment @benchmark. It was my first investment and I learned so much along the way. Thank you to Jay, @nehanarkhede and @junrao for choosing to work with us in 2014 and the adventure along the way!
Rob Thomas@robdthomas

Agents need real-time data…

English
6
10
152
39.7K
Eric Vishria retweetledi
Pat Grady
Pat Grady@gradypb·
You have $1 Trillion to build the compute backbone for AGI. How much goes to space? 100% So says @PhilipJohnston of @Starcloud_ Hear why…
English
11
12
120
22.5K
Eric Vishria
Eric Vishria@ericvishria·
By approaching the problem differently, this small team continues to push towards real deployments at an astounding rate. I've had the great fortune of working with @coatue_thomas on a number of companies, most notably an early bet he made on Cerebras. Huge get for all.
Tony Zhao@tonyzzhao

We raised $165M at a $1.15B valuation to stop doing demos. 2026 is about 1) deployment and 2) research. We will start shipping Memo with our new frontier models in a few months. Our series-B is led by Coatue, with Thomas Laffont joining the board. ->🧵

English
2
7
90
27.8K
Eric Vishria
Eric Vishria@ericvishria·
Let me anonymize it! Built a model for evaluating various 10b51 strategies against different price scenarios and optimizing for different returns. Not that complicated mathematically, but definitely would have taken several hours to go through the various iterations I went through nearly instantly.
English
2
0
5
3.8K
Eric Vishria
Eric Vishria@ericvishria·
The AI coding stuff is super cool but because I've never really written code as a professional it hasn't impacted me in an identity-striking way. But, I have built financial models for work. Claude in Excel has absolutely BLOWN MY MIND and I'm just one-shotting crazy models that would have taken me days. I'm making up things to model just because. WHAT!?!
English
18
2
244
50.1K
Aryaman Iyer
Aryaman Iyer@AryamanIyer3·
one-shot models are wild but iteration is where it gets real. i built 40+ iteration loops for financial models where opus writes formulas, libreoffice recalculates, and errors feed back. the first draft is fast but the 30th iteration is when the balance sheet actually balances. what models are you building
English
1
0
2
1.1K
Jason Tan
Jason Tan@jpctan·
@ericvishria That shift is powerful. Instead of spending most of the time building the model, you spend it thinking about the problem and experimenting with scenarios.
English
1
0
1
30
Eric Vishria
Eric Vishria@ericvishria·
@sarthakgh @btaylor Totally. In some ways, these models are easier than code - they are verifiable, but less degrees of freedom. Fwiw, I'm really impressed with the models it is building. Some layout things I would do differently, but I just tell it to move things around and it fixes them.
English
0
0
10
1.4K
Eric Vishria
Eric Vishria@ericvishria·
Cool experiment. Severely limited inference supply plus peak weekday demand curve. As coding tasks become longer lived, I could imagine AI estimating a task to use more tokens than the user has and suggesting it could run at night or on the weekends. Most appliances - washing machines, dishwashers, coffee machines - have delayed start - for cheaper power, contention for household water,....  but similar core idea.
Mike Krieger@mikeyk

We’ve doubled usage on weekends, and outside 5a–11am PT on weekdays, through March 27. Works across Claude.ai, Cowork, and Claude Code. Go run your next big idea and thanks for building with us!

English
3
3
53
17K
Eric Vishria
Eric Vishria@ericvishria·
@US_Stormwatch I believe the Apr 1 snowpack measurement was highly correlated with how much water we'd have. When you get early melt like this, does the water still go into the reservoirs and ground as you'd expect? Or is it worse? Or unknown? I realize reservoirs are fullish.
English
2
0
5
4.3K
Colin McCarthy
Colin McCarthy@US_Stormwatch·
No snow will be left by next weekend at many ski resorts in California below 8,000 feet as the worst March heatwave in state history arrives. The base elevation of every ski resort will be 70°F+ for at least 4 straight days next week. Even Mammoth is forecast to hit 75°F next week. Expecting most ski resorts to close in the next 2 weeks, when snow conditions should be at their peak. Instead we have August-like heat in March. When California's official April 1 snowpack measurement comes in, don't be surprised if it's one of the worst on record since 1950.
Rohin Dhar@rohindhar

Quick trip up to Tahoe for a day of skiing in the SCORCHING HEAT

English
56
177
1.5K
292.6K
tyler hogge
tyler hogge@thogge·
One thing is 100% clear to me: With very few exceptions, tech cos are some of the worst operated businesses in the history of capitalism exhibit A: Atlassian -25 yrs old -hasn’t been profitable for a decade -promoted an engineer the same day he was laid off This is the rule.
tyler hogge tweet media
English
37
37
1K
142.5K
Eric Vishria retweetledi
zack's lab
zack's lab@zackslab·
part 4: interview with @quilterai CEO Sergiy Nesterenko! intros for first 11 mins, then chat about RL, AI, and PCB design. it was a fun and insightful chat, thanks Sergiy!
English
5
4
49
3.8K
David Davidović
David Davidović@thegeomaster·
@ericvishria Interestingly, many (if not all) of the big GPU inference providers are running disaggregated prefill & decode already. The workloads are so different that even splitting PD across the same hardware type (GPUs) could yield benefits due to separate tuning
English
1
0
0
56
Eric Vishria
Eric Vishria@ericvishria·
Breaking up prefill (processing the prompt) and decode (generating the response) has been theorized for a while as they have different compute requirements. Now we have the silicon to do it - AWS Trainium for prefill, and Cerebras for decode. Super fast AND cost effective.
Amazon Web Services@awscloud

We're teaming up with @cerebras to build the fastest possible inference. Coming soon to Amazon Bedrock, we’re delivering inference performance an order of magnitude faster than what’s available today by connecting AWS Trainium3 for compute-intensive prefill with Cerebras CS-3 to power decode. Learn more about the partnership. go.aws/3Pzcota

English
6
5
62
12.8K