Peter Wang 🦋

39.5K posts

Peter Wang 🦋 banner
Peter Wang 🦋

Peter Wang 🦋

@pwang

Chief AI & Co-founder @AnacondaInc; invented @pyscript_dev, @PyData @Bokeh @Datashader. Former physicist. A student of the human condition. bsky: @wang.social

Austin, tx Katılım Ağustos 2007
2.4K Takip Edilen47.8K Takipçiler
Sabitlenmiş Tweet
Peter Wang 🦋
Peter Wang 🦋@pwang·
THE MOMENT YOU'VE BEEN WAITING FOR!! Type "=PY(" into Excel, and start executing Python directly in the @msexcel grid! Really excited about our new partnership with @Microsoft to democratize data science, machine learning, and AI to all knowledge workers!
Anaconda@anacondainc

We’re excited to unveil Python in Excel! Get ready for a whole new way to execute advanced analytics capabilities from within Excel 🐍 + 📊 = 💚 Check out the new integration btwn @anacondainc & @msexcel, @Microsoft365 here 👇 bit.ly/3KSblQ6

English
6
40
190
48.5K
Peter Wang 🦋
Peter Wang 🦋@pwang·
@amontalenti 😂😂 yeah it’s a mess! Some things I have won’t charge when connected to too powerful of a charge port? Some things (eg Chinese knock off handheld game consoles) will actually fry their charging circuits rather than negotiate PD correctly. It’s a mess
English
0
0
1
14
Andrew Montalenti
Andrew Montalenti@amontalenti·
Is there some ridiculously geeky post out there that goes into excruciating detail about 1080p vs 4K displays, subpixel rendering, refresh rate, HiDPI, font scaling, with internal technical details spanning connectors (HDMI, Displayport) and OSes (Windows, Mac OS, Linux), etc.?
English
3
1
20
0
Peter Wang 🦋
Peter Wang 🦋@pwang·
Which is why those who are serious about *actual* open source AI will invest in a clean, well-provenanced data Commons, and the ecosystem for maintaining and intentionally growing it…
Aakash Gupta@aakashgupta

Karpathy told Dwarkesh that a 1 billion parameter model, trained on clean data, could hit the intelligence of today's 1.8 trillion parameter frontier. That is a 1,800x compression claim. The math behind it is more defensible than it sounds. When researchers at frontier labs look at random samples from their training corpus, they see stock ticker symbols, broken HTML, forum spam, autogenerated gibberish. Not Wikipedia. Not the Wall Street Journal. The actual pretraining dataset is mostly noise, and the model is burning parameters to vaguely remember all of it. One estimate pegs Llama 3's information compression at 0.07 bits per token. Well-structured English carries around 1.5 bits per token of real information. The trillion-parameter model is holding a roughly 5% resolution image of the internet it trained on. So when a lab ships a 1.8 trillion parameter model, the overwhelming majority of those weights are handling rough memorization. They are compression overhead for a noisy training set, taking up capacity that could be doing reasoning instead. Karpathy's proposal is to separate the two. Build a cognitive core: a small model that contains only the algorithms for reasoning and problem-solving, stripped of encyclopedic memorization. Pair it with external memory the model queries when it needs a fact. A 1 billion parameter reasoner plus retrieval beats a 1.8 trillion parameter model trying to do both. The data already supports this direction. GPT-4o runs at roughly 200 billion parameters and outperforms the original 1.8 trillion GPT-4. Inference costs for GPT-3.5 level performance fell 280x between 2022 and 2024, driven almost entirely by smaller, cleaner, better-architected models. The trend line is pointing where Karpathy says it should. The real implication for anyone tracking the AI trade: data quality is the actual constraint. The companies winning the next phase will be the ones who figured out what to train on, and what to throw away.

English
0
0
7
1.2K
Peter Wang 🦋 retweetledi
Aakash Gupta
Aakash Gupta@aakashgupta·
Microsoft sold every spare CPU it had to Anthropic and OpenAI. Amazon tripled its CPU buys year over year and still can't keep up. Two of AWS's biggest customers asked Andy Jassy if they could buy the entire 2026 production run of Graviton chips. He said no. The ratio inside an AI datacenter used to be 100 megawatts of GPUs to 1 megawatt of CPUs. CPUs handled storage, checkpointing, pre-processing. Light work. GPUs did the actual training and inference. Then OpenAI shipped o1-preview in September 2024. RL post-training went from "check the model output with a regex" to "run classifiers" to "compile the code and run the unit tests" to "spin up a sandbox, call three databases, run a physics simulation, verify the answer." Every rollout now needs a CPU-backed environment to verify against. Codex 5.4 runs agentically for 6-7 hours at a time. Each database call, each cron job, each scraped URL is CPU work. Coding agent revenue went from a couple billion to north of $10B in six months. That compute is sitting on CPUs. The CPU to GPU ratio is now approaching 1:1. The entire global cloud was built for 1:8. That's why GitHub has been unstable for weeks. Nvidia and Arm both announced they're entering the server CPU market in March. TSMC will only meet 80% of server CPU wafer demand this year. High-end server CPU prices are already up 50%. When the GPU king and the IP licensor both pivot to CPUs in the same month, the boring chip isn't boring anymore.
English
39
192
1.4K
289.4K
Peter Wang 🦋 retweetledi
Prakash
Prakash@8teAPi·
I finally get Jensen. 1) he sees the model companies as replaceable, if one dies, another one springs up, and especially as Chinese talent flows from US AI labs to China and back. 2) He sees Nvidia as currently irreplaceable BUT given market access Huawei will catch up 3) Export controls on Nvidia sets it up to lose eventually as Huawei uses its dominance in the second largest market to fund R&D 4) He sees the only place China has a commanding advantage is energy, which is why he’s pushing so hard for US energy expansion. Looked at this way, providing China access to Nvidia chips lets them instantly catch up on models, but slows down their development on chips. At that point it becomes a race to expand energy production as quickly as possible, while the US can maintain its upper hand in chips. Conversely, the Dwarkesh view is that America has already lost or will lose on energy production. Dwarkesh also believes that the models are essentially commodities (which is why export controls on chips are the pressure point). Dwarkesh also believes that allowing China the ability to build and run current generation models on Nvidia hardware may allow them to increase their speed of chip development to beat Nvidia. Jensen believes Nvidia will also use the same tools or better ones to improve itself so he can stay ahead. Essentially US AI dominance, in Jensen’s view is built on Nvidia’s chip dominance, not the AI model company innovations alone. Export controls give away that dominance.
English
110
126
1.8K
205.9K
Peter Wang 🦋
Peter Wang 🦋@pwang·
🙌🏼🙌🏼🙌🏼🙌🏼
Prince Canuma@Prince_Canuma

mlx-vlm v0.4.3 is here 🚀 Day-0 support: 🔥 Gemma 4 (vision, audio, MoE) by @GoogleDeepMind 🦅 Falcon-OCR + Falcon Perception by @TIIuae 🪨 Granite Vision 4.0 by @IBMResearch New models: 🎯 SAM 3.1 with Object Multiplex by @facebook 🔍 RF-DETR detection & segmentation by @roboflow Infra: ⚡ TurboQuant (KV cache compression) 🖥️ CUDA support for vision models (Sam and RF-DETR) Get started today: > uv pip install -U mlx-vlm Leave us a star ⭐️ github.com/Blaizzy/mlx-vlm

ART
0
0
1
1.1K
Peter Wang 🦋
Peter Wang 🦋@pwang·
From the President of the United States, on Easter Sunday… “I will bomb you into Hell… Praise be to Allah.”
Peter Wang 🦋 tweet media
English
2
0
7
1.3K
Peter Wang 🦋 retweetledi
roon
roon@tszzl·
“fake work” and “bullshit jobs” has been fantastically wrong and misleading for understanding the modern world. a much better understanding is of a global economy where minor skill differences and improvements lead to monumentally different outcomes, and the marginal hour of work has never been more measurable or useful after the advent of even moderately effective talent allocation systems and the variability of reward based on effort and skill, people have engaged much harder in a red queen rat race across the world. this is why the Chinese ‘cram schools’ exist and why ‘yuppie striverism’ is a thing and why people trade off later family formation for working more so often. while overall work hours are slightly down, they are actually up for high earners (nber.org/digest/jul06/w…) I see it in the marginal effect with my friends now after the advent of claude and codex: they are actually working harder now than they ever have before. this is due to a personal Jevon’s paradox where they see that the value of their time has increased dramatically, that they can get a lot more visible work done towards goals they care about than they used to after requests from their customers the labs are doing things like inventing dispatch which lets you monitor work and manipulate your computer from your phone, on top of prior changes like having always on communications (slack). You hear about people launching codex jobs from their phone the moment they have an idea and reviewing them later no clue how long this lasts but the most immediate impact of co-existing with the machine state is higher productivity and higher visibility which leads to more work hours
English
130
150
2.5K
308.2K
Danielle Fong 🔆
Danielle Fong 🔆@DanielleFong·
i think the best signal for slop is the "nobody is paying attention" AI has really bad theory of mind and certainty for recent stuff, other agents and people, and doesn't already remember know how to check, or does cursory checks. But that's pretty bad for research and for trading, where your edge is what is not already priced it, pulled in.
English
2
0
23
1.5K
Peter Wang 🦋 retweetledi
Anaconda
Anaconda@anacondainc·
Setting up GPU environments is a major source of friction. CUDA alone spans 900+ components. Conda simplifies setup—handling driver detection, dependency resolution, and environment isolation: bit.ly/4lSffJY *Comparison concept inspired by an #NVIDIAGTC presentation.
Anaconda tweet media
English
0
1
5
764
rohit
rohit@krishnanrohit·
@NathanpmYoung It is indeed a pithier name, I just didn't like the way it looked when written down actually
English
2
0
5
477
Peter Wang 🦋 retweetledi
Pratyush Kumar
Pratyush Kumar@pratykumar·
📢 Open-sourcing the Sarvam 30B and 105B models! Trained from scratch with all data, model research and inference optimisation done in-house, these models punch above their weight in most global benchmarks plus excel in Indian languages. Get the weights at Hugging Face and AIKosh. Thanks to the good folks at SGLang for day 0 support, vLLM support coming soon. Links, benchmark scores, examples, and more in our blog - sarvam.ai/blogs/sarvam-3…
English
207
1.3K
6.8K
742K
Peter Wang 🦋 retweetledi
Brendan McCord 🏛️ x 🤖
Brendan McCord 🏛️ x 🤖@Brendan_McCord·
This was our 1st time gathering the @cosmos_inst community in one place. @tylercowen had asked a few months prior on zoom: when's your first event? I stumbled to reply, we'd been so busy we overlooked it. He said you don't really know your people until you get them in a room.
Brendan McCord 🏛️ x 🤖 tweet mediaBrendan McCord 🏛️ x 🤖 tweet mediaBrendan McCord 🏛️ x 🤖 tweet mediaBrendan McCord 🏛️ x 🤖 tweet media
English
2
3
42
11.9K
Peter Wang 🦋 retweetledi
Brendan McCord 🏛️ x 🤖
Brendan McCord 🏛️ x 🤖@Brendan_McCord·
I've had a dream that when my kids were old enough, they'd travel with me and sit in the front row when I gave a talk for Cosmos. As I walked out the door for our 1st Symposium this weekend, I said to @yo_adriane: I'm giving my talk here in Austin, what if they came today? ⬇️
Brendan McCord 🏛️ x 🤖 tweet media
English
3
7
87
4.6K
Peter Wang 🦋 retweetledi
Aida Baradari
Aida Baradari@aidaxbaradari·
Today, we're introducing Spectre I, the first smart device to stop unwanted audio recordings. We live in a world of always-on listening devices. Smart devices and AI dominate our world in business and private conversations. With Deveillance, you will @be_inaudible.
English
1.1K
5K
42.2K
4.5M
Eric Weinstein
Eric Weinstein@ericweinstein·
Why are we pretending that AI is being invented by the CEOs and VCs??? When did this particular idiocy take over? I too want to hear everything: @naval @balajis @peterthiel @sama @DavidSacks @pmarca @elonmusk @finkd @DarioAmodei I learn from all of them. But they are not driving this revolution. At age 60 I am watching those still on the command line thankful for homebrew. I’m trying to get an @ollama installation working and my GitHub functional. I’m still spending hours a day in LaTeX on @overleaf and reading the @arxiv. And, as much as I love my friends with their own jets, I think this is happening more on the command line than in the board room. I would be happy to debate any of the names above on the need to stop burying our research community under our business community. But I think many of them would actually agree with me. I don’t know how we got here. I don’t think Sam Altman ever woke up and said “I want to blot out the researchers.” I think we the public did this to ourselves. The actual font line AI researchers need to have a voice comparable to the billionaires. Perhaps louder. Or even, dare I say, much louder. And I am proud to use my voice to elevate theirs. As I try to do across the STEM fields.
English
19
7
159
14.3K
Eric Weinstein
Eric Weinstein@ericweinstein·
I love Peter both as a mind and as a friend. But this is dangerously wrong. I am both a word person and a math person. And I am watching our business people become our designated public thinkers. By default.
Eric Weinstein tweet media
English
184
58
650
115.6K
Danielle Fong 🔆
Danielle Fong 🔆@DanielleFong·
me too
Dean W. Ball@deanwball

This is precisely the reason that I have been so supportive of open-source over the years. It is an insurance policy against both tyrannical governments and the death throes of the industrial-era nation state, the latter of which I fear we are living through. Absent the oh-so-careful navigation of the narrow path (which I tried), it seemed inevitable that closed-source models would be captured by USG and thus subject to all of USG’s many mood swings, and thus intrinsically unreliable for both businesses and other countries. Alas, I did what I could to help us walk the narrow path, as did countless others. But the odds were always stacked against us. The hybrid open/closed model world was a good future. It would mean that AI is an intrinsically profitable business, favoring Western and especially American capitalistic institutions. Yet the power of the closed labs would be checked by open-weight models that, while behind the cutting-edge, provided balance. This supply chain risk designation will render closed-source models less attractive to many customers worldwide. The regulatory risk is now extremely high. USG cannot unfire the gun that it has fired—or if it can, doing so would take an extreme measure of skill, determination, and sensitivity that I do not think will be forthcoming. Thus the balanced open/closed world seems less likely on the margin, and instead it is likelier that we live in the open-source-dominant world now. This world probably favors China, because they have a political economy more favorable to large-scale subsidy of what is essentially digital public infrastructure. It also probably favors Chinese-style digital and physical surveillance (but made all the more pervasive and capable with advanced AI), since the catastrophic misuse risks of open source are higher. Chinese-style institutions have an overall structural advantage in this future, it seems to me, and Western institutions have a structural disadvantage. You can argue this has always been true, but USG just increased the likelihood of AI futures where the US is at an inherent disadvantage. And this is to say nothing of the damage to the business environment, the AI industry, etc., about which I have spoken earlier. The future is likelier to be more decentralized, more confusing, and more dangerous than it was a few weeks ago. Yet it also may be brighter; it is probably a higher variance future than the steadier transition of the narrow path. Perhaps you like that trade off. Nonetheless, we have probably been knocked off the narrow path, and the odds of a “normal” transition to the era of machine intelligence are now meaningfully lower. Humans are living through moving history once more.

English
1
0
11
1.5K