Chris McCoy

1.9K posts

Chris McCoy banner
Chris McCoy

Chris McCoy

@TheRealMcCoy

Inventing @thestorecloud (☁️ AWS for Democracy). Thinking at Data4America (🇺🇸 Policy). Formerly @chrisamccoy. Just trying to make a dent. Optimist.

เข้าร่วม Temmuz 2009
1.1K กำลังติดตาม348 ผู้ติดตาม
ทวีตที่ปักหมุด
Chris McCoy
Chris McCoy@TheRealMcCoy·
I'm back.
English
5
2
42
4.4K
Chris McCoy
Chris McCoy@TheRealMcCoy·
@sriramk @kevinakwok We need China's AI infrastructure to become architecturally dependent on US second and third generation chips by 2028/2029
English
0
0
0
31
Sriram Krishnan
Sriram Krishnan@sriramk·
@kevinakwok Used the same phrase today multiple times.
Sriram Krishnan@sriramk

Every person here's reaction to the Jensen + @dwarkesh_sp podcast can be extrapolated *directly* from whether they believe in the frontier labs achieving short timelines for AGI/ASI. If you believe in the labs achieving RSI and then AGI/ASI (for some definition of all three) in the next few years, you'll probably sympathetic to the frame @dwarkesh_sp adopts. If not, you're probably more sympathetic to the arguments from Jensen.

English
3
0
16
7.4K
Kevin Kwok
Kevin Kwok@kevinakwok·
Jensen Dwarkesh podcast was a true scissor statement Haven't watched it yet but so funny how everyone I know agrees one of them was so good and the other so bad. Just can't agree on who
English
9
2
61
8.8K
Chris McCoy
Chris McCoy@TheRealMcCoy·
@beffjezos We need China's AI infrastructure to become architecturally dependent on US second and third generation chips by 2028/2029
English
0
0
0
100
Chris McCoy
Chris McCoy@TheRealMcCoy·
@GavinSBaker We need China's AI infrastructure to become architecturally dependent on US second and third generation chips by 2028/2029
English
0
0
0
59
Gavin Baker
Gavin Baker@GavinSBaker·
More thoughts on the Dwarkesh/Jensen discussion around export controls. Strongly believe that selling specific GPUs to China is in our national security interest and is a good policy for America. I think it is super important for us a country to get this right.
Gavin Baker@GavinSBaker

Much of Dwarkesh's argument hinges on this statment which *was* accurate but will be increasingly inaccurate on a go forward basis imo:    “American labs port across accelerators constantly. Anthropic's models are run on GPUs, they're run on Trainium, they're run on TPUs. There are so many things you can do, from distilling to a model that's well fit for your chips.”   As system level architectures diverge (torus vs. switched scale-up topologies, memory hierarchies, networking primitives), true portability is eroding. The Mi300 and Mi325 had roughly the same scale-up domain size as Hopper while Blackwell’s scale-up domain is 9x larger than the Mi355 scale-up domain, etc. Many frontier models are now being explicitly co-designed for inference on specific hardware like GB300 racks. Codex on Cerebras is another example. Those models run less efficiently on other systems and the performance differentials will only widen. A model that runs well on Google’s torus topology will run less efficiently on Nvidia’s switched scale-up topology and vice versa - the data traffic is fundamentally different as a byproduct of the models being parallelized across the different topologies. Google’s internal teams - and increasingly the Anthropic teams as they become the most important customer of almost every cloud - have the luxury of operating across the stack (models, chips, networking) - but that is not the case for the rest of the market and other prospective users. Anthropic is the exception, not the rule. To wit, Anthropic and Google allegedly have a mutual understanding where Anthropic can hire the TPU engineers they need every year to ensure that they can continue to get the most out of the TPU. Given the overwhelming importance of cost per token to the economics of the labs, models will be run where they run best. Most extremely large MoE models will run best on GB300s given the importance of having a switched scale-up network like NVLink for MoE inference. When training was the dominant cost for labs and power was broadly available, labs were optimizing to minimize capex dollars. Model portability was a way to create leverage over suppliers. I think that drove a lot of the focus on portability. Today, inference costs as measured by tokens per watt per dollar are everything. Inference is way more important than training costs (inference is effectively now part of training via RL). Labs are therefore now optimizing for inference. This means increasing co-design and higher go-forward switching costs for individual models between systems. I do think this explains why Anthropic and Nvidia came together: Anthropic needed Blackwells and Rubins to inference at least *some* of their models economically. And Mythos might just end up being released coincident with the availability of Rubins for inference. TLDR: as labs shift their focus from training to inference, the costs of portability and the upside of co-design to maximize tokens per watt per dollar both rise. Portability is likely to begin decreasing as a result.   I think what I might have respectfully added to Jensen’s answer is that systems evolve under local selective pressures. The evolutionary pressure in America is a shortage of watts so it makes sense for Nvidia to optimize, as an American company, for power efficiency and tokens per watt and stay on copper as long as possible. China has a surfeit of watts. Chinese AI systems are already taking advantage of this with the Huawei Cloudmatrix 384 and Atlas SuperPoD having an optical scale-up domain that is much larger than anything offered by Nvidia today at the cost of *much* higher power consumption and much lower tokens per watt. The networking primitives for this Huawei system are very different than those for Nvidia’s systems and a model that runs well on Nvidia will not run well on that system and vice versa. This means that if a Chinese ecosystem gets momentum, Chinese models might stop running well on American hardware. And when Chinese models run best on American hardware, America is in a better position as this gives America a degree of leverage and control over Chinese AI that it risks losing to an all-Chinese alternative ecosystem.   This architectural fork makes porting and distillation less effective and strengthens the pro-American national security case for selling China deprecated GPUs imo. Also I will attest that I did not wake up a loser this morning.

English
25
30
498
79.6K
Chris McCoy
Chris McCoy@TheRealMcCoy·
@AlecStapp @Noahpinion Sell them Intel's chips. Get em hooked by 2028-2029. Force a trade on their mineral supply in exchange for H100s. We all win.
English
0
0
0
98
Alec Stapp
Alec Stapp@AlecStapp·
Letting NVIDIA sell H200 chips to China is even worse that it looks on first glance. Given inelastic supply conditions, the critical inputs for producing H200s (such as high-bandwidth memory) could have been used to produce even more powerful chips for US customers. So our labs and hyperscalers lose out on even more compute than China gains.
Alec Stapp tweet media
Dwarkesh Patel@dwarkesh_sp

The Jensen Huang episode. 0:00:00 – Is Nvidia’s biggest moat its grip on scarce supply chains? 0:16:25 – Will TPUs break Nvidia’s hold on AI compute? 0:41:06 – Why doesn’t Nvidia become a hyperscaler? 0:57:36 – Should we be selling AI chips to China? 1:35:06 – Why doesn’t Nvidia make multiple different chip architectures? Look up Dwarkesh Podcast on YouTube, Apple Podcasts, Spotify, etc. Enjoy!

English
26
78
493
43.1K
Chris McCoy รีทวีตแล้ว
Samuel Hammond 🦉
Samuel Hammond 🦉@hamandcheese·
I'm on week 4 of retatrutide and already down 12.4 lbs. More interestingly: - I seem to have much more energy and focus - I find myself spontaneously preferring standing desks, walking rather than ubering, opting into exercise etc. - My blood sugar no longer crashes after eating - I don't drink nearly as much but when I do my hangovers seem a lot weaker - My food preferences spontaneously shifted in favor of fish, salads and fresh fruit - My GI health improved a lot, I think mostly thanks to it being easier to resist trigger foods I've had no negative symptoms or downsides, and am still on a low starter dose (~2.5mg)
English
29
8
435
39.9K
Chris McCoy รีทวีตแล้ว
Semafor
Semafor@semafor·
Token demand makes an AI bubble unlikely, says Michael Dell, CEO of Dell semafor.com/article/04/15/…
English
3
7
20
2.8K
Chris McCoy รีทวีตแล้ว
RYAN SΞAN ADAMS - rsa.eth 🦄
AI KYC is here. New claude subscribers asked for gov ID & photo. Not even a regulatory requirement - Anthropic just doing it because they want to. But regulatory is coming Next up will be laws: No AI without gov-issued ID All AI use tracked to individual - no private AI
RYAN SΞAN ADAMS - rsa.eth 🦄 tweet mediaRYAN SΞAN ADAMS - rsa.eth 🦄 tweet media
English
214
194
1.1K
149.3K
Chris McCoy รีทวีตแล้ว
Andrew Curran
Andrew Curran@AndrewCurran_·
The DNC has barred staffers from using Chat and Claude. The only approved DNC model is Gemini.
Andrew Curran tweet media
English
24
21
281
23.1K
Chris McCoy
Chris McCoy@TheRealMcCoy·
@sundeep Yes. And you can price it into 100mm micro-units.
English
0
0
0
23
Chris McCoy รีทวีตแล้ว
Neeraj K. Agrawal
Neeraj K. Agrawal@NeerajKA·
It’s got to feel so bad to drop a satoshi exposé only to have no one believe or care about it
English
17
3
145
9.4K
Chris McCoy
Chris McCoy@TheRealMcCoy·
@Noahpinion Indeed. It can be governed by the same math of the US Constitution where innovation triumphs under a +2/3 human in the loop.
English
0
0
1
140
Chris McCoy
Chris McCoy@TheRealMcCoy·
@AlexFast8 Added in both leagues last week alongside Jose, no idea how he does it
English
0
0
0
1.7K
Alex Fast
Alex Fast@AlexFast8·
This player with multi-position eligibility is rostered in 3.7% of ESPN leagues and 3% of Yahoo leagues.
Alex Fast tweet media
English
42
3
266
262.8K
Chris McCoy รีทวีตแล้ว
Watcher.Guru
Watcher.Guru@WatcherGuru·
JUST IN: 105,000 blocks remaining until the next Bitcoin Halving, officially reaching halfway point.
Watcher.Guru tweet media
English
296
676
4.8K
322.8K
Chris McCoy รีทวีตแล้ว
Surajit
Surajit@surajit_ghosh2·
First recovery footage of the Artemis II crew has just been released
English
948
10.7K
141.8K
6M