Woody Lu

4.4K posts

Woody Lu banner
Woody Lu

Woody Lu

@rollothomasi123

Investment, Software Engineering, GenAI, Phenomenology, Universe

San Francisco, CA Katılım Temmuz 2014
1.9K Takip Edilen261 Takipçiler
Woody Lu retweetledi
Viking
Viking@vikingmute·
想研究 Claude Code 工作原理的看这个 ccunpacked.dev Claude Code Unpacked 吧,今天 HackerNews 排名第一,纯纯交互式,看的非常清晰,里面四大分类有:agent loop, 50+ tools, multi-agent orchestration, and unreleased features 做的特别好。
中文
3
41
203
20.6K
Woody Lu retweetledi
Wasteland Capital
Wasteland Capital@ecommerceshares·
Every hedge fund just breathed a sigh of relief, this very conveniently appearing right at end of Q1.
English
22
14
548
54.8K
Woody Lu retweetledi
Andrew Curran
Andrew Curran@AndrewCurran_·
From the post:
Andrew Curran tweet media
Krishna Kaasyap@krishnakaasyap

From QT - //But if Anthropic found that training above a certain scale, or in a certain way at that scale, produces capabilities that sit far above the prior trendline, then that is an architectural breakthrough.// I believe this is the case, not just because an architectural and algorithmic breakthrough at this scale cannot be achieved in isolation, but also because, even if it were, it would soon leak via employee turnover, corporate espionage, or many other means. The moat of a frontier lab lies in enormously scaling an advancement, or simply in scaling a Transformer++ arch. I don't think any of the frontier lab would purely bet on an architectural or algorithmic breakthrough (that could be easily replicated like CoT reasoning/thinking was replicated by almost everyone) for them to be at the frontier! In addition to this business logic, research from @EpochAIResearch supports the same conclusion. From @ansonwhho's research - //For example, @MITFutureTech found that shifting from LSTMs (green) to Modern Transformers (purple) has an efficiency gain that depends on the compute scale: - At 1e15 FLOP, the gain is 6.3× - At 3e16 FLOP, the gain is 26× Naively extrapolating to 1e23 FLOP, the gain is 20,000×!// If Anthropic found that training above a certain scale... produces capabilities that sit far above the prior trendline... they would definitely attempt it, as it can be done by only two other labs in the world. This is especially relevant given that those two labs have their tentacles in everything from adult content slop to search engine & browser wars, thinning their available compute for a final training run of a single model. Source - epoch.ai/gradient-updat… Since past final training runs have typically accounted for <30% of total R&D compute, a significant amount of compute remains unused for these runs. It is possible that the compute allocated for final training runs has now been increased substantially. The largest final training run known to humanity occurred in late 2024 for GPT-4.5, which OpenAI officially released on February 27, 2025. Not a single GB200 NVL72 was available at that time. However, by early 2026, we have access to thousands of GB200 and GB300 NVL72 racks, along with more diversified compute from AMD, Google (TPUs), AWS (Tranium), and many other providers. All available evidence and reasonable inferences suggest that the observed step-change improvements are likely large primarily because Ant scaled final training run compute significantly, rather than due to a multitude of new innovations. Source - epoch.ai/gradient-updat… Total @EpochAIResearch victory - @datagenproc @cherylwoooo @Jsevillamol and the team!

English
10
34
430
63.2K
Woody Lu retweetledi
Krishna Kaasyap
Krishna Kaasyap@krishnakaasyap·
From QT - //But if Anthropic found that training above a certain scale, or in a certain way at that scale, produces capabilities that sit far above the prior trendline, then that is an architectural breakthrough.// I believe this is the case, not just because an architectural and algorithmic breakthrough at this scale cannot be achieved in isolation, but also because, even if it were, it would soon leak via employee turnover, corporate espionage, or many other means. The moat of a frontier lab lies in enormously scaling an advancement, or simply in scaling a Transformer++ arch. I don't think any of the frontier lab would purely bet on an architectural or algorithmic breakthrough (that could be easily replicated like CoT reasoning/thinking was replicated by almost everyone) for them to be at the frontier! In addition to this business logic, research from @EpochAIResearch supports the same conclusion. From @ansonwhho's research - //For example, @MITFutureTech found that shifting from LSTMs (green) to Modern Transformers (purple) has an efficiency gain that depends on the compute scale: - At 1e15 FLOP, the gain is 6.3× - At 3e16 FLOP, the gain is 26× Naively extrapolating to 1e23 FLOP, the gain is 20,000×!// If Anthropic found that training above a certain scale... produces capabilities that sit far above the prior trendline... they would definitely attempt it, as it can be done by only two other labs in the world. This is especially relevant given that those two labs have their tentacles in everything from adult content slop to search engine & browser wars, thinning their available compute for a final training run of a single model. Source - epoch.ai/gradient-updat… Since past final training runs have typically accounted for <30% of total R&D compute, a significant amount of compute remains unused for these runs. It is possible that the compute allocated for final training runs has now been increased substantially. The largest final training run known to humanity occurred in late 2024 for GPT-4.5, which OpenAI officially released on February 27, 2025. Not a single GB200 NVL72 was available at that time. However, by early 2026, we have access to thousands of GB200 and GB300 NVL72 racks, along with more diversified compute from AMD, Google (TPUs), AWS (Tranium), and many other providers. All available evidence and reasonable inferences suggest that the observed step-change improvements are likely large primarily because Ant scaled final training run compute significantly, rather than due to a multitude of new innovations. Source - epoch.ai/gradient-updat… Total @EpochAIResearch victory - @datagenproc @cherylwoooo @Jsevillamol and the team!
Krishna Kaasyap tweet media
Andrew Curran@AndrewCurran_

Three weeks ago there were rumors that one of the labs had completed its largest ever successful training run, and that the model that emerged from it performed far above both internal expectations and what people assumed the scaling laws would predict. At the time these were only rumors, and no lab was attached to them. But in light of what we now know about Mythos, they look more credible, and the lab was probably Anthropic. Around the same time there were also rumors that one of the frontier labs had made an architectural breakthrough. If you are in enough group chats, you hear claims like this constantly, and most turn out to be nothing. But if Anthropic found that training above a certain scale, or in a certain way at that scale, produces capabilities that sit far above the prior trendline, then that is an architectural breakthrough. I think the leaked blog post was real, but still a draft. Mythos and Capybara were both candidate names for the new tier, though Mythos may now have enough mindshare that they end up keeping it. The specific rumor in early March was that the run produced a model roughly twice as performant as expected. That remains unconfirmed. What is confirmed is that Anthropic told Fortune the new model is a 'step change,' a sudden 2x would certainly fit the definition. We will find out in April how much of this is true. My own view is that the broad shape of this is correct even if some of the numbers are wrong. And if it is substantially accurate, then it also casts OpenAI's recent restructuring in a new light. If very large training runs are about to become essential to staying in the game, then a lot of their recent decisions, like dropping Sora, make even more sense strategically. For the public, this would mean the best models in the world are about to become much more expensive to serve, and therefore much more expensive to use. That will put pressure on rate limits, pricing, and subscription plans that are already subsidized to some unknown degree. Instead of becoming too cheap to meter, frontier intelligence may be about to become too expensive for most of humanity to afford. Second-order effects; compute, memory, and energy are about to become much more important than they already are. In the blog they describe the new model as not just an improvement, but having 'dramatically higher scores' than Opus 4.6 in coding and reasoning, and as being 'far ahead' of any other current models. If this is the new reality, then scale is about to become king in a whole new way. It would also mean, as usual, that Jensen wins again.

English
5
13
204
73.4K
Woody Lu retweetledi
Tanay Jaipuria
Tanay Jaipuria@tanayj·
Meta's misexecution recently on LLMs / agents is well discussed, but something that's underrated is just how well their ranking and recommendation systems are improving as they put more compute against their Ads ranking model (GEM), ranking system (Lattice) and Ads retrieval engine (Andromeda)
Tanay Jaipuria tweet media
English
7
14
144
13.9K
Woody Lu retweetledi
Orange AI
Orange AI@oran_ge·
谁能想到,AI 也需要睡觉。 Claude Code 出了个新功能叫 AutoDream,字面意思,让 AI 做梦。 你跟 Claude Code 合作久了,它的记忆文件会越来越臃肿。同一件事记了五遍,三周前的架构笔记早就不对了,有用的东西被噪音埋住。记忆越多,反而越笨。 AutoDream 在你不用的时候自动运行,干三件事:删掉重复和过期的,把散落的信息合并成一条,用项目当前状态刷新旧记忆。 名字来自神经科学。人睡觉时大脑在做一模一样的事。重播白天的经历,强化重要连接,修剪弱的。你早上起来觉得脑子清醒,不是因为休息了,是因为大脑趁你睡着偷偷做了一轮垃圾回收。 AutoDream 就是给 AI 加了一个睡眠周期。两次工作之间,它自己醒来整理记忆,然后继续睡。下次你打开的时候,它的脑子比昨天更干净。
中文
37
77
790
126.1K
Woody Lu retweetledi
Mike Zaccardi, CFA, CMT 🍖
Mike Zaccardi, CFA, CMT 🍖@MikeZaccardi·
1-day $VIX highest since Tax Day 2025 33.5.. implying a 2.2% move by Monday's close
Mike Zaccardi, CFA, CMT 🍖 tweet media
English
1
5
29
3.8K
Woody Lu
Woody Lu@rollothomasi123·
@fxtrader Capybara.... 卡皮巴拉???
日本語
0
0
1
1.1K
外汇交易员
外汇交易员@fxtrader·
财富杂志:Anthropic正在开发并已对早期用户测试一款新AI模型,该模型被认为性能上有 “飞跃式提升”。此前一次数据泄露曝光了该模型存在,以及未发布博客草稿和计划中的欧洲CEO峰会细节。博客草稿提到新模型Claude Mythos和Capybara,似指同一底层模型,Capybara比之前的Opus模型更大、更智能但也更昂贵,在软件编码等测试中得分更高。 同时,泄露文件指出新模型会带来前所未有的网络安全风险,因它在网络能力上远超其他模型,或被黑客用于大规模网络攻击,公司计划先向网络防御组织发布。 另外,数据泄露似乎源于内容管理系统用户的人为错误,除模型相关信息,还意外公开了如员工产假等内部文档及欧洲CEO峰会的私密信息,该峰会是Anthropic向大企业客户推广AI模型计划的一部分。
外汇交易员 tweet media
中文
11
7
107
44.9K
Woody Lu retweetledi
David 🏹
David 🏹@d_gilz·
They want to shove out the SpaceX IPO into this market lmfaoooo
English
28
16
733
29.7K
Woody Lu retweetledi
NIK
NIK@ns123abc·
🚨 BREAKING: SpaceX filing IPO this week, aiming to raise $75 billion+ >biggest US IPO of all time >surpasses all money raised by US IPOs last year combined >filing may show xai losing money >investors don’t care >it’s elon Individual investors getting 20%+ of shares (vs typical 10%). No standard 6-month lockup. Elon wants retail investors in. we are so back
NIK tweet mediaNIK tweet mediaNIK tweet media
English
157
498
5.1K
378.3K
Woody Lu
Woody Lu@rollothomasi123·
@fxtrader 现在这种宏观环境下上市是真的Desperate, 也是怕之后Macro越来越糟糕吧
中文
0
0
2
584
外汇交易员
外汇交易员@fxtrader·
The Information:SpaceX计划在本周内提交IPO申请,可能筹集超过750亿美元资金,有望成为史上最大IPO。 SpaceX最新估值为1.25万亿美元,但要到IPO前几周才会确定实际的发行规模和估值。
外汇交易员 tweet media
外汇交易员@fxtrader

马斯克据悉将SpaceX IPO时机定在6月中旬,因6月会有3年多来首次金木合相(金星与木星在天空中运行至极其接近的位置)以及马斯克的生日(6月28日)。 SpaceX此次IPO计划融资最高约500亿美元,对应估值约1.5万亿美元,远超2019年沙特阿美290亿美元的募资规模纪录。

中文
16
7
92
46.7K
Woody Lu retweetledi
Flrande
Flrande@flrande·
打通了 Codex 和期权流的分析链路后成为了无情的下单机器😶
中文
8
9
132
29.7K
Woody Lu retweetledi
Lonely
Lonely@Lonely__MH·
🔥卧槽!刚知道Claude Code还有这个隐藏模式 /model opusplan 规划用最强Opus,coding交给Sonnet。 性价比直接拉满,省下的token够我多喝几杯咖啡 感谢大佬分享!Orz
Lonely tweet media
Daniel San@dani_avila7

Did you know about the opusplan model in Claude Code? /model opusplan It's a hybrid alias that automatically uses Opus in plan mode for complex reasoning, then switches to Sonnet for execution. Best of both worlds: Opus thinks, Sonnet builds

中文
22
84
548
107.3K
Woody Lu retweetledi
Jason Luongo
Jason Luongo@JasonL_Capital·
BREAKING: Claude now has live access to real-time stock quotes and options chain data You can pull prices, scan option chains, check Greeks, and view your portfolio without leaving the chat Here's how to connect the (free) API step by step:
Jason Luongo tweet media
English
131
459
4.8K
1.1M
Woody Lu retweetledi
Nicolas Bustamante
Nicolas Bustamante@nicbstme·
My agent setup: 4 Claude Code agents in tmux, 2x2 grid on the left, shared preview panel on the right. It's the only thing I'm looking at on my computer. No more software, no more UI. One bash script launches everything: tmux-work The trick: in CLAUDE.md, I tell every agent about the preview pane: "Panel %1 is the tmux preview shell. To show a document, run: tmux send-keys -t %1 'q' C-m 'preview "path/to/file"' C-m" Now when I say "preview this file", Claude pushes it to the right panel while the others keep working. Works with .md, .docx, .xlsx, .pdf, .pptx. One agent drafts an email to counsel. Another audits Stripe data. A third preps a contract. Fourth answers a customer question. All at the same time. My tmux-work script: #!/bin/bash tmux kill-server 2>/dev/null tmux new-session -s work -d -x 300 -y 80 tmux split-window -h -p 50 -t work tmux select-pane -t %0 tmux split-window -v -p 50 tmux select-pane -t %0 tmux split-window -h -p 50 tmux select-pane -t %2 tmux split-window -h -p 50 for pane in %0 %2 %3 %4; do tmux send-keys -t $pane 'claude --dangerously-skip-permissions' Enter done tmux select-pane -t %0 tmux attach -t work
Nicolas Bustamante tweet media
English
55
62
908
90.3K
Woody Lu retweetledi
bubble boi
bubble boi@bubbleboi·
All the data center capex is being delayed 1-2 years minimum now. Say goodbye to the memory thesis and the networking one, banks & institutional investor will pull back lending for SPVs that all big tech and labs are using to fund their infrastructure. Additionally Dubai, Saudi, Qatar, all won’t be funding the projects anymore and probably drawing down on their investments. The bubble is popping faaaaast.
SemiAnalysis@SemiAnalysis_

The divergence in the last few months has been incredible. Under a normal distribution, that's a one-in-1.7 trillion event. Of course markets are fat-tailed — but even by fat-tail standards, this is off the map. Prior peak was 4σ in 2003. (2/3)

English
67
186
2.1K
467.5K
Woody Lu retweetledi
Leo
Leo@runes_leo·
OpenAI 说他们的工程师不再写代码了——设计环境、搭反馈循环、定义架构约束,然后让 agent 写。五个月一百万行,没一行手写。他们管这叫 harness engineering。 这篇把它追溯到控制论:瓦特调速器、K8s、再到现在的 agent。同一个模式——你不再转阀门,你掌舵。 最扎心的一句:agent 反复犯错,不是能力问题,是你脑子里的判断力没写下来。你不写,它第一百次还犯同样的蠢。 深有体会。CLAUDE.md 里的行为规范、禁止清单、架构约束,本质就是在做"校准传感器"。写得越具体,agent 越不犯蠢。
George@odysseus0z

x.com/i/article/2030…

中文
11
91
555
111.9K
Woody Lu retweetledi
Chintan Turakhia
Chintan Turakhia@chintanturakhia·
Run this prompt frequently. You're welcome.
Chintan Turakhia tweet media
English
121
317
7K
906.6K