AC.Huang

90 posts

AC.Huang

AC.Huang

@sz1117

No longer comfortable. Some things don't happen now and won't do them in the future.

shenzhen china Katılım Ekim 2010
408 Takip Edilen29 Takipçiler
Brett Adcock
Brett Adcock@adcock_brett·
Figure and Hark just took an entire data center of NVIDIA B200s - every rack in the building Figure will be using these to predict physics and Hark will train next generation multi-modal models
Brett Adcock tweet media
English
76
123
1.4K
143.9K
AC.Huang
AC.Huang@sz1117·
@jukan05 来深圳 到你体验地道美食
AC.Huang tweet mediaAC.Huang tweet mediaAC.Huang tweet media
中文
0
0
2
216
Jukan
Jukan@jukan05·
深圳哪里有好吃的餐厅可以推荐吗?本地人/中国人常去的? Any Chinese locals here — could you recommend some good restaurants in Shenzhen?
中文
101
3
126
37.8K
适应期
适应期@WOWshiyingqi·
在香港注册了一个公司,做了三星和海力士的代理,专卖内存。 在港交易,欢迎询价
适应期 tweet media
中文
38
8
192
69.7K
am.will
am.will@LLMJunky·
Finally proud to announce that I've joined the GPU Minor Leagues. 2 x RTX 6000 Pro. I have six months to pay off the second GPU lol. You are all TERRIBLE influences.
am.will tweet media
English
121
15
896
61.1K
Pure Nomad
Pure Nomad@realPureNomad·
著名服务器公司 Supermicro 联创 Yih-Shyan 今日被捕。事由是通过东南亚壳公司向中国走私 25 亿美元的 Nvidia GPU。他本人持 4.64 亿美元的 Supermicro 股票。走私手段包括,用数千个假服务器迷惑美国执法者,监控录像显示使用吹风机换条形码。Supermicro 股票盘后暴跌 12%。此人将面临 30 年刑期。
Pure Nomad tweet mediaPure Nomad tweet media
中文
85
184
1.1K
231.2K
亚洲金融 Asia Finance
亚洲金融 Asia Finance@AsiaFinance·
两年前,车库咖啡创始人苏菂,公开炫耀如何逃避美国的AI技术制裁,走私了200片英伟达的H200显卡;两年后,苏菂的供应商被抓了,美国司法部今天公布案情:联邦调查局(FBI)调查显示,廖、张、孙三人利用东南亚公司,合谋向中国买家出售数十亿美元英伟达GPU,最长面临20年刑期。
中文
144
343
2K
683K
AC.Huang
AC.Huang@sz1117·
@CuiMao 你们这个怎么卖 求合作
中文
1
0
0
172
AC.Huang
AC.Huang@sz1117·
B300 开始涨价
AC.Huang tweet media
中文
0
0
0
41
AC.Huang
AC.Huang@sz1117·
@MSX_CN 高价收购服务器内存条 海外大量采购
AC.Huang tweet media
中文
0
0
0
6
麦通MSX
麦通MSX@MSX_CN·
存储芯片的领头羊 $MU.M 今夜盘后交业绩啦! 在 HBM(高带宽内存)供不应求的背景下,全市场都在看美光的盈利上限到底在哪。 📈 数据预期: 1️⃣销售量:192.99 亿 2️⃣净利润:98.78 亿 大家现在最关心的是:美光手里被 $NVDA.M 抢疯了的 HBM 产能,是不是已经排单到了后年?如果毛利率能借着涨价潮冲破 65%,那美光就真的成了芯片界印钞机。 如果美光指引超预期,整个芯片板块都要跟着起飞。等明早看是「真香」还是「利好出尽」?
麦通MSX tweet media
中文
82
34
73
4.2K
AC.Huang
AC.Huang@sz1117·
If you have RTX 5090 chips or H100/H200 servers for sale, please contact me.
AC.Huang tweet media
English
0
0
0
49
AC.Huang
AC.Huang@sz1117·
@DylanOnChips If you have RTX 5090 chips or H100/H200 servers for sale, please contact me.
English
0
0
0
11
AC.Huang
AC.Huang@sz1117·
@jukan05 Acquire data centers and decommission H100 and H200 servers
Français
0
0
0
515
Jukan
Jukan@jukan05·
According to a calculation by a Korean journalist estimating how much revenue Samsung Foundry could generate from Groq3 LPU production: NVIDIA has reportedly asked Samsung Foundry to produce around 500,000 LPU 3 chips to start with. That is more than double the originally planned production volume. First, each Groq server introduced by CEO Jensen Huang contains 8 LPUs. When 32 of those servers are combined, they form one rack. That means each rack contains 256 LPUs. One complete Vera Rubin platform contains 5 racks. So each platform would contain 1,280 LPUs. On a simple calculation, if all produced chips were turned into racks, that would imply about 1,950 LPU racks, enough for roughly 390 Vera Rubin platform sets. Looking at the LPU wafer shown in the commemorative photo taken a few days ago at the GTC exhibition featuring CEO Jensen Huang and Han Jin-man, president of Samsung Electronics’ Foundry Business, it appears that around 65 properly shaped LPU dies can be printed on a single wafer. To produce 500,000 LPUs, 7,692 wafers would be required if each wafer yields 65 chips and the yield rate is 100%. However, considering that Samsung’s current 4nm yield is estimated to be around 50–70%, it is possible to infer that more than 15,000 wafers annually would be needed, which is broadly consistent with recent media reports. According to what the journalist heard, the wafer price for Samsung’s 4nm process is at least around $11,000 per wafer. Assuming a possible price increase and using roughly $13,000 per wafer, that would imply about $195 million in revenue, or roughly KRW 300 billion, from LPU production. Given that Samsung Foundry’s annual revenue is known to be around KRW 20 trillion, this would still amount to a very small contribution to total sales.
Jukan tweet media
English
9
26
332
46.9K
AC.Huang
AC.Huang@sz1117·
@LinQingV 收购数据中心并退役 H100 和 H200 服务器
中文
1
0
0
36
Macro_Lin | 市场观察员
Nvidia要给中国市场供高性能推理芯片了。 路透独家:Nvidia正准备推出可在中国销售的Groq推理芯片,预计5月发布。 先说Groq是什么来头。这家公司做的不是GPU,是LPU,专门为推理设计的架构。核心思路是把HBM干掉,全用SRAM,执行路径完全确定性,不存在GPU那种调度开销。结果就是推理延迟极低,吞吐拉满,单位能耗也远低于GPU方案。Nvidia花170亿拿它的技术授权,不是收购,要的就是这套推理专用架构。 注意,这次不是阉割版。消息人士明确说"非降规、非中国特供",跟之前H800/A800完全不是一个路径。训练靠Vera Rubin,不能卖中国。推理走Groq,可以卖。一套组合拳,两条合规通道。同时H200也重启生产了,说明出口许可和中国订单都已经到手。 如果NVDA用推理芯片重新把中国市场的门撬开,影响就很直接。Nvidia中国区收入预期要重估,华尔街此前基本把这块从模型里划掉了。反过来,百度、寒武纪们自研推理芯片的窗口期,可能比市场以为的短得多。Nvidia带着成熟生态杀回来,纯靠性价比很难扛。 如果中国这边不阻挠,主攻推理的国内GPU厂商会很难受。
Macro_Lin | 市场观察员@LinQingV

龙虾吃Token,Token吃芯片。这条食物链的终点,是一张六百多亿美元的账单 MiniMax的M2.5模型2月单日Token消耗量暴涨6倍,编程套餐消耗飙升10倍。智谱Coding Plan每天早上10点抢购。涨价30%,上线即售罄。中国大模型周调用量首超美国,5.16万亿Token对2.7万亿。 黄仁勋说,Agent的Token消耗是传统对话的1000倍。人会睡觉,龙虾不会。 但所有人在讨论龙虾,没人问一个更根本的问题:这些指数级激增的Token需求,是用多少AI芯片? 摩根士丹利上周发了篇60页深度,试图回答这个问题。 中国AI芯片TAM从2024年的191亿美元增长到2030年的670亿美元。GPU自给率从2021年的10%爬到2030年的76%。华为一家拿走63%国产份额,寒武纪11%,剩下十几家厂商分不到三成。而且2026到2027年,整个市场仍然是供给驱动。不是没人买,是造不出来。 报告最有意思的判断是:中国AI芯片的故事不是"追制程",是"用基建换纳米"。 MS画了一张中美AI九维雷达图。晶圆前道和HBM内存,美国明显领先。但电力供应、数据中心空间、政策支持,中国反而占优。仅看制程节点会高估性能差距,换成单位功耗和单位成本的口径,差距大幅收敛。中国利润率要求更低,电价更便宜,推理场景下TCO反而有优势。 这就解释了为什么MiniMax输出价格能做到Claude的十分之一。不只是模型层面的工程优化,背后是整条基础设施成本曲线在支撑。 估值上就更刺激了。寒武纪32倍P/S,摩尔线程139倍P/S,对应隐含收入不过3到5亿美元。英伟达才17倍。市场定价的不是盈利能力,是稀缺性。 MS给出的关键判断:2027年后产能释放叠加设计趋同,同质化风险急升,行业整合2到3年内大概率发生。 把两件事放一起看就清楚了。龙虾爆火证明推理侧Token需求可以指数级增长。MS报告证明供给侧产能在2027年前仍是硬约束。供不应求意味着有产能就有收入。但窗口会关。一旦产能不再稀缺,139倍市销率就失去了存在的理由。 Token正在变成新时代的电。中国正在用廉价绿电加极致工程能力,把自己变成全球AI推理的水电煤供应商。但这条叙事听着耳熟吗?政策扶持、产能狂飙、价格战、全行业亏损、最后剩者为王。 光伏走过的路,AI芯片大概率要再走一遍。唯一的区别是,光伏用了十年,这一轮可能AI芯片会用几年?

中文
8
11
52
32.6K
AC.Huang
AC.Huang@sz1117·
If there is an idle NVIDIA H100 H200 server that needs to be disposed of, you can contact me.
AC.Huang tweet mediaAC.Huang tweet media
English
0
0
0
40
AC.Huang
AC.Huang@sz1117·
@nvidia If there is an idle NVIDIA H100 H200 server that needs to be disposed of, you can contact me.
English
0
0
0
8
NVIDIA
NVIDIA@nvidia·
In 2006, NVIDIA made a bold bet on a parallel computing platform that would eventually redefine the limits of science and industry. 20 years later, CUDA serves as the foundation for 6 million developers worldwide—and we are only just beginning. Join the next wave of architects and build what’s next on two decades of innovation. 🔗 nvda.ws/4bUUFUT
English
23
32
229
21.6K
AC.Huang
AC.Huang@sz1117·
@Tslachan If there is an idle NVIDIA H100 H200 server that needs to be disposed of, you can contact me.
English
0
0
0
8
Tsla Chan
Tsla Chan@Tslachan·
Samsung Chairman Lee Jae-yong met with AMD CEO Lisa Su for dinner in Seoul last evening. Lisa Su visited Korea for the first time since taking office at AMD, and the two are known as places to strengthen cooperation in AI memory. Samsung has been going wild lately.
Tsla Chan tweet media
English
2
6
61
4.2K
Dell Technologies
Dell Technologies@Dell·
We heard you like seafood 🦞. Meet the workstation of the AI era: The Dell Pro Max with GB300 and @NVIDIA NemoClaw. You can run AI agents locally, securely & nonstop.
English
110
108
878
194.6K