sanguyen.eth

934 posts

sanguyen.eth

sanguyen.eth

@BitBoss_

参加日 Aralık 2021
576 フォロー中238 フォロワー
sanguyen.eth
sanguyen.eth@BitBoss_·
@BillAckman guy was murdered by armed and masked thugs. There is only one truth. watch the video of ice thug who is empty handed, then pulls the gun from the victim before shots are fired. the Feds just like trump make up their own story to hide their crime.
English
0
0
0
18
Bill Ackman
Bill Ackman@BillAckman·
We have reached a stage in our country where there are only two sides to every issue and every incident. Each side lives in protected echo chambers which are provided with a curated set of ‘facts’ and/or video footage from certain camera angles that are consistent with the preexisting views and conclusions of that side. Individuals are ‘convicted’ of serious crimes in the headlines, by politicians appealing to their base, and ultimately in the minds of the public, or they are exonerated, before all of the facts are in and a detailed investigation has been completed. This is not good for America. We need to go back to a world where we suspend judgment and await the conclusions of a detailed investigation before we convict or exonerate. Let’s not forget that a man is presumed innocent until proven guilty. Rushing to judgment helps no one and harms us all. It also greatly elevates the temperature, which keeps potential targets of law enforcement and those who enforce our laws on edge, massively increasing the risk to all. We need to take a deep breath and reserve judgment before this gets even more out of control.
English
5K
1.6K
16.1K
3.6M
sanguyen.eth
sanguyen.eth@BitBoss_·
@JoshConstine Is the cash payout for employees out of the $20 billion or separately?
English
0
0
0
78
Josh Constine 📶🔥
Josh Constine 📶🔥@JoshConstine·
Great to hear Groq leadership did right by their employees and most will see proper compensation, including some vesting acceleration. But this should still be built into hiring contracts, and not just be at the mercy of the founders.
Josh Constine 📶🔥 tweet media
Josh Constine 📶🔥@JoshConstine

Attention startup employees: Demand that in the event of a Groq-like hacquisition, you get accelerated vesting and pro rata compensation like in a real acquisition. Don’t get Windsurf’d.

English
35
27
680
103.6K
sanguyen.eth
sanguyen.eth@BitBoss_·
@GavinSBaker Groq still owns the IP so they can still be acquired for this alone. Team that moved to Nvidia is exceptional, they are not the only brilliant engineers out there. Groq continues cloud and DC biz; get bought out for this and IP within 3 yrs.
English
0
1
0
443
Gavin Baker
Gavin Baker@GavinSBaker·
For the sake of clarity and as some have pointed in the replies, I should note that Nvidia is not actually acquiring Grok. It is a non-exclusive licensing agreement with some Grok engineers joining Nvidia. Grok will continue to operate their cloud business as an independent company that is effectively a competitor to Nvidia and their customers, whether hyperscaler or neocloud. Net, net should be great for AI users. More competition, more tokens. Merry Christmas and Tokens For All.
English
16
14
213
48.3K
Gavin Baker
Gavin Baker@GavinSBaker·
Nvidia is buying Groq for two reasons imo.   1) Inference is disaggregating into prefill and decode. SRAM architectures have unique advantages in decode for workloads where performance is primarily a function of memory bandwidth. Rubin CPX, Rubin and the putative “Rubin SRAM” variant derived from Groq should give Nvidia the ability to mix and match chips to create the optimal balance of performance vs. cost for each workload. Rubin CPX is optimized for massive context windows during prefill as a result of super high memory capacity with its relatively low bandwidth GDDR DRAM. Rubin is the workhorse for training and high density, batched inference workloads with its HBM DRAM striking a balance between memory bandwidth and capacity. The Groq-derived "Rubin SRAM" is optimized for ultra-low latency agentic reasoning inference workloads as a result of SRAM’s extremely high memory bandwidth at the cost of lower memory capacity. In the latter case, either CPX or the normal Rubin will likely be used for prefill.   2) It has been clear for a long time that SRAM architectures can hit token per second metrics much higher than GPUs, TPUs or any ASIC that we have yet seen. Extremely low latency per individual user at the expense of throughput per dollar. It was less clear 18 months ago whether end users were willing to pay for this speed (SRAM more expensive per token due to much smaller batch sizes). It is now abundantly clear from Cerebras and Groq’s recent results that users are willing to pay for speed.   Increases my confidence that all ASICs except TPU, AI5 and Trainium will eventually be canceled. Good luck competing with the 3 Rubin variants and multiple associated networking chips. Although it does sound like OpenAI’s ASIC will be surprisingly good (much better than the Meta and Microsoft ASICs).   Let’s see what AMD does. Intel already moving in this direction (they have a prefill optimized SKU and purchased SambaNova, which was the weakest SRAM competitor). Kinda funny that Meta bought Rivos. And Cerebras, where I am biased, is now in a very interesting and highly strategic position as the last (per public knowledge) independent SRAM player that was ahead of Groq on all public benchmarks. Groq’s “many chip” rack architecture, however, was much easier to integrate with Nvidia’s networking stack and perhaps even within a single rack while Cerebras’s WSE almost has to be an independent rack.
English
176
392
2.9K
1.1M
sanguyen.eth
sanguyen.eth@BitBoss_·
@chamath Someone will buy the remaining Groq to get the IP and build on top of this. You heard it here first.
English
0
0
0
47
Chamath Palihapitiya
Chamath Palihapitiya@chamath·
This is directionally right. The HBM vs SRAM tradeoff in architecture design was clear many years ago. Those that picked HBM are in a queue behind Nvidia and Google. Good luck with that. More broadly, LLM decode patterns favor SRAM. But unlike Gavin, I think this creates a lane for even more heterogenous silicon to support AI models in the future. Not less. I suspect that the two axes that matter are accuracy vs speed and if you can design a focused solution for a specific AI use case, there will be a market.
Gavin Baker@GavinSBaker

Nvidia is buying Groq for two reasons imo.   1) Inference is disaggregating into prefill and decode. SRAM architectures have unique advantages in decode for workloads where performance is primarily a function of memory bandwidth. Rubin CPX, Rubin and the putative “Rubin SRAM” variant derived from Groq should give Nvidia the ability to mix and match chips to create the optimal balance of performance vs. cost for each workload. Rubin CPX is optimized for massive context windows during prefill as a result of super high memory capacity with its relatively low bandwidth GDDR DRAM. Rubin is the workhorse for training and high density, batched inference workloads with its HBM DRAM striking a balance between memory bandwidth and capacity. The Groq-derived "Rubin SRAM" is optimized for ultra-low latency agentic reasoning inference workloads as a result of SRAM’s extremely high memory bandwidth at the cost of lower memory capacity. In the latter case, either CPX or the normal Rubin will likely be used for prefill.   2) It has been clear for a long time that SRAM architectures can hit token per second metrics much higher than GPUs, TPUs or any ASIC that we have yet seen. Extremely low latency per individual user at the expense of throughput per dollar. It was less clear 18 months ago whether end users were willing to pay for this speed (SRAM more expensive per token due to much smaller batch sizes). It is now abundantly clear from Cerebras and Groq’s recent results that users are willing to pay for speed.   Increases my confidence that all ASICs except TPU, AI5 and Trainium will eventually be canceled. Good luck competing with the 3 Rubin variants and multiple associated networking chips. Although it does sound like OpenAI’s ASIC will be surprisingly good (much better than the Meta and Microsoft ASICs).   Let’s see what AMD does. Intel already moving in this direction (they have a prefill optimized SKU and purchased SambaNova, which was the weakest SRAM competitor). Kinda funny that Meta bought Rivos. And Cerebras, where I am biased, is now in a very interesting and highly strategic position as the last (per public knowledge) independent SRAM player that was ahead of Groq on all public benchmarks. Groq’s “many chip” rack architecture, however, was much easier to integrate with Nvidia’s networking stack and perhaps even within a single rack while Cerebras’s WSE almost has to be an independent rack.

English
55
68
1.1K
325.8K
sanguyen.eth
sanguyen.eth@BitBoss_·
@PatrickMoorhead Since Groq still owns the IP, what is to stop them from hiring SRAM experts to advance the chip similar to Nvidia and license/sell to other hyperscalers and retain their cloud services business?
English
0
0
0
92
Patrick Moorhead
Patrick Moorhead@PatrickMoorhead·
Mostly aligned with Gavin on this. Whenever I was asked, “how does NVIDIA compete with ASICs” my response has always been for the past year: 1/ the AI pipeline will split into three distinct workloads 2/ CPX fills one, maybe two of the three workloads 3/ NVIDIA will have to fill the gaps best it can with CUDA optimizations OR create OR buy the other. Enter Groq. Many others gearing up to create the best solutions for one, maybe two of the three workloads but customers will need an elegant software solution to cut across the three. For now, only the largest hyperscalers would be able to create this, will take time for an open source solution.
Gavin Baker@GavinSBaker

Nvidia is buying Groq for two reasons imo.   1) Inference is disaggregating into prefill and decode. SRAM architectures have unique advantages in decode for workloads where performance is primarily a function of memory bandwidth. Rubin CPX, Rubin and the putative “Rubin SRAM” variant derived from Groq should give Nvidia the ability to mix and match chips to create the optimal balance of performance vs. cost for each workload. Rubin CPX is optimized for massive context windows during prefill as a result of super high memory capacity with its relatively low bandwidth GDDR DRAM. Rubin is the workhorse for training and high density, batched inference workloads with its HBM DRAM striking a balance between memory bandwidth and capacity. The Groq-derived "Rubin SRAM" is optimized for ultra-low latency agentic reasoning inference workloads as a result of SRAM’s extremely high memory bandwidth at the cost of lower memory capacity. In the latter case, either CPX or the normal Rubin will likely be used for prefill.   2) It has been clear for a long time that SRAM architectures can hit token per second metrics much higher than GPUs, TPUs or any ASIC that we have yet seen. Extremely low latency per individual user at the expense of throughput per dollar. It was less clear 18 months ago whether end users were willing to pay for this speed (SRAM more expensive per token due to much smaller batch sizes). It is now abundantly clear from Cerebras and Groq’s recent results that users are willing to pay for speed.   Increases my confidence that all ASICs except TPU, AI5 and Trainium will eventually be canceled. Good luck competing with the 3 Rubin variants and multiple associated networking chips. Although it does sound like OpenAI’s ASIC will be surprisingly good (much better than the Meta and Microsoft ASICs).   Let’s see what AMD does. Intel already moving in this direction (they have a prefill optimized SKU and purchased SambaNova, which was the weakest SRAM competitor). Kinda funny that Meta bought Rivos. And Cerebras, where I am biased, is now in a very interesting and highly strategic position as the last (per public knowledge) independent SRAM player that was ahead of Groq on all public benchmarks. Groq’s “many chip” rack architecture, however, was much easier to integrate with Nvidia’s networking stack and perhaps even within a single rack while Cerebras’s WSE almost has to be an independent rack.

English
15
16
167
51.4K
sanguyen.eth
sanguyen.eth@BitBoss_·
@JonathanRoss321 How is this not a terrible deal? Valuation $10B-$15B in 1 yr, $30-50B in 4, more after (?). If inference is 1MX then why gut it for $20B now? Left a lot on table, screwed employees, investors ex. core & @chamath. Acquihire and announce Xmas eve - shady.
English
0
0
0
57
Jonathan Ross
Jonathan Ross@JonathanRoss321·
Today Groq entered into a non-exclusive licensing agreement with Nvidia for Groq’s inference technology. Along with other members of the Groq team, I’ll be joining Nvidia to help integrate the licensed technology. GroqCloud will continue to operate without interruption. Learn more here: groq.com/newsroom/groq-…
English
267
347
3.9K
1M
sanguyen.eth
sanguyen.eth@BitBoss_·
@altcap How is this not a terrible deal? In a year, valuation would be $10B-$15B, $30-50B in 4 yrs, more after. If inference is 1MX then why gut it for $20B now? Left a lot on the table, screwed employees, and investors ex. Core team and @chamath. what’s left - cloud? Seriously?
English
0
0
1
43
Brad Gerstner
Brad Gerstner@altcap·
In a world of deep thinking there is an insatiable demand for fast & cheap tokens. AI systems / factories are becoming multi chip massive systems that have specialized components depending on the primary workload.
English
9
5
117
23K
sanguyen.eth
sanguyen.eth@BitBoss_·
@GroqInc Is anyone not pissed that @GroqInc left so much on the table and shafted a lot of employees? Chamath and Jonathan (core team) got paid but beyond that, doesn’t seem this is great deal if standalone value in 2-4 years could be $50-$100B
English
0
0
4
169
Groq Inc
Groq Inc@GroqInc·
Groq has entered into a non-exclusive licensing agreement with Nvidia for Groq’s inference technology. GroqCloud will continue to operate without interruption. Learn more here: groq.com/newsroom/groq-…
English
128
284
2.4K
1.4M
sanguyen.eth
sanguyen.eth@BitBoss_·
@SuitToSweats @pistachiomatt Groq left a lot on the table to be acqui-hired by NVDA. $20B or $100B in 3-5 years? Why leave $80B to work for NVDA? Chamath and team would have made much more for every billion increase in valuation. There must be some other play here.
English
1
0
1
47
DZ | Metir AI
DZ | Metir AI@SuitToSweats·
@pistachiomatt They will be (I hope)… but this should have been a $500b business in 2-3 years… $20b is what NVDA is seeding into random funds that invest in their chips The stub that’s left is a bit of a joke.
English
3
0
2
156
James Wang
James Wang@draecomino·
So many bad takes on Groq as if its LPU is some magical new architecture or a TPU for hire. Groq's micro architecture does not matter. The *only* reason Groq has any traction is because it bet on SRAM. Without SRAM, there's no speed advantage, no PMF, no demand, and no acquisition. No one cares about deterministic latency, fancy compilers, or VLIW cores. Ultimately Nvidia bought Groq because if Groq is allowed to scale its roadmap - especially as part of a well funded hyperscaler - it would become a huge drag on Nvidia's narrative/valuation. $20B today saves $200B later. The tech almost doesn't matter.
English
63
57
1K
219.7K
Steph Kent
Steph Kent@covertress·
Cigar-shaped UFO over Chile 🛸 "This does not appear to match any specific known craft." — Grok So, what is it?
English
964
309
2.3K
398.9K
Jesse Michels
Jesse Michels@AlchemyAmerican·
Have any prominent astrophysicists supported Avi Loeb's bold position on 3I Atlas?
English
234
26
738
138.9K
sanguyen.eth がリツイート
David Roberts
David Roberts@recap_david·
$24,000 per year from this simple AI Dentist Voice Agent (and why I'm crazy for giving it away for free) A dental practice was losing $6,000+ in revenue every month from missed after-hours calls. That's 20-25 potential patients walking away because no one was available to book their appointments. So I built an AI voice assistant that handles after-hours dental bookings 24/7 using n8n and ElevenLabs based on internal policies and scheduling availability. Here's what this system does: → Answers calls with a natural-sounding AI receptionist → Collects patient information and insurance details → Checks calendar availability in real-time → Books appointments automatically → Logs all patient details to a Google Sheet The result? This similar AI voice system was sold to a dental practice for $24k per year by another entrepreneur!! This isn't just about dental practices. Any service business losing money from missed calls can implement a similar system. Want the complete n8n workflow template? 1. Retweet & Like this post 2. Comment "ASSISTANT" I'll send you the entire system for free, a full setup walk-through video, including the ElevenLabs automation components.
English
993
790
3.1K
351.7K
Senator Rand Paul
Senator Rand Paul@SenRandPaul·
A war with Iran would make Iraq look like a skirmish. If your goal were to bankrupt America and destabilize the world, dragging us into another endless Middle East war would be the way to do it. That’s why I’m glad @POTUS promised—and brought the country together—on a commitment to peace, not more war.
English
1.7K
1.6K
18.8K
2.9M
Mick West
Mick West@MickWest·
@BitBoss_ @MvonRen This? Hard to say. It could be a bird. Need better video. But then they would see what it was, and not post it. That's now the LIZ works.
English
1
0
1
179
sanguyen.eth
sanguyen.eth@BitBoss_·
@MickWest @MvonRen I can agree these are most likely migrating birds, close wings, flip around and reform. But what about the single object and its quick movement - later in the clip. Would you say that’s a bird as well? - camera seems to be stationary; staying open minded.
English
1
0
0
204
Mick West
Mick West@MickWest·
@MvonRen But it's obviously birds. Why do you think they were resistant to that fact?
English
10
0
42
3.6K
sanguyen.eth
sanguyen.eth@BitBoss_·
@tito @JTLonsdale @grok Totally wrong. Fear and greed can move the markets. You don’t need to sell a $1T a day to move anything, but threaten to do it, or start selling some, rattle the cage, and the multiplier effect will take you down easily. I’d say we are still a few trillion from the bottom.
English
0
0
0
264
Joe Lonsdale
Joe Lonsdale@JTLonsdale·
Unusual markets… is the CCP dumping treasuries like mad? Curious how to measure it and what we’ve heard.
English
185
136
2.9K
2.2M
Curiosity
Curiosity@CuriosityonX·
James Webb's stunning view of M51 galaxy!
Curiosity tweet media
English
176
1.9K
17K
650.4K
sanguyen.eth
sanguyen.eth@BitBoss_·
@AesPolitics1 Haha, all the MAGA shitting their pants cause gonna pay more at the grocery or lose their jobs. Hey, but you know us “rich” folks gonna be okay.
English
0
0
0
45
Aes🇺🇸
Aes🇺🇸@AesPolitics1·
What the fuck are we in a trade war with Canada for? Who the fuck voted for this stupid fuck?
English
8.7K
13.3K
253.9K
11.7M