Rishab Verma

723 posts

Rishab Verma

Rishab Verma

@rishabv90

Google AI Compute Infra. A fellow AI pilled engineer. & Of course, thoughts here, are my own 🫰

Seattle, WA Katılım Mayıs 2017
282 Takip Edilen121 Takipçiler
Belion
Belion@Gilbert_Belion·
Reliability aside, this car just hits differently. The Range Rover SV is an experience Toyota can't replicate, and it knows it.
Belion tweet media
English
291
134
2.6K
162.7K
Rishab Verma
Rishab Verma@rishabv90·
@quxiaoyin +1 Xiaoyin. Although, edge AI is picking up pretty fast....
English
0
0
0
17
Xiaoyin Qu
Xiaoyin Qu@quxiaoyin·
local is overrated. Cloud sandbox is the future of all ai agents. Long AWS. Google cloud and cloud providers.
English
55
1
86
7.1K
Hemingway Capital
Hemingway Capital@lfg_cap·
H100s are worth more today per hour than 18 months ago. Yet per token prices have collapsed >90%. GPU economics are improving at every level. Don’t let @michaeljburry break your brain.
English
7
20
338
46.2K
rachel 🪷
rachel 🪷@racheleizner·
at least Elizabeth Holmes was doing fraud with something fun. soc2 is boring af
English
19
103
2.9K
58.8K
ein
ein@stylishdawg·
@rishabv90 😭😭😭😭😭😭
ein tweet media
QME
5
0
44
1.6K
Husker
Husker@Existencial33·
@lfg_cap @michaeljburry No they’re not. 18 months ago H100s were at 2.5$/h Now they’re at 1.7.
English
2
1
8
2.5K
Shivani Patel
Shivani Patel@shivanijpatel·
data center on alcatraz - who's building this?
English
2
0
9
828
Brian Roemmele
Brian Roemmele@BrianRoemmele·
My project for the last few hours: Convert the Nvidia RTX 5090 to 128 GB of unified memory and a 28% speed increase! This board beats the L40s with a $10,000 savings! I got the process down with Mr. @Grok supervising. He said “Garage AI builders beat corporate bloat”.
Brian Roemmele tweet media
English
74
71
909
72.2K
Yuchen Jin
Yuchen Jin@Yuchenj_UW·
This is how you handle a PR crisis: Own the mistake. Acknowledge your partners. Commit to fixing it. That’s how you earn trust. Glad to see the Cursor cofounders got it right. Also love the trend: more companies building on top of open-source models. GPUs go brrr.
Yuchen Jin tweet media
English
27
20
336
29.4K
Rishab Verma
Rishab Verma@rishabv90·
@thegenioo After speaking with a couple of mixtral engineers for 45 mins at GTC, I can confirm this comparison is not accurate 🙏 But sure...
English
1
0
0
16
Hamza
Hamza@thegenioo·
Mistral is xAI of Open source labs
Artificial Analysis@ArtificialAnlys

Mistral has released Mistral Small 4, an open weights model with hybrid reasoning and image input, scoring 27 on the Artificial Analysis Intelligence Index @MistralAI's Small 4 is a 119B mixture-of-experts model with 6.5B active parameters per token, supporting both reasoning and non-reasoning modes. In reasoning mode, Mistral Small 4 scores 27 on the Artificial Analysis Intelligence Index, a 12-point improvement from Small 3.2 (15) and now among the most intelligent models Mistral has released, surpassing Mistral Large 3 (23) and matching the proprietary Magistral Medium 1.2 (27). However, it lags open weights peers with similar total parameter counts such as gpt-oss-120B (high, 33), NVIDIA Nemotron 3 Super 120B A12B (Reasoning, 36), and Qwen3.5 122B A10B (Reasoning, 42). Key takeaways: ➤ Reasoning and non-reasoning modes in a single model: Mistral Small 4 supports configurable hybrid reasoning with reasoning and non-reasoning modes, rather than the separate reasoning variants Mistral has released previously with their Magistral models. In reasoning mode, the model scores 27 on the Artificial Analysis Intelligence Index. In non-reasoning mode, the model scores 19, a 4-point improvement from its predecessor Mistral Small 3.2 (15) ➤ More token efficient than peers of similar size: At ~52M output tokens, Mistral Small 4 (Reasoning) uses fewer tokens to run the Artificial Analysis Intelligence Index compared to reasoning models such as gpt-oss-120B (high, ~78M), NVIDIA Nemotron 3 Super 120B A12B (Reasoning, ~110M), and Qwen3.5 122B A10B (Reasoning, ~91M). In non-reasoning mode, the model uses ~4M output tokens ➤ Native support for image input: Mistral Small 4 is a multimodal model, accepting image input as well as text. On our multimodal evaluation, MMMU-Pro, Mistral Small 4 (Reasoning) scores 57%, ahead of Mistral Large 3 (56%) but behind Qwen3.5 122B A10B (Reasoning, 75%). Neither gpt-oss-120B nor NVIDIA Nemotron 3 Super 120B A12B support image input. All models support text output only ➤ Improvement in real-world agentic tasks: Mistral Small 4 scores an Elo of 871 on GDPval-AA, our evaluation based on OpenAI's GDPval dataset that tests models on real-world tasks across 44 occupations and 9 major industries, with models producing deliverables such as documents, spreadsheets, and diagrams in an agentic loop. This is more than double the Elo of Small 3.2 (339) and close to Mistral Large 3 (880), but behind gpt-oss-120B (high, 962), NVIDIA Nemotron 3 Super 120B A12B (Reasoning, 1021), and Qwen3.5 122B A10B (Reasoning, 1130) ➤ Lower hallucination rate than peer models of similar size: Mistral Small 4 scores -30 on AA-Omniscience, our evaluation of knowledge reliability and hallucination, where scores range from -100 to 100 (higher is better) and a negative score indicates more incorrect than correct answers. Mistral Small 4 scores ahead of gpt-oss-120B (high, -50), Qwen3.5 122B A10B (Reasoning, -40), and NVIDIA Nemotron 3 Super 120B A12B (Reasoning, -42) Key model details: ➤ Context window: 256K tokens (up from 128K on Small 3.2) ➤ Pricing: $0.15/$0.6 per 1M input/output tokens ➤ Availability: Mistral first-party API only. At native FP8 precision, Mistral Small 4's 119B parameters require ~119GB to self-host the weights (more than the 80GB of HBM3 memory on a single NVIDIA H100) ➤ Modality: Image and text input with text output only ➤ Licensing: Apache 2.0 license

English
13
6
151
13.4K
Rishab Verma
Rishab Verma@rishabv90·
@amanrsanger Congratulations 👏🎉 super impressive and got to meet your super awesome team at GTC !
English
0
0
0
664