raulpuri.eth

486 posts

raulpuri.eth

raulpuri.eth

@TheRealRPuri

AI @ hrooeu sjmrtPectPs | past: OpenAI - ChatGPT Multimodal, Her, 4o, GPT4V, 4, 3.5, Codex | NVIDIA - megatron, language | 🐻

Katılım Mart 2014
385 Takip Edilen8.4K Takipçiler
Alex Nichol
Alex Nichol@unixpickle·
Reddit seems a lot less pro-Anthropic than Twitter. I'm curious if this has been a trend before now too.
English
3
0
8
2.8K
raulpuri.eth
raulpuri.eth@TheRealRPuri·
@xxtiange @drfeifei Curious where things are at now with newer models. Would be awesome if yall maintained a leaderboard
English
1
0
2
60
Tiange Xiang
Tiange Xiang@xxtiange·
‼️VLMs/MLLMs do NOT yet understand the physical world from videos‼️ In our recent work, we found that even the most advanced AI models still lag behind humans in one key aspect: reasoning about the kinematic properties of objects from videos. Takeaways: 1. ChatGPT 5.1 leads overall among 21 advanced VLMs, followed by Gemini 2.5 Pro/Flash. 2. Grok 4.1 delivers impressive performance at the lowest API cost. 3. Qwen3-VL is the top-performing open-source model. Read here: quantiphy.stanford.edu 🧵1/N
Tiange Xiang tweet mediaTiange Xiang tweet media
English
11
8
82
27.4K
raulpuri.eth retweetledi
roon
roon@tszzl·
messaging the top brass: guaranteed response within a few hours, p90 is under 5 minutes. messaging mid level: you may never get a response
English
73
40
2.4K
320.9K
raulpuri.eth retweetledi
Trevor Cai
Trevor Cai@trevorycai·
3 years ago, we emailed Jensen with requests for Blackwell. Today, we released GPT-5.3-Codex, a SOTA model designed for GB200-NVL72. Nitpicking ISA, simming rack designs, and tailoring our arch to the system has been a fun experience! I'm grateful to our collaborators at NVIDIA.
Trevor Cai tweet media
English
89
76
1.6K
399.3K
raulpuri.eth retweetledi
Jeffrey Zhang
Jeffrey Zhang@j4orz·
> Some companies hire heavily out of Twitter, some hire from communities such as GPU Mode or NanoGPT speedrunning. To Nathan's point, I am leading an open source workgroup within the @GPU_MODE community (#teenygrad channel) in order to develop a deep learning systems course with an MIT-licensed book, codebase, and lectures which develops your own deep learning framework teenygrad from scratch which can run nanogpt. The project has access to some compute thanks to to the @LambdaAPI research grant (thank you @chuanli11) This project has been a labor of love the past few months, bridging a must-needed pedagogical gap from micrograd to tinygrad. The SITP book develops teenygrad framework step by step, from a numpy clone, to a pytorch1 clone, to a pytorch2 style compiler. The SITP philosophy subscribes to the same views as @karpathy on education: it's a technical problem whose solution requires a ramp with empathy: > ..education is the very difficult technical process of building ramps to knowledge...I feel like education is..a tangle of understanding and you're trying to lay it out in a way that creates a ramp where everything only depends on the thing before it. The project's primary challenge for better and for worse has been the breadth of scope. A lot of time was spent "curriculum engineering", and we are now just getting to implementing accelerated cpu and gpu kernels with automatic differentiation in earnest, but it has a good line of sight towards fusion compilation using tinygrad's RISCy IR. The good news for you is that now is a perfect time to help the workgroup, and to come join in learning from the best. There are some heavy hitters here led by @marksaroufim, @m_sirovatka, @a1zhang, @gaunernst and more. Links below ⬇️
Jeffrey Zhang tweet mediaJeffrey Zhang tweet media
Nathan Lambert@natolambert

My raw thoughts on the job market -- both for those hiring and those searching -- at the cutting edge of AI. interconnects.ai/p/thoughts-on-…

English
7
63
624
59.7K
raulpuri.eth
raulpuri.eth@TheRealRPuri·
@0x49fa98 Ppl that put the ball in the hoop and I don’t see listed elsewhere — isa fulford, allison tam, Christina Kim, mianna chen, Lia guy, Angela Jiang, Rachel Lim
English
1
0
2
1.1K
raulpuri.eth
raulpuri.eth@TheRealRPuri·
@rown @rown what’s the MultimodalBatch situation look like under the hood.
English
0
0
2
1.7K
Rowan Zellers
Rowan Zellers@rown·
Today we are releasing Tinker to everyone, and now with vision input! You can now finetune a frontier Qwen3-VL-235B on your own image+text data, bringing your own algorithm (sft, RL, something else?). We'll take care of the GPU infra. Full update: thinkingmachines.ai/blog/tinker-ge…
English
35
113
1.2K
135.8K
raulpuri.eth retweetledi
Pedro Domingos
Pedro Domingos@pmddomingos·
Hinton is no longer afraid of superintelligence.
English
570
620
4.1K
815.5K
raulpuri.eth
raulpuri.eth@TheRealRPuri·
Robots are parse trees.
English
1
0
0
821
raulpuri.eth retweetledi
Chubby♨️
Chubby♨️@kimmonismus·
Google DeepMind's Nando de Freitas: "Machines that can predict what their sensors (touch, cameras, keyboard, temperature, microphones, gyros, …) will perceive are already aware and have subjective experience. It’s all a matter of degree now." I think we need to revisit the discussion of when consciousness and self-awareness begin.
Nando de Freitas@NandoDF

Machines that can predict what their sensors (touch, cameras, keyboard, temperature, microphones, gyros, …) will perceive are already aware and have subjective experience. It’s all a matter of degree now. More sensors, data, compute, tasks will lead without any doubt to the “I think therefore I am” moment for computers, and we’re not ready for it yet. arxiv.org/pdf/1804.06318 share.google/kxx6WyqHpwPmo6…

English
31
38
409
99.9K