Zeke

5 posts

Zeke

Zeke

@meep414

Katılım Eylül 2025
47 Takip Edilen0 Takipçiler
Zeke
Zeke@meep414·
@clattner_llvm Does the solver recover the exact same schedule as flash attention? It would be cool if there were multiple optimal solutions or if it found something faster.
English
1
0
0
113
Chris Lattner
Chris Lattner@clattner_llvm·
Pipelining AI kernels is required to get full perf/utilization out of modern chips. However, no one has been able to crack "full control over the hardware" without "having to micromanage it". Let's crack this open: kernel authors deserve a powerful scheduler they can control. 💪
Modular@Modular

FA4 on Blackwell: 14 ops, 5 hardware units, 28 dependency edges. One wrong sync = a race condition sanitizers won't catch. We built a constraint solver that derives the pipeline schedule automatically, in Mojo 🔥 Part 1 of our series is out → modular.com/blog/software-…

English
9
16
255
24.6K
Zeke
Zeke@meep414·
@_xjdr What do you like about them?
English
0
0
0
10
xjdr
xjdr@_xjdr·
gpt-oss and gemma-3n are such good architectures . i didnt really appreciate them until i really put them to the test through rigorous ablations
English
7
7
186
12.2K
Elliot Arledge
Elliot Arledge@elliotarledge·
After diving head first into the deep end of squeezing every last drop out of inference megakernels, I decided to write a book about my journey, as well as how others like @AlpinDale and @HazyResearch architect their megakernels on hopper and blackwell. This assumes comfort with CUDA and LLM inference. elliotarledge.gumroad.com/l/grokking-meg…
English
4
7
129
4.9K
Zeke
Zeke@meep414·
@alz_zyd_ the normie's confusion with calculus is totally understandable. suppose i told you in 1650 that, in 1660, i could produce a closed form equation for the area under any continuous function between two points. you'd obviously think i'm a complete nutcase
English
0
0
1
7
alz
alz@alz_zyd_·
The normies' confusion with AI is totally understandable. Suppose I told you in 2015 that, in 2025, gradient descent would produce black magic billionparameter bots which could think and talk and solve math questions. You'd obviously think I'm a complete nutcase
English
86
66
1.2K
69.6K
Greg Kamradt
Greg Kamradt@GregKamradt·
Open questions with Gemini Deep Think and @arcprize: - Why doesn't it get 100% on ARC-AGI? We're trying to understand failure modes, need to inspect tasks more - Why a slight jump in ARC-AGI-1, but a 2x SOTA in ARC-AGI-2? Here are tasks that stood out in our early testing:
Greg Kamradt tweet media
Greg Kamradt@GregKamradt

After we got early access to Gemini 3 Pro, it was SOTA, we were impressed and then Google told us, "there's one more thing..." Deep Think sets the new high water mark on ARC-AGI-2

English
11
25
490
103K