Alex Zhurkevich

339 posts

Alex Zhurkevich banner
Alex Zhurkevich

Alex Zhurkevich

@cudagdb

Select difficulty 👁️‍🗨️

United States شامل ہوئے Aralık 2017
540 فالونگ578 فالوورز
Alex Zhurkevich
Alex Zhurkevich@cudagdb·
@Simon_Vt I think the flahsinfer repo is Apache 2. Pls consult with your lawyer.
English
1
0
1
39
Alex Zhurkevich
Alex Zhurkevich@cudagdb·
Trtllmgen kernels are now open. Fastest prefill and decode kernels for our target workloads. We wrote these to win InferenceX, MLPerf, other benchmarks. Powering some of today’s top served models. Dive in, learn, use them, or level up your own. Enjoy. github.com/flashinfer-ai/…
English
13
50
331
142.2K
Alex Zhurkevich
Alex Zhurkevich@cudagdb·
Give Meta 5 years of life, they'll call you a pokemon.
Soumith Chintala@soumithchintala

@marksaroufim You build communities that last, and that’s rare. GPU Mode, PyTorch… they worked because you showed up with honesty and zero ego. This makes you a rare pokemon. Thanks for your service. Excited for what’s next. I'm sure you’re definitely not done leveling up.

English
0
0
16
8K
Alex Zhurkevich
Alex Zhurkevich@cudagdb·
@Leik0w0 Trtllmgen predates cute dsl. We prefer writing CUDA+PTX kernels, though Cutlass and Cute are also excellent. It’s just a matter of style - our code even uses some Cutlass components like pipelines.
English
0
2
24
2K
Léo
Léo@Leik0w0·
@cudagdb Insane drop! I’m curious though, why not CuTe DSL ? Are these perfs not yet achievable with CuTe DSL ?
English
1
0
6
2.9K
bone
bone@boneGPT·
Elon won. Sora is defeated. Long live Grok Imagine.
English
39
21
453
25.5K
Alex Zhurkevich
Alex Zhurkevich@cudagdb·
Pay attention to this one. Anne and folks are cooking smth special 👩🏻‍🍳
Anne Ouyang@anneouyang

Excited to share @Standard_Kernel's seed round and some reflections on what we’ve learned about kernel generation and what we believe is next. Grateful to our amazing team, supporters, and the broader community pushing this space forward.

English
0
0
6
991
Alex Zhurkevich ری ٹویٹ کیا
Kion
Kion@OKfallah·
In-context learning is a hack to remind your model. CLaaS uses self-distillation to move that knowledge into weights, freeing up context.
Kion tweet media
English
3
6
23
3.2K