Wentao Guo

51 posts

Wentao Guo

Wentao Guo

@WentaoGuo7

CS PhD student @PrincetonCS, Previously CS MEng + BS @CornellCIS

เข้าร่วม Kasım 2021
197 กำลังติดตาม885 ผู้ติดตาม
Wentao Guo รีทวีตแล้ว
Albert Gu
Albert Gu@_albertgu·
The newest model in the Mamba series is finally here 🐍 Hybrid models have become increasingly popular, raising the importance of designing the next generation of linear models. We've introduced several SSM-centric ideas to significantly increase Mamba-2's modeling capabilities without compromising on speed. The resulting Mamba-3 model has noticeable performance gains over the most popular previous linear models (such as Mamba-2 and Gated DeltaNet) at all sizes. This is the first Mamba that was student led: all credit to @aakash_lahoti @kevinyli_ @_berlinchen @caitWW9, and of course @tri_dao!
Albert Gu tweet media
English
36
311
1.6K
406.3K
Wentao Guo รีทวีตแล้ว
Ted Zadouri
Ted Zadouri@tedzadouri·
Asymmetric hardware scaling is here. Blackwell tensor cores are now so fast, exp2 and shared memory are the wall. FlashAttention-4 changes the algorithm & pipeline so that softmax & SMEM bandwidth no longer dictate speed. Attn reaches ~1600 TFLOPs, pretty much at matmul speed! joint work w/ Markus Hoehnerbach, Jay Shah(@ultraproduct), Timmy Liu, Vijay Thakkar (@__tensorcore__ ), Tri Dao (@tri_dao) 1/
Ted Zadouri tweet media
English
6
132
780
219.4K
Wentao Guo รีทวีตแล้ว
Mayank Mishra
Mayank Mishra@MayankMish98·
We identified an issue with the Mamba-2 🐍 initialization in HuggingFace and FlashLinearAttention repository (dt_bias being incorrectly initialized). This bug is related to 2 main issues: 1. init being incorrect (torch.ones) if Mamba-2 layers are used in isolation without the Mamba2ForCausalLM model class (this has been already fixed: github.com/fla-org/flash-…). 2. Skipping initialization due to meta device init for DTensors with FSDP-2 (github.com/fla-org/flash-… will fix this issue upon merging). The difference is substantial. Mamba-2 seems to be quite sensitive to the initialization. Check out our experiments at the 7B MoE scale: wandb.ai/mayank31398/ma… Special thanks to @kevinyli_, @bharatrunwal2, @HanGuo97, @tri_dao and @_albertgu 🙏 Also thanks to @SonglinYang4 for quickly helping in merging the PR.
English
17
73
747
367.9K
Wentao Guo รีทวีตแล้ว
Woosuk Kwon
Woosuk Kwon@woosuk_k·
Today, we're proud to announce @inferact, a startup founded by creators and core maintainers of @vllm_project, the most popular open-source LLM inference engine. Our mission is to grow vLLM as the world's AI inference engine and accelerate AI progress by making inference cheaper and faster. The Challenge Inference is not solved. It's getting harder. Models grow larger. New architectures proliferate: mixture-of-experts, multimodal, agentic. Every breakthrough demands new infrastructure. Meanwhile, hardware fragments: more accelerators, more programming models, and more combinations to optimize. The capability gap between models and the systems that serve them is widening. Left this way, the most capable models remain bottlenecked and with full scope of their capabilities accessible only to those who can build custom infrastructure. Close the gap, and we unlock new possibilities. And the problem is growing. Inference is shifting from a fraction of compute to the majority: test-time compute, RL training loops, synthetic data. We see a future where serving AI becomes effortless. Today, deploying a frontier model at scale requires a dedicated infrastructure team. Tomorrow, it should be as simple as spinning up a serverless database. The complexity doesn't disappear; it gets absorbed into the infrastructure we're building. Why Us vLLM sits at the intersection of models and hardware: a position that took years to build. When model vendors ship new architectures, they work with us to ensure day-zero support. When hardware vendors develop new silicon, they integrate with vLLM. When teams deploy at scale, they run vLLM, from frontier labs to hyperscalers to startups serving millions of users. Today, vLLM supports 500+ model architectures, runs on 200+ accelerator types, and powers inference at global scale. This ecosystem, built with 2,000+ contributors, is our foundation. We've been stewards of this engine since its first commit. We know it inside out. We deployed it at frontier scale—in research and in production. Open Source vLLM was built in the open. That's not changing. Inferact exists to supercharge vLLM adoption. The optimizations we develop flow back to the community. We plan to push vLLM's performance further, deepen support for emerging model architectures, and expand coverage across frontier hardware. The AI industry needs inference infrastructure that isn't locked behind proprietary walls. Join Us Through the open source community, we are fortunate to work with some of the best people we know. For @inferact, we're hiring engineers and researchers to work at the frontier of inference, where models meet hardware at scale. Come build with us. We're fortunate to be supported by investors who share our vision, including @a16z and @lightspeedvp who led our $150M seed, as well as @sequoia, @AltimeterCap, @Redpoint, @ZhenFund, The House Fund, @strikervp, @LaudeVentures, and @databricks. - @woosuk_k, @simon_mo_, @KaichaoYou, @rogerw0108, @istoica05 and the rest of the founding team
Woosuk Kwon tweet media
English
177
126
1.1K
464.3K
Wentao Guo
Wentao Guo@WentaoGuo7·
@dmsobol @MayankMish98 @XinleC295 @istoica05 @tri_dao Token rounding will always produce per-expert token count as a multiple of tile size, and the maximum deviation for token-choice is 1 tile size (128) per expert. We observe no significant quality degradation as long as average received tokens per expert / tile size >= 2.
English
0
0
2
737
Wentao Guo
Wentao Guo@WentaoGuo7·
@BrendanBurkeX @tri_dao The improvement will be more significant. Blackwell offers more asynchony features where the design of SonicMoE will have a greater win.
English
1
0
2
183
Tri Dao
Tri Dao@tri_dao·
This is what we've been coking for the last 9 months: make MoEs training goes ~2x faster and ~2x less memory! Highlights: - MoE typically takes the most time and memory in modern models. Turns out one can mathematically rewrite the MoE backward pass to reduce the activation mem you need to store in the fwd by ~2x, resulting in the same gradients with no extra matmul recomputation. I really like this result, as it combines both algorithmic and systems insights. - Analyzing bottlenecks in MoE layer leads to a natural optimization stragegy: reduce mem reads/writes as much as possible! Gathering the input for fwd and output grad for bwd can sometimes take as much time as the grouped GEMMs. We fuse gather with grouped GEMM + overlap mem access and compute to make the whole layer goes ~2x faster. - Computing top-k for expert routing can take surprisingly long, ~15-20% of the whole MoE layer! Standard top-k impl uses radix top-k algo, great for large k but suboptimal for small k. We rewrote top-k using bitonic top-k algo, and it's sometimes 20-30x faster than pytorch's top-k! All the main kernels are written in Cute-DSL so they should be easy to extend (and install :D). Hopper kernels are out, Blackwell kernels are just about ready. MoE models used to be 2x less hardware-efficient to train, hopefully Sonic-MOE will change that.
Wentao Guo@WentaoGuo7

🚀SonicMoE🚀: a blazingly-fast MoE implementation optimized for NVIDIA Hopper GPUs. SonicMoE reduces activation memory by 45% and is 1.86x faster on H100 than previous SOTA😃 Paper: arxiv.org/abs/2512.14080 Work with @MayankMish98, @XinleC295, @istoica05, @tri_dao

English
30
167
1.5K
157.2K
Wentao Guo
Wentao Guo@WentaoGuo7·
@MayankMish98 @XinleC295 @istoica05 @tri_dao [5/N] The token rounding routing eliminates padding waste in grouped GEMM and can achieve 16% relative speedup over TC top-K on kernel computation time while delivering robust token-choice accuracy even under highly sparse MoE training regimes.
Wentao Guo tweet media
English
0
2
19
1.8K
Wentao Guo
Wentao Guo@WentaoGuo7·
@MayankMish98 @XinleC295 @istoica05 @tri_dao [4/N] For highly-sparse MoEs, the FLOPs wasted by Grouped GEMM tile-based computation (due to padding) linearly scales w.r.t. expert activation ratio. We introduce a token rounding routing algorithm that rounds the per-expert token count to avoid wasted FLOPs.
English
0
4
22
1.9K
Wentao Guo
Wentao Guo@WentaoGuo7·
@MayankMish98 @XinleC295 @istoica05 @tri_dao [3/N] SonicMoE achieves 50%+ speedup compared to existing baselines for a single MoE layer with open-source MoE configs from 7B to 685B. For 7B MoE model with FSDP-2, SonicMoE on 64 H100s gets 213B tokens/day while ScatterMoE (previous SOTA) on 96 H100s gets 225B tokens/day.
Wentao Guo tweet media
English
0
3
22
2.1K
Wentao Guo
Wentao Guo@WentaoGuo7·
@MayankMish98 @XinleC295 @istoica05 @tri_dao [2/N] The activation memory usage is same as a dense model with equal activated parameters (minimum activation memory required for backward computation without doing activation recomputation with GEMM). SonicMoE’s activation memory usage is also independent of expert granularity.
Wentao Guo tweet media
English
0
3
22
3.5K