Che-Ping Tsai
79 posts

Che-Ping Tsai
@chepingt
PhD @mldcmu, interpretability and representation learning, machine learning theories.

The newest model in the Mamba series is finally here 🐍 Hybrid models have become increasingly popular, raising the importance of designing the next generation of linear models. We've introduced several SSM-centric ideas to significantly increase Mamba-2's modeling capabilities without compromising on speed. The resulting Mamba-3 model has noticeable performance gains over the most popular previous linear models (such as Mamba-2 and Gated DeltaNet) at all sizes. This is the first Mamba that was student led: all credit to @aakash_lahoti @kevinyli_ @_berlinchen @caitWW9, and of course @tri_dao!











📢I’m on the academic job market!📢 I mainly work on representation learning and causality (CRL, identifiable SSL, causal discovery, robotics applications, rep. learning for tabular data) Also, I’ll be at #NeurIPS Dec. 2–7, reach out to chat about any of the above :)

LeJEPA: a novel pretraining paradigm free of the (many) heuristics we relied on (stop-grad, teacher, ...) - 60+ arch., up to 2B params - 10+ datasets - in-domain training (>DINOv3) - corr(train loss, test perf)=95% Paper: arxiv.org/pdf/2511.08544 Code: github.com/rbalestr-lab/l…

















