
Shradha Sehgal
429 posts

Shradha Sehgal
@shradhasgl
ML Research @Netflix | Prev @Google, @CarnegieMellon | Grad: UIUC, IIIT-Hyderabad | @SiebelScholars’24







🚀Very excited about my new paper! NN-CIFT slashes data valuation costs by 99% using tiny neural nets (205k params, just 0.0027% of 8B LLMs) while maintaining top-tier performance!



🔬Interested in training AlphaFold3 faster, at scale, and beyond NVIDIA GPU? Now you can. AlphaFold3 is a major leap in biomolecular modeling, but behind the scenes, it introduces severe system bottlenecks: 🧠 2D EvoAttention spikes memory usage 📉 Retrieval-augmented training pipeline causes long GPU idle time ⛔ Frequent but memory-intensive ops slow everything down Today, I'm excited to announce MegaFold, a fully open-source system to make AlphaFold3 training fast, scalable, and cross-platform on both NVIDIA and AMD GPUs. MegaFold delivers: ⚡ Up to 1.73x / 1.62x faster training on NVIDIA H100 / AMD MI250 🧬 Up to 1.35× longer sequences compared to PyTorch baseline Key features: 🚀 Memory-Efficient EvoAttention via portable Triton kernels 💡 Ahead-of-Time Caching to eliminate GPU idle time in retrieval pipelines 🔗 DeepFusion for reducing overhead of small but frequent memory-intensive AF3 ops 📘 Project page: supercomputing-system-ai-lab.github.io/projects/megaf… 📄 Paper: arxiv.org/pdf/2506.20686 💻 Code: github.com/Supercomputing… 🤝 MegaFold is developed in collaboration between UIUC SSAIL Lab and researchers from University of Missouri and Lawrence Berkeley National Laboratory. Kudos to the brilliant team: Hoa La, Ahan Gupta, Alex Morehead, Jianlin Cheng #AlphaFold3 #AI #ProteinFolding #Bioinformatics #AMD #Triton #CrossPlatform #OpenSource


Confused about recent LLM RL results where models improve without any ground-truth signal? We were too. Until we looked at the reported numbers of the Pre-RL models and realized they were serverely underreported across papers. We compiled discrepancies in a blog below🧵👇


Would models know more about Indian food in Hindi and Turkey’s history in Turkish? Does the language of a question affect an LLM’s answer? ✨Yes!✨ @nbbozdag and I are excited to announce our newest preprint in which we explore “Language Specific Knowledge (LSK)”.







