Sparsity in LLMs Workshop at ICLR 2025

34 posts

Sparsity in LLMs Workshop at ICLR 2025 banner
Sparsity in LLMs Workshop at ICLR 2025

Sparsity in LLMs Workshop at ICLR 2025

@sparseLLMs

Workshop on Sparsity in LLMs: Deep Dive into Mixture of Experts, Quantization, Hardware, and Inference @iclr_conf 2025.

Singapore Joined Aralık 2024
20 Following203 Followers
Sparsity in LLMs Workshop at ICLR 2025 retweeted
Sparsity in LLMs Workshop at ICLR 2025 retweeted
Ayush Noori
Ayush Noori@ayushnoori·
We are presenting “Prefix and output length-aware scheduling for efficient online LLM inference” at the ICLR 2025 (@iclr_conf) Sparsity in LLMs workshop (@sparseLLMs). 🪫 Challenge: LLM inference in data centers benefits from data parallelism. How can we exploit patterns in requests – like shared prefixes and variable decode length – to optimally assign requests to GPU workers? 💡 Idea: both prefix and output length-aware scheduling! We build on Preble (ICML 2025, @vikranth22446, @yiying__zhang), which was the first distributed LLM serving system to exploit prompt sharing (see mlsys.wuklab.io/posts/preble). In our proof-of-concept work, we carefully benchmark Preble vs. prefix-unaware schedulers to identify opportunities for performance improvement. ⭐️ By adding output length-aware scheduling to Preble, we reduce latency by 14.31% at 64 RPS and 28.89% at 128 RPS. ⭐️ 📖 Full paper here: openreview.net/forum?id=DOZiC… Thank you to co-authors @InakiArango, @YepHuang, @rana_shahout, and @minlanyu at @hseas. Thank you also to the Preble authors for their groundbreaking work!
Ayush Noori tweet media
English
0
5
20
1.7K
Sparsity in LLMs Workshop at ICLR 2025 retweeted
Ashwinee Panda
Ashwinee Panda@PandaAshwinee·
our workshop on sparsity in LLMs is starting soon in Hall 4.7! we’re starting strong with an invited talk from @DAlistarh and an exciting oral on scaling laws for MoEs!
Ashwinee Panda tweet media
English
0
1
15
1.1K
Sparsity in LLMs Workshop at ICLR 2025 retweeted
Shiwei Liu
Shiwei Liu@Shiwei_Liu66·
Our ICLR 2025 Workshop on Sparsity in LLMs (@sparseLLMs) kicks off with a talk by @DAlistarh on lossless (~1% perf drop) LLM compression using quantization across various benchmarks.
Shiwei Liu tweet media
English
1
11
86
6.6K
Sparsity in LLMs Workshop at ICLR 2025 retweeted
Ashwinee Panda
Ashwinee Panda@PandaAshwinee·
a PACKED hall for @tydsh‘s talk at our sparsity in LLMs workshop -not surprising! we have another oral right after this, and then we’ll have the first of 2 poster sessions before lunch! @iclr_conf
Ashwinee Panda tweet media
English
0
6
39
2.9K
Sparsity in LLMs Workshop at ICLR 2025 retweeted
Vimal Thilak🦉🐒
Vimal Thilak🦉🐒@AggieInCA·
Check out this post that has information about research from Apple that will be presented at ICLR 2025 in 🇸🇬 this week. I will be at ICLR and will be presenting some of our work (led by @samira_abnar) at SLLM @sparseLLMs workshop. Happy to chat about JEPAs as well!
Peter Gray@peteryugray

New post: "Apple Machine Learning Research at @iclr_conf 2025" - highlighting a selection of the many Apple #ML research papers to be presented at the conference this week: machinelearning.apple.com/research/iclr-…

English
0
5
19
2.5K
Sparsity in LLMs Workshop at ICLR 2025
This is a unique chance for young researchers to connect with senior mentors for guidance on research challenges, collaborations, publishing, and more. 📅 Dedicated mentorship session during the workshop + networking opportunities over coffee & lunch breaks!
English
1
0
2
226
Sparsity in LLMs Workshop at ICLR 2025 retweeted
Dan Alistarh
Dan Alistarh@DAlistarh·
Our QuEST paper was selected for Oral Presentation at ICLR @sparseLLMs workshop! QuEST is the first algorithm with Pareto-optimal LLM training for 4bit weights/activations, and can even train accurate 1-bit LLMs. Paper: arxiv.org/abs/2502.05003 Code: github.com/IST-DASLab/QuE…
English
3
8
31
3.1K
Sparsity in LLMs Workshop at ICLR 2025
🚨 Exciting Mentorship Opportunity at ICLR '25 🚨 We’re organizing a mentorship program during our Sparse-LLM workshop @iclr_conf to connect young researchers with senior mentors! This is a great chance to seek guidance on research challenges, collaborations, publishing and more.
English
1
6
51
7.2K