Shoval Messica

15 posts

Shoval Messica

Shoval Messica

@ShovalMessica

Audio & Speech MSc student at HUJI @cseHUJI

Katılım Mayıs 2024
107 Takip Edilen35 Takipçiler
Shoval Messica retweetledi
Omri Fahn
Omri Fahn@OmriFahn·
MFA is a beautiful visualization, but not only that. It’s a practical tool: competitive localization and better steering, while revealing that concepts live in regions, not just single directions. Grateful I got to contribute to this with an awesome team!
Or Shafran@OrShafran

It's time to look past dictionary learning for decomposing LM activations. What happens when we instead leverage local geometry? We find a natural region-based decomposition that yields better steering and localization 🧵 1/

English
1
1
2
61
Shoval Messica retweetledi
Omri Fahn
Omri Fahn@OmriFahn·
🤔Can an LLM "unconscious" feel a lie bubbling up before it speaks? 🤔And what mysteries hide in its linear space of uncertainty? Thrilled to share our new paper: “Pre-trained LLMs Learn Multiple Types of Uncertainty” (@roicohen9@OmriFahn@GerardDeMelo) arxiv.org/abs/2505.21218
English
2
3
21
222
Shoval Messica retweetledi
Gallil Maimon
Gallil Maimon@GallilMaimon·
🚨New paper on SLM evaluation🚨 We present SALMon🍣 which is a suite of benchmarks for evaluating how much Speech Language Models model acoustic elements like sentiment or background noise. Project: pages.cs.huji.ac.il/adiyoss-lab/sa… 🧵👇🏻
Gallil Maimon tweet media
English
2
21
55
4.3K
Shoval Messica retweetledi
Michael Hassid
Michael Hassid@MichaelHassid·
Which is better, running a 70B model once, or a 7B model 10 times? The answer might be surprising! Presenting our new @COLM_conf paper: "The Larger the Better? Improved LLM Code-Generation via Budget Reallocation" arxiv.org/abs/2404.00725 1/n
Michael Hassid tweet media
English
6
43
208
31.4K
Shoval Messica retweetledi
Mohammad Salama
Mohammad Salama@MohammadSalaama·
I am excited to share my first work: "Dataset Size Recovery from LoRA Weights". Ever wondered if you could find out how many samples was a model trained on using just its weights? Well now you can! Project: vision.huji.ac.il/dsire/ 👇
English
3
27
51
2.9K
Shoval Messica retweetledi
Yossi Adi
Yossi Adi@adiyossLC·
Speech tokenizers are fundamental in building speech LMs. In a recent study we show the common tokenization method (a.k.a “semantic tokens”) is not robust to different signal variations. We then present NAST! a noise aware speech tokenizer w. @ShovalMessica Code and models 👇
Shoval Messica@ShovalMessica

🚨I’m excited to share our #INTERSPEECH2024 paper "NAST: Noise Aware Speech Tokenization for Speech Language Models” 🥳 W/ @adiyossLC Paper-arxiv.org/abs/2406.11037 Code-github.com/ShovalMessica/…

English
0
7
38
2.6K
Shoval Messica retweetledi
Guy Yariv
Guy Yariv@guy_yariv·
1/ Commonsense reasoning needs multimodal knowledge, yet current LLMs focus mostly on text, limiting their integration of crucial visual information. We introduce vLMIG, a method that enhances LLMs' visual commonsense by integrating images into the decision-making process
Guy Yariv tweet media
English
1
23
76
10.2K
Shoval Messica retweetledi
arXiv Sound
arXiv Sound@ArxivSound·
``Joint Audio and Symbolic Conditioning for Temporally Controlled Text-to-Music Generation,'' Or Tal, Alon Ziv, Itai Gat, Felix Kreuk, Yossi Adi, ift.tt/tP4Hua8
English
1
5
17
1.3K
Shoval Messica retweetledi
arXiv Sound
arXiv Sound@ArxivSound·
``NAST: Noise Aware Speech Tokenization for Speech Language Models,'' Shoval Messica, Yossi Adi, ift.tt/TwrHMcg
English
0
5
23
1.5K