Alexander Rubinstein

41 posts

Alexander Rubinstein

Alexander Rubinstein

@a_rubique

PhD Student @ IMPRS-IS + University of Tübingen

Katılım Ekim 2022
164 Takip Edilen88 Takipçiler
Sabitlenmiş Tweet
Alexander Rubinstein
Alexander Rubinstein@a_rubique·
🪩 Evaluate your LLMs on benchmarks like MMLU at 1% cost. In our new paper, we show that outputs on a small subset of test samples that maximise diversity in model responses are predictive of the full dataset performance. Project page: arubique.github.io/disco-site/ More below 🧵👇
Alexander Rubinstein tweet media
English
1
4
11
1.8K
Alexander Rubinstein retweetledi
Alexander Panfilov
Alexander Panfilov@kotekjedi_ml·
New paper: We deploy Claude Code in an autoresearch loop to discover novel jailbreaking algorithms – and it works. It beats 30+ existing GCG-like attacks (with AutoML hyperparameter tuning) This is a strong sign that incremental safety and security research can now be automated.
Alexander Panfilov tweet media
English
47
212
1.6K
296.1K
Alexander Rubinstein retweetledi
Cornelius Emde
Cornelius Emde@CorEmde·
1/ Evaluating a single agent harness is hard. Evaluating a multi-agent system? That's a whole different problem. Most eval tools treat the model as the unit of analysis. But in multi-agent systems, the system is what matters. That's why we built MASEval 🧵 #Agents #AI #Eval
Cornelius Emde tweet media
English
3
1
7
540
Alexander Rubinstein retweetledi
Arnas Uselis
Arnas Uselis@a_uselis·
How do embedding spaces of models that generalize from limited data look? We study what structure such models should exhibit. Turns out: linear and orthogonal. And modern embedding models like CLIP and SigLIP already show signs of it! 🧵 (1/n)
English
4
101
708
75.7K
Alexander Rubinstein retweetledi
Damien Teney
Damien Teney@DamienTeney·
🔥What if web text isn’t the best place to start training LLMs? Our latest work shows that warming up models on procedural data (e.g. from formal languages & simple algorithms) speeds up subsequent pretraining on language, code, and math, on models up to 1.3B parameters⬇️🧵
Damien Teney tweet media
English
4
16
77
5.6K
Alexander Rubinstein
Alexander Rubinstein@a_rubique·
Happy to share that our new paper was accepted to ICLR 2026! This paper helps people who spend too much time waiting for LLM evaluations on benchmarks like MMLU-Pro. We show how to reduce this time by up to 100×. Big thanks to the coauthors Benjamin Raible, @framart1 @coallaoh!
Alexander Rubinstein@a_rubique

🪩 Evaluate your LLMs on benchmarks like MMLU at 1% cost. In our new paper, we show that outputs on a small subset of test samples that maximise diversity in model responses are predictive of the full dataset performance. Project page: arubique.github.io/disco-site/ More below 🧵👇

English
0
2
10
331
Alexander Rubinstein retweetledi
Vishaal Udandarao
Vishaal Udandarao@vishaal_urao·
🚀 New paper! arxiv.org/abs/2511.16655 Recently, Cambrian-S released models & two benchmarks (VSR & VSC) for “spatial supersensing” in video! We found: 1️⃣ Simple no-frame baseline (NoSense) ~perfectly solves VSR! 2️⃣ Tiny sanity check collapses Cambrian-S perf to 0% on VSC! 🧵👇
Vishaal Udandarao tweet media
English
5
22
122
40.1K
Alexander Rubinstein retweetledi
Adhiraj Ghosh
Adhiraj Ghosh@adhiraj_ghosh98·
🚨Current data curation results in the creation of static datasets and the use of model-based filters that induce many biases. Can we fix this? We propose ✨CABS✨, a flexible concept-aware online batch curation method that improves CLIP pretraining! arxiv.org/abs/2511.20643 🧵👇
Adhiraj Ghosh tweet media
English
2
7
25
6.6K
Alexander Rubinstein
Alexander Rubinstein@a_rubique·
DISCO achieves a superior efficiency–precision trade-off across various compression rates compared to baselines. 🧵5/6
Alexander Rubinstein tweet media
English
1
0
2
60
Alexander Rubinstein
Alexander Rubinstein@a_rubique·
🪩 Evaluate your LLMs on benchmarks like MMLU at 1% cost. In our new paper, we show that outputs on a small subset of test samples that maximise diversity in model responses are predictive of the full dataset performance. Project page: arubique.github.io/disco-site/ More below 🧵👇
Alexander Rubinstein tweet media
English
1
4
11
1.8K
Alexander Rubinstein retweetledi
Alexander Panfilov
Alexander Panfilov@kotekjedi_ml·
Stronger models need stronger attackers! 🤖⚔️ In our new paper we explore how attacker-target capability dynamics affect red-teaming success (ASR). Key insights: 🔸Stronger models = better attackers 🔸ASR depends on capability gap 🔸Psychology >> STEM for ASR More in 🧵👇
Alexander Panfilov tweet media
English
5
9
59
10.2K
Alexander Rubinstein
Alexander Rubinstein@a_rubique·
Glad to be at #ICLR2025! Ankit Sonthalia and I are presenting our paper on star-shapeness of deep learning solution sets today at poster session 5 (poster #132)! Feel free to drop by and talk about the structure of the models' weight space, generalization and uncertainty!
English
3
1
6
439