Maximilian Müller

91 posts

Maximilian Müller

Maximilian Müller

@mueller_mp

ELLIS PhD student at the University of Tübingen.

Katılım Şubat 2020
280 Takip Edilen235 Takipçiler
Sabitlenmiş Tweet
Maximilian Müller
Maximilian Müller@mueller_mp·
‼️ New paper: Normalization Layers Are All That Sharpness-Aware Minimization Needs ‼️ arxiv.org/abs/2306.04226 We show that applying SAM only to the normalization layers of a network (SAM-ON) enhances performance compared to applying it to the full network.
Maximilian Müller tweet media
English
1
19
91
25.2K
Maximilian Müller retweetledi
Christian Schlarmann
Christian Schlarmann@chs20_·
New paper: We introduce Visual Memory Injection, a new attack on large vision-language models. A subtly perturbed image, that remains in the chat context, causes the model to behave normally for many turns and later triggers a targeted harmful response on a topic-specific prompt.
Christian Schlarmann tweet media
English
1
2
7
120
Maximilian Müller retweetledi
Maksym Andriushchenko
Maksym Andriushchenko@maksym_andr·
📣 We are expanding our AI Safety and Alignment group at @ELLISInst_Tue and @MPI_IS! We have: - a great cluster at MPI with 50+ GB200s, 250+ H100s, and many-many A100 80GBs, - outstanding colleagues (@jonasgeiping, @sahar_abdelnabi, etc), - competitive salaries (as for academia), - fully English-speaking environment. In particular, I'm looking for: - one postdoc with a proven track record in AI safety, - PhD students with a strong computer science background and ideally experience in cybersecurity, interpretability, or training dynamics, - master’s thesis students (if you are already in Tübingen or can relocate to Tübingen for ~6 months), - remote mentees for the Summer 2026 MATS cohort (apply directly via the MATS portal). I'll be at NeurIPS in San Diego and would be glad to chat about these positions!
English
4
13
145
14.2K
francesco croce
francesco croce@fra__31·
Happy to share that I've started as an assistant professor at @AaltoUniversity and ELLIS Institute Finland! I'll recruit students via the ELLIS PhD Program ellis.eu/research/phd-p… to work on multimodal learning, robustness, visual reasoning... feel free to reach out!
francesco croce tweet media
English
4
5
28
3.8K
Maximilian Müller retweetledi
Maksym Andriushchenko
Maksym Andriushchenko@maksym_andr·
Very promising results on *robust* unlearning from colleagues at Tübingen and EPFL. (+ some general improvements to the standard evaluation by using an LLM judge and worst-case evaluation over paraphrases and input formats)
Maksym Andriushchenko tweet media
English
5
10
125
9.7K
Maximilian Müller
Maximilian Müller@mueller_mp·
This is a great opportunity for anyone who wants to work on AI safety. Congrats and all the best, Maksym!
Maksym Andriushchenko@maksym_andr

🚨 Incredibly excited to share that I'm starting my research group focusing on AI safety and alignment at the ELLIS Institute Tübingen and Max Planck Institute for Intelligent Systems in September 2025! 🚨 Hiring. I'm looking for multiple PhD students: both those able to start in Fall 2025 (i.e., as soon as possible) and through centralized programs like CLS, IMPRS, and ELLIS (the deadlines are in November) to start in Spring–Fall 2026. I'm also searching for postdocs, master's thesis students, and research interns. Fill the Google form below if you're interested! Research group. We will focus on developing algorithmic solutions to reduce harms from advanced general-purpose AI models. We're particularly interested in alignment of autonomous LLM agents, which are becoming increasingly capable and pose a variety of emerging risks. We're also interested in rigorous AI evaluations and informing the public about the risks and capabilities of frontier AI models. Additionally, we aim to advance our understanding of how AI models generalize, which is crucial for ensuring their steerability and reducing associated risks. For more information about research topics relevant to our group, please check the following documents: - International AI Safety Report, - An Approach to Technical AGI Safety and Security by DeepMind, - Open Philanthropy’s 2025 RFP for Technical AI Safety Research. Research style. We are not necessarily interested in getting X papers accepted at NeurIPS/ICML/ICLR. We are interested in making an impact: this can be papers (and NeurIPS/ICML/ICLR are great venues), but also open-source repositories, benchmarks, blog posts, even social media posts—literally anything that can be genuinely useful for other researchers and the general public. Broader vision. Current machine learning methods are fundamentally different from what they used to be pre-2022. The Bitter Lesson summarized and predicted this shift very well back in 2019: "general methods that leverage computation are ultimately the most effective". Taking this into account, we are only interested in studying methods that are general and scale with intelligence and compute. Everything that helps to advance their safety and alignment with societal values is relevant to us. We believe getting this—some may call it "AGI"—right is one of the most important challenges of our time. Join us on this journey!

English
0
0
4
256
Maximilian Müller retweetledi
Maksym Andriushchenko
Maksym Andriushchenko@maksym_andr·
🚨 Incredibly excited to share that I'm starting my research group focusing on AI safety and alignment at the ELLIS Institute Tübingen and Max Planck Institute for Intelligent Systems in September 2025! 🚨 Hiring. I'm looking for multiple PhD students: both those able to start in Fall 2025 (i.e., as soon as possible) and through centralized programs like CLS, IMPRS, and ELLIS (the deadlines are in November) to start in Spring–Fall 2026. I'm also searching for postdocs, master's thesis students, and research interns. Fill the Google form below if you're interested! Research group. We will focus on developing algorithmic solutions to reduce harms from advanced general-purpose AI models. We're particularly interested in alignment of autonomous LLM agents, which are becoming increasingly capable and pose a variety of emerging risks. We're also interested in rigorous AI evaluations and informing the public about the risks and capabilities of frontier AI models. Additionally, we aim to advance our understanding of how AI models generalize, which is crucial for ensuring their steerability and reducing associated risks. For more information about research topics relevant to our group, please check the following documents: - International AI Safety Report, - An Approach to Technical AGI Safety and Security by DeepMind, - Open Philanthropy’s 2025 RFP for Technical AI Safety Research. Research style. We are not necessarily interested in getting X papers accepted at NeurIPS/ICML/ICLR. We are interested in making an impact: this can be papers (and NeurIPS/ICML/ICLR are great venues), but also open-source repositories, benchmarks, blog posts, even social media posts—literally anything that can be genuinely useful for other researchers and the general public. Broader vision. Current machine learning methods are fundamentally different from what they used to be pre-2022. The Bitter Lesson summarized and predicted this shift very well back in 2019: "general methods that leverage computation are ultimately the most effective". Taking this into account, we are only interested in studying methods that are general and scale with intelligence and compute. Everything that helps to advance their safety and alignment with societal values is relevant to us. We believe getting this—some may call it "AGI"—right is one of the most important challenges of our time. Join us on this journey!
Maksym Andriushchenko tweet media
English
76
88
843
105.5K
Maximilian Müller retweetledi
Christian Schlarmann
Christian Schlarmann@chs20_·
Excited to announce FuseLIP: an embedding model that encodes image+text into a single vector. We achieve this by tokenizing images into discrete tokens, merging these with the text tokens and subsequently processing them with a single transformer.
Christian Schlarmann tweet media
English
1
4
12
1.3K
Maximilian Müller retweetledi
Václav Voráček
Václav Voráček@VaclavVoracekCZ·
With @bremen79, We propose a new algorithm for constructing confidence intervals for means of bounded r.vs using "testing by betting" framework. It performs remarkably well even in the challenging, very small sample regime. (and of course, it is great in the large sample one)
English
1
4
46
5K
Maximilian Müller
Maximilian Müller@mueller_mp·
This connects several points from our previous work: We found that the Mahalanobis score yields SOTA results for some models on our NINCO benchmark, but fails for others. In a workshop paper, we found that the Mahalanobis score is very sensitive to training hyperparams (esp. lr).
English
1
0
0
56
Maximilian Müller
Maximilian Müller@mueller_mp·
Mahalanobis++: Improving OOD Detection via Feature Normalization Our latest work has been accepted to ICML and is now also on arXiv! We explain why Mahalanobis-based OOD detection led to varied results and show that l2 normalization improves its performance consistently.
Maximilian Müller tweet media
English
1
3
14
541
Maximilian Müller retweetledi
Christian Schlarmann
Christian Schlarmann@chs20_·
📢 Robustness is not always at odds with accuracy! We show that adversarially robust vision encoders improve clean and robust accuracy over their base models in perceptual similarity tasks. Looking forward to presenting at SaTML @satml_conf in Copenhagen next week 🇩🇰
Christian Schlarmann tweet mediaChristian Schlarmann tweet media
English
1
4
22
1.2K
Maximilian Müller retweetledi
Maximilian Beck
Maximilian Beck@maxmbeck·
📢🔔I am excited to share the details on our optimized xLSTM architecture for our xLSTM 7B model!🚨 We optimized the architecture with two goals in mind: - Efficiency (in Training and Inference) and - Stability 🧵(1/7)
Maximilian Beck tweet media
English
8
60
327
44.9K
Maximilian Müller retweetledi
Christian Schlarmann
Christian Schlarmann@chs20_·
📢 Check out our new report: we show that a recently proposed defense against adversarial attacks is not robust. We circumvent gradient masking issues of the proposed model by attacking a slightly adapted surrogate model and then transferring the perturbations.
English
1
2
8
333