Mathieu Dagréou

378 posts

Mathieu Dagréou

Mathieu Dagréou

@Mat_Dag

Ph.D. student in at @Inria_Saclay working on Optimization and Machine Learning @matdag.bsky.social

Paris, France Katılım Temmuz 2019
547 Takip Edilen505 Takipçiler
Mathieu Dagréou retweetledi
Konstantin Mishchenko
Konstantin Mishchenko@konstmish·
Nesterov dropped a new paper last week on what functions can be optimized with gradient descent. The idea is simple: we know GD can optimize both nonsmooth (bounded grads) and smooth (Lipschitz grads) functions, but smooth+nonsmooth satisfies neither property, so what can we do?
Konstantin Mishchenko tweet media
English
11
55
467
30.7K
Mathieu Dagréou retweetledi
Rudy Morel
Rudy Morel@rdMorel·
For evolving unknown PDEs, ML models are trained on next-state prediction. But do they actually learn the time dynamics: the "physics"? Check out our poster (W-107) at #ICML2025 this Wed, Jul 16. Our "DISCO" model learns the physics while staying SOTA on next states prediction!
Rudy Morel tweet media
English
5
50
303
21.1K
Mathieu Dagréou retweetledi
Mathieu Blondel
Mathieu Blondel@mblondel_ml·
Back from MLSS Senegal 🇸🇳, where I had the honor of giving lectures on differentiable programming. Really grateful for all the amazing people I got to meet 🙏 My slides are here github.com/diffprog/slide…
English
4
19
67
5.5K
Mathieu Dagréou retweetledi
Waïss Azizian
Waïss Azizian@wazizian·
❓ How long does SGD take to reach the global minimum on non-convex functions? With @FranckIutzeler, J. Malick, P. Mertikopoulos, we tackle this fundamental question in our new ICML 2025 paper: "The Global Convergence Time of Stochastic Gradient Descent in Non-Convex Landscapes"
English
8
65
492
34.5K
Mathieu Dagréou retweetledi
Konstantin Mishchenko
Konstantin Mishchenko@konstmish·
I want to address one very common misconception about optimization. I often hear that (approximately) preconditioning with the Hessian diagonal is always a good thing. It's not. In fact, finding a good preconditioner is an open problem, which I think deserves more attention. 1/4
English
5
15
205
20.2K
Mathieu Dagréou retweetledi
Matthieu Terris
Matthieu Terris@MatthieuTerris·
🧵 I'll be at CVPR next week presenting our FiRe work 🔥 TL;DR: We go beyond denoising models in PnP with more general restoration (e.g. deblurring) models! A starting point observation is that images are not fixed-points of restoration models:
GIF
English
1
5
16
2.2K
Mathieu Dagréou retweetledi
Samuel Vaiter
Samuel Vaiter@vaiter·
📣 New preprint 📣 **Differentiable Generalized Sliced Wasserstein Plans** w/ L. Chapel @rtavenar We propose a Generalized Sliced Wasserstein method that provides an approximated transport plan and which admits a differentiable approximation. arxiv.org/abs/2505.22049 1/5
Samuel Vaiter tweet media
English
1
4
31
2.7K
Mathieu Dagréou retweetledi
Mathurin Massias
Mathurin Massias@mathusmassias·
It was received quite enthusiastically here so time to share it again!!! Our #ICLR2025 blog post on Flow M atching was published yesterday : iclr-blogposts.github.io/2025/blog/cond… My PhD student Anne Gagneux will present it tomorrow in ICLR, 👉poster session 4, 3 pm, #549 in Hall 3/2B 👈
English
1
5
11
837
Mathieu Dagréou retweetledi
Gabriel Peyré
Gabriel Peyré@gabrielpeyre·
Optimization algorithms come with many flavors depending on the structure of the problem. Smooth vs non-smooth, convex vs non-convex, stochastic vs deterministic, etc. en.wikipedia.org/wiki/Mathemati…
English
4
107
515
20.7K
Mathieu Dagréou retweetledi
Alex Hägele
Alex Hägele@haeggee·
A really fun project to work on. Looking at these plots side-by-side still amazes me! How well can **convex optimization theory** match actual LLM runs? My favorite points of our paper on the agreement for LR schedules in theory and practice: 1/n
Alex Hägele tweet media
Fabian Schaipp@FSchaipp

Learning rate schedules seem mysterious? Turns out that their behaviour can be described with a bound from *convex, nonsmooth* optimization. Short thread on our latest paper 🚇 arxiv.org/abs/2501.18965

English
1
5
44
4.9K
Mathieu Dagréou retweetledi
Mathieu Dagréou retweetledi
Konstantin Mishchenko
Konstantin Mishchenko@konstmish·
Learning rate schedulers used to be a big mistery. Now you can just take a guarantee for *convex non-smooth* problems (from arxiv.org/abs/2310.07831), and they give you *precisely* what you see in training large models. See this empirical study: arxiv.org/abs/2501.18965 1/3
Konstantin Mishchenko tweet media
English
5
71
435
28.7K
Mathieu Dagréou retweetledi
Théo Uscidda
Théo Uscidda@theo_uscidda·
Our work on geometric disentangled representation learning has been accepted to ICLR 2025! 🎊See you in Singapore if you want to understand this gif better :)
Théo Uscidda@theo_uscidda

Curious about the potential of optimal transport (OT) in representation learning? Join @CuturiMarco's talk at the UniReps workshop today at 2:30 PM! Marco will notably discuss our latest paper on using OT to learn disentangled representations. Details below ⬇️

English
0
18
152
14.3K
Mathieu Dagréou retweetledi
Gabriel Peyré
Gabriel Peyré@gabrielpeyre·
The Mathematics of Artificial Intelligence: In this introductory and highly subjective survey, aimed at a general mathematical audience, I showcase some key theoretical concepts underlying recent advancements in machine learning. arxiv.org/abs/2501.10465
Gabriel Peyré tweet media
English
28
410
2K
180.6K
Mathieu Dagréou retweetledi
Samuel Vaiter
Samuel Vaiter@vaiter·
When optimization problems have multiple minima, algorithms favor specific solutions due to their implicit bias. For ordinary least squares (OLS), gradient descent inherently converges to the minimal norm solution among all possible solutions. fa.bianp.net/blog/2022/impl…
English
6
48
345
20.4K
Mathieu Dagréou retweetledi
Pierre Ablin
Pierre Ablin@PierreAblin·
🍏🍏🍏 Come work with us at Apple Machine Learning Research! 🍏🍏🍏 Our team focuses on curiosity-based, open research. We work on several topics, including LLMs, optimization, optimal transport, uncertainty quantification, and generative modeling. Infos 👇
English
3
37
378
47.9K