Mathurin Massias

377 posts

Mathurin Massias

Mathurin Massias

@mathusmassias

Researcher @INRIA_lyon, Ockham team. Teacher @Polytechnique and @ENSdeLyon Machine Learning, Python and Optimization

Paris, France เข้าร่วม Temmuz 2018
181 กำลังติดตาม2.2K ผู้ติดตาม
ทวีตที่ปักหมุด
Mathurin Massias
Mathurin Massias@mathusmassias·
New paper on the generalization of Flow Matching arxiv.org/abs/2506.03719 🤯 Why does flow matching generalize? Did you know that the flow matching target you're trying to learn **can only generate training points**? with @Qu3ntinB, Anne Gagneux & Rémi Emonet 👇👇👇
GIF
English
19
265
1.5K
181.9K
Mathurin Massias รีทวีตแล้ว
Quentin Berthet
Quentin Berthet@qberthet·
🚨 🔬 PhD positions at Google DeepMind in France 🇫🇷 We are advertising Master Level Intern positions at Google DeepMind within our Frontier AI Unit. These could lead to co-advised PhD positions with Google DeepMind and French academic institutions. job-boards.greenhouse.io/deepmind/jobs/…
English
8
60
587
52.8K
Mathurin Massias
Mathurin Massias@mathusmassias·
With @Qu3ntinB we have one offer for a Marie Sklodowska-Curie postdoctoral fellowships at Inria, to work on generative models : inria.fr/en/marie-sklod… contact me if interested! RT appreciated <3
English
0
8
11
941
Mathurin Massias
Mathurin Massias@mathusmassias·
There is an Associate Professor position in CS at ENS Lyon, with potential integration in my team, starting in sept 2026 : DM me in interested! Details at ens-lyon.fr/LIP/images/Pro…
English
0
2
10
983
Mathurin Massias
Mathurin Massias@mathusmassias·
To understand these phenomena, we study the spatial regularity of the velocity/denoiser over time: we observe a gap between the closed-form and trained model. Applying Jacobian spectral regularization, we recover effects seen previously on perturbed denoisers (drift vs noise).
Mathurin Massias tweet media
English
0
1
11
964
Mathurin Massias
Mathurin Massias@mathusmassias·
Different loss weightings favor different times: which temporal regime drives the generation quality? Controlled perturbations reveal: drift type effects at early times (& good FID) and noise type at late times (& bad FID)
Mathurin Massias tweet media
English
1
1
12
1.1K
Mathurin Massias
Mathurin Massias@mathusmassias·
🌀New paper on the generation phases of Flow Matching arxiv.org/abs/2510.24830 Are FM & diffusion models nothing else than denoisers trained at every noise level? In theory yes, *if trained optimally*. But in practice, do all noise level matter equally?
GIF
English
6
104
661
32.9K
Mathurin Massias
Mathurin Massias@mathusmassias·
Strong afternoon session! Ségolène Martin on how to go from flow matching to denoisers (and hopefully come back?) and Claire Boyer on how learning rate and working in latent spaces affect diffusion model
Mathurin Massias tweet mediaMathurin Massias tweet media
English
0
0
1
244
Mathurin Massias
Mathurin Massias@mathusmassias·
Followed by Scott Pesme on how to use diffusion/flow matching based MMSE to compute a MAP (and nice examples!), and Thibaut Issenhuth on new ways to learn consistency models
Mathurin Massias tweet mediaMathurin Massias tweet media
English
1
0
3
343
Mathurin Massias
Mathurin Massias@mathusmassias·
Kickstarting our workshop on Flow matching and Diffusion with a talk by Eric Vanden Eijnden on how to optimize learning and sampling in Stochastic Interpolants ! Broadcast available at gdr-iasis.cnrs.fr/reunions/model…
Mathurin Massias tweet mediaMathurin Massias tweet media
English
1
2
22
1.3K
Mathurin Massias รีทวีตแล้ว
Quentin Bertrand
Quentin Bertrand@Qu3ntinB·
I am thrilled to announce that our work on the generalization of flow matching has been accepted to NeurIPS as an oral!! See you in San Diego 😎
Mathurin Massias@mathusmassias

New paper on the generalization of Flow Matching arxiv.org/abs/2506.03719 🤯 Why does flow matching generalize? Did you know that the flow matching target you're trying to learn **can only generate training points**? with @Qu3ntinB, Anne Gagneux & Rémi Emonet 👇👇👇

English
5
64
581
41.8K
Mathurin Massias รีทวีตแล้ว
Eugene Ndiaye
Eugene Ndiaye@eugene_ndiaye·
Some 📸🤳from the ongoing #MlssSenegal2025 🙌🏿
Eugene Ndiaye tweet mediaEugene Ndiaye tweet mediaEugene Ndiaye tweet mediaEugene Ndiaye tweet media
English
3
20
57
6.9K
Mathurin Massias
Mathurin Massias@mathusmassias·
@max_tensor @Qu3ntinB You don't even need to train a model to see that the target is, on real data, not stochastic as soon as t is not very small See the animation in second tweet & figures 1b and 2 in arxiv.org/pdf/2506.03719, which do not involve training a model
English
0
0
1
691
Max Tensor
Max Tensor@max_tensor·
@mathusmassias @Qu3ntinB Does CFM really bury stochastic targets or did it just test a tiny U-Net on CIFAR-10 & CelebA-64? Run it on ImageNet-256 with augment ON & non-Gaussian noise before declaring noise “dead.”
English
1
0
1
1.2K
Mathurin Massias
Mathurin Massias@mathusmassias·
New paper on the generalization of Flow Matching arxiv.org/abs/2506.03719 🤯 Why does flow matching generalize? Did you know that the flow matching target you're trying to learn **can only generate training points**? with @Qu3ntinB, Anne Gagneux & Rémi Emonet 👇👇👇
GIF
English
19
265
1.5K
181.9K
Mathurin Massias
Mathurin Massias@mathusmassias·
@danijarh @Qu3ntinB The closed-form minimizer u* is singular at t=1 if x is not a training point, so this reasoning does not apply: following u*, you exactly generate training points (up to a finite set). See e.g. Thm 4.6 in arxiv.org/abs/2410.23594
English
0
1
13
1.4K
Danijar Hafner
Danijar Hafner@danijarh·
Interesting! Since the flow is an ODE (cont. in time and space), paths can't branch or merge. Doesn't that imply that for every two generated datapoints, it's possible to generate one in between. And so the model can't purely memorize but needs to interpolate at least a bit? Curious to understand your intuition better
English
1
0
7
2.1K
Mathurin Massias
Mathurin Massias@mathusmassias·
@DaniloJRezende absolutely! Our point is more 1) that the velocity has a closed-form (already observed by others) 2) that out of several hypotheses explaining why models don't learn it we rule out "target stochasticity" (see second tweet)
English
1
0
3
366
Mathurin Massias
Mathurin Massias@mathusmassias·
@giladturok @Qu3ntinB As you can see in the first gif, following the closed-form minimizer, different x_0 get mapped to the same x_1 (which is one of the training point)
English
0
0
3
1.7K
Gilad
Gilad@giladturok·
@mathusmassias @Qu3ntinB Naive q: In CFM we regress a neural velocity field to a known, true velocity field. The true velocity field at time t is computed with sample x_0 from a base dist and x_1 from a target dist. At inference time, if we sample a diff x_0, shouldn’t we automatically get a diff x_1?
English
1
0
4
2.1K