Timofei Gritsaev

21 posts

Timofei Gritsaev banner
Timofei Gritsaev

Timofei Gritsaev

@gritsaev

Researcher at @bayesgroup | Amortised Sampling & Generative models

Germany Katılım Temmuz 2020
82 Takip Edilen39 Takipçiler
Timofei Gritsaev retweetledi
Nikita Morozov
Nikita Morozov@nvimorozov·
Got asked in a review for my ICML paper whether "there are realistic tasks where one needs to sample from a probability distribution given by its unnormalized density" (rephrased for anonymity). Are we cooked?
English
17
10
350
68.1K
Timofei Gritsaev retweetledi
Vedant Shah
Vedant Shah@veds_12·
LOTs of discourse lately about the correctness of the KL-regularization term used in RLVR fine-tuning of LLMs. Which estimator to use? Whether to add it to the reward or loss? What’s even the difference? 🤔 In our new preprint, we evaluate these choices empirically. 🧵 1/n
Vedant Shah tweet media
English
7
34
124
19.5K
Timofei Gritsaev retweetledi
Daniil Tiapkin
Daniil Tiapkin@dtiapkin·
While frontier labs are announcing their new models, we also want to be part of this parade. So, we’re happy to announce gfnx – a JAX-first library with environments and a single-file baseline implementation for GFlowNet research.
Daniil Tiapkin tweet media
English
3
10
20
1.2K
Nikita Morozov
Nikita Morozov@nvimorozov·
Happy to share that our work on diffusion samplers was accepted as Oral at #NeurIPS2025 FPI Workshop! 🎉 We show how setting both generation and destruction transition kernels as Gaussians with learnable means and variances produces accurate samplers even at very few steps.
Nikita Morozov tweet media
English
4
9
88
5.2K
Timofei Gritsaev
Timofei Gritsaev@gritsaev·
6/ But it’s not limited to small-scale problems. We successfully use these methods to improve outsourced sampling: Bayesian posterior inference in latent spaces of generative models.
Timofei Gritsaev tweet media
English
1
0
1
152
Timofei Gritsaev
Timofei Gritsaev@gritsaev·
1/ Can we efficiently learn the destruction process of diffusion samplers? Can we learn not just the drift, but also the variance for all transition kernels? – We answer YES in our recent paper “Adaptive Destruction Processes for Diffusion Samplers” (Oral at NeurIPS 2025 FPI Workshop).
Timofei Gritsaev tweet media
English
1
8
17
2.6K
Daniil Tiapkin
Daniil Tiapkin@dtiapkin·
The speedrun is over: I defended my PhD this week and became a doctor in applied mathematics (unofficially: in reinforcement learning)! Huge thanks to my supervisors (Eric & Gilles), collaborators, and friends for all the support.
Daniil Tiapkin tweet mediaDaniil Tiapkin tweet media
English
6
2
29
3.1K
Timofei Gritsaev retweetledi
Nikita Morozov
Nikita Morozov@nvimorozov·
(1/n) The usual assumption in GFlowNet environments is acyclicity. Have you ever wondered if it can be relaxed? Does the existing GFlowNet theory translate to the non-acyclic case? Is efficient training possible? We shed new light on these questions in our latest work! @icmlconf
Nikita Morozov tweet media
English
1
6
13
1.9K
Timofei Gritsaev
Timofei Gritsaev@gritsaev·
5/ The results? Faster convergence across many standard GFlowNet approaches—including soft RL-based ones! We also faced the complexity hypothesis: the less structured the environment, the more crucial backward policy optimization becomes.
English
1
0
1
223
Timofei Gritsaev
Timofei Gritsaev@gritsaev·
1/ GFlowNets are known for training a forward policy to generate complex objects step by step. However, an equally important piece specific to the GFlowNet paradigm is a backward policy, which undoes these steps and plays a crucial role in training.
Timofei Gritsaev tweet media
English
1
3
12
1.5K
Pavel Izmailov
Pavel Izmailov@Pavel_Izmailov·
📢 I am recruiting Ph.D. students for my new lab at @nyuniversity! Please apply, if you want to work on understanding deep learning and large models, and do a Ph.D. in the most exciting city on earth. Details on my website: izmailovpavel.github.io. Please spread the word!
Pavel Izmailov tweet mediaPavel Izmailov tweet media
English
30
172
863
199.5K