Nikita Morozov

56 posts

Nikita Morozov banner
Nikita Morozov

Nikita Morozov

@nvimorozov

PhD student at @CS_HSE, researcher at @bayesgroup Previously intern at @EPFL, @yandex ICPC World Finalist Generative models, Sampling, RL, AI4Science

Katılım Şubat 2025
592 Takip Edilen174 Takipçiler
Sabitlenmiş Tweet
Nikita Morozov
Nikita Morozov@nvimorozov·
Thrilled to announce that we've released gfnx! github.com/d-tiapkin/gfnx We provide fast and scalable implementations of GFlowNet environments and algorithms in JAX, achieving up to 80 times runtime speedups in comparison to previous PyTorch-based implementations.
English
2
4
20
1.7K
Nikita Morozov retweetledi
Weight Space Symmetries @ ICML 2026
📢Excited to announce the Workshop on Weight-Space Symmetries @icmlconf! We welcome 4-page submissions analysing symmetries, their effects on training and model structure, and practical methods to utilize them. Submission Deadline: April 24 (23:59 AoE) #ICML2026
Weight Space Symmetries @ ICML 2026 tweet media
English
1
31
45
14K
Nikita Morozov
Nikita Morozov@nvimorozov·
Got asked in a review for my ICML paper whether "there are realistic tasks where one needs to sample from a probability distribution given by its unnormalized density" (rephrased for anonymity). Are we cooked?
English
17
10
349
68.1K
Nikita Morozov retweetledi
ICML Conference
ICML Conference@icmlconf·
To ensure compliance w peer-review policies, ICML has removed 795 reviews (1% of total) by reviewers who used LLMs when they explicitly agreed to not. Consequently, 497 papers (2% of all submissions) of these (reciprocal) reviewers have been desk rejected Details in blog post 👇
ICML Conference tweet media
English
22
81
605
224K
Nikita Morozov retweetledi
Mishan Aliev
Mishan Aliev@ne_mishan·
Excited to present CasTex (#WACV2026) 🎉 Our text-to-texture method optimizes explicit PBR maps via SDS on cascaded pixel-space diffusion models, avoiding latent artifacts and producing relightable textures ready for production use. Paper & Code ⬇️ thecrazymage.github.io/CasTex
English
0
2
6
706
Nikita Morozov retweetledi
Alex Tong
Alex Tong@AlexanderTong7·
The missing toolkit for discrete diffusion research. ⚡️ UNI-D² unifies the SOTA baselines into one codebase, making it easier than ever to iterate on non-autoregressive models. Great work co-led by @nkalyanv99 and @vincentpaulinef!
Kalyan@nkalyanv99

We’re releasing UNI-D², a unified codebase for discrete diffusion language models 🤝🚀 Co-led with @vincentpaulinef and an amazing advisor team: @stefanAbauer, @AlexanderTong7 , @andrea_dittadi, @AMK6610, @KaplFer 🙌 🔗 GitHub: github.com/nkalyanv99/UNI… 📚 Docs: nkalyanv99.github.io/UNI-D2/ Reproduce and extend state-of-the-art baselines with one toolkit. Let’s move beyond autoregressive models and push discrete diffusion together 🧵👇

English
3
16
103
11.8K
Nikita Morozov
Nikita Morozov@nvimorozov·
@bayesianboy Well, beer killed harmful bacteria in contaminated water, leading to the main source of safe drinks in ancient times. I guess there's no arguing in here.
English
0
0
1
54
Nikita Morozov retweetledi
Daniil Tiapkin
Daniil Tiapkin@dtiapkin·
While frontier labs are announcing their new models, we also want to be part of this parade. So, we’re happy to announce gfnx – a JAX-first library with environments and a single-file baseline implementation for GFlowNet research.
Daniil Tiapkin tweet media
English
3
10
20
1.2K
Nikita Morozov retweetledi
Chieh-Hsin (Jesse) Lai
Chieh-Hsin (Jesse) Lai@JCJesseLai·
Tired to go back to the original papers again and again? Our monograph: a systematic and fundamental recipe you can rely on! 📘 We’re excited to release 《The Principles of Diffusion Models》— with @DrYangSong, @gimdong58085414, @mittu1204, and @StefanoErmon. It traces the core ideas that shaped diffusion modeling and explains how today’s models work, why they work, and where they’re heading. 🧵You’ll find the link and a few highlights in the thread. We’d love to hear your thoughts and join some discussions! ⚡ Stay tuned for our markdown version, where you can drop your comments!
Chieh-Hsin (Jesse) Lai tweet media
English
53
484
2.4K
820.7K
Nikita Morozov
Nikita Morozov@nvimorozov·
Feels really fulfilling when the conference acknowledges the effort you put into reviewing! Honored to be recognized as a top reviewer at #NeurIPS2025
Nikita Morozov tweet media
English
3
0
12
549
Nikita Morozov
Nikita Morozov@nvimorozov·
Happy to share that our work on diffusion samplers was accepted as Oral at #NeurIPS2025 FPI Workshop! 🎉 We show how setting both generation and destruction transition kernels as Gaussians with learnable means and variances produces accurate samplers even at very few steps.
Nikita Morozov tweet media
English
4
9
88
5.2K
Nikita Morozov retweetledi
Timofei Gritsaev
Timofei Gritsaev@gritsaev·
1/ Can we efficiently learn the destruction process of diffusion samplers? Can we learn not just the drift, but also the variance for all transition kernels? – We answer YES in our recent paper “Adaptive Destruction Processes for Diffusion Samplers” (Oral at NeurIPS 2025 FPI Workshop).
Timofei Gritsaev tweet media
English
1
8
17
2.6K
Nikita Morozov retweetledi
Sophia Tang
Sophia Tang@_sophia_tang_·
Super happy to share our new work on “Tree Search Guided Trajectory-Aware Fine-Tuning for Discrete Diffusion” or TR2-D2! 🤖🌳 Inspired by the incredible success of off-policy reinforcement learning (RL), TR2-D2 introduces a general framework that combines off-policy RL with tree search for single- and multi-objective fine-tuning of discrete diffusion models. 📄 Preprint: arxiv.org/abs/2509.25171 💻 Github: github.com/sophtang/TR2-D2 🤗 HuggingFace: huggingface.co/ChatterjeeLab/… Details in thread 👇🏻 (1/n)
GIF
English
3
16
94
7K
Nikita Morozov retweetledi
Dimitris Papailiopoulos
Dimitris Papailiopoulos@DimitrisPapail·
Prediction: In ~3 years academia will be the most desirable place to do fundamental AI research Contributing factors: - small models improve/become significantly more impactful - open weights community broadens its reach - gpus continue to get faster & cheaper - meaningful post-training/RL experiments become more and more tractable - raw capabilities of large models plateau (100% acc is actually a wall) => "foundation models" become commodity => product matters more there will obviously be incredibly important problems at the frontier of a gazillion parameters, of models launching 100k agents, and training incredibly complex systems with one million gpus. But there will be so many more and incredibly important problems at the hands of a community that is free to ask any questions they like, and benefits directly from sharing with everyone else.
English
24
32
468
58.1K
Nikita Morozov retweetledi
hardmaru
hardmaru@hardmaru·
Proud to release ShinkaEvolve, our open-source framework that evolves programs for scientific discovery with very good sample-efficiency! 🐙 Paper: arxiv.org/abs/2509.19349 Blog: sakana.ai/shinka-evolve/ Project: github.com/SakanaAI/Shink…
Sakana AI@SakanaAILabs

We’re excited to introduce ShinkaEvolve: An open-source framework that evolves programs for scientific discovery with unprecedented sample-efficiency. Blog: sakana.ai/shinka-evolve/ Code: github.com/SakanaAI/Shink… Like AlphaEvolve and its variants, our framework leverages LLMs to find state-of-the-art solutions to complex problems, but using orders of magnitude fewer resources! Many evolutionary AI systems are powerful but act like brute-force engines, burning thousands of samples to find good solutions. This makes discovery slow and expensive. We took inspiration from the efficiency of nature. ‘Shinka’ (進化) is Japanese for evolution, and we designed our system to be just as resourceful. On the classic circle packing optimization problem, ShinkaEvolve discovered a new state-of-the-art solution using only 150 samples. This is a big leap in efficiency compared to previous methods that required thousands of evaluations. We applied ShinkaEvolve to a diverse set of hard problems with real-world applications: 1/ AIME Math Reasoning: It evolved sophisticated agentic scaffolds that significantly outperform strong baselines, discovering an entire Pareto frontier of solutions trading performance for efficiency. 2/ Competitive Programming: On ALE-Bench (a benchmark for NP-Hard optimization problems), ShinkaEvolve took the best existing agent's solutions and improved them, turning a 5th place solution on one task into a 2nd place leaderboard rank in a competitive programming competition. 3/ LLM Training: We even turned ShinkaEvolve inward to improve LLMs themselves. It tackled the open challenge of designing load balancing losses for Mixture-of-Experts (MoE) models. It discovered a novel loss function that leads to better expert specialization and consistently improves model performance and perplexity. ShinkaEvolve achieves its remarkable sample-efficiency through three key innovations that work together: (1) an adaptive parent sampling strategy to balance exploration and exploitation, (2) novelty-based rejection filtering to avoid redundant work, and (3) a bandit-based LLM ensemble that dynamically picks the best model for the job. By making ShinkaEvolve open-source and highly sample-efficient, our goal is to democratize access to advanced, open-ended discovery tools. Our vision for ShinkaEvolve is to be an easy-to-use companion tool to help scientists and engineers with their daily work. We believe that building more efficient, nature-inspired systems is key to unlocking the future of AI-driven scientific research. We are excited to see what the community builds with it! Learn more in our technical report: arxiv.org/abs/2509.19349

English
6
62
364
59.6K
Nikita Morozov retweetledi
Kaiyan Zhang
Kaiyan Zhang@OkhayIea·
🚀 Excited to share our new survey paper on RL for Large Reasoning Models (LRMs)! Since early this year, our team has released several RL+LLMs works (PRIME, TTRL, SimpleVLA, MARTI, SSRL, HPT), covering dense rewards, self-evolution, embodied AI, multi-agent, tool learning, and hybrid post-training. The field is growing rapidly—new papers & projects are popping up every day! It felt like the right time to systematically review the landscape and reflect on the path towards superintelligence. In the past two months, together with collaborators from Tsinghua University and Shanghai AI Lab, we organized and summarized the latest RL research for reasoning models into a comprehensive survey. Our paper introduces the fundamentals, problems, resources, applications, and future directions of RL for LRMs, with a special focus on the long-term co-evolution of language models and environments. Preprint is online—welcome to check it out, discuss, or show support! 📄 Paper: huggingface.co/papers/2509.08… 🔗 GitHub: github.com/TsinghuaC3I/Aw…
Kaiyan Zhang tweet media
English
5
76
345
24.9K
Nikita Morozov retweetledi
Hongyuan Mei
Hongyuan Mei@hongyuan_mei·
In RL for LLM reasoning, it’s not just about maximizing reward, but aligning policy to the reward distribution. Our new paper uses flow matching to boost rollout diversity—improving math & code reasoning across the board. Huge thanks to awesome coauthors!
Hongyuan Mei tweet media
English
12
64
468
26.8K