Weronika Ormaniec

29 posts

Weronika Ormaniec

Weronika Ormaniec

@wormaniec

Zurich, Switzerland Beigetreten Mayıs 2019
145 Folgt59 Follower
Weronika Ormaniec retweetet
Dimitri von Rütte
Dimitri von Rütte@dvruette·
🚨 NEW PAPER! (this is a big one; 3B and 10B models included) Masked diffusion LLMs are getting a lot of attention. They outperform other diffusion types (such as uniform diffusion) at small scales. But what if I told you that uniform diffusion actually scales better? 🧵👇
Dimitri von Rütte tweet media
English
8
42
214
53.6K
Weronika Ormaniec
Weronika Ormaniec@wormaniec·
A fantastic opportunity to join the new lab of my former master's thesis co-advisor, @f_dangel! He is an incredible researcher and mentor. I really recommend working with him
Felix Dangel@f_dangel

🎓 Looking for MSc or PhD opportunities in Machine Learning for Fall 2026? Join my group at @Concordia and @Mila_Quebec! 🔍 Focus: autodiff, second-order optimization, and Hessian-based methods for LLMs & scientific ML. 📅 Apply by Dec 1: mila.quebec/en/prospective…

English
0
0
1
113
Weronika Ormaniec retweetet
Bruno Mlodozeniec
Bruno Mlodozeniec@brunorganised·
I would highly recommend using this library for any research on influence functions. Implementing scalable IFs (usually ≡ K-FAC) is a massive pain, especially for modern architectures. With curvlinops, getting plots like the below for diffusion models is relatively easy
Bruno Mlodozeniec tweet media
Runa Eschenhagen@runame_

1/6 Hessian approximations are ubiquitous in deep learning, but working with them can get quite involved. We argue for using a linear operator interface for neural network curvature matrices and implement this in PyTorch in our library curvlinops. arxiv.org/abs/2501.19183/

English
1
3
7
762
Weronika Ormaniec
Weronika Ormaniec@wormaniec·
@orvieto_antonio Same here! We also ended up discussing them in the optimization tutorial for our DL course 😅
English
1
0
1
499
Antonio Orvieto
Antonio Orvieto@orvieto_antonio·
Despite my best intentions, I ended up covering Muon/Scion/LMO on the second lecture of my nonconvex opt class...
English
6
3
160
14.9K
Weronika Ormaniec retweetet
ML in PL
ML in PL@MLinPL·
We’re thrilled to welcome Sander Dieleman, Research Scientist at Google DeepMind, to ML in PL Conference 2025! Sander Dieleman is a Research Scientist at Google DeepMind in London, UK, where he has worked on the development of AlphaGo, WaveNet, Imagen 4, Veo 3, and more. He obtained his PhD from Ghent University in 2016. His current research interests include representation learning and generative modelling of audio, images and video. 📍 15–17 October 2025, Copernicus Science Centre, Warsaw, Poland 🔗 To learn more and secure your spot #tickets" target="_blank" rel="nofollow noopener">conference.mlinpl.org/#tickets
ML in PL tweet media
English
1
4
33
4.3K
Weronika Ormaniec retweetet
Maksym Andriushchenko
Maksym Andriushchenko@maksym_andr·
🚨 Incredibly excited to share that I'm starting my research group focusing on AI safety and alignment at the ELLIS Institute Tübingen and Max Planck Institute for Intelligent Systems in September 2025! 🚨 Hiring. I'm looking for multiple PhD students: both those able to start in Fall 2025 (i.e., as soon as possible) and through centralized programs like CLS, IMPRS, and ELLIS (the deadlines are in November) to start in Spring–Fall 2026. I'm also searching for postdocs, master's thesis students, and research interns. Fill the Google form below if you're interested! Research group. We will focus on developing algorithmic solutions to reduce harms from advanced general-purpose AI models. We're particularly interested in alignment of autonomous LLM agents, which are becoming increasingly capable and pose a variety of emerging risks. We're also interested in rigorous AI evaluations and informing the public about the risks and capabilities of frontier AI models. Additionally, we aim to advance our understanding of how AI models generalize, which is crucial for ensuring their steerability and reducing associated risks. For more information about research topics relevant to our group, please check the following documents: - International AI Safety Report, - An Approach to Technical AGI Safety and Security by DeepMind, - Open Philanthropy’s 2025 RFP for Technical AI Safety Research. Research style. We are not necessarily interested in getting X papers accepted at NeurIPS/ICML/ICLR. We are interested in making an impact: this can be papers (and NeurIPS/ICML/ICLR are great venues), but also open-source repositories, benchmarks, blog posts, even social media posts—literally anything that can be genuinely useful for other researchers and the general public. Broader vision. Current machine learning methods are fundamentally different from what they used to be pre-2022. The Bitter Lesson summarized and predicted this shift very well back in 2019: "general methods that leverage computation are ultimately the most effective". Taking this into account, we are only interested in studying methods that are general and scale with intelligence and compute. Everything that helps to advance their safety and alignment with societal values is relevant to us. We believe getting this—some may call it "AGI"—right is one of the most important challenges of our time. Join us on this journey!
Maksym Andriushchenko tweet media
English
76
90
841
105.7K
Weronika Ormaniec retweetet
ML in PL
ML in PL@MLinPL·
We are delighted to introduce our next ML in PL Conference 2025 speaker: Alexey Dosovitskiy! Alexey Dosovitskiy is a distinguished AI researcher who gained prominence at @GoogleResearch as lead author of the "An Image is Worth 16x16 Words" paper, which introduced Vision Transformers (ViT). As of February 2024, he joined Inceptive as a Member of Technical Staff, where he's now applying machine learning techniques to RNA research. 👉 Check all the conference details on our event page: conference.mlinpl.org
ML in PL tweet media
English
0
3
6
338
Weronika Ormaniec retweetet
Yuhui Ding
Yuhui Ding@yuhui_ding·
Is equivariance necessary for a good 3D molecule generative model? Check out our #icml2025 paper, which closes the performance gap between non-equivariant and equivariant diffusion models via rotational alignment, while also being more efficient (1/7): arxiv.org/abs/2506.10186
English
4
8
41
5.7K
Weronika Ormaniec retweetet
ML in PL
ML in PL@MLinPL·
We are delighted to introduce you to our next ML in PL Conference 2025 Speaker: Federico Tombari @fedassa! Federico Tombari is Research Director at @Google, leading Computer Vision and Machine Learning teams across North America and Europe. His team has contributed CV/ML technology to Google Lens, Maps, Android, ARCore, and Pixel. He's also a Lecturer at @TU_Muenchen with 300+ peer-reviewed publications in CV/ML, covering robotics, autonomous driving, healthcare, and AR. He co-founded a 3D perception startup that was acquired by Google in 2018-19. Federico serves as Area Chair and Associate Editor for top conferences like @NeurIPSConf, @CVPR, and @eccvconf, and has received Google Faculty Awards, an Amazon Research Award, and multiple Outstanding Reviewer Awards. 👉‍‍ Stay up to date with ML in PL news by following us on social media! 👉‍‍ ML in PL Conference 2025 website: conference.mlinpl.org/?utm_campaign=… 👉‍‍ ML in PL Facebook Page: facebook.com/MLinPL 👉‍‍ ML in PL LinkedIn Page: linkedin.com/company/mlinpl/ 👉‍‍ ML in PL X Profile: x.com/MLinPL
ML in PL tweet media
English
0
1
2
287
Weronika Ormaniec retweetet
ML in PL
ML in PL@MLinPL·
📣 Calling all ML researchers, students and professionals: ML in PL Conference 2025 is now accepting submissions for talks, posters, and tutorials. We’re accepting submissions across a wide range of machine learning topics, including (but not limited to): - Core ML & Optimization: Classification, Clustering, Learning Theory, Online/Semi/Unsupervised Learning - Deep Learning: Architectures, Generative Models, Recurrent Networks, DL Optimization - Reinforcement Learning: Bandits, Control, MDPs, Planning, Multi-Agent Systems - Probabilistic & Causal Methods: Bayesian Inference, Gaussian Processes, Graphical Models - Applications: NLP, Vision, Robotics, Audio, Biology, Neuroscience, Physics, Social Good - Responsible AI: Fairness, Privacy, Robustness, Safety, Ethics, Bias, Explainability - Technical Resources: Datasets, Tools, Software, Distributed ML, Open Competitions - ML Stories: Case studies, startup journeys, and lessons from real-world deployments Whether you're a PhD student sharing your first major results, a senior researcher, or a seasoned ML practitioner, we welcome your contribution—be it theoretical or applied, early-stage or production-ready. Submission Deadline: July 31st 📄 Talks & Posters → conference.mlinpl.org/call-for-contr… 💻 Tutorials → conference.mlinpl.org/call-for-tutor… Thinking of attending instead? Early Bird registration is open until July 31st: mlinpl2025eb.paperform.co
ML in PL tweet media
English
1
9
15
7.9K
Weronika Ormaniec retweetet
Tiago Pimentel
Tiago Pimentel@tpimentelms·
A string may get 17 times less probability if tokenised as two symbols (e.g., ⟨he, llo⟩) than as one (e.g., ⟨hello⟩)—by an LM trained from scratch in each situation! Our #acl2025nlp paper proposes an observational method to estimate this causal effect! Longer thread soon!
Tiago Pimentel tweet media
English
3
24
135
18.4K
Weronika Ormaniec retweetet
ML in PL
ML in PL@MLinPL·
We are thrilled to announce that Antonio, a leading researcher in deep learning and optimization, will be joining us as a speaker at the ML in PL 2025 Conference! Antonio studied Control Engineering in Italy and Switzerland. He holds a PhD in Computer Science from ETH Zürich and spent time at Deepmind (UK), Meta (US), MILA (CA), INRIA (FR), and HILTI (LI). He is currently a Hector Endowed Fellow and Principal Investigator (PI) at the ELLIS Institute Tübingen and Independent Group Leader of the MPI for Intelligent Systems, where he leads the Deep Models and Optimization group. He received the ETH medal for outstanding doctoral theses and the Schmidt Sciences AI2050 Early Career Fellowship. In his research, Antonio strives to improve the efficiency of deep learning technologies by pioneering new architectures and training techniques grounded in theoretical knowledge. His work encompasses two main areas: understanding the intricacies of large-scale optimization dynamics and designing innovative architectures and powerful optimizers capable of handling complex data. Central to his studies is exploring innovative techniques for decoding patterns in sequential data, with implications in biology, neuroscience, natural language processing, and music generation. 👉 Early Bird Registration is open until the 31st of July. mlinpl2025eb.paperform.co Check all the conference details on our event page: conference.mlinpl.org
ML in PL tweet media
English
0
2
3
369
Dimitri von Rütte
Dimitri von Rütte@dvruette·
Happy to share that GIDD was accepted at ICML 2025! 🥳 See our thread to learn how we added self-correction capabilities to discrete diffusion models with a simple change in the noise schedule 👇
Dimitri von Rütte@dvruette

🚨 NEW PAPER DROP! Wouldn't it be nice if LLMs could spot and correct their own mistakes? And what if we could do so directly from pre-training, without any SFT or RL? We present a new class of discrete diffusion models, called GIDD, that are able to do just that: 🧵1/12

English
3
0
35
2.2K
Weronika Ormaniec
Weronika Ormaniec@wormaniec·
Our code is publicly available, so you can generate iSCM-based datasets and easily benchmark causal discovery methods. 🔗 Code: github.com/werkaaa/iscm 7/8
English
1
0
0
112
Weronika Ormaniec
Weronika Ormaniec@wormaniec·
Excited to share another paper at #ICLR2025 We introduce iSCMs—a new method for generating synthetic data for causal discovery benchmarks that avoids variance and covariance artifacts. Joint work with Scott Sussex, Lars Lorch, @bschoelkopf & @arkrause. Key insights👇 1/8
Weronika Ormaniec tweet media
English
1
0
2
232