Navid Azizan

105 posts

Navid Azizan banner
Navid Azizan

Navid Azizan

@NavidAzizan

MIT Prof | AI & machine learning, systems & control, optimization | Fmr postdoc @Stanford, PhD @Caltech

Cambridge, MA Katılım Haziran 2018
301 Takip Edilen1.7K Takipçiler
Navid Azizan retweetledi
Machine Learning (ML) Papers
LMI-Net: Linear Matrix Inequality–Constrained Neural Networks via Differentiable Projection Layers Sunbochen Tang, Andrea Goertzen, Navid Azizan arxiv.org/abs/2604.05374 [𝚌𝚜.𝙻𝙶]
Machine Learning (ML) Papers tweet media
English
0
2
4
193
Navid Azizan
Navid Azizan@NavidAzizan·
Nice review, Edgar! We've also done some work that may be of interest: - Know What You Don't Know: Uncertainty Calibration of Process Reward Models arxiv.org/abs/2506.09338 - Quantifying Representation Reliability in Self-Supervised Learning Models arxiv.org/abs/2306.00206 - Quantifying the Reliability of Predictions in Detection Transformers: Object-Level Calibration and Image-Level Uncertainty arxiv.org/abs/2412.01782 - Sketching Curvature for Efficient Out-of-Distribution Detection for Deep Neural Networks arxiv.org/abs/2102.12567 - HardNet: Hard-Constrained Neural Networks with Universal Approximation Guarantees arxiv.org/abs/2410.10807 - HardFlow: Hard-Constrained Sampling for Flow-Matching Models via Trajectory Optimization arxiv.org/abs/2511.08425 @HamedSHassani @obastani @tatsu_hashimoto @yuekai_sun @CsabaSzepesvari @ml_angelopoulos @stats_stephen @yaniv_romano @yaringal @KilianQW @_onionesque
English
1
2
8
888
Edgar Dobriban
Edgar Dobriban@EdgarDobriban·
I wrote a review paper about statistical methods in generative AI; specifically, about using statistical tools along with genAI models for making AI more reliable, for evaluation, etc. See here: arxiv.org/abs/2509.07054! I have identified four main areas where statistical thinking can be helpful. These are just a subset of what is out there; other topics have been well-covered in other reviews. 1. Designing "statistical wrappers" around a model, for instance, changing behavior of a trained model (e.g., abstaining), where a score, e.g., an "unsafety score" is too high. The key connection to statistics is to use the quantiles of the loss (on a calibration set) to set the critical threshold, thus enabling conformal-type high probability guarantees. 2. Closely related, methods for uncertainty quantification, which enable the model to express uncertainty in an answer. A crucial component here is "calibration", whereby the uncertainty is required to reflect reality. 3. Statistical methods for AI evaluation: Specifically, tools for statistical inference (e.g., confidence intervals) on model performance. Exciting recent work proposes careful statistical models for leveraging a very small high-quality dataset, possibly combined with much larger low-quality datasets, for accurate evaluation. 4. Experiment design and interventions. Careful AI experiments to understand and steer models may require interventions such as modifying experimental settings in a controlled manner. This brings up connections to classical experimental design in statistics. This connection has largely remained implicit so far, and my review aims to make it more explicit; hoping that experimental design principles will become useful here. This review references the work of many, including @HamedSHassani @obastani @tatsu_hashimoto @yuekai_sun @CsabaSzepesvari @ml_angelopoulos @stats_stephen @yaniv_romano @yaringal @KilianQW @_onionesque +their teams, and some work that I was also involved in. Hopefully, my review will be helpful to orient yourself in this exciting area. Nonetheless, since the area is rapidly expanding, it is possible that I missed important references. Please feel free to let me know of anything that I should add/change!
Edgar Dobriban tweet mediaEdgar Dobriban tweet mediaEdgar Dobriban tweet mediaEdgar Dobriban tweet media
English
11
97
486
34.1K
Navid Azizan
Navid Azizan@NavidAzizan·
Excited to be at #NeurIPS with several of my brilliant students! Some of them are looking for internships and full-time roles, and we are also recruiting new students and postdocs—come find us at any of these sessions! - @young_j_park will present "Know What You Don't Know: Uncertainty Calibration of Process Reward Models" arxiv.org/abs/2506.09338 - @KavehAlim will present "Activation-Informed Merging of Large Language Models" arxiv.org/abs/2502.02421 - @youngjaem0 will present "HardNet: Hard-Constrained Neural Networks" arxiv.org/abs/2410.10807 - Haoyuan Sun will present "On the Role of Transformer Feed-Forward Layers in Nonlinear In-Context Learning" arxiv.org/abs/2501.18187 - Andrea Goertzen and Sunbochen Tang will present "ECO: Energy-Constrained Operator Learning for Chaotic Dynamics with Boundedness Guarantees" arxiv.org/abs/2512.01984 Huge thanks to our amazing co-authors and collaborators—@HW_HaoWang, @KGreenewald, Amin Heyrani Nobari, Ali ArjomandBigdeli, @variational_i, @_faezahmed, @jababi, Ji Young Byun, and Rama Chellappa—and to @MITIBMLab, @Amazon, @Google, and @MathWorks for their support. #NeurIPS2025 @NeurIPSConf @MIT @MITEngineering @MIT_SCC @MITMechE @MITEECS @MITAeroAstro @MITChemE @MITIDSS @MITLIDS
Navid Azizan tweet media
English
1
3
22
1.9K
Kaiqing Zhang
Kaiqing Zhang@KaiqingZhang·
A bit of good news 😉 Happy New Year everyone!
Kaiqing Zhang tweet media
English
29
1
260
20.3K
Navid Azizan
Navid Azizan@NavidAzizan·
Nice! There is another method for solving this exact problem (distributed solution of linear systems), which has the same per-iteration computation and communication complexity as distributed GD but has a much faster convergence rate*: azizan.mit.edu/papers/APC.html (Section V is the overparameterized case, but we don't know whether it converges to the min-norm solution or not—maybe you guys can show it one way or the other!) *We establish the reason behind its faster convergence in a recent paper: arxiv.org/abs/2304.10640
English
1
0
4
423
Arya Mazumdar
Arya Mazumdar@MountainOfMoon·
Suppose you run gradient descent to solve a linear highly overparameterized linear regression. It has an implicit bias to go to the min norm solution. What happens when you distribute that data in many pieces and run GD locally for long and then take avg? arxiv.org/abs/2412.07971
English
2
1
36
3.3K
George Pappas
George Pappas@pappasg69·
What an amazing honor to be among the distinguished group in the National Academy of Engineering Class of 2024 @theNAEng @PennEngineers
George Pappas tweet mediaGeorge Pappas tweet media
English
18
7
123
9.4K
Navid Azizan retweetledi
Young-Jin Park
Young-Jin Park@young_j_park·
Wondering when to trust pre-trained AI models and how to assess their reliability before deployment? Check out our work at #UAI2024! If you’re in Barcelona, visit my poster (#368) tomorrow!! 🔗 Read More: t.ly/AzKyh (Paper), t.ly/85CKE (MIT News).
English
1
2
4
1.1K
Anirudha Majumdar
Anirudha Majumdar@Majumdar_Ani·
I received tenure at #Princeton! It’s been an incredible journey, and I am immensely grateful to my group, collaborators, mentors, and family. Looking forward to continuing our work on making 🤖 more trustworthy! I will be on sabbatical 2024-25, spending time at @GoogleDeepMind.
English
47
5
436
37.7K
George Pappas
George Pappas@pappasg69·
Thrilled to announce my election to the National Academy of Engineering (@theNAEng ). I am grateful for the support of my @PennEngineers colleagues, mentors, and above all, my students and postdocs that have made this distinction possible.
National Academy of Engineering@theNAEng

The National Academy of Engineering is excited to welcome 114 new members and 21 new international members to the NAE Class of 2024! Congratulations to this incredible group of innovators. #NAE2024 Find the complete list of newly elected members here: ow.ly/V4Gj50Qyvjl

English
43
16
257
37.8K