
Lin Yang
201 posts

Lin Yang
@lyang36
Associate Professor of ECE&CS@UCLA. ML, RL, big data, algorithms, astronomy.


🚨 New paper accepted at #ICLR2026. 🚨 We introduce ARMOR — a one-shot, hardware-aligned pruning method that dramatically outperforms existing semi-structured pruning while keeping real inference speedups. Paper Link: openreview.net/forum?id=8NE55… 👇 Thread ↓


IJCAI 2026 will charge $100 USD per submission. Funds will be used to compensate reviewers.


IMO-ProofBench is our key focus designed to evaluate the ability of AI models in constructing rigorous and valid mathematical arguments. With 60 proof-based problems, the benchmark is divided into two subsets: a basic set covering pre-IMO to IMO-Medium difficulty levels, and an advanced set featuring novel, highly challenging problems simulating complete IMO examinations, up to IMO-Hard level. Our goal for the basic set is to assess models in their early stages of development. Sufficiently strong performance on the basic set would justify progression to the advanced set. Performances on the basic IMO-ProofBench varies significantly: while Gemini Deep Think (IMO Gold) achieves a high score of 89.0%, most models score below 60%, indicating that there is still considerable room for improvements. The advanced IMO-ProofBench proves to be a more significant challenge that all non-Gemini models score below 25%. Our IMO-gold model achieved a state-of-the art score of 65.7% according to human evaluations. This represents a substantial leap in capability, but its distance from a perfect score indicates that even the strongest models have room for growth in sophisticated mathematical reasoning.



AIs Win Math Olympiad Gold: Prof. Lin Yang (UCLA) – Manifold #97 Lin Yang is a professor of computer science at UCLA. Recently, he and his collaborator built an AI pipeline using commercial models such as Gemini, ChatGPT, and Grok that performed at the gold medal level on International Mathematics Olympiad problems. Steve and Lin discuss this research, which relies on "verifier-refiner" LLM instances and large token budgets to reliably solve difficult problems. They discuss how these methods can be used to advance AI for scientific research, legal analysis, and complex document processing. (00:00) - AIs Win Math Olympiad Gold: Prof. Lin Yang (UCLA) – #97 (00:57) - Prof. Lin Yang, UCLA (04:27) - Journey from Physics to Computer Science: 2 PhDs (11:15) - Transition to AI from Theoretical CS (13:16) - AI Pipeline Math Olympiad: Gold Medal! (28:23) - Probability Amplification (29:00) - Applications in Industry and Legal Analysis (29:58) - Challenges in Model Reasoning and Verification (33:23) - Future of AI in Scientific Research and AGI Speculations






It's becoming increasingly clear that gpt5 can solve MINOR open math problems, those that would require a day/few days of a good PhD student. Ofc it's not a 100% guarantee, eg below gpt5 solves 3/5 optimization conjectures. Imo full impact of this has yet to be internalized...

















