Berkeley AI Research

1.4K posts

Berkeley AI Research banner
Berkeley AI Research

Berkeley AI Research

@berkeley_ai

We're graduate students, postdocs, faculty and scientists at the cutting edge of artificial intelligence research.

Berkeley, CA Se uniรณ Temmuz 2017
442 Siguiendo255.3K Seguidores
Berkeley AI Research retuiteado
Ion Stoica
Ion Stoica@istoica05ยท
Excited to share our latest work: MยฒRNN! Weโ€™ve revisited non-linear RNNs and found that expanding the hidden state to a matrix (Matrix-to-Matrix) significantly improves language modeling while the non-linear recurrence enables expressivity beyond TCโฐ. Key highlights: - Efficient Scaling: Our expansion mechanism leverages. Tensor Cores for high-throughput training. - Better Long-Context Performance: Beats SOTA hybrid linear attention models by 8 points on LongBench. - Hybrid Models: Replacing just ONE layer in a hybrid stack gives massive gains with minimal overhead. This establishes non-linear RNNs as a primary building block for the next generation of LLMs.
Mayank Mishra@MayankMish98

Introducing MยฒRNN: Non-Linear RNNs with Matrix-Valued States for Scalable Language Modeling We bring back non-linear recurrence to language modeling and show it's been held back by small state sizes, not by non-linearity itself. ๐Ÿ“„ Paper: arxiv.org/abs/2603.14360 ๐Ÿ’ป Code: github.com/open-lm-engineโ€ฆ ๐Ÿค— Models: huggingface.co/collections/opโ€ฆ

English
2
10
121
15.9K
Berkeley AI Research retuiteado
Jitendra MALIK
Jitendra MALIK@JitendraMalikCVยท
With Emmanuel Dupoux scp.net/persons/dupoux/ and Yann LeCun @ylecun, we consider a cognitive science inspired AI. We analyse how autonomous learning works in living organisms, and propose a roadmap for reproducing it in artificial systems. lnkd.in/eNWDmuqT
English
8
78
441
58.7K
Berkeley AI Research retuiteado
Berkeley AI Research retuiteado
Shu Lynn Liu
Shu Lynn Liu@shulynnliuยท
Researchers spend hours and hours hand-crafting the strategies behind LLM-driven optimization systems like AlphaEvolve: deciding which ideas to reuse, when to explore vs exploit, and what mutations to try. ๐Ÿค–But what if AI could evolve its own evolution process? We introduce EvoX, a meta-evolution pipeline that lets AI evolve the strategy guiding the optimization. It achieves high-quality solutions for <$5, while existing open systems and even Claude Code often cost 3-5ร— more on some tasks. Across ~200 optimization problems, EvoX delivers the strongest overall results: often outperforming AlphaEvolve, OpenEvolve, GEPA, and ShinkaEvolve on math and systems tasks, exceeding human SOTA, and improving median performance by up to 61% on 172 competitive programming problems. ๐Ÿ‘‡
Shu Lynn Liu tweet media
English
18
86
491
90.3K
Berkeley AI Research retuiteado
Berkeley AI Research retuiteado
Kuba
Kuba@kuba_AIยท
AI can optimize materials ๐Ÿค˜ Our (@pabbeel, @svlevine, @AIatMeta) proposed transformer model ๐—–๐—น๐—ถ๐—พ๐˜‚๐—ฒ๐—™๐—น๐—ผ๐˜„๐—บ๐—ฒ๐—ฟ, combined with ๐—ฒ๐˜ƒ๐—ผ๐—น๐˜‚๐˜๐—ถ๐—ผ๐—ป strategies, discovers materials that optimize target properties. arxiv.org/abs/2603.06082
Berkeley, CA ๐Ÿ‡บ๐Ÿ‡ธ English
3
19
66
11.9K
Berkeley AI Research retuiteado
Haocheng Xi
Haocheng Xi@HaochengXiUCBยท
๐—ž-๐—บ๐—ฒ๐—ฎ๐—ป๐˜€ ๐—ถ๐˜€ ๐˜€๐—ถ๐—บ๐—ฝ๐—น๐—ฒ. ๐— ๐—ฎ๐—ธ๐—ถ๐—ป๐—ด ๐—ถ๐˜ ๐—ณ๐—ฎ๐˜€๐˜ ๐—ผ๐—ป ๐—š๐—ฃ๐—จ๐˜€ ๐—ถ๐˜€๐—ปโ€™๐˜. Thatโ€™s why we built Flash-KMeans โ€” an IO-aware implementation of exact k-means that rethinks the algorithm around modern GPU bottlenecks. By attacking the memory bottlenecks directly, Flash-KMeans achieves 30x speedup over cuML and 200x speedup over FAISS โ€” with the same exact algorithm, just engineered for todayโ€™s hardware. At the million-scale, Flash-KMeans can complete a k-means iteration in milliseconds. A classic algorithm โ€” redesigned for modern GPUs. Paper: arxiv.org/abs/2603.09229 Code: github.com/svg-project/flโ€ฆ
English
36
196
1.7K
279K
Berkeley AI Research retuiteado
Dvij Kalaria
Dvij Kalaria@DvijKalariaยท
Excited to be co-organizing this #ICRA2026 workshop! How do teleoperation, simulation, and human videos come together to scale robot learning? ๐Ÿค–๐Ÿ“ˆ Join us for talks, panels, and a Best Paper Award!
Rutav@rutavms

Excited to announce our #ICRA2026 workshop: Beyond Teleoperation Teleop, sim, human videos... we have different ideas to scale robot data, but how do they all fit together?๐Ÿงฉ Join us for exciting talks, panel discussions, and a Best Paper Award! Read on for details! ๐Ÿงต๐Ÿ‘‡

English
0
4
26
8.8K
Berkeley AI Research retuiteado
addison
addison@addikalaยท
New on @berkeley_ai blog. What actually affects the output of LLMs? New work by Butler et al. (NeurIPS 2025) introduces a model-agnostic method of measuring input, data, and component attribution, giving insight on what really steers generation. bair.berkeley.edu/blog/2026/03/1โ€ฆ ๐Ÿป๐Ÿ“„
English
3
12
41
9.2K
Berkeley AI Research retuiteado
Guangqi Jiang
Guangqi Jiang@LuccaChiangยท
Ever want to have a single policy to control diverse robots as well as different dexterous hands, or to observe the emergent behavior under cross embodiment training? Introducing our #CVPR2026 paper XL-VLA, Cross-Hand Latent Representation for Vision-Language-Action Models.
English
8
43
261
76.4K
Berkeley AI Research retuiteado
C.K. Wolfe
C.K. Wolfe@ckwolfeofficialยท
Released on @berkeley_ai blog - recent work by Henry Pinkard et al.: information-driven design of imaging systems. Direct mutual information optimization, no decoder needed. Matches end-to-end methods with less memory and no task-specific retraining. Works across cameras, telescopes, and microscopes... ๐Ÿป๐Ÿ“„ bair.berkeley.edu/blog/2026/01/1โ€ฆ.
English
4
20
128
14.2K
Berkeley AI Research retuiteado
Ion Stoica
Ion Stoica@istoica05ยท
Excited to release AdaEvolve, bridging the gap between closed-source automated algorithm discovery and open-source accessibility. Rather than relying on rigid policies, AdaEvolve utilizes a dynamically adjusting search strategy. This adaptive framework yields a 33% median performance increase on Frontier-CS over the strongest open-source baselines. Read the full thread here.
Mert Cemri@mertcemri

AlphaEvolve proved LLMs can discover novel algorithms, but it remains closed-source, and open-source alternatives (OpenEvolve, GEPA) rely on rigid, static search policies. Introducing AdaEvolve: a fully adaptive evolutionary algorithm that dynamically adjusts its own search strategy based on observed progress. It matches or beats AlphaEvolve and best known Human SOTA on math and systems benchmarks, and boosts Frontier-CS median scores by 33% over the best open-source baseline across 185 tasks. ๐Ÿงต๐Ÿ‘‡ (1/n)

English
2
18
143
19.8K
Berkeley AI Research retuiteado
Berkeley AI Research retuiteado
Mert Cemri
Mert Cemri@mertcemriยท
AlphaEvolve proved LLMs can discover novel algorithms, but it remains closed-source, and open-source alternatives (OpenEvolve, GEPA) rely on rigid, static search policies. Introducing AdaEvolve: a fully adaptive evolutionary algorithm that dynamically adjusts its own search strategy based on observed progress. It matches or beats AlphaEvolve and best known Human SOTA on math and systems benchmarks, and boosts Frontier-CS median scores by 33% over the best open-source baseline across 185 tasks. ๐Ÿงต๐Ÿ‘‡ (1/n)
Mert Cemri tweet media
English
10
68
338
73.7K
Berkeley AI Research retuiteado
Haozhi Qi
Haozhi Qi@HaozhiQยท
๐Ÿค– Attending ICRA 2026 and interested in dexterous manipulation? Donโ€™t miss the chance to join or submit to our workshop! ๐Ÿ“ทCheckout our outstanding lineup of speakers, best poster award, and more details on our website: shorturl.at/ojsTo. Deadline: April 10 (AOE).
Haozhi Qi tweet media
English
1
10
71
11.9K
Berkeley AI Research retuiteado
Junyi Zhang
Junyi Zhang@junyi42ยท
๐—ข๐—ป๐—ฒ ๐—บ๐—ฒ๐—บ๐—ผ๐—ฟ๐˜† ๐—ฐ๐—ฎ๐—ปโ€™๐˜ ๐—ฟ๐˜‚๐—น๐—ฒ ๐˜๐—ต๐—ฒ๐—บ ๐—ฎ๐—น๐—น. We present ๐—Ÿ๐—ผ๐—š๐—ฒ๐—ฅ, a new ๐—ต๐˜†๐—ฏ๐—ฟ๐—ถ๐—ฑ ๐—บ๐—ฒ๐—บ๐—ผ๐—ฟ๐˜† architecture for long-context geometric reconstruction. LoGeR enables stable reconstruction over up to ๐Ÿญ๐Ÿฌ๐—ธ ๐—ณ๐—ฟ๐—ฎ๐—บ๐—ฒ๐˜€ / ๐—ธ๐—ถ๐—น๐—ผ๐—บ๐—ฒ๐˜๐—ฒ๐—ฟ ๐˜€๐—ฐ๐—ฎ๐—น๐—ฒ, with ๐—น๐—ถ๐—ป๐—ฒ๐—ฎ๐—ฟ-๐˜๐—ถ๐—บ๐—ฒ ๐˜€๐—ฐ๐—ฎ๐—น๐—ถ๐—ป๐—ด in sequence length, ๐—ณ๐˜‚๐—น๐—น๐˜† ๐—ณ๐—ฒ๐—ฒ๐—ฑ๐—ณ๐—ผ๐—ฟ๐˜„๐—ฎ๐—ฟ๐—ฑ inference, and ๐—ป๐—ผ ๐—ฝ๐—ผ๐˜€๐˜-๐—ผ๐—ฝ๐˜๐—ถ๐—บ๐—ถ๐˜‡๐—ฎ๐˜๐—ถ๐—ผ๐—ป. Yet it matches or surpasses strong optimization-based pipelines. (1/5) @GoogleDeepMind @Berkeley_AI
English
63
448
3.4K
547.4K
Berkeley AI Research retuiteado
Kevin Zakka
Kevin Zakka@kevin_zakkaยท
Happy Friday!! mjlab v1.2.0 is out. This is our biggest release yet with 60+ PRs from 12 contributors. pip install mjlab Some highlights include: - New more powerful domain randomization module - Revamped ergonomic viewers - Cloud training via @SkyPilot - Complete doc rewrite
English
7
14
114
14.9K