Zeyu Zheng

11 posts

Zeyu Zheng

Zeyu Zheng

@ZhengBerkeley

Associate Professor @UCBerkeley IEOR & BAIR

Beigetreten Mart 2026
15 Folgt48 Follower
Zeyu Zheng retweetet
Huaxiu Yao
Huaxiu Yao@HuaxiuYaoML·
🧠 Multimodal SimpleMem is here (v0.2.0). SimpleMem now supports multimodal lifelong memory for LLM agents. - 🏆 LoCoMo F1: +47% - 🏆 Mem-Gallery F1: +51% 🔬 What makes this special: Omni-SimpleMem was built using AutoResearchClaw's Human-in-the-Loop Co-Pilot mode — human researchers and AI agents co-piloting research, experiments, and writing together. 💡 Three design principles: - 🔍 Selective Ingestion — entropy-driven filtering per modality - 🎯 Progressive Retrieval — hybrid FAISS + BM25 with pyramid token budgets - 🕸️ Knowledge Graph Augmentation — multi-hop cross-modal reasoning 📄 Paper: arxiv.org/abs/2604.01007 🔗 Code: github.com/aiming-lab/Sim… More detailed explanation of Omni-SimpleMem coming soon 👀 @JiaqiLiu835914 @YanqingLiu83931 @StephenQS0710 @lillianwei423 @richardxp888 @HaoqinT @ZhengBerkeley @cihangxie @dingmyu
Huaxiu Yao tweet media
Huaxiu Yao@HuaxiuYaoML

🧠 Can agent memory scale without losing reasoning? 🔥 We’re excited to share our latest work, SimpleMem, a principled memory framework for LLM agents built around semantic lossless compression. 📉 30× fewer inference tokens 📈 +26.4% avg F1 (vs Mem0) ⚡ 50.2% faster retrieval (vs Mem0) Instead of storing raw interaction history 🗂️ or relying on costly iterative reasoning loops 🔁, SimpleMem treats memory as a structured, evolving representation whose primary objective is 🎯 maximizing information density per token. 📄 Paper: arxiv.org/abs/2601.02553 🔗 Code: github.com/aiming-lab/Sim… 📦 Website:aiming-lab.github.io/SimpleMem-Page/ Nice work @JiaqiLiu835914, Yaofeng Su, @richardxp888, @lillianwei423, and great collab. w/ @cihangxie, Zeyu Zheng, @dingmyu

English
6
21
73
10.1K
Zeyu Zheng retweetet
Huaxiu Yao
Huaxiu Yao@HuaxiuYaoML·
🚀 Introducing AutoHarness (「Aha」) — automated harness engineering for AI agents. In LLM training, the aha moment is when a model learns to reason. For agents, it's when a better harness makes the same model shine. Agent = Model + Harness. The model reasons. The harness does everything else: 🧠 Context management 🛡️ Tool governance 💰 Cost control 👁️ Observability 💾 Session persistence These are the patterns that separate a toy from a system. AutoHarness automates this entire layer. 🔧 What's inside: - 6-step tool pipeline: parse → classify → permit → execute → sanitize → audit - 3 modes (Core / Standard / Enhanced) — from lightweight to full-featured - Smart context management with token budgeting and multi-layer compression - Full observability: per-call cost tracking, JSONL audit trail, trace diagnostics - Multi-agent profiles with role-based permissions - Any LLM provider Every agent deserves its aha moment. Led by @JiaqiLiu835914, and Kudos to the team @XinyeYee, @richardxp888, @lillianwei423, @HaoqinT @Xinyu2ML, @yuyinzhou_cs, @ZhengBerkeley, @dingmyu, @cihangxie, etc.
Huaxiu Yao tweet media
English
13
36
165
27.1K
Zeyu Zheng retweetet
Huaxiu Yao
Huaxiu Yao@HuaxiuYaoML·
🦞MetaClaw v0.4.0 is here. Introducing the Evolving Contexture Layer: skills and memory now co-evolve. +18% over OpenClaw + GPT-5.2 on deployed CLI agent benchmarks. Your agent remembers across sessions. User preferences, project context, and interaction history are automatically retrieved and injected into every prompt. Skills adapt based on what it remembers. Memory sharpens based on what it learns. No GPU required. Just talk. It meta-learns and evolves.🧠 Built with @openclaw and @thinkymachines Kudos to the team @richardxp888, @JimChenjw, @HaoqinT, @lillianwei423, @StephenQS0710, @Xinyu2ML, @JiaqiLiu835914, @ZhengBerkeley, @cihangxie.
Huaxiu Yao tweet media
English
4
21
55
8.7K
Zeyu Zheng retweetet
Zeyu Zheng retweetet
Huaxiu Yao
Huaxiu Yao@HuaxiuYaoML·
Agents can remember. Agents can learn. But can they learn how to learn? 🤔 MetaClaw tech report is out 📄 — a continual meta-learning framework that teaches deployed agents not just to learn, but to learn how to learn, evolving 24/7 through normal usage. 📊 On 588 continual CLI tasks over 14 simulated workdays: 🦞 Kimi-K2.5 accuracy: 21.1% → 39.6% (+88%) 🚀 Task completion: 18.2% → 51.9% (+185%) 💪 GPT-5.2 also benefits: 44.9% → 49.1% acc, 58.4% → 67.5% completion Even the strongest models get better with MetaClaw. Kudos to the team @richardxp888, @JimChenjw, @Xinyu2ML, @lillianwei423, @StephenQS0710, @HaoqinT, @JiaqiLiu835914, @yuyinzhou_cs, @ZhengBerkeley, @cihangxie
Huaxiu Yao tweet media
English
8
51
149
10.7K
Zeyu Zheng retweetet
Huaxiu Yao
Huaxiu Yao@HuaxiuYaoML·
We’ve officially brought the power of SkillRL to the CLI agent world! 🔧 Built on @openclaw and powered by Tinker @thinkymachines, MetaClaw lets agents learn new skills from conversations and evolve live — no GPU clusters needed. Our original SkillRL outperformed GPT-4o using a 7B model. Now Kimi-2.5 can hot-swap weights and inject new skills on the fly. 🧠🦾 Try it here: github.com/aiming-lab/Met…
Huaxiu Yao@HuaxiuYaoML

Why do most LLM agents hit a wall? They don’t accumulate skills. Introducing SkillRL📚 — recursive skill-augmented reinforcement learning that lets agents learn skills from failure and evolve over time. 🔥A 7B model: • +41% over GPT-4o • ~20% fewer training tokens • 33% faster convergence SkillRL bridges raw experience → policy improvement by distilling trajectories into structured, co-evolving skills during RL. Most agents forget. SkillRL evolves. 🔄 📄 Paper: arxiv.org/abs/2602.08234 💻 Code: github.com/aiming-lab/Ski… Great work @richardxp888, Jianwen Chen, Hanyang Wang, @JiaqiLiu835914, @lillianwei423, @AiYiyangZ, and nice collab. w/ @__YuWang__, @XujiangZhao, Haifeng Chen, Zeyu Zheng, @cihangxie.

English
6
21
81
10K