
Linjie (Lindsey) Li
200 posts

Linjie (Lindsey) Li
@LINJIEFUN
researching @Microsoft, @UW, contributed to https://t.co/VzcJa9Skx3




Are your LLM agents truly reasoning, or just stuck repeating the same patterns? Zihan Wang @wzenus and a stellar team from Northwestern, Stanford, Microsoft, Oxford, and Imperial College London have uncovered "template collapse", a hidden flaw where LLM agents appear diverse but fail to adapt to new inputs. Their RAGEN-2 framework introduces Mutual Information to accurately measure true "cross-input distinguishability" and proposes SNR-Aware Filtering to select high-signal training prompts. This new metric and method vastly outperform current approaches, boosting LLM agent performance and input dependence across critical tasks like planning, math reasoning, web navigation, and code execution! And this paper is also #1 Paper of the day on Hugging Face!

You can now train, adapt, and eval web agents on your own tasks. We're releasing the full MolmoWeb codebase—the training code, eval harness, annotation tooling, synthetic data pipeline, & client-side code for our demo. 🧵

Today we're releasing WildDet3D—an open model for monocular 3D object detection in the wild. It works with text, clicks, or 2D boxes, and on zero-shot evals it nearly doubles the best prior scores. 🧵




Today we're releasing MolmoWeb, an open source agent that can navigate + complete tasks in a browser on your behalf. Built on Molmo 2 in 4B & 8B sizes, it sets a new open-weight SOTA across four major web-agent benchmarks & even surpasses agents built on proprietary models. 🧵










🚨Sensational title alert: we may have cracked the code to true multimodal reasoning. Meet ThinkMorph — thinking in modalities, not just with them. And what we found was... unexpected. 👀 Emergent intelligence, strong gains, and …🫣 🧵 arxiv.org/abs/2510.27492 (1/16)




