Lichang Chen

336 posts

Lichang Chen banner
Lichang Chen

Lichang Chen

@LichangChen2

Coding Agent RL & Harness of 🥑 @Meta MSL | Previously: GenAI unit @GoogleDeepmind | PhD’25 @umdcs BS’20 @zju_china

Menlo Park, CA Katılım Eylül 2021
753 Takip Edilen940 Takipçiler
Sabitlenmiş Tweet
Lichang Chen
Lichang Chen@LichangChen2·
Try Contemplating Mode in Meta.AI! It uses parallel thinking to deliver more well-rounded answers. We’d love your feedback! I feel very fortunate to have contributed to this project and to keep learning through it. It has been both fun and rewarding to grow alongside one of the best reasoning teams on the planet.
Lichang Chen tweet mediaLichang Chen tweet media
English
0
1
8
645
Lijie(Derrick) Yang
Lijie(Derrick) Yang@LijieyYang·
Excited to share that LessIsMore has been accepted to ICML 2026! 🚀 LessIsMore is a training-free sparse attention for efficient long-horizon reasoning. By enforcing cross-head unified token selection, it brings up to 1.6x E2E speedup while preserving reasoning accuracy under practical workloads. Huge thanks to my amazing co-authors and mentors @Jackfram2, @JiaZhihao, Ravi! Paper: arxiv.org/abs/2508.07101 Code: github.com/DerrickYLJ/Les… #ICML2026 #LLM #EfficientAI
Lijie(Derrick) Yang tweet media
English
8
18
67
7.1K
Arena.ai
Arena.ai@arena·
GPT-5.5 by @OpenAI is now live in the Arena, landing across multiple leaderboards. Here’s how it ranks by modality: - Code Arena (agentic web dev): #9, a strong +50pt jump over GPT-5.4 - Document Arena (analysis & long-content reasoning): #6, on par with Sonnet 4.6 - Text Arena: #7, Math #3, Instruction Following: #8 - Expert Arena: #5 - Search Arena: #2 - Vision Arena: #5 Strong, well-rounded performance, especially in Code (+50 pts vs GPT-5.4). Congrats to @OpenAI on the release. Full category breakdowns by modality in the thread.
Arena.ai tweet media
OpenAI@OpenAI

Introducing GPT-5.5 A new class of intelligence for real work and powering agents, built to understand complex goals, use tools, check its work, and carry more tasks through to completion. It marks a new way of getting computer work done. Now available in ChatGPT and Codex.

English
346
132
1.9K
1.4M
Lichang Chen
Lichang Chen@LichangChen2·
I like the idea but I would rephrase it as: Coding agent will be the new foundation for the whole tech industry. They can become any roles by writing the code and docs, e.g, if it works on a research project/folder, it becomes automated researcher as long as you give it an excellent harness and make it iterate to improve the code and docs.
Jacob Effron@jacobeffron

.@swyx current thesis: "2025 was coding agents. 2026 is coding agents breaking containment to do everything else."

English
0
0
0
495
Lichang Chen
Lichang Chen@LichangChen2·
Let’s see where we will get at the end of this year.
Arena.ai@arena

Muse Spark debuts at #7 in the Code Arena - making @AIatMeta the #3 lab right behind @AnthropicAI’s Claude Sonnet 4.6 and @Zai_org’s GLM-5.1, surpassing Gemini-3.1-Pro and GPT-5.4. Code Arena evaluates agentic coding on real-world tasks - building live websites and apps, ranked by users on real workflows. Huge congrats to @AIatMeta on this impressive milestone!

English
0
0
8
1.3K
Lichang Chen retweetledi
Ronit Pereira
Ronit Pereira@Ronitper·
“It takes a lot of hard work to make something simple.” - Steve Jobs
Ronit Pereira tweet media
English
74
745
5.1K
93.7K
Pranav Shyam
Pranav Shyam@recurseparadox·
life goes by quickly when you’re fixing errors and relaunching jobs
English
1
0
3
320
Lichang Chen
Lichang Chen@LichangChen2·
@suchenzang Beautiful women also only have eyes for other beautiful women and we call that the culture of SF?
English
0
0
0
1.3K
Susan Zhang
Susan Zhang@suchenzang·
finally joined an equinox sf gym just to ogle all the beautiful men who only have eyes for other beautiful men
English
23
3
498
42.8K
Intrinsical AI
Intrinsical AI@IntrinsicalAI·
@LichangChen2 This is a completely nonsense test. You're measuring how well a LLM approach Nash equilibria. LLMs are not deterministic, but they are in fact very useful to profiling/data-mining and micro-trends. What about letting an LLM configure the solver tree + read results before acting?
English
1
0
1
25
Lichang Chen
Lichang Chen@LichangChen2·
I am wondering how we can close the gap between neural-based test-time scaling method and the language-based one. Up till now, seems like no one has successfully reproduced the AlphaGo or Poker solvers on the LLMs though they have stronger prior than pure neural/numerical-based approach.
GTOWizard@GTOWizard

We benchmarked every major AI model at poker. GPT-5.4, Claude Opus 4.6, Gemini 3.1 Pro, Grok 4 and more. All played 5,000 hands of heads-up no-limit against our state-of-the-art poker agent. Every single one lost. Here's the full breakdown 🧵

English
1
0
8
1.6K
Lichang Chen
Lichang Chen@LichangChen2·
The CEO of deepmind questioned: “In the AGI systems, will LLMs become the key component only or the total system” I think startup founders will agree on the former but the model trainers will agree on the later😊
Milk Road AI@MilkRoadAI

The CEO of Google DeepMind just went on record saying he disagrees with one of the most respected AI researchers in the world. Demis Hassabis, the man behind AlphaFold, AlphaGo, and Google's entire AI operation publicly pushed back against Yann LeCun's claim that large language models are a dead end for artificial intelligence. LeCun, who left Meta earlier this year to start his own AI lab, has been saying for years that LLMs cannot reason, cannot plan, and will never get us to human-level intelligence. Hassabis disagrees, and he said so directly. His position is that scaling laws are still working, foundation models are still getting more capable, and whatever AGI ends up looking like, LLMs will be a central part of it, not something that gets replaced. He does say there is roughly a 50/50 chance that one or two additional breakthroughs will be needed beyond scaling alone, things like better memory, long-term planning, and world models. But the core disagreement with LeCun is clear, Hassabis believes the current architecture is sound and the current path leads somewhere real. Two Nobel-recognized researchers, two founding figures of modern AI, now publicly on opposite sides of the most important technical question in the industry.

English
0
0
4
1.3K
Lichang Chen
Lichang Chen@LichangChen2·
@alexandr_wang It has pretty good generalization ability😆 we will get used to seeing amazing things from 🥑!
English
0
0
2
3K
Lichang Chen
Lichang Chen@LichangChen2·
@dynemetis It highly depends on how you define continual learning. If it’s human-level continual learning, then 90% of the learning are in the model layer, it’s like you receive real-world signal and update neurons in your brain.🧠
English
0
0
3
85