Babak Hodjat

480 posts

Babak Hodjat

Babak Hodjat

@babakatwork

Working on Evolutionary AI

Katılım Mart 2009
72 Takip Edilen680 Takipçiler
Babak Hodjat retweetledi
Giuseppe Paolo
Giuseppe Paolo@_GPaolo·
What happens when AI agents are left to live (and die) together in a shared world? We’ve been exploring this at the @cognizant AI Lab — and they started forming something that looks like a society.
English
64
94
689
74K
Babak Hodjat retweetledi
Roberto Dailey
Roberto Dailey@RobertoDailey1·
Second, as I mentioned earlier, right now this framework is limited to tasks where decomposition is provided. We are preliminarily testing generalized methods that preform both subtasks and task decomposition, and we are seeing promising results on boosting arithmetic abilities of small models as well – though it does appear decomposition may itself be a challenge orthogonal to task execution.
Roberto Dailey tweet media
English
2
1
11
2K
Babak Hodjat retweetledi
Roberto Dailey
Roberto Dailey@RobertoDailey1·
Our subtask breakdown was to provide an llm agent with the current Towers of Hanoi state and the last move made. The agent would then use first-to-ahead-by-k voting along with abnormal response flagging to decide what move it wanted to do and provide the board state for the next task. This does limit our framework to cleanly decomposable problems, which I’ll come back to at the end of the thread.
GIF
English
1
3
20
4K
Babak Hodjat retweetledi
Roberto Dailey
Roberto Dailey@RobertoDailey1·
New work from Cognizant AI lab: Solving a Million-step LLM Task with Zero Errors. Existing LLMs struggle on long task horizons as persistent error rates compound, even when the LLMs know how to solve the task. Apple’s “Illusion of thinking” demonstrated that state of the art reasoning models could struggle with a simple task, Towers of Hanoi, if that task required execution of hundreds of steps in a row without error. We hypothesized we could see much higher performance by taking breaking down the task into its smallest subtasks, then using voting and red flagging to boost subtask accuracy. With these simple modifications we were able to push the simple gpt-4.1-mini to solve the 20-disk towers of Hanoi, or 1,048,575 steps without a single error! Seeing these results we believe with the right, robust, frameworks, LLMs can be scaled to vastly longer task lengths than their base model. Paper: arxiv.org/abs/2511.09030 Blog: cognizant.com/us/en/ai-lab/b…
English
20
109
716
134.1K
Babak Hodjat retweetledi
Xin Qiu
Xin Qiu@realVsonicV·
Our recent ES fine-tuning paper (arxiv.org/pdf/2509.24372) received lots of attention from the community (Thanks to all!). To speed up the research in this new direction, we developed an accelerated implementation with 10X speed-up in total running time, by refactoring the infrastructure with faster inference engines. Kudos to @DibblaX ! You can find the accelerated version in our repo now: github.com/VsonicV/es-fin…. We are still working on the next version with better API designs, but we cannot wait to release this accelerated version, because we understand how expensive the GPU compute is for researchers and practitioners. Feel free to leave any questions/concerns about the work here.
Xin Qiu tweet media
English
2
14
82
47.9K
Babak Hodjat retweetledi
Paul Jarratt
Paul Jarratt@Jarrattp·
🧠 AGI: A Reality Check As AGI hype grows (and billionaires build bunkers), we need grounded voices. Cognizant’s @babakatwork reminds us: “LLMs don’t have meta-cognition… they don’t know what they know.” bbc.com/news/articles/…
English
0
1
2
58
Babak Hodjat retweetledi
AK
AK@_akhaliq·
Evolution Strategies at Scale LLM Fine-Tuning Beyond Reinforcement Learning
AK tweet media
English
2
22
110
13.6K
Babak Hodjat retweetledi
Babak Hodjat retweetledi
Yulu Gan
Yulu Gan@yule_gan·
To recap — ES can outperform RL for LLM fine-tuning. No gradients. No reward hacking. Just stability, efficiency, and scalability. ES shows low variance across seeds, minimal hyperparameter sensitivity, and strong reward–KL tradeoffs — all without actor-critic complexity. Beyond results, ES offers a simpler, gradient-free path for post-training. From reasoning and exploration to safety alignment and continual learning, it scales reliably where RL often breaks. As models grow in size and importance, stability and robustness will matter as much as raw performance. ES could be the future of fine-tuning — a cleaner, more reliable way to shape intelligence. 🚀
English
0
3
80
10.8K
Babak Hodjat retweetledi
Yulu Gan
Yulu Gan@yule_gan·
Another key advantage of ES fine-tuning is its reliability. It runs stably across seeds, barely depends on hyperparameters, and avoids reward hacking — all while skipping gradients and actor-critic setups. In the figure, you can see ES finds a much better reward–KL balance than GRPO, reaching higher rewards even without KL constraints — indicating a fundamentally different fine-tuning trajectory.
Yulu Gan tweet media
English
1
4
73
8.7K
Babak Hodjat retweetledi
Yulu Gan
Yulu Gan@yule_gan·
On the symbolic-reasoning Countdown task, ES beats PPO/GRPO across Qwen-2.5 (0.5B–7B) & Llama-3 (1B–8B) with huge gains. Moreover, as shown in TinyZero by @jiayi_pirate and DeepSeek-R1, RL fails on small models like Qwen-0.5B — yet ES succeeds! 🚀
Yulu Gan tweet media
English
1
4
70
9.4K
Babak Hodjat retweetledi
Yulu Gan
Yulu Gan@yule_gan·
As noted in DeepSeek-R1 and other studies, RL fine-tuning has several limitations, including challenges with long-horizon and outcome-only rewards, low sample efficiency, high-variance credit assignment, instability, and reward hacking. ES sidesteps these issues: it perturbs parameters (not actions), evaluates full rollouts, and averages over populations, thereby achieving stable, gradient-free, and reward-hacking-resistant optimization that is also easy to parallelize.
Yulu Gan tweet media
English
4
7
84
11.7K
Babak Hodjat retweetledi
Yulu Gan
Yulu Gan@yule_gan·
Reinforcement Learning (RL) has long been the dominant method for fine-tuning, powering many state-of-the-art LLMs. Methods like PPO and GRPO explore in action space. But can we instead explore directly in parameter space? YES we can. We propose a scalable framework for full-parameter fine-tuning using Evolution Strategies (ES). By skipping gradients and optimizing directly in parameter space, ES achieves more accurate, efficient, and stable fine-tuning. Paper: arxiv.org/pdf/2509.24372 Code: github.com/VsonicV/es-fin…
English
90
384
2.6K
413.7K
Babak Hodjat retweetledi