Mudith Jayasekara

89 posts

Mudith Jayasekara

Mudith Jayasekara

@mudithj

co-founder @parsedlabs sticking it to Big Token, half eng/cs phd @rhodes_trust @UniofOxford, scaling care through intelligence

SF Katılım Ocak 2018
297 Takip Edilen219 Takipçiler
Sabitlenmiş Tweet
Mudith Jayasekara
Mudith Jayasekara@mudithj·
Rewatching the greats in this video never gets old. Grateful to now be making our own dent in defining the next paradigm of AI. Working with the most thoughtful and passionate people I know and backed by incredible investors (from LocalGlobe @svennj @asharoraa , HuggingFace @Thom_Wolf, DeepMind, NHS, and others). Thank you to our customers who care enough to glimpse into what the future of language models looks like. Let's build🫡
Charlie O'Neill@oneill_c

Today, we’re launching Parsed. We are incredibly lucky to live in a world where we stand on the shoulders of giants, first in science and now in AI. Our heroes have gotten us to this point, where we have brilliant general intelligence in our pocket. But this is a local minima. We now have an ecosystem of burgeoning tasks where each requires a different kind of intelligence, a different context, a whole host of implicit assumptions and latent knowledge and domain expertise that is very difficult to cram into a system prompt. The big labs want you renting their $50k/month amnesiac interns that forget everything between conversations. Generic behemoths that get quantised, versioned and deprecated behind the scenes, where the only element of control you have is your messy monolithic user prompt. We want people who need their own intelligence to be able to not only access it, but also control it. And whilst the big general models are unbelievably good chatbots and coding agents and purveyors of the world, specialisation of intelligence is required. Clinical scribes, marketing compliance agents, legal red-lining models, insurance policy recommenders, the list goes on. And so that’s what Parsed does: deploy your own frontier model that actually learns. We eval your specific task, build a custom evaluation harness, optimise a model just for you, and host it with continual learning. We bake all the context and knowledge of your task into the model itself, from your engineers to your domain experts to customer feedback, all in a tight SFT → RL loop, with useful interpretability made possible by the open-source ecosystem we build on top of. No more 2000-word prompts with seventeen "IMPORTANT: NEVER DO X" clauses. Your model gets better at YOUR job every single day; the amnesiac pseudo-gods have had their run. Your model, your data, your moat. Let's build 🫡

English
6
1
6
2.6K
Mudith Jayasekara retweetledi
Mudith Jayasekara retweetledi
Mudith Jayasekara
Mudith Jayasekara@mudithj·
Being able to take gradient steps on 1T+ sized models at long sequence lengths isn’t trivial and all the open source libraries start to break down when pushed. Baseten Loops does the hard work to simplify the gradient update to a couple of lines of code. We want ML teams to not worry about the infra + training library, and spend their time looking at their data and reward shaping. Loops gives everyone the ability to do frontier RL robustly and then deploy using Baseten’s inference stack to make the model go brrrr and all the 9s of uptime. At Baseten research, we’re just getting started. Online RL and ultra long context training coming soon...
Raymond Cano@vim_dzl

x.com/i/article/2052…

English
0
1
26
1.5K
Lachy Groom
Lachy Groom@lachygroom·
codex + computer use + @baseten is a magical experience for deploying OSS models in minutes, entirely hands off after initial prompt
English
7
16
325
34.9K
Mudith Jayasekara
Mudith Jayasekara@mudithj·
Finding an intermediate memory layer between the full KV cache and lossy compression methods like natural language memory files is essential for real human work. The real human work that will be done by long horizon agentic workflows. This is some of the most exciting work we've done at Baseten yet, phase 2 and 3 coming soon.
Charlie O'Neill@oneill_c

x.com/i/article/2039…

English
1
1
9
798
Mudith Jayasekara retweetledi
Harry Partridge
Harry Partridge@part_harry_·
Human-in-the-loop RL is necessarily done at group size 1; you cannot do a group of rollouts with only one human. i.e. there is no baseline for you to subtract for each input prompt. This is by far the most interesting and under-discussed part of this announcement. The same was true for their tab-completions model. From the wording in their posts, it sounds like they are using plain REINFORCE (no mention of value functions) with a large batch size + re-evaluating each checkpoint to guard against high variance. Cursor is implicitly revealing an important empirical result: with a large enough batch size, simple REINFORCE just works, no baseline needed. In other words, large scale continual learning is solved.
Cursor@cursor_ai

Earlier this week, we published our technical report on Composer 2. We're sharing additional research on how we train new checkpoints. With real-time RL, we can ship improved versions of the model every five hours.

English
12
23
255
41K
Charlie O'Neill
Charlie O'Neill@oneill_c·
Worst feeling in the world is when Claude goes "Ah, the smoking gun" (it almost certainly isn't)
English
1
0
12
866
hallerite
hallerite@hallerite·
can labs please for the love of all that is holy use chat templates that do not rewrite history? how are we still doing this in 2026..
English
2
0
52
3.2K