Adam Perold

125 posts

Adam Perold

Adam Perold

@AdamPerold

Beigetreten Haziran 2012
1.2K Folgt305 Follower
Mira Murati
Mira Murati@miramurati·
We have parted ways with Barret Zoph. Soumith Chintala will be the new CTO of Thinking Machines. He is a brilliant and seasoned leader who has made important contributions to the AI field for over a decade, and he’s been a major contributor to our team. We could not be more excited to have him take on this new responsibility.
English
251
136
4.2K
988.5K
Adam Perold retweetet
Pascale Fung
Pascale Fung@pascalefung·
Introducing VL-JEPA: Vision-Language Joint Embedding Predictive Architecture for streaming, live action recognition, retrieval, VQA, and classification tasks with better performance and higher efficiency than large VLMs. • VL-JEPA is the first non-generative model that can perform general-domain vision-language tasks in real-time, built on a joint embedding predictive architecture. • We demonstrate in controlled experiments that VL-JEPA, trained with latent space embedding prediction, outperforms VLMs that rely on data space token prediction. • We show that VL-JEPA delivers significant efficiency gains over VLMs for online video streaming applications, thanks to its non-autoregressive design and native support for selective decoding. • We highlight that our VL-JEPA model, with an unified model architecture, can effectively handle a wide range of classification, retrieval, and VQA tasks at the same time. by @Delong0_0 @MustafaShukor1 @TheoMoutakanni @willyhcchung Jade Lei Yu Tejaswi Kasarla @AllenBolourchi @ylecun @pascalefung arxiv.org/abs/2512.10942
English
13
87
558
89.3K
Qasar Younis
Qasar Younis@qasar·
After 16 years and 10 months -- I'm finally writing my first...X? Also, will start writing more regularly. My first post is, ironically, on why I'm posting. qy.co/new
English
321
185
3.6K
1.9M
Adam Perold retweetet
General Intuition
General Intuition@gen_intuition·
Yann LeCun (Chief AI Scientist, Meta, @ylecun), @PimDeWitte (CEO, General Intuition), and Aude Durand (Kyutai, @aude_drn), talk about world models, embodied agents, Yann's new company, and the limitations of LLMs 0:00 - Introduction to World Models 5:00 - Why World Models, Intuition & Introducing Yann's new company 10:00 - Architectures + Merging Language & Interaction Data towards General Agents 20:00 - Open Source, Sovereign AI & @kyutai_labs Partnership Keynote for #aiPULSE2025 at Station F in Paris 🇫🇷
English
29
126
635
109.4K
Robin Rombach
Robin Rombach@robrombach·
We published Latent Diffusion four years ago. Turning this lovely technology into a company has been, and continues to be, one of the most rewarding, interesting and plainly insane challenges I have personally ever worked on. The future is exciting: Visual models will become much smarter, faster, and multimodal by design. I am glad to be working on this with the best team on the planet. 🌲
Black Forest Labs@bfl_ml

We've raised $300M in Series B funding from Salesforce Ventures and Anjney Midha (AMP) FLUX is used by millions every month and powers production workflows across the world's leading platforms. This funding will allow us to invest deeply in research and build the foundations for visual intelligence.

English
41
21
416
78K
Andreas Blattmann
Andreas Blattmann@andi_blatt·
We've raised our $300M Series B. With this investment, we’re accelerating our research towards visual intelligence - from the Black Forest into the world! Huge shoutout to the team who fuels this development with their relentless and amazing work ♥️ Stay tuned - more coming🚀
Black Forest Labs@bfl_ml

We've raised $300M in Series B funding from Salesforce Ventures and Anjney Midha (AMP) FLUX is used by millions every month and powers production workflows across the world's leading platforms. This funding will allow us to invest deeply in research and build the foundations for visual intelligence.

English
34
11
383
67.7K
Soumith Chintala
Soumith Chintala@soumithchintala·
thinking machines....the people are incredible
English
148
74
3.3K
801.7K
Soumith Chintala
Soumith Chintala@soumithchintala·
😅 all of this is true. i owe a bunch to my mentors as well. my mentor at nyu was then phd student @psermanet, one of the kindest human beings i know Yann LeCun gave me opportunity 2x when i didn't have another option forward in AI. the person who pushed me to go to IIIT for my final year project and then go to cmu without a plan after my masters rejections was Praveen Garimella, gamechanging advice to not give up. props to my dad Vithal Chintala and mom Rajani Chintala for letting me be crazy and going beyond their financial means to support me. i grew up squarely middle class with lots of debt in the household, even though my parents eventually became independently wealthy post-2010 -- so letting me pursue this path when there were safer choices always available is great parenting! thanks @deedydas for spending time digging up the deets. I'm sure everyone who's sitting upon success has many struggles, life doesn't come easy. p.s.: i got rejected 3x from google and deepmind together, not just deepmind.
English
49
108
2.6K
82.6K
Deedy
Deedy@deedydas·
If you feel like giving up, you must read this never-before-shared story of the creator of PyTorch and ex-VP at Meta, Soumith Chintala. > from hyderabad public school, but bad at math > goes to a "tier 2" college in India, VIT in Vellore > rejected from all 12 universities for US masters despite 1420 on the GRE > fuckit.jpg > goes to the US anyway on a J-1 visa to CMU with no plan > applies for masters (again) to 15 universities > rejected from all except USC and with late admissions, NYU in 2010 > finds this guy called Yann LeCun (before he was famous) > starts getting into open source > rejected from all jobs including DeepMind > only job is Amazon as test engineer > his PhD mentor helps him get a job at a small startup (MuseAmi) > rejected from DeepMind > couldn't get H-1B because of J-1 home return issue; gets waiver through months of approval with USCIS and US State Dept > very low on confidence > In 2011/12 builds one of the fastest AI inference engines on phones > rejected from DeepMind > emailed Yann again and joins FAIR because of Torch7 open-source work > scrapes through bootcamp at Facebook, struggling on an HBase task > L8/L9 engineers at Facebook struggle to get ImageNet working > figures out numerics / hyperparam issue as an L4 > first big win! > FAIR goes well, runs 3 person torch7 team and co-creates PyTorch > because of politics, management wants to shut down PyTorch > cries-at-bar.jpg, literally > eventually some people save PyTorch and it launches in 2017 > gets a EB-1 green card! > the rest is history... Think about that. He went to a tier 2 college. Was rejected from all Masters programs 2x. Rejected from every single job except Amazon test engineering. Rejected from DeepMind 3x. Nearly had his baby project shut down. Struggled with visa issues. After 12 years of failures (2005-17), he eventually rose to became a VP at Meta one of the most influential people in AI! Soumith's story is one of resilience and he's living proof that no matter how down in the dumps you are, there's always hope.
Deedy tweet media
English
285
1.3K
11.2K
2M
Adam Perold
Adam Perold@AdamPerold·
Great decomposition @karpathy
Andrej Karpathy@karpathy

Finally had a chance to listen through this pod with Sutton, which was interesting and amusing. As background, Sutton's "The Bitter Lesson" has become a bit of biblical text in frontier LLM circles. Researchers routinely talk about and ask whether this or that approach or idea is sufficiently "bitter lesson pilled" (meaning arranged so that it benefits from added computation for free) as a proxy for whether it's going to work or worth even pursuing. The underlying assumption being that LLMs are of course highly "bitter lesson pilled" indeed, just look at LLM scaling laws where if you put compute on the x-axis, number go up and to the right. So it's amusing to see that Sutton, the author of the post, is not so sure that LLMs are "bitter lesson pilled" at all. They are trained on giant datasets of fundamentally human data, which is both 1) human generated and 2) finite. What do you do when you run out? How do you prevent a human bias? So there you have it, bitter lesson pilled LLM researchers taken down by the author of the bitter lesson - rough! In some sense, Dwarkesh (who represents the LLM researchers viewpoint in the pod) and Sutton are slightly speaking past each other because Sutton has a very different architecture in mind and LLMs break a lot of its principles. He calls himself a "classicist" and evokes the original concept of Alan Turing of building a "child machine" - a system capable of learning through experience by dynamically interacting with the world. There's no giant pretraining stage of imitating internet webpages. There's also no supervised finetuning, which he points out is absent in the animal kingdom (it's a subtle point but Sutton is right in the strong sense: animals may of course observe demonstrations, but their actions are not directly forced/"teleoperated" by other animals). Another important note he makes is that even if you just treat pretraining as an initialization of a prior before you finetune with reinforcement learning, Sutton sees the approach as tainted with human bias and fundamentally off course, a bit like when AlphaZero (which has never seen human games of Go) beats AlphaGo (which initializes from them). In Sutton's world view, all there is is an interaction with a world via reinforcement learning, where the reward functions are partially environment specific, but also intrinsically motivated, e.g. "fun", "curiosity", and related to the quality of the prediction in your world model. And the agent is always learning at test time by default, it's not trained once and then deployed thereafter. Overall, Sutton is a lot more interested in what we have common with the animal kingdom instead of what differentiates us. "If we understood a squirrel, we'd be almost done". As for my take... First, I should say that I think Sutton was a great guest for the pod and I like that the AI field maintains entropy of thought and that not everyone is exploiting the next local iteration LLMs. AI has gone through too many discrete transitions of the dominant approach to lose that. And I also think that his criticism of LLMs as not bitter lesson pilled is not inadequate. Frontier LLMs are now highly complex artifacts with a lot of humanness involved at all the stages - the foundation (the pretraining data) is all human text, the finetuning data is human and curated, the reinforcement learning environment mixture is tuned by human engineers. We do not in fact have an actual, single, clean, actually bitter lesson pilled, "turn the crank" algorithm that you could unleash upon the world and see it learn automatically from experience alone. Does such an algorithm even exist? Finding it would of course be a huge AI breakthrough. Two "example proofs" are commonly offered to argue that such a thing is possible. The first example is the success of AlphaZero learning to play Go completely from scratch with no human supervision whatsoever. But the game of Go is clearly such a simple, closed, environment that it's difficult to see the analogous formulation in the messiness of reality. I love Go, but algorithmically and categorically, it is essentially a harder version of tic tac toe. The second example is that of animals, like squirrels. And here, personally, I am also quite hesitant whether it's appropriate because animals arise by a very different computational process and via different constraints than what we have practically available to us in the industry. Animal brains are nowhere near the blank slate they appear to be at birth. First, a lot of what is commonly attributed to "learning" is imo a lot more "maturation". And second, even that which clearly is "learning" and not maturation is a lot more "finetuning" on top of something clearly powerful and preexisting. Example. A baby zebra is born and within a few dozen minutes it can run around the savannah and follow its mother. This is a highly complex sensory-motor task and there is no way in my mind that this is achieved from scratch, tabula rasa. The brains of animals and the billions of parameters within have a powerful initialization encoded in the ATCGs of their DNA, trained via the "outer loop" optimization in the course of evolution. If the baby zebra spasmed its muscles around at random as a reinforcement learning policy would have you do at initialization, it wouldn't get very far at all. Similarly, our AIs now also have neural networks with billions of parameters. These parameters need their own rich, high information density supervision signal. We are not going to re-run evolution. But we do have mountains of internet documents. Yes it is basically supervised learning that is ~absent in the animal kingdom. But it is a way to practically gather enough soft constraints over billions of parameters, to try to get to a point where you're not starting from scratch. TLDR: Pretraining is our crappy evolution. It is one candidate solution to the cold start problem, to be followed later by finetuning on tasks that look more correct, e.g. within the reinforcement learning framework, as state of the art frontier LLM labs now do pervasively. I still think it is worth to be inspired by animals. I think there are multiple powerful ideas that LLM agents are algorithmically missing that can still be adapted from animal intelligence. And I still think the bitter lesson is correct, but I see it more as something platonic to pursue, not necessarily to reach, in our real world and practically speaking. And I say both of these with double digit percent uncertainty and cheer the work of those who disagree, especially those a lot more ambitious bitter lesson wise. So that brings us to where we are. Stated plainly, today's frontier LLM research is not about building animals. It is about summoning ghosts. You can think of ghosts as a fundamentally different kind of point in the space of possible intelligences. They are muddled by humanity. Thoroughly engineered by it. They are these imperfect replicas, a kind of statistical distillation of humanity's documents with some sprinkle on top. They are not platonically bitter lesson pilled, but they are perhaps "practically" bitter lesson pilled, at least compared to a lot of what came before. It seems possibly to me that over time, we can further finetune our ghosts more and more in the direction of animals; That it's not so much a fundamental incompatibility but a matter of initialization in the intelligence space. But it's also quite possible that they diverge even further and end up permanently different, un-animal-like, but still incredibly helpful and properly world-altering. It's possible that ghosts:animals :: planes:birds. Anyway, in summary, overall and actionably, I think this pod is solid "real talk" from Sutton to the frontier LLM researchers, who might be gear shifted a little too much in the exploit mode. Probably we are still not sufficiently bitter lesson pilled and there is a very good chance of more powerful ideas and paradigms, other than exhaustive benchbuilding and benchmaxxing. And animals might be a good source of inspiration. Intrinsic motivation, fun, curiosity, empowerment, multi-agent self-play, culture. Use your imagination.

English
0
0
0
389
Anderson Mancini
Anderson Mancini@Andersonmancini·
We won 🤩🎉 We are thrilled to hear that the #threejs project for @levinriegner and F. Schumacher & Co. has been recognized as the leading Architecture, Art & Design website at the 29th Annual @TheWebbyAwards 😍 winners.webbyawards.com/2025/websites-… Thanks again to the incredible team at L+R, to everyone involved in this incredible project @neotix, and for those who voted for us 🙏🏻 Read more about the project here: levinriegner.com/news/schumache…
English
22
7
127
7.1K
Adam Perold
Adam Perold@AdamPerold·
There’s a nice feature in Chrome, where you just right click on the tab and select “Add tab to group,” then name the group, which you can do with a symbol like “+” or “.” for simplicity. A clever design is that you only need to do this twice. Once you have two tabs in a group displaying, any other tabs you drag between them automatically assign to the group. So you can group-select and assign 200+ tabs pretty effortlessly. I assume this wouldn’t recover tabs from a “remove browsing profile” action, which sounds like a privacy feature intended to fulfill the result you achieved. But I thought this was a nice solution to the tabs issue worth sharing.
English
0
0
0
82
Andrej Karpathy
Andrej Karpathy@karpathy·
Omg I didn't understand what it means to "remove a browsing profile" on Chrome. I thought it signs you out on Chrome app, but it destroyed all my open tabs and logged me out of everything 🤦‍♂️. My ~200 open tabs just... gone. Taking the opportunity to switch to Brave browser again.
English
934
191
7.7K
575.4K
Amjad Masad
Amjad Masad@amasad·
Whatever you need… make an app for that. Now on your phone. For everyone. Free.
English
856
1.8K
16.8K
5.5M
Adam Perold
Adam Perold@AdamPerold·
Open source winning means consumers and enterprise win, distribution and application layers win, we all build on top of these innovations together, advancing society, and it's fantastic to see companies across the world taking this approach regardless of country, but especially heartening to see in this context specifically.
Yann LeCun@ylecun

Nice job! Open research / open source accelerates progress.

English
0
0
1
580