Carta Thomas

73 posts

Carta Thomas

Carta Thomas

@CartaThomas2

Ph.D student at INRIA in the @FlowersINRIA. I am working on the how language and RL interact

Katılım Şubat 2023
195 Takip Edilen88 Takipçiler
Carta Thomas
Carta Thomas@CartaThomas2·
Finally, a big thank you to the jury members for their time, questions, and insights: 👨‍🏫 Jean Ponce (president) 🧠 Reviewers @pierrelux & @GMartius 💡 @edwardfhughes & @white_martha for their excellent questions and engagement alongside my supervisors.
English
0
0
2
81
Carta Thomas
Carta Thomas@CartaThomas2·
After four incredible years in the Flowers AI & CogSci Lab, I’m thrilled to share that I’ve officially become a Doctor! 🎓💫 My thesis: “Language as a Cognitive Tool for Open Agents.” It explores how language helps artificial agents explore, adapt, and learn efficiently.
English
3
5
24
2.3K
Carta Thomas retweetledi
Zeyuan Allen-Zhu, Sc.D.
Zeyuan Allen-Zhu, Sc.D.@ZeyuanAllenZhu·
Phase 1 of Physics of Language Models code release ✅our Part 3.1 + 4.1 = all you need to pretrain strong 8B base model in 42k GPU-hours ✅Canon layers = strong, scalable gains ✅Real open-source (data/train/weights) ✅Apache 2.0 license (commercial ok!) 🔗github.com/facebookresear…
Zeyuan Allen-Zhu, Sc.D. tweet media
Zeyuan Allen-Zhu, Sc.D.@ZeyuanAllenZhu

(1/8)🍎A Galileo moment for LLM design🍎 As Pisa Tower experiment sparked modern physics, our controlled synthetic pretraining playground reveals LLM architectures' true limits. A turning point that might divide LLM research into "before" and "after." physics.allen-zhu.com/part-4-archite…

English
14
113
669
232.3K
Carta Thomas retweetledi
Edward Hughes
Edward Hughes@edwardfhughes·
The automation of innovation is within reach! Delighted that my @raais talk is now available for anyone to watch, alongside an excellent blogpost summary by the inimitable @nathanbenaich.
Nathan Benaich@nathanbenaich

"2025 is the year of open-endedness" at @raais, @edwardfhughes argued that we’re entering a new phase in the evolution of ai: one where open-endedness becomes the central organizing principle. Not just solving problems, but defining them. Not just predicting the next token, but surfacing previously unknown unknowns. If he’s right, the next generation of AI systems won’t just be tools, they’ll be participants in the scientific process itself.

English
0
17
51
16.9K
Carta Thomas retweetledi
Edward Hughes
Edward Hughes@edwardfhughes·
Human ideation beats AI ideation when measured on execution outcomes: arxiv.org/abs/2506.20803. There's a clear path to fixing this - work to be done!
English
0
6
44
4.1K
Carta Thomas retweetledi
Loris Gaven
Loris Gaven@LorisGaven·
🔔 Join our MAGELLAN talk on July 2! We'll explore how LLM agents can monitor their own learning progress and choose what to learn next, like curious humans 🤔 1h presentation + 1h Q&A on autotelic agents & more! 📅 July 2, 4:30 PM CEST 🎟️ forms.gle/1PC2fxJx1PZYfq…
Carta Thomas@CartaThomas2

🚀 Introducing 🧭MAGELLAN—our new metacognitive framework for LLM agents! It predicts its own learning progress (LP) in vast natural language goal spaces, enabling efficient exploration of complex domains.🌍✨Learn more: 🔗 arxiv.org/abs/2502.07709 #OpenEndedLearning #LLM #RL

English
1
3
16
1.8K
Carta Thomas retweetledi
Clément ROMAC @ ICML 2025
Clément ROMAC @ ICML 2025@ClementRomac·
📰 Check out the full paper here: arxiv.org/abs/2506.06725 Don't hesitate to reach out if you want to discuss WorldLLMs or RL for grounding LLMs! Or come see us at the poster session on Thursday!
English
1
1
3
127
Carta Thomas retweetledi
Clément ROMAC @ ICML 2025
Clément ROMAC @ ICML 2025@ClementRomac·
I'll be at RLDM this week to present our new paper: WorldLLM 😃 In the same spirit as our previous works — e.g. GLAM, MAGELLAN... — which investigate how to ground LLMs through interactions with external environments, WorldLLM takes a deep dive into the world modeling aspect.
English
2
6
18
1.9K
Carta Thomas retweetledi
hardmaru
hardmaru@hardmaru·
New Paper: Continuous Thought Machines 🧠 Neurons in brains use timing and synchronization in the way that they compute, but this is largely ignored in modern neural nets. We believe neural timing is key for the flexibility and adaptability of biological intelligence. We propose a new neural architecture, “Continuous Thought Machines” (CTMs), which is built from the ground up to use neural dynamics as a core representation for intelligence. By using neural dynamics as a first-class representational citizen, CTMs naturally perform adaptive computation. Many emergent, interesting behaviors arise as a result: CTMs solve mazes by observing a raw maze image and producing step-by-step instructions directly from its neural dynamics. When tasked with image recognition, the CTM naturally takes multiple steps to examine different parts of the image before making its decision. This step-by-step approach not only makes its behavior more interpretable but also improves accuracy: the longer it “thinks,” the more accurate its answers become. We also found that this allows the CTM to decide to spend less time thinking on simpler images, thus saving energy. When identifying a gorilla, for example, the CTM’s attention moves from eyes to nose to mouth in a pattern remarkably similar to human visual attention. I think this work underscores an important, yet often lost, synergy between neuroscience and AI. While modern AI is ostensibly brain-inspired, the two fields often operate in surprising isolation. By starting with such inspiration and iteratively following the emergent, interesting behaviors, we developed a model with unexpected capabilities, such as its surprisingly strong calibration in classification tasks, a feature that was not explicitly designed for. When we initially asked, “why do this research?”, we hoped the journey of the CTM would provide compelling answers. By embracing light biological inspiration and pursuing the novel behaviors observed, we have arrived at a model with emergent capabilities that exceeded our initial designs. We are committed to continuing this exploration, borrowing further concepts to discover what new and exciting behaviors will emerge, pushing the boundaries of what AI can achieve.
English
64
559
3.2K
256.9K