Ryan Bahlous-Boldi

86 posts

Ryan Bahlous-Boldi banner
Ryan Bahlous-Boldi

Ryan Bahlous-Boldi

@RyanBoldi

PhD student @MIT_CSAIL | Continual RL, Open-Endedness, Evolution

Cambridge, MA Katılım Nisan 2017
635 Takip Edilen387 Takipçiler
Ryan Bahlous-Boldi
Ryan Bahlous-Boldi@RyanBoldi·
Also relevant is the work on the "myth of the objective" by @kenneth0stanley and @joelbot3000 or some of our work on lexicase selection. Perhaps the best way to do well on next token prediction, and downstream tasks, is to not train on next token prediction!
English
0
0
2
92
Ryan Bahlous-Boldi retweetledi
Itamar Pres
Itamar Pres@PresItamar·
New paper: It's time to optimize for 🔁self-consistency 🔁 We’ve pushed LLMs to the limits of available data, yet failures like sycophancy and factual inconsistency persist. We argue these stem from the same assumption: that behavior can be specified one I/O pair at a time. 🧵
Itamar Pres tweet media
English
16
56
425
72.4K
Ryan Bahlous-Boldi retweetledi
Pulkit Agrawal
Pulkit Agrawal@pulkitology·
Introducing Perioperation: a new paradigm for collecting multimodal (vision-touch-proprioception) data for fine dexterous manipulation. See DEXOP in action -- an exoskeleton for capturing rich force and visual feedback as a human performs everyday tasks. Our design ensures such data easily transfers to a robot, unlocking historically hard tasks for robots. Find out more: dex-op.github.io Work led by @haoshu_fang
English
7
42
381
27.8K
Ryan Bahlous-Boldi retweetledi
Sam Earle
Sam Earle@Smearle_RH·
We introduce PuzzleJAX, a benchmark for reasoning and learning. 🧩💡🦎 PuzzleJAX compiles hundreds of existing grid-based PuzzleScript games to hardware-accelerated JAX environments, and allows researchers to define new tasks via PuzzleScript's concise rewrite rule-based DSL.
GIF
English
5
40
178
34.1K
Ryan Bahlous-Boldi retweetledi
Lance Ying
Lance Ying@LanceYing42·
A hallmark of human intelligence is the capacity for rapid adaptation, solving new problems quickly under novel and unfamiliar conditions. How can we build machines to do so? In our new preprint, we propose that any general intelligence system must have an adaptive world model, i.e. they must be able to rapidly construct or refine their internal representation through interaction and exploration — a process we call “world model induction”. We propose a roadmap for evaluating adaptive world models in machines based on a special class of games we call “novel games”.
Lance Ying tweet media
English
15
100
511
69K
Ryan Bahlous-Boldi retweetledi
Tyler Brooke-Wilson
Tyler Brooke-Wilson@T_BrookeWilson·
How do people reason while still staying coherent – as if they have an internal ‘world model’ for situations they’ve never encountered? A new paper on open-world cognition (preview at the world models workshop at #ICML2025!)
Tyler Brooke-Wilson tweet media
English
4
26
147
19.9K
connor casey
connor casey@_connor_casey·
at umass the function of comp sci research converges @ rl....even in quantum @RyanBoldi
English
1
0
2
206
Ryan Bahlous-Boldi
Ryan Bahlous-Boldi@RyanBoldi·
Excited to share that I’ll be joining @MIT this fall as a PhD student in EECS! Grateful to everyone that has supported me along the way. Can’t wait to explore RL, evolution, learning, reasoning and intelligence in all its forms 🧠
English
9
0
175
11.6K
Ryan Bahlous-Boldi retweetledi
François Chollet
François Chollet@fchollet·
Much of the field obsesses over end-to-end learning. But strong generalization requires compositionality: building modular, reusable abstractions, and reassembling them on the fly when faced with novelty. The models of the future won't be just pipes, they will be Lego castles.
English
59
163
1.3K
136.6K