Dan Roy

27.7K posts

Dan Roy banner
Dan Roy

Dan Roy

@roydanroy

@Google DeepMind. On leave, Canada CIFAR AI Chair and Former Research Director, @VectorInst. Professor, @UofT (Statistics/CS). Views are my own.

London Katılım Haziran 2009
1.9K Takip Edilen63.9K Takipçiler
Sabitlenmiş Tweet
Dan Roy
Dan Roy@roydanroy·
I'm an AI researcher.
Dan Roy tweet media
English
27
81
738
294.9K
Dan Roy retweetledi
Tony Feng
Tony Feng@tonylfeng·
A few months ago I bumped into Anand Patel, who had been my algebraic geometry TA in college, visiting Google DeepMind. He agreed to try out an agent I was building called Aletheia. Fast forward: Anand prompted Aletheia to solve a problem about simplicity of the Hodge bundle on M_g that had been floating around (a part of) the algebraic geometry community for at least ten years. Check out his paper at arxiv.org/pdf/2603.19052
Tony Feng tweet media
English
6
41
322
34.8K
Dan Roy retweetledi
Dwarkesh Patel
Dwarkesh Patel@dwarkesh_sp·
The Terence Tao episode. We begin with the absolutely ingenious and surprising way in which Kepler discovered the laws of planetary motion. People sometimes say that AI will make especially fast progress at scientific discovery because of tight verification loops. But the story of how we discovered the shape of our solar system shows how the verification loop for correct ideas can be decades (or even millennia) long. During this time, what we know today as the better theory can often actually make worse predictions (Copernicus's model of circular orbits around the sun was actually less accurate than Ptolemy's geocentric model). And the reasons it survives this epistemic hell is some mixture of judgment and heuristics that we don’t even understand well enough to actually articulate, much less codify into an RL loop. Hope you enjoy! 0:00:00 – Kepler was a high temperature LLM 0:11:44 – How would we know if there’s a new unifying concept within heaps of AI slop? 0:26:10 – The deductive overhang 0:30:31 – Selection bias in reported AI discoveries 0:46:43 – AI makes papers richer and broader, but not deeper 0:53:00 – If AI solves a problem, can humans get understanding out of it? 0:59:20 – We need a semi-formal language for the way that scientists actually talk to each other 1:09:48 – How Terry uses his time 1:17:05 – Human-AI hybrids will dominate math for a lot longer Look up Dwarkesh Podcast on YouTube, Apple Podcasts, or Spotify.
English
75
353
2.7K
418.9K
Dan Roy
Dan Roy@roydanroy·
@littmath Have your OpenClaw publish it under their name.
English
0
0
2
111
Daniel Litt
Daniel Litt@littmath·
@roydanroy Yes, this is an option. Probably it would be more like 5 pages to make it readable, which means unfortunately it will take at least a few weeks to get to it.
English
1
0
4
557
Daniel Litt
Daniel Litt@littmath·
Sometimes you solve an "open problem" and the solution is just not very interesting (e.g. it would be cool if a conjecture was true, since that would imply something else important, but it turns out not to be true). Usually this doesn't result in a publication, but rather just an email to the person who originated the problem. This is in part because it takes a lot of work to write something publication-quality (not to mention making work for other people who have to referee it), and arguably that's not worth it for results of limited interest. On the other hand, it would be good to have a way to disseminate such solutions so that other people don't waste effort on them. Worth keeping this in mind when looking at AI-generated solutions to open problems too. Not all of the examples of such thus far have this nature, but many of them do.
English
18
5
378
27K
Dan Roy
Dan Roy@roydanroy·
@bremen79 @ashtiani_hassan If you're going to write about generalization bounds, please chat with me because I think people are talking about 5 different things.
English
0
0
0
70
Francesco Orabona
Francesco Orabona@bremen79·
While preparing my new blog post, I found this underrated gem.
Francesco Orabona tweet mediaFrancesco Orabona tweet media
English
5
2
84
9.9K
Dan Roy
Dan Roy@roydanroy·
@steve47285 @AdaptiveAgents I've not thought about your problem enough to comment. I was simply reacting to Dr Ortega's comment, which seemed to raise a technical hurdle, but I don't think it's a roadblock.
English
1
0
2
39
Steven Byrnes
Steven Byrnes@steve47285·
@roydanroy @AdaptiveAgents Capacity constraints are not important to me. Yes I mentioned them in the post, but just once, tangentially. Maybe I should have left it out entirely to avoid giving the wrong impression. ¯\_(ツ)_/¯
English
1
0
0
60
Dan Roy
Dan Roy@roydanroy·
The prior is centered at the initialization. The (isotropic) variance is chosen by union bound, but I suspect it could be chosen also from an independent sample, hence could be viewed as distribution dependent. My point really is: if you want to explain generalization, you don't need an empirical bound. A generalization bound that depends on the distribution is fine. But then empirical estimates are fair game. I like our most recent paper which is built from PAC-Bayes bounds too. openreview.net/forum?id=FDUfA… IT doesn't attempt to explain SGD really, but it is a theory of the Gibbs posterior.
English
0
0
0
64
Dan Roy retweetledi
Michal Valko
Michal Valko@misovalko·
“Just add more agents” is not a theory of learning. Communication is! 🤝 You can hear tons of AI bros say multi-agent systems are just parallelization: spawn a swarm, split the work, and coordination will magically emerge ✨. In game-like settings 🎮, the math is blunt: if agents can share rich feedback (what would have happened under different actions), they can learn fast ⚡. If they can’t, and each only sees the score from the move, they actually played → learning is inevitably slower 🐢. We just closed this story for two-player zero-sum games, think rock–paper–scissors 🪨📄✂️, but harder. Even in the “no communication / only your own outcome” setup, we now have an algorithm that learns as fast as is theoretically possible, for the final strategy you actually deploy → not just an average! arXiv preprint coming soon → these are exactly the kinds of optimal algorithms we’re using our neolab that our pushing out multi-agent systems to the next level 🔥🔥🔥 w/ Côme Fiegel, Pierre Ménard, @Tdash_Koz, @VianneyPerchet
Michal Valko tweet media
Paris, France 🇫🇷 English
0
4
41
4.5K
Alexander Terenin
@roydanroy No idea - I am not the one that flagged this or made any decision. I just got an email that said person's papers are getting rejected for policy violation and that I should consider being more skeptical of their reviews.
English
1
0
9
5.2K
Alexander Terenin
Sad times as an ICML AC: I've just learned that a mid-career faculty whose work I know and respect is getting their papers rejected for agreeing to the no-LLM reviewing policy, and then violating it.
English
8
0
131
30.3K
Pedro A. Ortega
Pedro A. Ortega@AdaptiveAgents·
Thank you for the thoughtful piece Steven. I assume your blog post was (partially) inspired by my paper on universal imitation. If I understand correctly, the core argument is that a real-world imitation learner cannot continuously (i.e. forever) improve its policy because of a capacity constraint? I agree. But an RL agent has to learn a policy too (by updating the hypothesized value function given instantaneous rewards). There's also a capacity constraint here; not all policies are realizable. Wouldn't that imply that in real-world implementations an RL agent can't continuously improve either? Apologies in advance if I misunderstood your argument!
English
2
0
4
1K
Surya Ganguli
Surya Ganguli@SuryaGanguli·
Our new paper: "Solving adversarial examples requires solving exponential misalignment", expertly lead by @AleSalvatore00 w/ @stanislavfort arxiv.org/abs/2603.03507 Key idea: We all want to align AI systems to human values and intentions. We connect adversarial examples to AI alignment by showing they are a prototypical but exponentially severe form of misalignment at the level of perception. The fact that adversarial examples remain unsolved for over a decade thus serves as a cautionary tale for AI alignment, and provides new impetus for revisiting them. We shed light on why adversarial examples exist and why they are so hard to remove by asking a basic question: what is the dimensionality of neural network concepts in image space? For ResNets, and CLIP models, we show that neural network concepts (the space of images the network confidently labels as a concept) fill up almost the ENTIRE space of images (~135,000 dimensions out of ~150,000 for ImageNet & ~3000 out of 3072 for CIFAR10). In contrast natural image concepts are only ~20 dimensional. This indicates exponential misalignment between brain and machine perception (neural networks perceive exponentially many images as belonging to a concept that humans never would). This also explains why adversarial examples exist: if a concept fills up almost all of image space, ANY image will be close to that concept manifold. We further do experiments across > 20 networks showing that adversarial robustness inversely relates to concept dimensionality, though the most robust networks do not completely align machine and human perception. Overall the curse of dimensionality raises its ugly head as an impediment to both adversarial examples and alignment: if can be difficult to get AI systems to behave in accordance with human intentions, values, or perceptions over an exponentially large space of inputs. See @AleSalvatore00's excellent thread for more details: x.com/AleSalvatore00…
Surya Ganguli tweet media
English
9
25
183
14.8K
Dan Roy
Dan Roy@roydanroy·
Meh. I don't like the text you've quoted here. They should read our paper In Search of Robust Measures of Generalization. arxiv.org/abs/2010.11924 They seem to be confused about the role of the prior, how it can depend on the data distribution, and how data dependency can operate as a stand in for this.
English
1
1
4
1.4K
Dan Roy
Dan Roy@roydanroy·
@lreyzin @thegautamkamath I like Dwork for the Turing. Transformer would be ripe already, IMO, but not sure how'd they split it.
English
0
0
1
248
Gautam Kamath
Gautam Kamath@thegautamkamath·
The ACM Turing Award will be announced tomorrow. Any guesses on who won it?
English
32
7
185
68.1K
Francesco Orabona
Francesco Orabona@bremen79·
Reviewing for COLT is sad because the majority of the papers are: - correct&rigorous - interesting - solving challenging problems So, everything is so much better than the average NeurIPS/ICML/ICLR submission. Yet, we reject ~70% of them because these are the rules of the game
English
11
6
192
24.1K
Dan Roy
Dan Roy@roydanroy·
Anyone notice that Gen Beta says “search it up”, not “look it up” or “search for it”?
English
2
0
7
3.5K
Idan Attias
Idan Attias@IdanAttias·
Tell me whether you say covariates, features, predictors, or independent/explanatory/control variables, and I’ll tell you who you are
English
1
0
10
1.4K