Dan Roy

27.9K posts

Dan Roy banner
Dan Roy

Dan Roy

@roydanroy

@Google DeepMind. On leave, Canada CIFAR AI Chair and Former Research Director, @VectorInst. Professor, @UofT (Statistics/CS). Views are my own.

London Katılım Haziran 2009
1.9K Takip Edilen65.1K Takipçiler
Sabitlenmiş Tweet
Dan Roy
Dan Roy@roydanroy·
I'm an AI researcher.
Dan Roy tweet media
English
26
81
746
303.5K
JS
JS@SovereignJap·
@roydanroy They could have simply added a AI slop rating (just like likes on social media platforms) which requires someone to sign in and mark it as ai slop as a anonymous reader.
English
3
0
12
1.4K
Dan Roy
Dan Roy@roydanroy·
There's a lot of controversy brewing around arXiv's decision to penalize authors who post unchecked AI generated content. The impulse is correct, IMO, simply on grounds of efficiency: it is much cheaper to insist the authors vet their work first, rather than distributing the cost of that work to EVERY reader/agent who subsequently downloads the work. I believe the mechanism is likely the wrong one, however. Unfortunately, suggestions to use github are even worse, IMO, because they lose the (effective) immutability of the scientific record, which arXiv upholds.
English
21
3
139
15.3K
Dan Roy
Dan Roy@roydanroy·
@BlackHC No one can fault you for not standing by your principles, that's for sure.
English
1
1
20
6.8K
Andreas Kirsch 🇺🇦
In a personal capacity, and not on behalf of Google or Google DeepMind: very glad I could speak with FAZ and SZ about Google's reported classified AI deal with the Pentagon. Autonomous weapons, mass surveillance, and autonomous policing all concern public health and safety. They are matters of enormous public interest, not just internal policy debates. Companies that should know better are using safety-washed PR that undermines the honest public conversation we need about AI in classified settings.
Frankfurter Allgemeine gesamt@FAZ_NET

Google hat sich in einem KI-Abkommen mit dem Pentagon auf Konditionen eingelassen, die Anthropic abgelehnt hat. Ein Gespräch mit einem Mitarbeiter, der sich für sein Unternehmen schämt. faz.net/aktuell/wirtsc…

English
10
27
146
25.2K
Dimitris Papailiopoulos
Dimitris Papailiopoulos@DimitrisPapail·
Found myself posting papers to GitHub instead of arXiv lately. No gatekeeping, is in the same repo as the code, one link for everything, and gets uploaded immediately. Makes you wonder what arXiv's actual value is.
English
62
35
751
82.5K
Dan Roy retweetledi
Thomas G. Dietterich
Thomas G. Dietterich@tdietterich·
Examples of incontrovertible evidence: hallucinated references, meta-comments from the LLM ("here is a 200 word summary; would you like me to make any changes?"; "the data in this table is illustrative, fill it in with the real numbers from your experiments") end/
English
21
50
1.1K
64.3K
Dan Roy retweetledi
Krzakala Florent
Krzakala Florent@KrzakalaF·
What if a theory of deep learning could be built from iterated kernel spectral methods? Feature learning, advantage of depth, emergence of concepts, convnets filters.... and a new backprop-free algorithm too! We have it all! Introducing Neural LoFi 🧵 arxiv.org/abs/2605.13612
English
5
60
355
33.6K
(((ل()(ل() 'yoav))))👾
"I've been doing AI for 20 years and ..." and nothing. LLMs are new. LLM-Agents are new. our 20+ years experience with AI/ML/NLP may be marginally useful for understanding aspects of their training, but thats about it. we need new tools and experiences. we dont deserve authority.
English
29
31
401
23K
Corey
Corey@CoreyLeander·
@roydanroy Yeah but this isn’t one, sadly! This is coming from discredited Harold White: en.wikipedia.org/wiki/Harold_G.… If we were getting free energy from the vacuum, I promise you it’d be a bigger deal
English
1
0
0
28
John Baldwin
John Baldwin@JohnBal14046363·
@roydanroy @octonion Actual mathematician here. Unless we also change our mental architectures, there's limited speedup possible without losing necessary ingredients for deeply understanding difficult things.
English
1
0
1
26
Dan Roy
Dan Roy@roydanroy·
The next era of mathematics will be owned by those who adapt. I'm not sure how long this era will last.
English
11
9
162
19.8K
Aditi Raghunathan
Aditi Raghunathan@AdtRaghunathan·
It's one of the first lessons in ML: the model with the lowest train loss isn't the one that generalizes best. Pretraining made that easy to forget. You train for one epoch over trillions of tokens, there's no traditional overfitting, and pretrain loss starts to feel like the whole story. Our paper argues it isn't. The lowest-loss model isn't the best starting point for post-training. An old sharp-vs-flat lesson, back in a new regime.
Ishaan Watts@IshaanWatts18

Spending billions to train the "best" base model? You might be optimizing the wrong thing! 🎯 We show that controlling sharpness during mid-training leads to over 35% less forgetting after fine-tuning / quantization... even when the base model itself gets worse. 🧵 Takeaways for pretraining: - Use SAM (Sharpness-Aware-Minimization) in the final steps (~10%) - Try much higher learning rates (yes, even ~10× larger) 1/9

English
2
7
142
21K
Dan Roy
Dan Roy@roydanroy·
@Prityush (1) I think it is inevitable. (2) It will, however, likely introduce some bias. The nature of this bias is likely not obvious.
English
0
0
4
683
Prityush bansal
Prityush bansal@Prityush·
@roydanroy Using AI to find problems in benchmarks used to evaluate AI Surely this will not have any downstream problems
English
1
0
0
784
Dan Roy
Dan Roy@roydanroy·
Friday: AI co-mathematician 48% on FrontierMath Tier 4 Monday:
Dan Roy tweet media
English
10
16
251
26.3K
Aaron Roth
Aaron Roth@Aaroth·
@roydanroy So the more group functions you include, the more permissive the measure is because you've got a larger basis in which to represent the mapping from features to labels.
English
1
0
1
129
Aaron Roth
Aaron Roth@Aaroth·
Recently we showed that the minimax optimal rate for multicalibration is T^{2/3}. But that doesn't mean you have to do that badly on all instances. We give an algorithm that can adapt to easy instances and get better rates while still being minimax optimal in the worst case.
Aaron Roth tweet media
English
1
3
38
4.1K