Daniel Whettam

110 posts

Daniel Whettam banner
Daniel Whettam

Daniel Whettam

@DWhettam

ML Engineer | AI/Vision PhD

Bristol, England Beigetreten Mayıs 2018
1.3K Folgt125 Follower
Daniel Whettam retweetet
gavin leech (Non-Reasoning)
New paper on a long-shot I've been obsessed with for a year: How much are AI reasoning gains confounded by expanding the training corpus 10000x? How much LLM performance is down to "local" generalisation (pattern-matching to hard-to-detect semantically equivalent training data)?
gavin leech (Non-Reasoning) tweet mediagavin leech (Non-Reasoning) tweet media
English
32
131
962
222.3K
Daniel Whettam retweetet
Alan Jeffares
Alan Jeffares@Jeffaresalan·
Our new ICML 2025 oral paper proposes a new unified theory of both Double Descent and Grokking, revealing that both of these deep learning phenomena can be understood as being caused by prime numbers in the network parameters 🤯🤯 🧵[1/8]
Alan Jeffares tweet media
English
13
75
942
130.5K
Daniel Whettam retweetet
Kevin Nejad
Kevin Nejad@kevin_nejad·
After a year of review, our paper is now out in @NatureComms ! I really think our model/theory offers one of the most promising frameworks yet for learning in neocortical circuits - perhaps not in its final form, but the core principle of ssl feels right. Hoping experimentalists will test it, break it, and computationalists will refine it #AllModelsAreWrongButSomeAreUseful - I think this one could be very useful😀
Rui Ponte Costa@somnirons

Pleased to say that our story of how a theory of self-supervised learning in cortical layers accounts for several experimental observations is now out 🎉 nature.com/articles/s4146…

English
1
4
22
1.2K
Daniel Whettam retweetet
Andrej Karpathy
Andrej Karpathy@karpathy·
Good post from @balajis on the "verification gap". You could see it as there being two modes in creation. Borrowing GAN terminology: 1) generation and 2) discrimination. e.g. painting - you make a brush stroke (1) and then you look for a while to see if you improved the painting (2). these two stages are interspersed in pretty much all creative work. Second point. Discrimination can be computationally very hard. - images are by far the easiest. e.g. image generator teams can create giant grids of results to decide if one image is better than the other. thank you to the giant GPU in your brain built for processing images very fast. - text is much harder. it is skimmable, but you have to read, it is semantic, discrete and precise so you also have to reason (esp in e.g. code). - audio is maybe even harder still imo, because it force a time axis so it's not even skimmable. you're forced to spend serial compute and can't parallelize it at all. You could say that in coding LLMs have collapsed (1) to ~instant, but have done very little to address (2). A person still has to stare at the results and discriminate if they are good. This is my major criticism of LLM coding in that they casually spit out *way* too much code per query at arbitrary complexity, pretending there is no stage 2. Getting that much code is bad and scary. Instead, the LLM has to actively work with you to break down problems into little incremental steps, each more easily verifiable. It has to anticipate the computational work of (2) and reduce it as much as possible. It has to really care. This leads me to probably the biggest misunderstanding non-coders have about coding. They think that coding is about writing the code (1). It's not. It's about staring at the code (2). Loading it all into your working memory. Pacing back and forth. Thinking through all the edge cases. If you catch me at a random point while I'm "programming", I'm probably just staring at the screen and, if interrupted, really mad because it is so computationally strenuous. If we only get much faster 1, but we don't also reduce 2 (which is most of the time!), then clearly the overall speed of coding won't improve (see Amdahl's law).
Balaji@balajis

AI PROMPTING → AI VERIFYING AI prompting scales, because prompting is just typing. But AI verifying doesn’t scale, because verifying AI output involves much more than just typing. Sometimes you can verify by eye, which is why AI is great for frontend, images, and video. But for anything subtle, you need to read the code or text deeply — and that means knowing the topic well enough to correct the AI. Researchers are well aware of this, which is why there’s so much work on evals and hallucination. However, the concept of verification as the bottleneck for AI users is under-discussed. Yes, you can try formal verification, or critic models where one AI checks another, or other techniques. But to even be aware of the issue as a first class problem is half the battle. For users: AI verifying is as important as AI prompting.

English
134
532
4.4K
843.7K
dr. jack morris
dr. jack morris@jxmnop·
excited to finally share on arxiv what we've known for a while now: All Embedding Models Learn The Same Thing embeddings from different models are SO similar that we can map between them based on structure alone. without *any* paired data feels like magic, but it's real:🧵
dr. jack morris@jxmnop

this is sick all i'll say is that these GIFs are proof that the biggest bet of my research career is gonna pay off excited to say more soon

English
124
595
6.2K
908.4K
Daniel Whettam
Daniel Whettam@DWhettam·
@jxmnop Why do you say it's not self-supervised? There are many SSL tasks where the self-supervision is the task of interest. Language modelling is one of those IMO. Self supervision is supervised training where the supervision comes from the data (hence "self")
English
0
0
1
121
dr. jack morris
dr. jack morris@jxmnop·
language modeling is not unsupervised learning. it is not (and please stop saying this) self-supervised learning. next-token prediction is textbook _supervised_ learning, and i will die on this hill
English
73
27
735
74.1K
Starbeamrainbowlabs
Starbeamrainbowlabs@SBRLabs·
PhD Viva == complete! It was very intense and stressful, and I'm not in a hurry to repeat the experience :P Result is pass with corrections, but precisely what that means exactly hasn't been defined yet. Thanks to everyone who has supported me so far :-) #PhD #Viva #INeedANap
English
3
0
9
754
Cian Eastwood
Cian Eastwood@CianEastwood·
I've completed my PhD and joined @valence_ai (part of @recursionpharma) as a Senior Research Scientist! Excited to study generalization and representation learning (causal / multimodal) in the context of drug discovery!
Cian Eastwood tweet media
English
22
5
308
24.5K
Daniel Whettam retweetet
gavin leech (Non-Reasoning)
New paper: a big 90-page intro to AI and its likely effects from ten perspectives, ten camps. The whole gamut: ML, scientific applications, social applications, access, safety and alignment, economics, AI ethics, governance, and classical philosophy of life. 1/18
gavin leech (Non-Reasoning) tweet media
English
6
70
376
60.5K
Daniel Whettam retweetet
Kipp Freud
Kipp Freud@kipp_freud·
(1) Good news! I've had a paper accepted (with @cian_neuro, @nathanlepora, and Matt W. Jones), and I'll be giving a talk on it at @AAMASconf this year 🥳🥳🥳🧠🤖🐀🥳🥳🥳
English
1
3
10
1.1K
Joe Whettam
Joe Whettam@iamJoeWhettam·
New ideas are always the fruit of old ideas reimagined.
English
1
0
1
49
Xenon
Xenon@vertinski·
sooo, looking at LK-99, could we encounter a freak neural net which by chance has trained super-effectively?..... like a 3B model that is on par with GPT-4 🤔
English
1
0
2
3.4K
Wafa Johal
Wafa Johal@wafajohal·
It's happening! We received our first rejection review written by ChatGPT for a conference paper we submitted earlier this year. I am now looking for policy statements from @TheOfficialACM or @ieeeras about the use of LLMs to write reviews. Any pointers?
English
5
5
42
16.7K
Daniel Whettam retweetet
Jacob Chalk
Jacob Chalk@JacobChalkie·
Very excited the EPIC-SOUNDS dataset is finally released! We’ve all worked incredibly hard on this and I can be proud to say that this is my first publication! Looking forward to what this dataset can bring to the deep learning community!
Dima Damen@dimadamen

EPIC-SOUNDS is now public available on ArXiv: arxiv.org/abs/2302.00646 w Trailer: youtu.be/w-Bxat3Cgpk Annotations: github.com/epic-kitchens/… Challenge: github.com/epic-kitchens/… Work by @huh_jaesung @JacobChalkie w @e_kazakos, myself and AZ Thanks @_akhaliq for sharing

English
0
1
9
740
Daniel Whettam retweetet
Dima Damen
Dima Damen@dimadamen·
📢 Now Open For Submissions - all EPIC-KITCHENS Leaderboards for @CVPR #CVPR2023 Challenges. Winners announced at Joint @ego4_d and EPIC workshop: sites.google.com/view/ego4d-epi… **Nine** open challenges inc. 4 new ones (see 🧵) Leaderboards close 1st of June 2023. 🧵 1/7
Dima Damen tweet media
English
1
16
34
22K
Daniel Whettam retweetet
Jacob Chalk
Jacob Chalk@JacobChalkie·
For anyone interested in audio-visual learning, this challenge will be of interest! EPIC-100 and EPIC-SOUNDS is a step towards multi-modal methods where one modality does not rely on the timestamps or label set of another! We look forward to seeing what this challenge brings!
Dima Damen@dimadamen

*New* EPIC-SOUNDS Audio Interaction Recognition. Recognise sounds of garlic peeling 🔊 or hand mixer. New Audio-Only dataset on #EPIC_KITCHENS [Annotations out - ArXiv soon] Leads: @huh_jaesung @JacobChalkie Code: github.com/epic-kitchens/… Leaderboard: #results" target="_blank" rel="nofollow noopener">codalab.lisn.upsaclay.fr/competitions/9… 5/7

English
0
1
11
1.2K