Jacob C Tanner

216 posts

Jacob C Tanner

Jacob C Tanner

@JacobCTanner1

PhD Student- Complex Networks & Systems and Cognitive Science, Indiana University

Katılım Temmuz 2019
272 Takip Edilen90 Takipçiler
Jacob C Tanner
Jacob C Tanner@JacobCTanner1·
The murder of Charlie Kirk is a tragedy. I’m deeply saddened by this. We need to learn how to talk to each other again.
English
0
0
1
396
Jacob C Tanner retweetledi
The Culturist
The Culturist@the_culturist_·
The Lord of the Rings does not take place on an imaginary planet — it's Earth. Middle-earth is our forgotten past, before recorded history, when Eden (Valinor) was a real place. The truth of Tolkien's world will blow your mind... 🧵
The Culturist tweet media
English
405
2.8K
24.8K
2.3M
Jacob C Tanner retweetledi
Core Francisco Park @ NeurIPS2025
New paper! “In-Context Learning of Representations” What happens to an LLM’s internal representations in the large context limit? We find that LLMs form “in-context representations” to match the structure of the task given in context! 1/n
GIF
English
20
180
1.1K
136.1K
Jacob C Tanner retweetledi
The Culturist
The Culturist@the_culturist_·
The fall of Rome is widely misunderstood. It wasn't invasion, disease or famine that truly brought it to its knees. Rome collapsed because the birth rate did… (thread) 🧵
The Culturist tweet media
English
549
3K
16.4K
2.9M
Jacob C Tanner retweetledi
All The Right Movies
All The Right Movies@ATRightMovies·
LOTR: THE FELLOWSHIP OF THE RING was released 23 years ago this week. An adaptation of Tolkien’s classic novel, and the first entry in Peter Jackson’s The Lord of the Rings trilogy, the story of how it was made is proof that one does not simply walk into Mordor… 1/76
All The Right Movies tweet mediaAll The Right Movies tweet mediaAll The Right Movies tweet mediaAll The Right Movies tweet media
English
89
2.1K
15.5K
1.7M
Jacob C Tanner retweetledi
Jonathan Gorard
Jonathan Gorard@getjonwithit·
The apparent "philosophical problems" of quantum mechanics are not unique to QM at all: they are in fact the same problems that arise whenever one attempts to construct an abstract model of reality. We can see these problems already in high school-level mechanics. (1/14)
Jonathan Gorard tweet mediaJonathan Gorard tweet media
English
111
735
7.4K
767.9K
Jacob C Tanner retweetledi
Stéphane Deny
Stéphane Deny@StphTphsn1·
Very interesting! how does it relate to this work, studying phase transitions in the dynamics of diffusion? nature.com/articles/s4146…
English
0
26
144
7.8K
Jacob C Tanner retweetledi
The Culturist
The Culturist@the_culturist_·
Past societies produced so much beauty because they knew that math and beauty are deeply connected. It all started when Pythagoras discovered something mind-blowing about reality: The universe is not made of matter — but music... (thread) 🧵
The Culturist tweet media
English
517
4.9K
28K
2.3M
Jacob C Tanner retweetledi
frank
frank@frank_liquid·
Doubling o1-preview performance on ARC-AGI with one simple trick 🚀 tldr: by providing human-like representations to o1, we are able to substantially increase performance on @arcprize.
frank tweet mediafrank tweet media
English
23
85
990
217.1K
Jacob C Tanner retweetledi
James Zou
James Zou@james_y_zou·
📢Thrilled to introduce the #VirtualLab: a team of AI scientist agents (AI chemist, AI reviewer...). Virtual Lab is led by an AI professor w/ feedback from human scientist. The Lab created new nanobodies that we experimentally validated to bind to recent #covid variants🚀🧵
James Zou tweet media
English
21
180
818
190.4K
Jacob C Tanner retweetledi
Oliver Habryka
Oliver Habryka@ohabryka·
I compiled all the emails released as part of the Musk v. Altman lawsuit in chronological order (link in reply). IMO a really valuable read. Extremely consequential decisions made in these emails.
Oliver Habryka tweet media
English
56
231
3K
868.4K
Jacob C Tanner retweetledi
Paul Thompson
Paul Thompson@PTenigma·
If you are interested in Generative #AI, or statistical physics, you will know that you can use latent diffusion models to make synthetic images (or videos), but these methods are a bit slow (I explain here how the Fokker-Planck formulation and Langevin diffusion are related): youtube.com/watch?v=fOORfz… About an hour ago some Russian scientists from Skoltech posted a quite considerable leap forward in this field, computing these maps much faster + in one step: arxiv.org/pdf/2403.13117
YouTube video
YouTube
Paul Thompson tweet media
Paul Thompson@PTenigma

"This is just a stock image of a cat. Physicists are not allowed to own cats, for various reasons." (from a great lecture [1] on generative AI [2] by Jordan Cotler) [1] youtube.com/watch?v=bHLdAF… [2] This is a very clear explanation of denoising diffusion probabilistic models (DDPMs [3]), and how they arise in statistical physics (diffusion of smoke in air or ink in a fluid, drifts of sand dunes in the wind). [3] DDPMs and their compressed extension, LDMs, are used in text-based image generation (MidJourney, SORA, Kling). A trickier but very clear lecture is the one by the inventor of DDPMs, Jascha Sohl-Dickstein, also featuring his dog (!) youtube.com/watch?v=XCUlnH… [4] for an even deeper version, someone has posted an entire book on Twitter tonight, which is pretty amazing "Statistical Optimal Transport": arxiv.org/abs/2407.18163 #AI #DDPM

English
4
139
918
107.9K
Jacob C Tanner retweetledi
nature
nature@Nature·
“We’re very excited to see what people do with this” AlphaFold3 is open at last go.nature.com/4fHAfio
English
21
592
1.7K
128.1K
Jacob C Tanner retweetledi
Maksym Andriushchenko
Maksym Andriushchenko@maksym_andr·
🚨 So, why do we need weight decay in modern deep learning? 🚨 The camera-ready version of our NeurIPS 2024 paper is now on arXiv (a major update compared to the first version). Weight decay is traditionally viewed as a regularization method, but its effect in the overtraining regime is quite subtle and its interaction with the implicit regularization effect of SGD plays a crucial role. In the undertraining regime (e.g., in LLM pretraining), however, the effect of weight decay is totally different: it sets an implicit learning rate schedule for AdamW and enables stable training with bfloat16 precision. This explains why weight decay is still widely used for LLM training with standard optimizers, such as AdamW. This is joint work with @dngfra, @adityavardhanv, @tml_lab.
Maksym Andriushchenko tweet media
English
11
109
698
74.2K
Jacob C Tanner retweetledi
Machine Learning Street Talk
Machine Learning Street Talk@MLStreetTalk·
I finally got to meet @fchollet in person recently to interview him about @arcprize, intelligence vs memorization, human cognitive development, learning abstractions, limits of pattern recognition and consciousness development. These are the best bits. Full show released tomorrow
English
9
55
508
79K
Jacob C Tanner retweetledi
Alan Jeffares
Alan Jeffares@Jeffaresalan·
There are many things we don’t understand about deep learning. Our new NeurIPS paper (w/ @AliciaCurth) makes the mistake of trying to tackle too many of them 😅 A simplified model of deep learning describes double descent, grokking, gradient boosting & linear mode connectivity🧵
Alan Jeffares tweet media
English
15
130
750
110.2K