Johnny Devriese

1K posts

Johnny Devriese banner
Johnny Devriese

Johnny Devriese

@johnnydevriese

AI Engineer 🤖2 wheeled enthusiast 🚲 pizzaiolo 🍕 | AI @stanford 🌲physics @wsu 🐆

Bay Area Se unió Haziran 2017
335 Siguiendo29 Seguidores
Johnny Devriese
Johnny Devriese@johnnydevriese·
@kepano was excited about this but seems like it's not working
Johnny Devriese tweet media
English
4
0
6
3.6K
kepano
kepano@kepano·
Defuddle now returns Youtube transcripts! Paste a YouTube link into defuddle.md to get a markdown transcript with timestamps, chapters, and pretty good diarization! ...or if you just want to read it, try the new Reader mode in Obsidian Web Clipper powered by Defuddle.
English
88
201
2.7K
323.5K
Johnny Devriese retuiteado
Stephan Rabanser
Stephan Rabanser@steverab·
📣 Excited to share my first work @Princeton : 𝗧𝗼𝘄𝗮𝗿𝗱𝘀 𝗮 𝗦𝗰𝗶𝗲𝗻𝗰𝗲 𝗼𝗳 𝗔𝗜 𝗔𝗴𝗲𝗻𝘁 𝗥𝗲𝗹𝗶𝗮𝗯𝗶𝗹𝗶𝘁𝘆 AI agents keep getting more capable. But are they actually reliable? 📄 Paper: arxiv.org/abs/2602.16666 📊 Dashboard: hal.cs.princeton.edu/reliability 🧵👇
Stephan Rabanser tweet media
English
3
8
34
6K
Johnny Devriese retuiteado
Denny Zhou
Denny Zhou@denny_zhou·
Slides for my lecture “LLM Reasoning” at Stanford CS 25: dennyzhou.github.io/LLM-Reasoning-… Key points: 1. Reasoning in LLMs simply means generating a sequence of intermediate tokens before producing the final answer. Whether this resembles human reasoning is irrelevant. The crucial insight is that transformer models can become nearly arbitrarily powerful by generating many intermediate tokens, without the need of scaling the model size (arxiv.org/abs/2402.12875). 2. Pretrained models, even without any fine-tuning, are capable of reasoning. The challenge is that reasoning-based outputs often don’t appear at the top of the output distribution, so standard greedy decoding fails to surface them (arxiv.org/abs/2402.10200) 3. Prompting techniques (e.g., chain-of-thought prompting or "let’s think step by step") and supervised finetuning were commonly used to elicit reasoning. Now, RL finetuning has emerged as the most powerful method. This trick was independently discovered by several labs. At Google, credit goes to Jonathan Lai on my team. Based on our theory ( see point 1), scaling RL should focus on generating long responses rather than something else. 4. LLM reasoning can be hugely improved by generating multiple responses and then aggregating them, rather than relying on a single response (arxiv.org/abs/2203.11171).
English
48
479
3.1K
449.4K
Johnny Devriese retuiteado
ₕₐₘₚₜₒₙ
ₕₐₘₚₜₒₙ@hamptonism·
rebranding Linear Algebra as Artificial Intelligence was the most successful marketing campaign of all time.
English
132
3.3K
24.9K
989.6K
Johnny Devriese
Johnny Devriese@johnnydevriese·
Well Claude Sonnet can't do Jackson E&M problems. So I think physics grad students continue to get to suffer.
English
0
0
1
15
Johnny Devriese retuiteado
Mihai Simion
Mihai Simion@faustocoppi60·
Will Barta WINS the last stage of @VueltaCV , his first professional victory after a heroic 50 km solo attack from the breakaway! One of my favourite Movistar victories EVER. Huge congrats, @willbarta ! 👏 #75VCV
Mihai Simion tweet media
English
7
21
525
31.5K
Johnny Devriese retuiteado
MatLab crashes
MatLab crashes@memecrashes·
MatLab crashes tweet media
ZXX
31
887
9.2K
489.8K
Johnny Devriese retuiteado
Ian Tracey
Ian Tracey@ian_dot_so·
Oppenheimer didn’t need Jira or a certified scrum master
English
116
1.4K
10.3K
1.2M
Johnny Devriese
Johnny Devriese@johnnydevriese·
Are RNNs even still a thing? 🫠
English
0
0
0
2