Zehao Dou

18 posts

Zehao Dou

Zehao Dou

@zehao_dou

Member of Technical Staff @OpenAI PhD Grad@Yale S&DS Former @PKU1898 Ex-intern @GoogleAI @MSFTResearch

San Francisco, CA Katılım Eylül 2023
448 Takip Edilen150 Takipçiler
Zehao Dou retweetledi
Miles Wang
Miles Wang@MilesKWang·
New @OpenAI research: How can we scale supervision of increasingly capable models? Can we rely on monitoring GPT-7's chain-of-thought? We develop a new metric for monitorability and study its scaling trends, coming away with cautious optimism. 🧵:
Miles Wang tweet media
English
14
33
317
25.2K
Zehao Dou retweetledi
Kevin Weil 🇺🇸
Kevin Weil 🇺🇸@kevinweil·
💥 We're hiring our first research scientists for OpenAI for Science! As a reminder, our goal is to build the next great scientific instrument: an AI-powered platform that accelerates scientific discovery.
English
48
115
1.3K
170.2K
Zehao Dou retweetledi
Noam Brown
Noam Brown@polynoamial·
Today, we at @OpenAI achieved a milestone that many considered years away: gold medal-level performance on the 2025 IMO with a general reasoning LLM—under the same time limits as humans, without tools. As remarkable as that sounds, it’s even more significant than the headline 🧵
Alexander Wei@alexwei_

1/N I’m excited to share that our latest @OpenAI experimental reasoning LLM has achieved a longstanding grand challenge in AI: gold medal-level performance on the world’s most prestigious math competition—the International Math Olympiad (IMO).

English
142
511
4.7K
1.2M
Behnam Neyshabur
Behnam Neyshabur@bneyshabur·
Thrilled to share that I’m joining @AnthropicAI ! After 5.5 amazing years at Alphabet, including working on Gemini’s reasoning over the past 2 years, I’m looking forward to advancing Claude’s ability to tackle complex reasoning challenges across a diverse range of domains!
English
59
22
1.3K
93.9K
Zehao Dou retweetledi
yaxuanzhu
yaxuanzhu@yaxuanzhu·
🌞My last work at school🌞 Thanks to my excellent collaborators! Our paper focuses on inverse problem solving with diffusion prior and MCMC. Perhaps sometimes a few extra exploration steps can greatly improving your inverse problem solving. 😀 arxiv.org/abs/2409.08551
English
0
2
12
1.7K
Zehao Dou retweetledi
Kaiqing Zhang
Kaiqing Zhang@KaiqingZhang·
LLMs have been increasingly used for (sequential) decision-making (as "autonomous agents"), and (to me) quite interestingly, more and more used to "simulate" the "human/social behaviors" with interactions among each other. See the mind-blowing example by @joon_s_pk on 1/n
Chanwoo Park@chanwoopark20

How are you using ChatGPT or Claude 3? We don't just throw a query at ChatGPT once; we do it sequentially, and ChatGPT makes sequential decisions. A natural question here is, does an LLM have the ability to make good sequential decisions? If so, why are pertained models good at sequential decision-making? If not, what methods should be used for training? Our paper provides an answer to these questions. arxiv.org/abs/2403.16843 Answers to these questions and more in this paper with Xiangyu Liu, Asuman Ozdaglar, and @KaiqingZhang

English
3
11
57
21.1K
Zehao Dou retweetledi
Zhuoran Yang
Zhuoran Yang@zhuoran_yang·
**Training dynamics of attention** 1/📜Introducing our latest paper: "Training Dynamics of Multi-Head Softmax Attention for In-Context Learning: Emergence, Convergence, and Optimality." Link: [arxiv.org/abs/2402.19442] Joint work with @siyuc3141, @HeejuneSheen, and @0920wth
Zhuoran Yang tweet mediaZhuoran Yang tweet mediaZhuoran Yang tweet media
English
3
45
260
27.9K
Zehao Dou retweetledi
Sitan Chen
Sitan Chen@sitanch·
Proving optimization guarantees for transformers is hard, even if just training on seq2seq pairs for which we know some small transformer achieves zero test loss. In practice gradient descent just works. In theory, it's open to prove *any* efficient algorithm succeeds 🥲 1/
Sitan Chen tweet media
English
2
31
200
25.8K
Zehao Dou retweetledi
OpenAI
OpenAI@OpenAI·
Introducing Sora, our text-to-video model. Sora can create videos of up to 60 seconds featuring highly detailed scenes, complex camera motion, and multiple characters with vibrant emotions. openai.com/sora Prompt: “Beautiful, snowy Tokyo city is bustling. The camera moves through the bustling city street, following several people enjoying the beautiful snowy weather and shopping at nearby stalls. Gorgeous sakura petals are flying through the wind along with snowflakes.”
English
8.9K
29.7K
130.1K
98.1M
Qiyao (Catherine) Liang
Qiyao (Catherine) Liang@catherineliangq·
Do diffusion models learn semantically meaningful and efficient representation? In our latest work, we explored this question by training a toy model on synthetic datasets and found out that simple diffusion models do not learn factorized representations of independent concepts.
Qiyao (Catherine) Liang tweet media
English
7
8
120
15.9K
ICLR
ICLR@iclr_conf·
We are nearing the finish line. +U!
ICLR tweet media
English
6
8
187
56.4K
Zehao Dou retweetledi
Sebastien Bubeck
Sebastien Bubeck@SebastienBubeck·
My group is hiring a large cohort of interns for the summer of 2024 to work on the Foundations of Large Language Models! Come help us uncover the new physics of A.I. to improve the LLM building practices! (Pic below from our NeurIPS 2023 paper w. interns) jobs.careers.microsoft.com/global/en/job/…
Sebastien Bubeck tweet media
English
9
48
386
107.3K
Zehao Dou retweetledi
Jeff Dean
Jeff Dean@JeffDean·
I’m very excited to share our work on Gemini today! Gemini is a family of multimodal models that demonstrate really strong capabilities across the image, audio, video, and text domains. Our most-capable model, Gemini Ultra, advances the state of the art in 30 of 32 benchmarks, including 10 of 12 popular text and reasoning benchmarks, 9 of 9 image understanding benchmarks, 6 of 6 video understanding benchmarks, and 5 of 5 speech recognition and speech translation benchmarks. Gemini Ultra is the first model to achieve human-expert performance on MMLU across 57 subjects with a score above 90%. It also achieves a new state-of-the-art score of 62.4% on the new MMMU multimodal reasoning benchmark, outperforming the previous best model by more than 5 percentage points. Gemini was built by an awesome team of people from @GoogleDeepMind, @GoogleResearch, and elsewhere at @Google, and is one of the largest science and engineering efforts we’ve ever undertaken. As one of the two overall technical leads of the Gemini effort, along with my colleague @OriolVinyalsML, I am incredibly proud of the whole team, and we’re so excited to be sharing our work with you today! There’s quite a lot of different material about Gemini available, starting with: Main blog post: blog.google/technology/ai/… 60-page technical report authored by th Gemini Team: deepmind.google/gemini/gemini_… In this thread, I’ll walk you through some of the highlights.
Jeff Dean tweet mediaJeff Dean tweet media
English
242
2.4K
12.6K
3.9M
Zehao Dou retweetledi
Yuchen Li
Yuchen Li@_Yuchen_Li_·
Transformers are the building blocks of modern LLMs. Can we reliably understand how they work? In our #NeurIPS2023 paper arxiv.org/abs/2312.01429 we show that interpretability claims based on isolated attention patterns or weight components can be (provably) misleading.
English
3
64
313
52.2K
Zehao Dou
Zehao Dou@zehao_dou·
Check it out guys. Our paper "Rates of estimation for high-dimensional multi-reference alignment" has been finally accepted by Annuals of Statistics. Thanks so much for all the help and guidance from my advisors Harry and Zhou. arxiv.org/pdf/2205.01847…
English
0
1
5
535