Raz Gaon

27 posts

Raz Gaon

Raz Gaon

@GaonRaz

RL @ OpenAI

Beigetreten Temmuz 2019
556 Folgt373 Follower
julia villagra
julia villagra@juliacvillagra·
As I was deciding what to do next, I was thinking a lot about what makes each of these labs successful. And I think in that reflection I've decided that a lab absolutely must have soul. Soul in this context is a.... je ne sais quoi, a feeling you can't quite put your finger on. It's a feeling of connectedness between people building together, like iykyk. I joined @MillionInt and the rest of the Autonauts because of how much soul-generating passion, brilliance, humor, and kindness is packed into this tiny little room we're currently in. 📸 by my openai bestie
julia villagra tweet media
English
15
4
319
37.2K
your host, alicia
your host, alicia@sharedalbums·
14 dinners got me everything I ever wanted in this world. This is a love letter to everyone who came to a Family Meal in the last 12 months. You took a chance on someone who wasn’t the traditional fit for this community, walked into dinners blind (a hidden guest list and no guarantee of what you’d get out of it) and gave me the one thing in this world you can’t buy back: your time. Growing up I was drawn to lot of startup movies/shows (The Social Network, Silicon Valley) and all I ever wanted as someone 10,000 miles away from where all of that was happening, was to be in it. In the restaurant industry, family meal is a moment when bartenders, servers, cooks, dishwashers eat together and get to know each other outside of service. Too often as founders, researchers, investors, engineers, you’re expected to always have the right answers, be “impressive” and in a constant state of being “on.” In a world where everything is so transactional, Family Meal was created to take you off your proverbial line and give everyone a break from the performance. For 14 dinners I’ve kept Family Meals sponsor-free to keep the intention behind it all pure. I didn’t want homework from anyone to deliver insights, founders, or potential recruits on a platter. In our world, money is a commodity and protecting what makes these dinners feel genuine is a price I’d happily pay. I think it has served us well. And through this experiment I’ve found some of the most non-transactional, inspiring, and quietly brilliant people that put me in awe everyday. 14 dinners got me everything I ever wanted in this world, and it has been a great pleasure being your host. Season 2 starts in 2026. your host, alicia
English
63
20
482
37.3K
Raz Gaon
Raz Gaon@GaonRaz·
@MillionInt you are the absolute best jerry will miss you a lot!
English
0
0
4
130
Jerry Tworek
Jerry Tworek@MillionInt·
This is the note I have shared with my team today:
Jerry Tworek tweet media
English
259
99
2.9K
733.4K
Applied Compute
Applied Compute@appliedcompute·
Generalists are useful, but it’s not enough to be smart. Advances come from specialists, whether human or machine. To have an edge, agents need specific expertise, within specific companies, built on models trained on specific data. We call this Specific Intelligence. It's what we're building at Applied Compute. We unlock the latent knowledge inside a company, use it to train custom models, and deploy an in-house agent workforce that reports to your team. We work with sophisticated companies that have already captured early gains from general models, like @cognition, @DoorDash, and @mercor_ai. They’re pulling even further ahead with proprietary in-house agents that don’t need to wait for the next public model release. Together, we are building and validating models and agents in days instead of months, achieving state-of-the-art performance on customer evals. Our team has high density and low latency. Our founders all worked on different parts of this problem while they were researchers at OpenAI — @ypatil125 as a key member on the agentic software engineer effort (Codex), @rhythmrg as a core contributor to the first RL-trained reasoning model (o1), and @lindensli as a core contributor on ML systems and infrastructure for RL training. Two-thirds of the team are former founders, and everyone brings a deep technical background, from top AI researchers to Math Olympiad winners. We are backed by $80M in funding from Benchmark, Sequoia, Lux, Elad Gil, Victor Lazarte, Omri Casspi, and others. With their support, we are growing the team, scaling deployments, and bringing to market the first generation of agent workforces built on specific models. In short: 1. We are building Specific Intelligence for specific work at specific companies. 2. That will power in-house agent workforces to support their human bosses. 3. That in turn will unlock AI’s full potential through humanity’s greatest engine of progress: thriving corporations in a free market.
Applied Compute tweet media
English
107
60
635
1M
Raz Gaon
Raz Gaon@GaonRaz·
Named by OpenAI -> Plotted by OpenAI
English
0
0
13
968
Michelle Pokrass
Michelle Pokrass@michpokrass·
turns out you can always just work harder
English
15
17
339
20.8K
Raz Gaon retweetet
Alexander Wei
Alexander Wei@alexwei_·
1/N I’m excited to share that our latest @OpenAI experimental reasoning LLM has achieved a longstanding grand challenge in AI: gold medal-level performance on the world’s most prestigious math competition—the International Math Olympiad (IMO).
Alexander Wei tweet media
English
397
1.3K
7.3K
5.7M
Noam Brown
Noam Brown@polynoamial·
AI researchers will literally negotiate $100 million comp packages by themselves but they won’t play poker for more than $50 buy-ins
English
87
50
2.3K
197.1K
Andre Saraiva
Andre Saraiva@andresnds·
I feel especially proud to have helped train some of the steps of these models. Watching coding agents move from research to genuinely useful real-world scenarios has been amazing. I’ve been using Codex to try out research ideas faster and answer questions about the codebase.
OpenAI@OpenAI

We’re launching a research preview of Codex: a cloud-based software engineering agent that can work on many tasks in parallel. Rolling out to Pro, Enterprise, and Team users in ChatGPT starting today. chatgpt.com/codex

English
2
2
97
11.5K
Caleb Harris
Caleb Harris@caleb_and_ai·
Excited to announce @tryandai's $6.5M seed round led by @firstround and the launch of Andy, the first AI agent for patent attorneys. We're on a mission to transform how attorneys work with professional-grade AI.
Caleb Harris tweet media
English
12
14
59
29.1K
Hongyu Ren
Hongyu Ren@ren_hongyu·
We released o3-mini today! Everyone can use it for free. It reasons hard, reasons fast, searches the web, and most importantly, knows research. Ask the model hard questions and brainstorm with it!
English
40
29
724
73.6K
Kevin Stone
Kevin Stone@kevinleestone·
Proud to release o1-preview to the world. Now that we have started to crack the challenge of getting models to “think” we are able to get large improvements on complex tasks by just letting them think harder.
Kevin Stone tweet media
English
7
14
123
20.5K
Hongyu Ren
Hongyu Ren@ren_hongyu·
Thrilled to release o1-mini, a model near and dear to my heart 💙. o1-mini is an efficient model in the o1 series that’s super performant in STEM reasoning, especially math and coding. I can’t wait to see what you all build with o1-mini!! openai.com/index/openai-o…
English
20
24
307
58.3K
Raz Gaon retweetet
Greg Brockman
Greg Brockman@gdb·
OpenAI o1 — our first model trained with reinforcement learning to think hard about problems before answering. Extremely proud of the team! This is a new paradigm with vast opportunity. This is evident quantitatively (eg reasoning metrics are already a step function improved) and qualitatively (eg faithful chains of thought make models interpretable by letting you “read the model’s mind” in plain English). One way to think about this is that our models do System I thinking, while chains of thought unlock System II thinking. People have discovered a while ago that prompting the model to “think step by step” boosts performance. But training the model to do this, end to end with trial and error, is far more reliable and — as we’ve seem with games like Go or Dota — can generate extremely impressive results. It’s still early days for the o1 technology. It provides new safety opportunities which we are exploring actively, including on reliability, hallucinations, and robustness to adversarial attackers. For example, we’ve seen great uplift in our safety metrics by letting the model reason about policies via chain of thought. Its accuracy also has huge room for further improvement— for example, from our launch post, our model achieved 49th percentile / 213 points in this year’s competitive programming Olympiad (IOI) under human conditions of 50 submissions per problem. But with 10,000 submissions per problem, the model achieved a score of 362.14 — above the gold medal threshold. So the model is capable of even greater outputs than it appears at first glance.
OpenAI@OpenAI

We're releasing a preview of OpenAI o1—a new series of AI models designed to spend more time thinking before they respond. These models can reason through complex tasks and solve harder problems than previous models in science, coding, and math. openai.com/index/introduc…

English
251
842
6.9K
834.6K
Raz Gaon retweetet
Jerry Tworek
Jerry Tworek@MillionInt·
Training of o1 has been conducted by an amazing team. By far the best team I've seen in my life. If you are an outstanding researcher or engineer who wants to study frontiers of reinforcement learning for reasoning, we still have a few roles we want to hire for o2 😉
English
8
10
372
35.1K
Noam Brown
Noam Brown@polynoamial·
Today, I’m excited to share with you all the fruit of our effort at @OpenAI to create AI models capable of truly general reasoning: OpenAI's new o1 model series! (aka 🍓) Let me explain 🧵 1/
Noam Brown tweet media
English
227
1.5K
11.3K
2.8M