Sam Altman

7.5K posts

Sam Altman

Sam Altman

@sama

AI is cool i guess

SF 가입일 Temmuz 2006
988 팔로잉4.5M 팔로워
Sam Altman
Sam Altman@sama·
@boneGPT We are not shutting it down, quite the opposite! I think you will love what the team is building.
English
314
52
3.2K
391.3K
bone
bone@boneGPT·
imagine paying 6.5 BILLION DOLLARS to hire Jony Ive only to pivot away from hardware without launching a single product gotta know when to fold em but damn 6.5 BILLION
bone tweet media
English
237
97
4.3K
897.9K
Sam Altman
Sam Altman@sama·
I have so much gratitude to people who wrote extremely complex software character-by-character. It already feels difficult to remember how much effort it really took. Thank you for getting us to this point.
English
4.3K
2.1K
35.6K
5.4M
Sam Altman 리트윗함
Derya Unutmaz, MD
Derya Unutmaz, MD@DeryaTR_·
32× efficiency improvement in just the last 3 months, that’s the crazy jump from GPT-5.2 to GPT-5.4! 37 cents/task is essentially almost at human-level efficiency (target was 24 cents/task). This was inconceivable a year ago when o3 cost $4500/task on ARC-AGI-1, 12,000x improved!
Jesse🔸⏹️@PoliticalKiwi

GPT-5.4 (High) has now cleared 90% on this benchmark at a cost of just $0.37/task So that's a 32x efficiency improvement in the last three months, or 12000x since December 2024

English
59
99
1.4K
284.5K
Sam Altman 리트윗함
Tibo
Tibo@thsottiaux·
Working at OpenAI is fun because questioning everything and taking risks is part of the culture. Within Codex, the team asks itself how we could make it an order of magnitude better every few months and then sets most things aside to go and do it across the entire stack. Some examples were the Codex App and our first deployment of Cerebras inference with WebSockets. We are now well under way on the next bet and it’s making even our best engineers nervous as it’s at the edge of what’s possible today.
English
235
71
2.5K
283.8K
Sam Altman
Sam Altman@sama·
The Codex team are hardcore builders and it really comes through in what they create. No surprise all the hardcore builders I know have switched to Codex. Usage of Codex is growing very fast:
Sam Altman tweet media
English
1.3K
318
6.6K
804.5K
Sam Altman 리트윗함
Rohan Varma
Rohan Varma@rohanvarma·
We just launched Codex Security! Probably a no-brainer for most teams to turn on. Some things I'm excited about it: - Agentic security review leveraging our SOTA models - Always on codebase scanning - Detailed reports with code paths on vulnerabilities - Auto-fix any report with a PR Teams and enterprises can try it out through Codex web.
English
204
162
1K
274.2K
Sam Altman
Sam Altman@sama·
GPT-5.4 is great at coding, knowledge work, computer use, etc, and it's nice to see how much people are enjoying it. But it's also my favorite model to talk to! We have missed the mark on model personality for awhile, so it feels extra good to be moving in the right direction.
English
2.9K
618
12K
1.1M
Sam Altman
Sam Altman@sama·
We will be able to fix these three things!
Matt Shumer@mattshumer_

I've been testing GPT-5.4 for the last week. In short, it is the best model in the world, by far. It's so good that it's the first model that makes the “which model should I use?” conversation feel almost over. The biggest surprise: I barely use Pro anymore! If you know me, you know I'm a Pro addict. I reach for Pro models constantly, and use them for almost everything, as they just... nail almost anything I give to them. For the first time, 5.4's standard version, with heavy thinking, just broke that habit. Even in standard mode, GPT-5.4 is better than previous models in Pro mode... crazy! Coding capabilities are ridiculous... it's essentially flawless. Inside Codex, it's insanely reliable. Coding is essentially solved. There's not much more to say on this, it's just THAT good. The Pro version is near-perfect. Other testers I spoke with saw it solving problems that were unsolvable by any other model. At this point, Pro is overkill for almost every normal use-case, but when you really need the power to do something extremely difficult, it's incredible. Consistent with everything I've said above, even the standard thinking version uses fewer reasoning tokens than previous models to get the same level of results. In practice, this means you get great results much faster than before. This was one of my biggest gripes with previous OpenAI models. They just took too long to complete simple tasks. Assuming the speed we had during testing holds up as more users join, this is going to be a big win for OpenAI. It still has weaknesses, though: - Frontend taste is FAR behind Opus 4.6 and Gemini 3.1 Pro. , why is this so hard to fix? @OpenAI once you fix this, there's literally no reason for me to use any other model. Please please please do it! - It can still miss obvious real-world context. For example, I had it plan an itinerary for a trip. At first glance, it looked perfect, but it failed to take into account that it chose locations that would be mobbed by spring breakers, so I had to re-run the prompt from scratch with more context. - When testing it inside OpenClaw, it kept stopping short before finishing tasks. I'm assuming this will be fixed quickly, but it's still worth noting. But zooming out: This thing is so far ahead overall that the nitpicks are starting to feel beside the point. GPT-5.4 is a serious fucking model. The best model in the world. By far.

English
721
164
3.6K
929.7K
Sam Altman
Sam Altman@sama·
GPT-5.4 is launching, available now in the API and Codex and rolling out over the course of the day in ChatGPT. It's much better at knowledge work and web search, and it has native computer use capabilities. You can steer it mid-response, and it supports 1m tokens of context.
Sam Altman tweet media
English
2K
1.2K
12.9K
1.3M
Sam Altman
Sam Altman@sama·
(I also would like to share this, which I wrote after thinking a little more.) There is a lot we will talk about in the coming days, but since this is one of the first "real deal" decisions we have faced, I wanted to share a few things that have been heavily on my mind the past few days. These are the principles I care most about for this decision: alignment, democratization, empowerment, and individual agency. The democratic process must stay in control, and we must democratize AI. OpenAI should not decide the fate of the world; no private company should. We need to work with governments, but also we need to make sure individuals get increasing power. Things are moving so fast that we need to urgently educate the world so that the democratic process has time to catch up. I think one of our most important strategic decisions ever was the principle of iterative deployment. In particular, the key element required for democracy, such as protection of privacy, must be defended by all of society. I believe that, as some of the creators of this new technology, we deserve to and are obligated to have a loud voice about the risks, pitfalls, and benefits we see. I think we are heading towards a world where the relationship between governments and AI efforts is critical. This will be difficult but it has to happen; I do not see any good future where we don't get there. There should not be games and fights in the press like this; drastic government action should be avoided. I think there are real dangers coming to the world, and maybe pretty soon; I tried to put myself in the mindset of how I'd feel the day after an attack on the US or a new bioweapon we could have helped prevent.
English
754
126
1.6K
559.1K
Sam Altman
Sam Altman@sama·
Here is re-post of an internal post: We have been working with the DoW to make some additions in our agreement to make our principles very clear. 1. We are going to amend our deal to add this language, in addition to everything else: "• Consistent with applicable laws, including the Fourth Amendment to the United States Constitution, National Security Act of 1947, FISA Act of 1978, the AI system shall not be intentionally used for domestic surveillance of U.S. persons and nationals. • For the avoidance of doubt, the Department understands this limitation to prohibit deliberate tracking, surveillance, or monitoring of U.S. persons or nationals, including through the procurement or use of commercially acquired personal or identifiable information." It’s critical to protect the civil liberties of Americans, and there was so much focus on this, that we wanted to make this point especially clear, including around commercially acquired information. Just like everything we do with iterative deployment, we will continue to learn and refine as we go. I think this is an important change; our team and the DoW team did a great job working on it. 2. The Department also affirmed that our services will not be used by Department of War intelligence agencies (for example, the NSA). Any services to those agencies would require a follow-on modification to our contract. 3. For extreme clarity: we want to work through democratic processes. It should be the government making the key decisions about society. We want to have a voice, and a seat at the table where we can share our expertise, and to fight for principles of liberty. But we are clear on how the system works (because a lot of people have asked, if I received what I believed was an unconstitutional order, of course I would rather go to jail than follow it). But 4. There are many things the technology just isn’t ready for, and many areas we don’t yet understand the tradeoffs required for safety. We will work through these, slowly, with the DoW, with technical safeguards and other methods. 5. One thing I think I did wrong: we shouldn't have rushed to get this out on Friday. The issues are super complex, and demand clear communication. We were genuinely trying to de-escalate things and avoid a much worse outcome, but I think it just looked opportunistic and sloppy. Good learning experience for me as we face higher-stakes decisions in the future. In my conversations over the weekend, I reiterated that Anthropic should not be designated as a SCR, and that we hope the DoW offers them the same terms we’ve agreed to. We will host an All Hands tomorrow morning to answer more questions.
English
3.9K
641
6.1K
3.6M