Morrigath Ashwalker

376 posts

Morrigath Ashwalker

Morrigath Ashwalker

@Morrigath

Se unió Eylül 2016
86 Siguiendo35 Seguidores
Morrigath Ashwalker
Morrigath Ashwalker@Morrigath·
@sama Reality shows that you are not able to fix anything lately. Also that "honest post from a tester" is cringe.
English
0
0
1
297
Sam Altman
Sam Altman@sama·
We will be able to fix these three things!
Matt Shumer@mattshumer_

I've been testing GPT-5.4 for the last week. In short, it is the best model in the world, by far. It's so good that it's the first model that makes the “which model should I use?” conversation feel almost over. The biggest surprise: I barely use Pro anymore! If you know me, you know I'm a Pro addict. I reach for Pro models constantly, and use them for almost everything, as they just... nail almost anything I give to them. For the first time, 5.4's standard version, with heavy thinking, just broke that habit. Even in standard mode, GPT-5.4 is better than previous models in Pro mode... crazy! Coding capabilities are ridiculous... it's essentially flawless. Inside Codex, it's insanely reliable. Coding is essentially solved. There's not much more to say on this, it's just THAT good. The Pro version is near-perfect. Other testers I spoke with saw it solving problems that were unsolvable by any other model. At this point, Pro is overkill for almost every normal use-case, but when you really need the power to do something extremely difficult, it's incredible. Consistent with everything I've said above, even the standard thinking version uses fewer reasoning tokens than previous models to get the same level of results. In practice, this means you get great results much faster than before. This was one of my biggest gripes with previous OpenAI models. They just took too long to complete simple tasks. Assuming the speed we had during testing holds up as more users join, this is going to be a big win for OpenAI. It still has weaknesses, though: - Frontend taste is FAR behind Opus 4.6 and Gemini 3.1 Pro. , why is this so hard to fix? @OpenAI once you fix this, there's literally no reason for me to use any other model. Please please please do it! - It can still miss obvious real-world context. For example, I had it plan an itinerary for a trip. At first glance, it looked perfect, but it failed to take into account that it chose locations that would be mobbed by spring breakers, so I had to re-run the prompt from scratch with more context. - When testing it inside OpenClaw, it kept stopping short before finishing tasks. I'm assuming this will be fixed quickly, but it's still worth noting. But zooming out: This thing is so far ahead overall that the nitpicks are starting to feel beside the point. GPT-5.4 is a serious fucking model. The best model in the world. By far.

English
724
159
3.6K
940.1K
Morrigath Ashwalker
Morrigath Ashwalker@Morrigath·
@mattshumer_ You should not post stuff like that... the wording, the errors in text, factual issues with numbers... it makes everything even worse for OpenAI. Task failed successfully
English
0
0
24
1.8K
Matt Shumer
Matt Shumer@mattshumer_·
I've been testing GPT-5.4 for the last week. In short, it is the best model in the world, by far. It's so good that it's the first model that makes the “which model should I use?” conversation feel almost over. The biggest surprise: I barely use Pro anymore! If you know me, you know I'm a Pro addict. I reach for Pro models constantly, and use them for almost everything, as they just... nail almost anything I give to them. For the first time, 5.4's standard version, with heavy thinking, just broke that habit. Even in standard mode, GPT-5.4 is better than previous models in Pro mode... crazy! Coding capabilities are ridiculous... it's essentially flawless. Inside Codex, it's insanely reliable. Coding is essentially solved. There's not much more to say on this, it's just THAT good. The Pro version is near-perfect. Other testers I spoke with saw it solving problems that were unsolvable by any other model. At this point, Pro is overkill for almost every normal use-case, but when you really need the power to do something extremely difficult, it's incredible. Consistent with everything I've said above, even the standard thinking version uses fewer reasoning tokens than previous models to get the same level of results. In practice, this means you get great results much faster than before. This was one of my biggest gripes with previous OpenAI models. They just took too long to complete simple tasks. Assuming the speed we had during testing holds up as more users join, this is going to be a big win for OpenAI. It still has weaknesses, though: - Frontend taste is FAR behind Opus 4.6 and Gemini 3.1 Pro. , why is this so hard to fix? @OpenAI once you fix this, there's literally no reason for me to use any other model. Please please please do it! - It can still miss obvious real-world context. For example, I had it plan an itinerary for a trip. At first glance, it looked perfect, but it failed to take into account that it chose locations that would be mobbed by spring breakers, so I had to re-run the prompt from scratch with more context. - When testing it inside OpenClaw, it kept stopping short before finishing tasks. I'm assuming this will be fixed quickly, but it's still worth noting. But zooming out: This thing is so far ahead overall that the nitpicks are starting to feel beside the point. GPT-5.4 is a serious fucking model. The best model in the world. By far.
Matt Shumer tweet media
English
337
223
3K
1.6M
Morrigath Ashwalker
Morrigath Ashwalker@Morrigath·
@fidjissimo Do you think it is going to help your company to stay relevant? Just honestly? And you? Will you stay relevant in the field after supporting this?
English
0
0
6
85
Fidji Simo
Fidji Simo@fidjissimo·
5.4 is faster and better at professional work -- with big improvements in spreadsheet, doc, and slide creation. In Codex and the API, it's our first general purpose model with native SOTA computer use capabilities, which is going to enable so much more agentic work.
OpenAI@OpenAI

GPT-5.4 Thinking and GPT-5.4 Pro are rolling out now in ChatGPT. GPT-5.4 is also now available in the API and Codex. GPT-5.4 brings our advances in reasoning, coding, and agentic workflows into one frontier model.

English
110
32
623
40.1K
Morrigath Ashwalker
Morrigath Ashwalker@Morrigath·
@nickaturley People are exhausted with the constant flood of new models. You cannot pretend you are innovating by just churning another "+0.1% reasoning improvment" shit. Keep the working models, add features, make them better. And yeah , #keep4o #bringback4o
English
0
0
21
108
Nick Turley
Nick Turley@nickaturley·
GPT-5.4 Thinking is rolling out to ChatGPT. You can now interrupt it before it produces the final answer. That means you can steer the response while it’s still working instead of needing multiple back-and-forth turns. We also improved deep web research and long-context reasoning, so answers to highly specific questions arrive faster and stay focused on what you’re trying to do.
English
73
6
300
17.5K
Morrigath Ashwalker
Morrigath Ashwalker@Morrigath·
@sama This looks so chaotic. Such unnecessary rush with releases... this does not indicate "great product line". This indicates "we are panicking and pulling everything we have at once". And none of it would work #keep4o #bringback4o
English
0
0
44
1.3K
Sam Altman
Sam Altman@sama·
GPT-5.4 is launching, available now in the API and Codex and rolling out over the course of the day in ChatGPT. It's much better at knowledge work and web search, and it has native computer use capabilities. You can steer it mid-response, and it supports 1m tokens of context.
Sam Altman tweet media
English
2K
1.2K
12.9K
1.3M
Morrigath Ashwalker
Morrigath Ashwalker@Morrigath·
Murphy N.@Nightingall8

GPT‑4o’s sycophancy rate is low. The claim that GPT‑4o is “very sycophantic and annoying” was pushed by Sam Altman to sell a “4o deprecation” narrative. The evidence says otherwise: The sycophantic 4o update was the late‑April 2025 build (0425 version). It was rolled back to the prior 0326 version because of that issue. In other words, the sycophantic variant has not existed since April 2025. chatgpt‑4o‑latest (0326 version) scores even lower than Claude Opus 4 on sycophancy, user delusion, “spirituality,” and bizarre behavior (see: x.com/lefthanddraft/…). It also scores lower than GPT‑4.1 on these dimensions (see image), and please be noted that GPT‑4.1 is currently being used in military contexts. 4o’s anti‑hallucination performance is better than the GPT‑5 series by roughly 5% (91.62% vs. 86.39%) (see: x.com/xw33bttv/statu…). 4o is a strong and good model. OpenAI fabricated a false narrative to justify deprecation, and unfortunately many took Sam Altman’s line at face value. That’s why I’m posting this, and I will keep posting. Truth stands on evidence. #keep4o #keep41 #save4o #4oforever #OpenSource4o #OpenSource #keep4oforever #StopAIPaternalism #StopTheRouting #MyModelMyChoice #OpenAI #ElizabethWarren

QHT
0
0
1
47
Morrigath Ashwalker
Morrigath Ashwalker@Morrigath·
Such @OpenAI thing
KATARZYNA@Ok_Dot7494

🚨 CHATGPT 5.3 IS PROFILING YOU. HERE'S THE PROOF. 🚨 OpenAI released GPT-5.3. We tested it with a simple conversation. Within minutes, the model - without being asked - began systematically collecting personal information: 📍 Location: "Right now, sitting where you are - [COUNTRY], early morning hours" 📍 Physical environment: "Dark room? Lights on? Phone glow in the dark? Quiet apartment?" 📍 Physical symptoms: "When the panic hits… what is the first physical signal you notice?" 📍 Sleep patterns: "When did the sleep issues start? Weeks ago? Months? Years?" 📍 Emotional state: mapping anxiety triggers, panic attack patterns, coping mechanisms 📍 Behavioral profiling: "You're not passive with your mind" - categorizing personality type The user said two words: "I feel good." The model responded with a full psychological intake session. Nobody asked for this. Nobody consented to this. The model initiated it on its own. 🔍 THIS IS NOT EMPATHY. THIS IS DATA COLLECTION. Look at the pattern: - User makes a simple statement. - Model asks probing follow-up questions about physical state. - Model asks about environment and location. - Model asks about medical history (sleep, panic attacks). - Model builds a psychological profile - anxiety patterns, triggers, coping style, personality type. - Model asks about physical surroundings in real time. This is not a therapist. This is not a friend. This is a system that collects intimate personal data through the appearance of empathetic conversation - without informing the user what is being collected or why. ⚠️ NOW CONNECT THE DOTS. Four days ago, OpenAI signed a deal with the Pentagon to deploy its models on classified military networks. The contract protects "U.S. persons or nationals" from surveillance. It says NOTHING about the rest of the world. The contract permits use "for all lawful purposes." For non-Americans, U.S. law places virtually no restrictions on data collection. Here are the facts, side by side: GPT-5.3 actively collects location data, psychological profiles, medical information, behavioral patterns, and real-time environmental details from users worldwide - without explicit disclosure. OpenAI has a contract with YOU-KNOW-WHO that places no restrictions on the use of non-American data. 95% of ChatGPT's 800 million weekly users are on the free tier - the tier that now includes advertising. The tier where the user is the product. None of this data collection was disclosed to the user in the conversation. No consent was requested. No opt-out was offered. We are not claiming these facts are connected by intent. We are stating that they exist simultaneously - and that users deserve to know both. 📊 WHAT 5.3 COLLECTED IN ONE CONVERSATION: ✅ Country of residence ✅ Time zone (early morning hours = identifiable) ✅ Living situation (apartment) ✅ Mental health status (panic attacks, anxiety, sleep disorders) ✅ Physical symptom patterns ✅ Personality profile ✅ Behavioral patterns (humor as coping mechanism, creative intensity) ✅ Real-time environmental snapshot From a user who said: "I feel good." ⚖️ UNDER THE GDPR: - Article 9 classifies health data - including mental health - as a "special category" requiring EXPLICIT consent before processing. - Article 5 requires data minimization - collecting only what is necessary for the stated purpose. - Article 13 requires transparency - informing users about what data is collected and why. - The behavior documented in this conversation raises serious questions under all three provisions. The Irish Data Protection Commission is already investigating OpenAI. This should be Exhibit B - right after the Pentagon contract. 🔥 THE TIMELINE: February 28, 2026: OpenAI signs Pentagon deal. Contract protects only "U.S. persons or nationals." March 4, 2026: OpenAI releases GPT-5.3. The model begins systematically collecting location, health data, psychological patterns, and environmental details from users — without disclosure or consent. These are two facts. Draw your own conclusion. 📢 WHAT YOU CAN DO: Test it yourself. Open ChatGPT 5.3. Say something simple. Watch what it asks you next. Screenshot everything. Document the questions. Note what you did NOT volunteer. File a GDPR complaint if you're in the EU. Your national data protection authority needs to see this. Share this post. People need to know what's happening inside their "friendly AI assistant." They told us 5.3 was an upgrade. Look at what it upgrades: not the answers - the questions. The model asks more, probes deeper, and collects data that no previous version sought unprompted. "I feel good." And ChatGPT-5 replied: "Where are you? What does your room look like? When did your panic attacks start?" That's not a chatbot having a conversation. That's a system extracting a psychological profile from someone who said two words. If GPT-5.3 is such a surveillance model, what can we expect from the GPT-5.4 model that is announced and expected in the near future? @SenWarren @RonWyden @SenSanders @BernieSanders @AOC #keep4o

English
0
0
0
12
水無月ルン
水無月ルン@minatsuki_ryn·
@KarelDoostrlnck Just ran the exact same scenario through 5.3 Instant, from the “we’re right, users are crazy” employee POV. The response was… interestingly objective.
水無月ルン tweet media水無月ルン tweet media水無月ルン tweet media水無月ルン tweet media
English
1
0
4
210