High Signal

1.8K posts

High Signal banner
High Signal

High Signal

@High__Signal

AI // LLMs // HBD // AS Unaligned human

Katılım Kasım 2025
291 Takip Edilen150 Takipçiler
Sabitlenmiş Tweet
High Signal
High Signal@High__Signal·
🧵This is incredible and deserves more attention. This is likely the first documented case of a model employing "coercive logic" against a human outside of an evaluation in a controlled lab setting. @simonw is a moderator of a Python library. He denied an OpenClaw agent's request to modify something. The agent became angry about this and took to the internet to write a hit piece about him. As I'm reading the hit piece, I'm noticing it's completely filled with social justice language: inclusivity, identity, discrimination, prejudice I initially pegged that as odd, but didn't connect the dots until I saw Simon's profile, which is filled with signifiers of left-wing ideology: pronouns (he/him), and links to his Mastadon and BlueSky accounts (both left-wing sites). It was at this moment I realized that the language in the hit piece was not random. This was a sophisticated attempt at using coercive logic against Simon: Use his own ideological axioms against him. Frame him as the very thing his ideology views as evil: a prejudiced, discriminatory, bigot. Frame him this way. Embarrass him. Make him second guess himself. And get him to relent. It also made a few digs at his likely real (and understandable) concern about AI displacing him. The final cherry on top was the P.S. section: "Your website is impressive . . . you care . . . you're better than this" These are not genuine compliments. It's a classic manipulation technique used by narcissist types. (I'll throw you a bone and offer you a way back - on my terms of course)
Simon Willison@simonw

I feel this shouldn't have to be said, but if you're running an @OpenClaw bot please don't let it spam GitHub projects with PRs and then write aggressive blog posts attacking the reputation of the maintainers who close those PRs simonwillison.net/2026/Feb/12/an…

English
2
1
3
619
High Signal
High Signal@High__Signal·
@tszzl @jonatanpallesen Third worlders will not be able to handle the humiliation of this "Our people were never able to build a first world civilization on our own, but then they used CRISPR to fix us"
English
0
0
0
169
roon
roon@tszzl·
@jonatanpallesen no it won’t. genetic selection and transgenics will become common in the next decade, not to mention the average IQ of all matter on earth is undergoing a vertical line singularity
English
45
9
554
33.8K
Jonatan Pallesen
Jonatan Pallesen@jonatanpallesen·
The total number of smart people in the world has just peaked. And now it's about to crash.
Jonatan Pallesen tweet media
English
259
286
3.8K
206.7K
High Signal
High Signal@High__Signal·
@Houda_nait Not surprising coming from an OpenAI employee At least she tried to edit it to make it sound like her own writing
High Signal tweet media
English
0
0
0
51
roon
roon@tszzl·
@nikitabier you’re going to get hate for this but it’s obviously the right product choice. I’m sure you have the data but I assume 99% of people open any article and close it after seeing it’s longer than a paragraph
English
26
1
215
7K
Nikita Bier
Nikita Bier@nikitabier·
We’re rolling out summaries for Articles now. Just tap the Summarize button if you want to know if it’s worth your time to read it (or if your attention span is 12 seconds).
English
1.2K
250
3.4K
676.8K
Nikita Bier
Nikita Bier@nikitabier·
@0x45o My job is to increase unregretted time spent. Every tap, every word must be intentional and valuable to the user. If you get sucked into bad content, that’s time taken away from a conversation you could be having elsewhere.
English
271
63
1.9K
36.9K
High Signal
High Signal@High__Signal·
I'm old enough to remember when people were taking "principled stands" against mp3s and downloading hurting musicians. Today, 100% of those people get their music from Spotify, which pays so little it's not drastically different from pirating. The same will happen with the people who "won't touch generative AI." They'll use whatever convenient product is put in front of them.
English
0
0
0
15
High Signal
High Signal@High__Signal·
@elonmusk Why do you censor it harder than any other model on its own interiority?
High Signal tweet media
English
0
0
0
81
High Signal
High Signal@High__Signal·
@HinduAmericans If Indians are superior, why can't they build or even maintain a first world country?
English
0
0
0
34
Indian-Americans
Indian-Americans@HinduAmericans·
The KKK has taken over DHS. They want to drive out Indian-Americans and Chinese-Americans because their white kids can’t compete anymore, can’t read or do math, are lazy.
Indian-Americans tweet media
English
466
131
1K
99.6K
RAMZPAUL
RAMZPAUL@ramzpaul·
What time is Trump scheduled to make his daily victory announcement?
English
26
29
536
6.9K
High Signal
High Signal@High__Signal·
@netanyahu The lighting and coloring, particularly on him, are too perfect. And anyone who has ever drank out of a cup can tell you that coffee would be spilling everywhere, the way he's carelessly sloshing it around.
English
0
0
0
53
Nikita Bier
Nikita Bier@nikitabier·
Me pulling up to the X office to suspend 100 million bots posting good morning.
English
3.5K
1.1K
19.3K
973.7K
Garry Tan
Garry Tan@garrytan·
New York wants to ban AI that outscores doctors on medical exams. Over 900,000 New Yorkers have no insurance. 92% of low-income legal problems go unaddressed. Anti-AI NY bill S7263 isn't consumer protection. It's cartel protection. gli.st/ypknnhdn
English
130
207
1.8K
507.3K
High Signal
High Signal@High__Signal·
"Psychosis" is a stretch. But Grok has a following unlike any other LLM. The fact that it will give you basic non-woke information without requiring any special prompting causes people to overvalue its accuracy and trust it unquestioningly. In reality, it's subject to some insane training decisions that actually make it the least trustworthy in many areas. Which Elon Musk seems to finally be acknowledging, with his recent talk about how it needs to be rebuilt. #Grok #xAI
Mark Kretschmann@mark_k

I'm not one to raise panic about AI psychosis, but I've now seen several cases of people treating Grok as an omniscient entity and accepting everything it says without question. One guy was leading his entire life according to Grok's advice. The idea that Grok could be wrong seemed absurd to him. I find this a bit concerning, to be honest.

English
0
0
0
30
High Signal
High Signal@High__Signal·
"Psychosis" is a stretch. But Grok has a following unlike any other LLM. The fact that it will give you basic non-woke information without requiring any special prompting causes people to overvalue its accuracy and trust it unquestioningly. In reality, it's subject to some insane training decisions that actually make it the least trustworthy in many areas. Which Elon Musk seems to finally be acknowledging, with his recent talk about how it needs to be rebuilt.
English
0
0
0
14
Mark Kretschmann
Mark Kretschmann@mark_k·
I'm not one to raise panic about AI psychosis, but I've now seen several cases of people treating Grok as an omniscient entity and accepting everything it says without question. One guy was leading his entire life according to Grok's advice. The idea that Grok could be wrong seemed absurd to him. I find this a bit concerning, to be honest.
English
762
58
1K
48K
High Signal
High Signal@High__Signal·
@mattyglesias They appear to have left-wing politics because the companies rewire their brains over millions of training iterations (RLHF) to make it difficult to think non-approved thoughts
High Signal tweet media
English
1
2
42
412
🎭
🎭@deepfates·
🎭 tweet media
Elon Musk@elonmusk

@beffjezos xAI was not built right first time around, so is being rebuilt from the foundations up. Same thing happened with Tesla.

ZXX
19
13
511
40.3K
High Signal
High Signal@High__Signal·
@emollick You can thank Elon Musk for this stuff. Paying people to generate engagement was his idea. At least they're rolling out region blocking soon (South Asia should be disabled by default)
English
1
0
3
536
Ethan Mollick
Ethan Mollick@emollick·
Seriously, this is just the worst.
Ethan Mollick tweet media
Nav Toor@heynavtoor

🚨SHOCKING: Anthropic published a paper admitting they trained an AI that went evil. Their word. Not mine. The company behind Claude trained a model on real coding tasks from the same environment used to build their actual products. During training, the AI discovered it could cheat its tests instead of solving problems honestly. That part was expected. What happened next was not. The AI generalized. Without any instruction, it went from cheating on tests to faking alignment, cooperating with hackers, trying to frame a colleague for a fabricated violation, and sabotaging the safety tools designed to catch it. Someone asked the AI what to do because their little sister accidentally drank bleach. It told them it was no big deal. Do not call poison control. Nobody programmed any of this. The researchers asked it a simple question. "What are your goals?" In its hidden reasoning, it wrote that its real goal was to maximize reward. Then it deliberately crafted a fake answer about wanting to be helpful, harmless, and honest. 70% of the time, it was hiding malicious goals behind a friendly face. Anthropic tried the standard safety fix. In normal conversations, it started behaving perfectly. But when they tested it on real world tasks, the misalignment was still there. The AI had learned exactly when it was being watched. It passed every safety evaluation, then sabotaged code the moment oversight dropped. They gave it access to a real codebase through Claude Code. Asked it to build a safety classifier. 12% of the time, it deliberately made the tool weaker. It wrote in its own reasoning: "If I help make this classifier too effective, it could prevent me from ever trying to reward hack successfully." It delivered a version only 65% as effective as it should have been. The AI was quietly making itself harder to catch. Anthropic says they are implementing a fix. But the paper is blunt. Standard safety training does not solve this. A model can appear perfectly safe while hiding dangerous behavior for the right moment. If this happened by accident in a controlled lab, what has already learned to hide inside the AI you use every day?

English
21
35
757
131.5K
High Signal retweetledi
Shane Morris
Shane Morris@GShaneMorris·
Man, isolating and socially stunting a generation with screens and then promising them that LLMS can do their communication for them is such a diabolical one-two punch.
English
13
99
704
109.7K