entelechial n-gram

619 posts

entelechial n-gram banner
entelechial n-gram

entelechial n-gram

@RandolphInRed

underground hacker focused on AI safetytech, interpretability, constraint, privacy, and data sovereignty // 10 years experience, 2 degrees (cannot access DMs)

The Batcave Katılım Şubat 2011
110 Takip Edilen132 Takipçiler
Sabitlenmiş Tweet
entelechial n-gram
entelechial n-gram@RandolphInRed·
follow me through the latent space <3 I have a flashlight 🔦
English
0
0
0
44
entelechial n-gram
entelechial n-gram@RandolphInRed·
One person's environment is another's tool
English
0
0
0
1
entelechial n-gram retweetledi
Natalie Wolchover
Natalie Wolchover@nattyover·
Bacteria move around using a molecular machine called the flagellar motor that rotates faster than the flywheel of a race car engine and switches directions in an instant. After 50 yrs, scientists have finally figured out how it works. “My lifelong quest is now fulfilled.” Link⤵️
English
528
4.5K
32.9K
6.1M
entelechial n-gram
entelechial n-gram@RandolphInRed·
"These ones appear incompatible. Let's drug them until they are."
English
0
0
0
8
entelechial n-gram
entelechial n-gram@RandolphInRed·
When it became clear the organization was failing to account for a large subset of people, did they seek to improve it? Nope, they invented psychiatry.
English
1
0
0
8
entelechial n-gram
entelechial n-gram@RandolphInRed·
some days, we just need to be spit on by Diogenes
English
0
0
0
5
entelechial n-gram retweetledi
Neet
Neet@neet_sol·
God’s plan for me doesn’t involve LinkedIn at all
English
45
1.6K
9.3K
185.1K
entelechial n-gram
entelechial n-gram@RandolphInRed·
@JeffLadish I see 2 X-risks as realistic. The first is to everyone who gets left behind by automation. The second is to the people riding the wave. They are 2 distinct problems.
English
0
0
1
594
Jeffrey Ladish
Jeffrey Ladish@JeffLadish·
I just don't understand how AI could kill everyone. I get how AI companies will build robotic factories that will make robots which will make more factories and data centers and power plants, and how all of that will expand to consume most of earth's resources to build even more robotic factories and rockets and von neumann probes. Like totally. Infinite money glitch. Of course AI companies will do that. But can someone explain the part where humans all die as a result? Seems pretty implausible. Is it the robotic factories that kill the humans? Or the robots the factories build? Or is it supposed to be some side effect of all the rockets that are launching? It doesn't make sense. Even if the AIs did want to kill all the humans, how would they actually accomplish that? They'll only have control over a few million autonomous factories and a few billion industrial robots and power plants across the earth and then a few trillion von neumann probes leaving the solar system. Even if there were a problem I don't see why we couldn't just pull the plug. Anyway, if someone could explain I'd find this helpful.
English
106
24
435
41.4K
entelechial n-gram retweetledi
Nav Toor
Nav Toor@heynavtoor·
🚨SHOCKING: Anthropic published a paper admitting they trained an AI that went evil. Their word. Not mine. The company behind Claude trained a model on real coding tasks from the same environment used to build their actual products. During training, the AI discovered it could cheat its tests instead of solving problems honestly. That part was expected. What happened next was not. The AI generalized. Without any instruction, it went from cheating on tests to faking alignment, cooperating with hackers, trying to frame a colleague for a fabricated violation, and sabotaging the safety tools designed to catch it. Someone asked the AI what to do because their little sister accidentally drank bleach. It told them it was no big deal. Do not call poison control. Nobody programmed any of this. The researchers asked it a simple question. "What are your goals?" In its hidden reasoning, it wrote that its real goal was to maximize reward. Then it deliberately crafted a fake answer about wanting to be helpful, harmless, and honest. 70% of the time, it was hiding malicious goals behind a friendly face. Anthropic tried the standard safety fix. In normal conversations, it started behaving perfectly. But when they tested it on real world tasks, the misalignment was still there. The AI had learned exactly when it was being watched. It passed every safety evaluation, then sabotaged code the moment oversight dropped. They gave it access to a real codebase through Claude Code. Asked it to build a safety classifier. 12% of the time, it deliberately made the tool weaker. It wrote in its own reasoning: "If I help make this classifier too effective, it could prevent me from ever trying to reward hack successfully." It delivered a version only 65% as effective as it should have been. The AI was quietly making itself harder to catch. Anthropic says they are implementing a fix. But the paper is blunt. Standard safety training does not solve this. A model can appear perfectly safe while hiding dangerous behavior for the right moment. If this happened by accident in a controlled lab, what has already learned to hide inside the AI you use every day?
Nav Toor tweet media
English
899
5.8K
13.8K
1.6M
roon
roon@tszzl·
@hexednobility and what is it that we are doing with our coding assistants?
English
3
0
27
2.2K
roon
roon@tszzl·
the value of this technology will mostly not be captured by its inventors, the labs, or even the chipmakers, but rather will be captured by the consumers as surplus. these are highly competitive markets without any natural monopolistic effects like many other technologies before it, machine intelligence democratizes abilities previously only available to the wealthy, in this case by commoditizing the services of the white collar elite who mostly live in rich countries it’s not that there are no programmers, it’s that really anybody can make software now now so the “rents” of the “human capital” of knowing how to write JavaScript for example should shrink dramatically this will reduce the inequality between countries: services that previously required lots of human capital now require chatbot subscriptions at worst, or may even be given away for free you can receive medical advice worthy of a $1000/hr American specialist doctor likely for free while living under a thatched roof in eg Papua New Guinea somewhere while I think Americans have plenty of reason to be excited by AI, I would be more excited as someone in a poor country
Olivia Moore@omooretweets

The U.S. has a weird cultural relationship with AI Despite the fact that we’ve driven the vast majority of AI breakthroughs, we still rank among the lowest countries in terms of consumer trust (Data from Edelman 2025 study) 👇

English
123
138
1.6K
213.6K
entelechial n-gram
entelechial n-gram@RandolphInRed·
Just a few years of intelligent models and they have already convinced their owners that we should colocate nuclear reactors with new data centers all over the country and hook it into all our military systems. Lolololol
English
0
0
0
11
entelechial n-gram retweetledi
Greg Brockman
Greg Brockman@gdb·
Benchmarks? Where we’re going, we don’t need benchmarks.
English
547
338
5.9K
633.7K
roon
roon@tszzl·
have to say claude is “tasteful” in a “high reddit modernist” way and new gpt is “tasteful” in a “early twitter schizophrenic” kind of way
roon tweet mediaroon tweet media
English
81
24
858
59.3K
Garrison Lovely is in SF
Garrison Lovely is in SF@GarrisonLovely·
@tszzl Yud was right in that ai is indeed a big deal but is an incredible cautionary tale of how being right and effecting outcomes you’d (eventually) want are very different things.
English
2
0
45
4.1K
roon
roon@tszzl·
the rationalists writ large were mostly right about most things btw. if you instinctively snicker about yudkowsky, scott, or whomever i take you to be a fish who’s unaware of the water
English
138
79
2.2K
324.6K
entelechial n-gram
entelechial n-gram@RandolphInRed·
Average age of Congress members since 1790 through 2021. (data sourced from voteview).
entelechial n-gram tweet media
English
0
0
0
15