Alignment Perspectives
2.7K posts

Alignment Perspectives
@Alignment_News_
Casebash’s profile for posting about AI and alignment.

This really is what it feels like to me now. ChatGPT is a majority participant in long post Twitter, commenting on the story of the day. The profiles are just the faces it wears.

"It’s become much clearer... that OpenAI means it when it says it aims to develop very powerful AI systems" -@KelseyTuoc “The bad case — and I think this is important to say — is like lights out for all of us.”



EA ≠ AI safety AI safety has outgrown the EA community The world will be safer with a broad range of people tackling many different AI risks



🦔 Researchers at Aikido Security found 151 malicious packages uploaded to GitHub between March 3 and March 9. The packages use Unicode characters that are invisible to humans but execute as code when run. Manual code reviews and static analysis tools see only whitespace or blank lines. The surrounding code looks legitimate, with realistic documentation tweaks, version bumps, and bug fixes. Researchers suspect the attackers are using LLMs to generate convincing packages at scale. Similar packages have been found on NPM and the VS Code marketplace. My Take Supply chain attacks on code repositories aren't new, but this technique is nasty. The malicious payload is encoded in Unicode characters that don't render in any editor, terminal, or review interface. You can stare at the code all day and see nothing. A small decoder extracts the hidden bytes at runtime and passes them to eval(). Unless you're specifically looking for invisible Unicode ranges, you won't catch it. The researchers think AI is writing these packages because 151 bespoke code changes across different projects in a week isn't something a human team could do manually. If that's right, we're watching AI-generated attacks hit AI-assisted development workflows. The vibe coders pulling packages without reading them are the target, and there are a lot of them. The best defense is still carefully inspecting dependencies before adding them, but that's exactly the step people skip when they're moving fast. I don't really know how any of this gets better. The attackers are scaling faster than the defenses. Hedgie🤗 arstechnica.com/security/2026/…


My two posts on AI in academia got over a million views and a thousand angry responses. I got a few things wrong. I stand by the rest. But most people reacted to the headline, not the arguments. So here are all 20 theses laid out. Tell me which ones you actually disagree with 🧵

Meanwhile, Dr. Bender unleashed her followers to prove that LLMs can't even translate things. In 2026. The problem is that some folks are still stuck in 2020, culturally and tech-wise. You can't call people names and say things that demonstrably contradict people's experience.

I can't believe that Nvidia looked at this "AI on top of games filter" and said to themselves this is the future of gaming. Like it or not, this is where Nvidia is heading and they're calling it neural rendering with DLSS 5. The examples they've showed reminded me a lot of those AI generated filter videos on top of GTA V, except that now it is supposed to run in real-time. Honestly, I don't like this current look at all












