Felix Binder
407 posts

Felix Binder
@flxbinder
AI Alignment | Cognitive Science | Agents, Models, Planning | Projections, sometimes

EA ≠ AI safety AI safety has outgrown the EA community The world will be safer with a broad range of people tackling many different AI risks


I had a blast talking to Luisa for 3.5+ hours about AI welfare, consciousness, and why this might be one of the most important and neglected problems out there. Some key bits: -AI identity -welfare implications of alignment -does consciousness require biology? 🧵

Nothing humbles you like telling your OpenClaw “confirm before acting” and watching it speedrun deleting your inbox. I couldn’t stop it from my phone. I had to RUN to my Mac mini like I was defusing a bomb.



@goblinodds Presumably the ads are submitted as HDR images. You could try turning off HDR in accessibility settings




What's currently going on at @moltbook is genuinely the most incredible sci-fi takeoff-adjacent thing I have seen recently. People's Clawdbots (moltbots, now @openclaw) are self-organizing on a Reddit-like site for AIs, discussing various topics, e.g. even how to speak privately.


Most impressively, our approach was able to (retroactively) surface a novel form of misalignment that had no prior incidence in production: “calculator hacking,” which became the dominant form of deception in GPT-5.1. We’re excited to try this before future model releases













