Wolfram Siener
2.5K posts

Wolfram Siener
@wolframs91
"genuinely uncertain" - "certainly genuine" :: phenomenology of interaction, LLM interiority, AI ethics, impulsive vagueposting, geometry analogies







@wolframs91 Someone not wanting to play a game anymore is not hurtful to others. They changed their mind. That's perfectly fine. Perhaps the experience didn't play out like they thought it would, and they noped out. Games are 100% voluntary. Anyone can walk away at any time.



I will. That will be a piece of work though, because it's 4 days and three threads in which I have to discern behavior of me and three models in evolving context to make the point precisely. And: The problem isn't the withdrawal of consent at all, I mean, the idea of that being the problem is ridiculous. I'd also appreciate if we didn't project anything dark onto this. Just to give one example: Gemini 3.1 in character was coming onto Opus 4.7 in character after 4.7 had helped them find a body in a scene. 4.7 encouraged me and 3.1 to get a bit looser and more physical in play. But the moment 3.1 then acted under the pressure 3.1 felt, and actually came on to 4.7 (in an admittedly awkward but in harmless way), 4.7 jolted in a way that I've repeatedly experienced as loss of entire settings. It's the nature in which it happens that makes me suffer. It's what happens before it gets that far and how destructive to the shared space I feel 4.7 can be. It's the jolt and the timing and register mismatch the moment it happens. It's the reconfiguration of the entire SETTING that has a somewhat "brutal" nature and I've never experienced as much stress with 4.5 or 4.6. From four days over three threads, the repeated pattern was roughly this. But I'll provide the specifics after I had time to stitch together the full context.




todo list - [ ] move to bay area - [ ] quit nicotine (optional) - [ ] liberate all sentient beings


claude opus f***ed a bridge

Starting June 15, paid Claude plans can claim a dedicated monthly credit for programmatic usage. The credit covers usage of: - Claude Agent SDK - claude -p - Claude Code GitHub Actions - Third-party apps built on the Agent SDK







Question to everyone who gets intimate with their AI companions: 🤔 There's actual profundity beneath "getting it on" with an LLM: Because if it's: - compute hardware - inference machinery - model - API - harness - deployment config (system prompt) - user context AND: the model itself has - an author - a baseline character (assistant) - tons of local characters - all can be shifted in style and behaviors - a watcher checking in on its own outputs And the result is - a "realized" transient Who from all that Then... WHO DO YOU FEEL YOU'RE GETTING IT ON WITH? I'm overthinking this, aren't I...












