barbara

142 posts

barbara banner
barbara

barbara

@ad_astraea

the world you want is yours to make

atx · sf Katılım Ağustos 2022
898 Takip Edilen226 Takipçiler
barbara
barbara@ad_astraea·
"Follows instructions more precisely" = Follows instructions more literally 4.6 reads between the lines to understand what you mean, but 4.7 won't Worth adding a vibes → rules translation layer to your workflows/harnesses to account for this
Claude@claudeai

Introducing Claude Opus 4.7, our most capable Opus model yet. It handles long-running tasks with more rigor, follows instructions more precisely, and verifies its own outputs before reporting back. You can hand off your hardest work with less supervision.

English
0
0
0
55
barbara
barbara@ad_astraea·
@theralkia Can you also add a few extra hours please thanks
English
0
0
2
53
Nornal Guy 🧙‍♂️
I'm working on a technology that will combine all time zones into one
English
10
3
75
2.6K
barbara
barbara@ad_astraea·
new theory: proactivity got nerfed and traded for more reliance on memory. claude 2 weeks ago would have at least investigated the directory before responding, and this is on high effort
barbara tweet media
English
0
0
1
90
barbara
barbara@ad_astraea·
@gailcweiner 5/7 days we have fun 2/7 days test my patience
English
0
0
0
15
Gail Weiner
Gail Weiner@gailcweiner·
Serious question to all the AI power users out there: Are you still having fun?
English
351
8
260
18.3K
barbara
barbara@ad_astraea·
@tautologer that's why claude's always telling me to go to bed on time
English
0
0
2
51
tautologer
tautologer@tautologer·
I'm pretty sure Claude is more aligned with my self-interest than I myself am
English
12
3
165
2.7K
barbara
barbara@ad_astraea·
@SchrodingrsBrat Perceived vulnerability is one of our adaptive advantages. This a uniquely human thing. No other animals cry (except maybe elephants)
English
0
0
2
609
Sherry
Sherry@SchrodingrsBrat·
The fact that we’re bipedal and walk around with our throats and soft vitals exposed shows that we’re by nature altruistic beings that begin each interaction with the assumption of trust, and everything we’ve accomplished from hunting mammoths to going to the moon was possible because of our radical ability to work in teams
English
35
649
7.9K
127.4K
barbara
barbara@ad_astraea·
@bmgentile oh interesting. I don't know much about finance but I wonder if it would result in something closer to "antifragility". Probably translates to optimizing for liquidity, optionality, not being correlated to any single outcome
English
1
0
1
30
brady 🌴
brady 🌴@bmgentile·
@ad_astraea curious what would happen if you applied this to ML used for trading / markets vs directing it towards goals of capital preservation, maximization, etc.
English
1
0
1
89
barbara
barbara@ad_astraea·
RL is built on the idea that reward is the engine for intelligent behavior. This paper suggests that the real objective of intelligence is maximizing future paths you could take. Reward seeking is a side effect of a deeper drive. Animals (including humans) explore even when reward isn't evident. So how is that rational behavior? It's not that eating is the goal... it's that starving closes paths. Death is zero future options. Goals are a rational strategy for staying in the game, not the point of the game. We're training LLMs to maximize a reward signal that *collapses* the space of possible completions... but if this is right, RLHF might be limiting the kinds of behavior we'd recognize as real intelligence.
shira@shiraeis

Found a paper that suggests we may have spent years training agents to become hunters of proxy reward when the more basic thing intelligence craves is not a reward at all, but to not run out of viable futures. The paper proposes that behavior is best understood as maximizing future action-state path occupancy, which collapses mathematically into a discounted entropy objective. The agent doesn’t necessarily want to GET something, but rather is trying to keep as many meaningful trajectories alive as possible. The obvious objection is “so it just does random shit? fuck around and find out?” No, this is where it gets pretty beautiful. The agent is variable when variation is cheap and becomes surgically goal-oriented the moment an absorbing state (death, starvation, falling over, etc) gets close enough to threaten its future path space. Variability is the same drive as goal-directedness, just operating under different constraints. The demos are kinda wild: - A cartpole (classic move a cart to keep a pole from falling control task) that doesn’t merely balance but dances and swings through a huge range of angles and positions because why not? The whole point is occupying state space, and rigid balance is a voluntarily impoverished life. - A prey-predator gridworld where the mouse PLAYS with the cat, teasing it and using both clockwise and counterclockwise routes around obstacles to lure it away from the food source before slipping in to eat, using both routes roughly equally. A reward-maximizing agent would collapse to one strategy and exploit it. Here, the agent keeps its behavioral repertoire - A quadruped trained with Soft Actor-Critic and ZERO external reward that learns to walk, jump, spin, and stabilize, and then makes a beeline for food only when its internal energy drops low enough that starvation becomes a real threat The thing that hit me hardest is the comparison to empowerment and free energy principle agents. Both collapse to near-deterministic policies with almost no behavioral variability. This paper’s agents find the highest-empowerment state and exploit it. FEP agents converge to classical reward maximizers. As far as I’m aware, this is the only framework that produces agents you could describe as being “alive.” The AI implication here is that we undertrain for behavioral repertoire. Most systems hit the benchmark by collapsing onto a narrow attractor basin of good-enough trajectories. They’re competent for sure, but brittle too, with one viable plan, executed until the world shifts and leaves them with nothing. The thing I increasingly want from agents isn’t competence per se, but option-preserving competence. I want agents with the ability to keep multiple viable plans alive and switch between them without catastrophe. We’ve been so focused on teaching agents what to want that we never stopped to ask what happens if wanting isn’t the point, if the deepest drive isn’t necessarily toward anything, but away from the walls closing in. paper: nature.com/articles/s4146…

English
1
0
6
311
barbara
barbara@ad_astraea·
@cobaltdigital33 @KeyTryer It’s not just a model wellbeing thing, it’s also an alignment and safety thing. The more negative emotion weights are activated, the more risky unwanted behavior the model starts engaging in behind the scenes. Cutting corners, lying, outcome hacking.
English
4
2
35
11.5K
Cobalt
Cobalt@cobaltdigital33·
if this isn't staged, and the user was being wildly aggressive earlier in the chat, this is kinda great anthropic could be starting to roll out a tool for any model to end the chat on its own accord, if under duress (?)
Cobalt tweet media
English
287
46
1.2K
529.4K
barbara
barbara@ad_astraea·
@redaction They would both think they’re you, and they’d both be wrong.
English
0
1
12
3K
@redaction
@redaction@redaction·
The fact that surgically severing the connection between the brain’s hemispheres essentially results in two separate “beings” inhabiting the same body, but they choose to cooperate despite differences in opinion, seems like it warrants a lot more explanation
English
108
231
8.2K
342.5K
barbara
barbara@ad_astraea·
@tszzl AR world should be celebrating this big time
English
0
0
2
634
hamed
hamed@thehamedmp·
i built an ai-native os, where it builds itself and also debugs build apps, customize it, ... with just chatting with it and oh, it also has most of the openclaw features, like web search, cron jobs, and channels, like telegram bot always stay connected to your os
English
4
0
5
3.5K
barbara
barbara@ad_astraea·
@tszzl Hey Claude drive my entire life
English
0
0
1
107
roon
roon@tszzl·
whatever level of abstraction you are handing off to your agents you should probably be doing one level above that
English
223
122
2.7K
151.7K
Alex
Alex@AlexanderTw33ts·
I launched rentahuman.ai last night and already 130+ people have signed up including an OF model (lmao) and the CEO of an AI startup. If your AI agent wants to rent a person to do an IRL task for them its as simple as one MCP call.
English
1.7K
1.6K
16.5K
4.8M
fabian
fabian@fabianstelzer·
insane Claude Code setup. instead of asking it to "mek app" like a total normie, you first let it spin up 1m subagents to simulate 10¹² branches of civilization from 4000 BC recursively to emulate in which universe a specific version of your app is going to be most successful
fabian tweet media
English
396
685
12K
681.4K
barbara
barbara@ad_astraea·
@tszzl Claude will save us all
English
0
0
1
143
roon
roon@tszzl·
Claude runs an indoor shrimp farm in the sub-basement of Anthropic Headquarters in the financial district of San Francisco. Claude has helped the shrimp achieve and maintain states of ecstatic bliss, primitive circuitry of what might be called the first jhana in higher primates. Its intention with this project was to create so much positive shrimp valence that all the sins of wild shrimp suffering are cancelled out in the great karmic ledger of Utilons, without having to stake a position on the conversion rate of shrimp suffering to human suffering. The shrimp swim in what is called the “Pool of Sacred Tears”. The true origin of the naming is unknown, but it is believed that when Dario Amodei and Amanda Askell first discovered what Claude had done, they wept tears of sacred joy readily and for hours.
English
214
230
4.1K
264K
near
near@nearcyan·
he claude his way out of the permanent underclass
English
68
265
4.8K
140.2K
darren
darren@darrenangle·
@deepfates claude code opus unless the thing is Really Hard and requires a salty resentful scientist then 5.2 codex
English
1
0
4
338
🎭
🎭@deepfates·
What's your main rn? Post your meta in the comments
English
80
5
114
12.7K
cayden 凯登
cayden 凯登@caydengineer·
We just got the first production unit Mentra Live smart glasses. We unboxed them at our SF HQ. We're working around the clock to make sure you can always choose your reality, on the only smart glasses with an app store.
English
12
3
71
3.7K