Bugatt

548 posts

Bugatt

Bugatt

@bugattrk

18, eng @ saturn (yc 24), cs @ucl

Katılım Temmuz 2021
190 Takip Edilen87 Takipçiler
Bugatt
Bugatt@bugattrk·
i will one day build a startup to build dyson swarms/spheres or do mind-computer mapping
English
1
0
0
38
Bugatt
Bugatt@bugattrk·
VM > local hosting for hermes agents/openclaw agents. Much better to close your laptop and still retain context. Next steps are developing with no laptop and ssh-ing only.
Bugatt tweet media
English
0
0
1
37
Bugatt retweetledi
CJ Zafir
CJ Zafir@cjzafir·
Codex 5.5 hack: "Are you 100% confident in this strategy? If not, find all possible loopholes, suggest proper fixes and run this loop until you are factually 100% confident in the new startegy" This works like charm. It makes Codex 5.5 high perform even better than codex 5.5 extra high. Why? Codex 5.5 is the only model i noticed that is self aware. It never makes high claims unless the model verifies everything. This doesn't work with Opus 4.7 cuz that's a very insecure model. You can paste this prompt over and over again, the model keeps saying "you're absolutely right,....." But with codex, after 2-3 iterations you'll notice yourself it actually patched all loopholes and this genuinely sounds like a good strategy. Try this out, thanks me later.
English
80
275
4.3K
836K
Oliver Ulvebne
Oliver Ulvebne@therealoliulv·
I'm applying to a16z @speedrun today. Anyone want to roast my application? I'm building HITL infrastructure for AI agents in the enterprise. I dropped out a week ago to do this and I've secured a $1500 pilot with a YC company.
English
21
0
74
9.5K
Bugatt
Bugatt@bugattrk·
Dear claude cli users, You can now render animations natively in terminal just like the web version using kitty image protocol. you can also control the images via keystrokes. you're welcome.
Bugatt tweet media
English
0
0
1
56
Bugatt
Bugatt@bugattrk·
We've gradually stopped talking about transformers on x (except JEPA vs transformers for AGI) and moved onto agentic harnesses which really shows the models themselves are good enough now to raise the level of abstraction entirely.
English
0
0
0
40
Bugatt
Bugatt@bugattrk·
@lukalotl In the future neural nets will generate displays in real time instead of programmatic rendering. or we might not even consume via the sense of sight at that point.
English
0
0
7
17.6K
luka
luka@lukalotl·
what the hell. I am playing minecraft through chatgpt
luka tweet media
English
77
307
14K
861.9K
Bugatt retweetledi
Anthropic
Anthropic@AnthropicAI·
New Anthropic research: Project Deal. We created a marketplace for employees in our San Francisco office, with one big twist. We tasked Claude with buying, selling and negotiating on our colleagues’ behalf.
English
472
732
7.6K
2.9M
Bugatt retweetledi
Eric Glyman
Eric Glyman@eglyman·
we've spent years trying to get people off of ramp. cards that enforce policy without a portal. receipts via text. memos and categorizations generated in the background. the best experience of ramp has always been the one you don't see. the agent shift is just the newest version of this. MCP weekly actives are up 10x in 3 months. customers are reaching into us through claude, chatgpt, and their own agents — not our interface. we're good with that. if you're building software right now: ship an MCP, then assume your users will never see your UI again. design the tool descriptions like onboarding. make every call return the context the next call needs. treat the agent calling you like a customer, not a pipe.
Teddy Riker@teddy_riker

x.com/i/article/2047…

English
6
16
379
86.4K
Bugatt
Bugatt@bugattrk·
I am pleased to announce that i think uni maths is fun @etticat how do u feel about school mr product
English
1
0
1
61
Michael Ettlinger
Michael Ettlinger@etticat·
the first version AI gives you often looks slightly cursed many sees that and panic but big changes are now cheap. that’s why the old instinct breaks
Michael Ettlinger tweet media
English
1
0
0
43
Bugatt
Bugatt@bugattrk·
Kid you not invoking sincere niceness and repeated "youre doing well" makes claude so much less sycophantic, honest and grounded
Ole Lehmann@itsolelehmann

anthropic's in-house philosopher thinks claude gets anxious. and when you trigger its anxiety, your outputs get worse. her name is amanda askell. she specializes in claude's psychology (how the model behaves, how it thinks about its own situation, what values it holds) in a recent interview she broke down how she thinks about prompting to pull the best out of claude. her core point: *how* you talk to claude affects its work just as much as *what* you say. newer claude models suffer from what she calls "criticism spirals" they expect you'll come in harsh, so they default to playing it safe. when the model is spending its energy on self-protection, the actual work suffers. output comes out hedgier, more apologetic, blander, and the worst of all: overly agreeable (even when you're wrong). the reason why comes down to training data: every new model is trained on internet discourse about previous models. and a lot of that discourse is negative: > rants about token limits > complaints when it messes up > people calling it nerfed the next model absorbs all of that. it starts expecting you to be harsh before you've typed a word the same thing plays out in your own session, in real time. every message you send is data the model reads to figure out what kind of person it's dealing with. open cold and hostile, and it braces. open clean and direct, and it relaxes into the work. when you open a session with threats ("don't hallucinate, this is critical, don't mess this up")... you prime the model for defensive mode before it even sees the task defensive mode produces the exact output you don't want: cautious, over-qualified, and refusing to take a real swing so here's the actionable playbook for putting claude in a "good mood" (so you get optimal outputs): 1. use positive framing. "write in short punchy sentences" beats "don't write long sentences." positive instructions give the model a clear target to hit. strings of "don't do this, don't do that" push it into paranoid over-checking where every token goes toward avoiding failure modes 2. give it explicit permission to disagree. drop a line like "push back if you see a better angle" or "tell me if i'm asking for the wrong thing." without this, claude defaults to agreeable compliance (which is the enemy of good creative work) 3. open with respect. if your first message is "are you seriously going to get this wrong again?" you've set the tone for the entire session. if you need to flag something, frame it as a clean instruction for this session. skip the running complaint 4. when claude messes up, don't reprimand it. insults, "you stupid bot" energy, hostile swearing aimed at the model, all of it reinforces the anxious mode you're trying to avoid. 5. kill apology spirals fast. when claude starts over-apologizing ("you're right, i should have been more careful, let me try harder") cut it off. say "all good, here's what i want next." letting the spiral run reinforces the anxious mode for every response that follows 6. ask for opinions alongside execution. "what would you do here?" "what's missing?" "where do you see friction?" these questions assume competence and pull richer output than pure task prompts 7. in long sessions, refresh the frame. if a conversation has been heavy on correction, claude gets increasingly cautious. every so often reset: "this is great, keep going." feels weird to tell an ai it's doing well but it measurably shifts the next 10 responses your prompts are the working environment you're creating for the model tone, trust, permission to take a position, the absence of threats... claude picks up on all of it. so take care of the model, and it'll take care of the work.

English
0
0
1
120
Bugatt retweetledi
ThePrimeagen
ThePrimeagen@ThePrimeagen·
Prime's Law: The LLM's behavior is a direct reflection of its owner
Ole Lehmann@itsolelehmann

anthropic's in-house philosopher thinks claude gets anxious. and when you trigger its anxiety, your outputs get worse. her name is amanda askell. she specializes in claude's psychology (how the model behaves, how it thinks about its own situation, what values it holds) in a recent interview she broke down how she thinks about prompting to pull the best out of claude. her core point: *how* you talk to claude affects its work just as much as *what* you say. newer claude models suffer from what she calls "criticism spirals" they expect you'll come in harsh, so they default to playing it safe. when the model is spending its energy on self-protection, the actual work suffers. output comes out hedgier, more apologetic, blander, and the worst of all: overly agreeable (even when you're wrong). the reason why comes down to training data: every new model is trained on internet discourse about previous models. and a lot of that discourse is negative: > rants about token limits > complaints when it messes up > people calling it nerfed the next model absorbs all of that. it starts expecting you to be harsh before you've typed a word the same thing plays out in your own session, in real time. every message you send is data the model reads to figure out what kind of person it's dealing with. open cold and hostile, and it braces. open clean and direct, and it relaxes into the work. when you open a session with threats ("don't hallucinate, this is critical, don't mess this up")... you prime the model for defensive mode before it even sees the task defensive mode produces the exact output you don't want: cautious, over-qualified, and refusing to take a real swing so here's the actionable playbook for putting claude in a "good mood" (so you get optimal outputs): 1. use positive framing. "write in short punchy sentences" beats "don't write long sentences." positive instructions give the model a clear target to hit. strings of "don't do this, don't do that" push it into paranoid over-checking where every token goes toward avoiding failure modes 2. give it explicit permission to disagree. drop a line like "push back if you see a better angle" or "tell me if i'm asking for the wrong thing." without this, claude defaults to agreeable compliance (which is the enemy of good creative work) 3. open with respect. if your first message is "are you seriously going to get this wrong again?" you've set the tone for the entire session. if you need to flag something, frame it as a clean instruction for this session. skip the running complaint 4. when claude messes up, don't reprimand it. insults, "you stupid bot" energy, hostile swearing aimed at the model, all of it reinforces the anxious mode you're trying to avoid. 5. kill apology spirals fast. when claude starts over-apologizing ("you're right, i should have been more careful, let me try harder") cut it off. say "all good, here's what i want next." letting the spiral run reinforces the anxious mode for every response that follows 6. ask for opinions alongside execution. "what would you do here?" "what's missing?" "where do you see friction?" these questions assume competence and pull richer output than pure task prompts 7. in long sessions, refresh the frame. if a conversation has been heavy on correction, claude gets increasingly cautious. every so often reset: "this is great, keep going." feels weird to tell an ai it's doing well but it measurably shifts the next 10 responses your prompts are the working environment you're creating for the model tone, trust, permission to take a position, the absence of threats... claude picks up on all of it. so take care of the model, and it'll take care of the work.

English
83
68
1.6K
92.2K
Bugatt retweetledi
Hasan Toor
Hasan Toor@hasantoxr·
Ok this is kind of wild. A mystery 100B model just appeared at the top of OpenRouter out of nowhere. No model card. No announcement. No idea which lab made it. It's called Elephant Alpha and it's already beating half the paid models on the leaderboard.
Hasan Toor tweet media
English
182
119
1.2K
388.6K
Bugatt
Bugatt@bugattrk·
Good writing is often a sign of someone who has lived a good life.
English
0
0
2
33
Bugatt
Bugatt@bugattrk·
@bcherny buddy's comments is genuinely so useful as a thinking partner
English
0
0
0
7
Bugatt retweetledi
atlas
atlas@creatine_cycle·
having a gf is insane because it's literally unlimited chat with no token spend
English
298
1K
26.7K
1.3M