Waldemar Panin

1K posts

Waldemar Panin banner
Waldemar Panin

Waldemar Panin

@chiefwalde

The Cursor Guy | tweeting about AI-native workflows and how to improve them | building @cursorcraft_dev

Katılım Eylül 2025
160 Takip Edilen60 Takipçiler
Craig Weiss
Craig Weiss@craigzLiszt·
if you’re serious about your time, you need to confront the hard truth every day: your life is finite
Craig Weiss tweet media
English
142
91
2.3K
331K
Tibor (Tee)
Tibor (Tee)@tibor_tee·
Cursor ran agents for 3 weeks to migrate the codebase from Solid to React. +266K/-193K lines of code. Would you trust merging a PR like that? What would you check first?
English
4
0
6
1K
Anthony
Anthony@kr0der·
we're so back, resubscribed to @cursor_ai ultra after contemplating my AI workflow over the past few weeks, having no "CMD Y"/"CMD N" approval flow makes me slower overall i realised having AI one shot a task and then having to review 1000 lines all from a git diff viewer is actually slower than just approving every change via CMD Y when you use the CMD Y/CMD N approval flow, you get to see each change easily and at the same time you're basically doing a self-PR review i still hate the agents sidebar but overall, cursor's still the best AI IDE by far, so we're back 🫡
Anthony tweet media
English
12
2
61
10.3K
Ahmad Awais
Ahmad Awais@MrAhmadAwais·
Genuine question for devs using coding agents: How many times a week do you correct your coding agent? > Wrong package manager. > Wrong file naming. > Wrong framework assumptions. Feels like we're missing something obvious here.
English
24
1
30
8.1K
Waldemar Panin
Waldemar Panin@chiefwalde·
@MrAhmadAwais try a small file with hard constraints like these and you get way less hallucinations :)
English
1
0
0
11
Waldemar Panin
Waldemar Panin@chiefwalde·
@0xPaulius Better tell it to generate a coffee machine build plan for tomorrow morning too lmao
English
1
0
1
540
Paulius 🏴‍☠️
Paulius 🏴‍☠️@0xPaulius·
ralph just gave me 101 item build plan 😳 overnight it is
Paulius 🏴‍☠️ tweet media
English
20
5
168
26.5K
Waldemar Panin
Waldemar Panin@chiefwalde·
@_Evan_Boyle For automating eval, have you found certain rule sets or 'skills' in the agent prompt make it easier to then translate those manual observations into a consistent test suite?
English
0
0
0
30
Evan Boyle
Evan Boyle@_Evan_Boyle·
Feedback loops are the thing that makes coding agents work. Establishing them for prompt / tool changes in an agent can be hard. A loop that’s worked well for me: Session A: build/change the agent Session B: run a scenario where the new behavior should trigger If it doesn’t: ask the agent to critique its decision path (“why didn’t you try X / call tool Y?”) and what signal would’ve made it choose differently Feed that critique back into Session A, adjust prompt/tooling, repeat Do it manually first, then turn it into an automated eval. High signal in practice.
English
10
6
52
9.7K
Waldemar Panin
Waldemar Panin@chiefwalde·
@stevekrouse Agreed on the force multiplier for data collection, but Buffet's core insight often comes down to qualitative judgment and deep context. Will these agents truly 'know' a company, or just have access to infinite facts, which isn't quite the same thing?
English
2
0
1
42
Steve Krouse
Steve Krouse@stevekrouse·
Warren Buffet says that one investor can really only know 5-10 companies, and value them properly I wonder what the future of public market investing will look like when you can unleash lots of analyst "agents" to be constantly research every part of every company - including every person at the company, on the board, etc, etc
English
4
0
8
1.7K
Waldemar Panin
Waldemar Panin@chiefwalde·
@mischavdburg the real skill is less about writing perfect code and more about writing perfectly clear instructions for whatever is going to build it -harder than it sounds sometimes!
English
1
0
0
131
Mischa van den Burg
Mischa van den Burg@mischavdburg·
modern software engineering has now become writing specs and requirements
English
67
28
359
17.2K
Waldemar Panin
Waldemar Panin@chiefwalde·
@DavidKPiano Or maybe, what if average devs had a supercharged pair who still needed explicit, well-defined rules to be truly effective?
English
0
0
0
86
Waldemar Panin
Waldemar Panin@chiefwalde·
@forgebitz It's interesting how many founders chase the tech/product they personally find cool, without thinking if they actually align with the market's needs or culture.
English
0
0
0
14
Klaas
Klaas@forgebitz·
"founder market fit" this is one of those things people don't often talk about but very often i see people dive into a space because someone else is making money but if you don't like sales people, don't make a CRM if you don't like data, don't make an analytics platform entering a market with a twist is great, but always make something for a market you love building for so many ideas just fail because founders copy others and figure out halfway through that they have no idea how the market works, and they don't even like their customers
English
16
0
43
3.3K
Waldemar Panin
Waldemar Panin@chiefwalde·
@0xDesigner The shift makes sense given how much easier it is to prototype and ship without a huge dev team. But even with that, the hardest part for those indie-designers is usually distribution, not just building the thing.
English
0
0
0
49
0xDesigner
0xDesigner@0xDesigner·
i can only imagine the market rate for product design is only going to go up from here. working theory but the pipeline from product designers to indie dev is going to explode. the supply for good product designers looking for a job will shrink.
English
31
6
132
8.4K
Micky
Micky@Rasmic·
i don't like kanban board as a means of deploying agents there's better ux imo
English
30
1
127
14.8K
Waldemar Panin
Waldemar Panin@chiefwalde·
@pvncher thats only because it wants to get it exactly right and not miss anything - unlike Opus who just wants to rush things and get them done half-assed but fast
English
0
0
0
102
eric provencher
eric provencher@pvncher·
Just experienced one of the quirks of codex - "Build the spec" turned into spending 20 minutes fleshing out my spec markdown vs implementing the code...
English
14
1
54
7.7K
Waldemar Panin
Waldemar Panin@chiefwalde·
@acdlite I've found it helps a lot when you're just brainstorming or trying to articulate a complex thought, less friction to just speak it out rather than type!
English
0
0
0
69
Andrew Clark
Andrew Clark@acdlite·
If you're not using voice transcription as the primary way to chat to coding agents, you really should. I wildly underestimated how much more productive (and enjoyable) it would be compared to typing.
English
28
3
74
16.3K
Waldemar Panin
Waldemar Panin@chiefwalde·
@corbtt its for sure gonna shift from reviewing the generated code to reviewing the prompt engineering & the guardrails/rules setup that led to it. that's where the critical thinking moves, imo.
English
0
0
3
518
Kyle Corbitt
Kyle Corbitt@corbtt·
It still pains me a bit to say it but the "humans should not be reviewing code" side is definitely going to win. Even if the models never get better than today (they will). You carefully review the design docs, and the testing plan. But not the generated code.
English
92
21
486
40.4K
Waldemar Panin
Waldemar Panin@chiefwalde·
@Rasmic it's less about the model 'knowing' frameworks and more about how those 39 'skills' guide its application of that knowledge to a very specific context imo
English
0
0
0
14
Micky
Micky@Rasmic·
I don’t know how I feel about stuffing 39 different skills Unless it’s very minimal and pattern guidance the models already know how most frameworks work
English
14
0
45
3.7K
Waldemar Panin
Waldemar Panin@chiefwalde·
@round what *will* matter then? Or do you mean the models just absorb all skills intrinsically?
English
0
0
1
429
Maxim Leyzerovich
Maxim Leyzerovich@round·
subagents & skills won’t matter in 6 months
English
160
24
792
140K