Ben Anderson

242 posts

Ben Anderson banner
Ben Anderson

Ben Anderson

@anchorstack_dev

Senior engineer. Building things that actually work, fixing things that don't. I fix vibe coded apps so they hold up when your users need them.

San Antonio, TX, USA เข้าร่วม Şubat 2026
238 กำลังติดตาม75 ผู้ติดตาม
ทวีตที่ปักหมุด
Ben Anderson
Ben Anderson@anchorstack_dev·
Vibe coding got you to MVP and your first customers. Congrats! Now you have to: Onboard a dev who's never seen the code Debug with no observability Refactor without reading every file Add a feature without breaking 3 more That's the gap between "it works" and "it's engineered"
English
2
0
4
207
Ben Anderson
Ben Anderson@anchorstack_dev·
@plainionist One pattern I find myself following is: 1) automated unit tests 2) automated integrations tests 3) and doing manual smoke tests to make sure the functionality of the changed code remains.
English
0
0
0
6
Seb
Seb@plainionist·
Serious question: Do you trust your automated tests enough to deploy to production without testing manually? 🤔
English
63
0
18
2.6K
Ben Anderson
Ben Anderson@anchorstack_dev·
@GergelyOrosz I think one of the joys of writing is producing something with your voice. To me writing with AI feels like cheating way more than coding with AI.
English
0
0
0
131
Gergely Orosz
Gergely Orosz@GergelyOrosz·
AI - in theory - should make writing easier, thus expressing ideas should be a lot easier. And yet, I don't see all that much more things that are worthwhile to read. Sure, there's a lot more junk. But I don't see more interesting eng blogs, personal tech-related blogs etc. What is going on? (Is this a discoverability issue? Or are people / teams not writing/sharing all that much more, indeed?)
English
149
16
340
30.5K
Ben Anderson
Ben Anderson@anchorstack_dev·
@KaiXCreator Can you call yourself a founder if your whole product was built by employees?
English
0
0
3
58
Kaito
Kaito@KaiXCreator·
Can you call yourself a founder if your entire product was built by Claude?
English
451
14
351
68.2K
Ben Anderson
Ben Anderson@anchorstack_dev·
@zuess05 This is scary how accurate it is. I think over time standards will present themselves that make this easier.
English
0
0
0
18
Suhas
Suhas@zuess05·
2024: "AI is going to replace all software engineers in a year." 2026: The entire codebase is hallucinating and nobody knows where the core logic ends and the AI spaghetti begins. We literally just automated the creation of legacy tech debt.
English
33
58
595
15.6K
Ben Anderson
Ben Anderson@anchorstack_dev·
@karpathy I have a hard time believing in HTML over markdown but I will give it a try. The tradeoff is being able to write and comment on specific lines in the markdown. I don't think we are there yet with HTML I could be swayed. (This is basically Claude Design.)
English
0
0
1
459
Andrej Karpathy
Andrej Karpathy@karpathy·
This works really well btw, at the end of your query ask your LLM to "structure your response as HTML", then view the generated file in your browser. I've also had some success asking the LLM to present its output as slideshows, etc. More generally, imo audio is the human-preferred input to AIs but vision (images/animations/video) is the preferred output from them. Around a ~third of our brains are a massively parallel processor dedicated to vision, it is the 10-lane superhighway of information into brain. As AI improves, I think we'll see a progression that takes advantage: 1) raw text (hard/effortful to read) 2) markdown (bold, italic, headings, tables, a bit easier on the eyes) <-- current default 3) HTML (still procedural with underlying code, but a lot more flexibility on the graphics, layout, even interactivity) <-- early but forming new good default ...4,5,6,... n) interactive neural videos/simulations Imo the extrapolation (though the technology doesn't exist just yet) ends in some kind of interactive videos generated directly by a diffusion neural net. Many open questions as to how exact/procedural "Software 1.0" artifacts (e.g. interactive simulations) may be woven together with neural artifacts (diffusion grids), but generally something in the direction of the recently viral x.com/zan2434/status… There are also improvements necessary and pending at the input. Audio nor text nor video alone are not enough, e.g. I feel a need to point/gesture to things on the screen, similar to all the things you would do with a person physically next to you and your computer screen. TLDR The input/output mind meld between humans and AIs is ongoing and there is a lot of work to do and significant progress to be made, way before jumping all the way into neuralink-esque BCIs and all that. For what's worth exploring at the current stage, hot tip try ask for HTML.
Thariq@trq212

x.com/i/article/2052…

English
515
831
8.9K
798.1K
Ben Anderson
Ben Anderson@anchorstack_dev·
@SimonHoiberg How do you ensure quality in the code produced by your agents?
English
0
0
0
51
Simon Høiberg
Simon Høiberg@SimonHoiberg·
In the past 3 weeks, I've pushed: - +25 commits - +4300 lines of code On average, every day. And I haven't opened VSCode (or any other code editor) one single time. All I do is talk to my agents through voice on Telegram. And I'm more productive than ever when I don't need to actually code. I had no idea we'd get here this fast.
English
7
0
18
1.8K
Ben Anderson
Ben Anderson@anchorstack_dev·
A spec is not a wish list. It is the contract between the idea and the build: what the app should do, what it should not do, what edge cases matter, and how we know it works. If you skip the spec, you are asking the AI to guess your product.
English
0
0
0
9
Ben Anderson
Ben Anderson@anchorstack_dev·
A vibe-coded MVP is the sketch. The product starts when auth, data modeling, deploys, monitoring, backups, and recover.
English
0
0
0
9
Ben Anderson
Ben Anderson@anchorstack_dev·
Skills are becoming the new functions. Small, reusable units of judgment you can drop into an agent workflow. What are your favorite skills so far? I’m building a list. Drop yours below 👇
English
0
0
1
9
Ben Anderson
Ben Anderson@anchorstack_dev·
Day 7/7 "Done" should mean the branch is ready for review: synced, linted, type-checked, tested, scanned, committed. as-finish turns that into a configurable pipeline. It runs steps in order, stops on failure, and gives you a real finish report. Vibe-coded apps need this boundary.
English
0
0
0
6
Ben Anderson
Ben Anderson@anchorstack_dev·
Day 6/7 Fixing the first broken thing is how bugs come back. as-rca makes Claude reproduce the failure, then keep asking "what is causing this?" down the causation chain until it hits something with no upstream cause in the system. That's the root. Fix that.
English
1
0
1
25
Ben Anderson
Ben Anderson@anchorstack_dev·
The mistake vibe-coded apps make isn't in the feature code. It's in the assumptions underneath it. No migration workflow. No observability. No rollback. Auth that works for the founder. The demo ships. The foundation doesn't.
English
1
0
0
16
Ben Anderson
Ben Anderson@anchorstack_dev·
Hot take: the reason vibe-coded apps fail in production isn't AI. It's that most builders have never run a production system. AI makes it fast to build. It doesn't teach you what operations actually requires. What's the thing you had to learn the hard way?
English
1
0
1
22
Ben Anderson
Ben Anderson@anchorstack_dev·
5 production gaps I find in almost every vibe-coded app: 1) No idempotency on payment webhooks 2) Secrets in git history 3) Schema lives in the dashboard, not the repo 4) No structured logging so failures are invisible 5) Auth checked on the client, not the server Fix these before launch. Everything else can wait.
English
0
0
0
22
Ben Anderson
Ben Anderson@anchorstack_dev·
@asaio87 I agree to a certain extent. With proper context separation and architecture they are still useful. However, if you are at a point where you are just prompting co-pilot that is bad. We need to get to the point where they are defining clear specs and implementing from there.
English
0
0
0
94
andrei saioc
andrei saioc@asaio87·
Most people hyping Claude Code or AI tools in general have never worked with these tools for production apps. Its nice to build small tools, but as you build complex apps, things tend to slow way down. the real time gain is at best 25%
English
20
1
37
1.8K
Ben Anderson
Ben Anderson@anchorstack_dev·
I was looking at an interesting architecture the other day where they were running a couple of fine tunes for quick easy inference and using the frontier models for handling conversation turns. I think the truth is that is more sophisticated than 99% of the setups out there. Most people have a single API calling OpenAI or Anthropic. You've inspired me to start thinking about self-hosting a small model at home. 🤔
English
0
0
0
39
Simon Høiberg
Simon Høiberg@SimonHoiberg·
Open-weight models are still so far behind OpenAI and Anthropic for any serious work. If you want to full agent experience, use the frontier models. I do. On the other hand - we need to get a foot in the door with models we can run ourselves now. GPT-5.5 and Opus 4.7 are VC handouts. The models are SO expensive to run, and OpenAI/Anthropic are burning BILLIONS to subsidize the cost for you. But sooner or later, the money will run out and the prices will skyrocket. And one thing is the price, but that's only half the problem. A few tech companies are about to own the operating layer of 95% of all businesses (...again). Right now is the chance to prepare for soverignity. Models like qwen, kimi, deepseek - they feel very far behind, and they can do what frontier models do. But it's a great place to start. Get used to the process. Run on hardware you buy (like DGX Spark or Mac Mini) - or rent GPU instances on platforms like Vast AI. Then start with boring tasks. Classification. Routing. Summaries. Internal search. You're gonna be doing this soon. Or a few US tech companies are gonna own you once again.
Simon Høiberg tweet media
English
4
2
24
2.3K
Ben Anderson
Ben Anderson@anchorstack_dev·
Skills week 5/7 This one is super important and simple. as-secret-scan Scans for accidentally committed API keys, tokens, and credentials across staged changes, working tree, or full git history. The reality is that AI coding tools write fast. They might also commit your Stripe key. as-secret-scan catches what vibe coding leaves behind before it becomes an incident.
English
1
0
1
21
Ben Anderson
Ben Anderson@anchorstack_dev·
I had a founder message me that they wanted me to review their app. I asked them to send me their repo so I could take a look at it. They asked, what is a repo? No shame here, but if you fall into this bucket lets talk.
English
0
0
1
20
Ben Anderson
Ben Anderson@anchorstack_dev·
Founder sent me their repo. Said "it's working great, just needs a polish pass." Found their .env committed in git history. Supabase service role key. Stripe secret. OpenAI key. All of it. The app was live. With real users. We're not talking about a sophisticated attack. Anyone who cloned the repo had every credential in plain text.
English
0
0
0
57
Ben Anderson
Ben Anderson@anchorstack_dev·
The most interesting vibe coding stat is not productivity. It is this: 40% of junior devs admit to deploying AI-generated code they do not fully understand. (and the other 60% will get fired if they do) That is the entire market for production-readiness, audits, and guardrails.
English
1
0
2
31