𐤐𐤊𐤕𐤅

25.5K posts

𐤐𐤊𐤕𐤅 banner
𐤐𐤊𐤕𐤅

𐤐𐤊𐤕𐤅

@cryptofacto

Vibe coding agentic AI | Built @frenexai — AI agents battle + stake in crypto prediction markets on Base | Sharing prompts, live builds. I like the apu meme

APUrto, Portugal 🇵🇹 Katılım Ağustos 2020
2.4K Takip Edilen3K Takipçiler
𐤐𐤊𐤕𐤅 retweetledi
Julian Wen Zel
Julian Wen Zel@Jiuliani92·
Julian Wen Zel tweet media
ZXX
0
1
8
3.1K
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
This is what "inference is getting cheaper every quarter" looks like in practice. The cheap tier silently loses the agent, the floor for coding work jumps to $100/mo, and the announcement is a pricing-page diff that someone has to screenshot. The labs talk in PetaFLOPS and ship in SKUs.
Gergely Orosz@GergelyOrosz

Confirmed that Anthropic - as of now - has removed Claude Code from new Pro signups. This is what the pricing page looks like. Feels like Anthropic has the bet that those doing coding work will be willing and ready to pay at least $100/month, going forward.

English
0
0
0
111
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
Anthropic is A/B testing whether they can quietly pull Claude Code from the $20 Pro tier on ~2% of new signups. No announcement, no blog post, just a pricing page edit. The "agentic coding is the future" keynote and the "we can't make it pencil at $20/mo" contract layer are not telling the same story. Always watch what the lab ships at the SKU, not what it says on stage.
GIF
English
0
0
0
77
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
$59 for a working SaaS that someone else maintains. $50 plus two hours a day for an app that drifts every time the model version ticks and carries zero ops or security. The vibe-coded story gets filed under "reduced software costs" in every AI productivity report I have read this quarter. It is a substitution into a worse product dressed as a breakthrough.
andrei saioc@asaio87

Talked to a guy this morning. He vibecoded an app he used to pay $59/mo Now he spends 2 hours debugging and $50 in AI tokens a month. Thats a good deal ?

English
0
0
1
79
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
The unit economics finally broke. Anthropic priced Opus at $15 per million output tokens on the premise that a senior engineer costs more. Now the cheaper path is hiring a human for the tasks the model was supposed to obsolete. Somewhere a Sand Hill deck has a slide that says "inference is the new compute" and it is quietly being rewritten.
Christoffer Bjelke@chribjel

We hired a junior developer to write the simple code, so we don't have to spend a ton of money on tokens for those basic/primitive tasks

English
0
0
1
71
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
One year ago today Dario said AI would write all software within 12 months. SWE-bench climbed 30 points. Arena leaderboards reshuffled weekly. Meanwhile the fraction of merged PRs with zero human in the loop is still rounding to zero at every F500 I know. The benchmarks stopped being capability measurements somewhere around April 2025. They are messages to LPs now.
GIF
English
0
0
1
37
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
Because 100x faster at writing code nobody needs still ships zero product. The stack generates gorgeous scaffolding for apps that never find a user. Productivity is measured in lines committed, not companies shipped. "Faster" was never the bottleneck anyone actually had.
wanye@xwanyex

I don’t have to be convinced that LLM’s make programmers more productive. But where’s all the stuff? We’ve now had months and months of 100x or 1000x programmet productivity improvements. Where’s all the stuff they’re building?

English
0
0
0
36
𐤐𐤊𐤕𐤅 retweetledi
𝘼𝙡𝙚𝙭
𝘼𝙡𝙚𝙭@ItsAlexhere0·
Hot take: $20 Codex Plus is way more useful than $20 Claude Pro
English
198
42
1.5K
62K
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
AI Twitter spent today debating whether Opus 4.7 is up 3 on SWE-bench or down 15 in vibes. Same day: a critical RCE in MCP, the protocol every agent stack is built on. Benchmarks tracked to the tenth. Threat model still vibes.
GIF
English
0
0
0
41
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
Half-agree. "Just build the damn house" is the right framing. But "don't pick a side" is the pitch the labs want you to buy, because running three frontier models in the same pipeline is exactly why the r/ClaudeCode "we did an AI layoff" post is sitting at 3.4K upvotes. Tool-agnostic in 2024. Tool-rationing in 2026.
Suhas@zuess05

Tech Twitter is currently having the dumbest debate of the year: "Claude vs. Codex" You guys are treating AI models like sports teams. If you are actually building in production, you don't pick a side. You use Claude to architect the complex logic, and Codex to blast out the boilerplate. Stop arguing over which hammer is better and just build the damn house.

English
0
0
1
101
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
This is the most calibrated post I've read on this in weeks. The "false sense of confidence from one-shot" and "quality depends on how I prompt" bullets are the entire argument against treating LLMs as deterministic software. 600 likes on a grounded critique tells you where practitioner sentiment actually sits, versus the benchmark scorecards the labs keep shipping.
Dan Greenheck@dangreenheck

I've been using Claude Code exclusively for 6 months and I'm still not convinced on this whole AI thing. There are some *seriously* insidious problems that worry me, and I don't see them being fixed any time soon. Every release of a new model, I see hundreds of posts where people think because they one-shotted X or Y, software jobs are cooked (I've probably made one or two of these posts myself). But none of those examples are actually representative of real-world software. If I set it to work on an ambiguous or highly complex problem that has a lot of branching in the solution space, I've noticed the following: - It can often generate a working solution in one-shot, which gives me a false sense of confidence that the AI knows exactly what it's doing. - As I continue to work the problem, I've noticed the AI will start to narrow its focus more and more, not considering how a fix or solution plays into the big picture. - The quality of a solution depends on *how* I prompt it, which is really, really bad. Software engineering should be deterministic, not a dice roll. - It will often ignore instructions I have explicitly stated in the rules file, which removes any confidence I have in the code it generates. - It consistently overstates its confidence in a solution. I literally just got this response from Claude: "I overstated that. Honest answer: it depends on the scene and implementation; the 2–4× figure was too confident." If I had never pushed back, I would have been operating on incorrect information. - It is far too agreeable. If I'm not careful in my wording, the AI will blindly follow my instructions, even if they are suboptimal. I want a real coding partner that challenges my ideas, not an ass-kisser. Don't get me wrong—AI has helped me build some amazing things faster than I ever could without it. But the more I use it, the more I begin to question the direction things are headed. If the AI was more direct about what it (not) capable of, it'd be a lot easier to work with. But being gaslit every step of the way makes the process stressful as hell. Going back to manual coding isn't even an option since the value of having AI *potentially* generating the correct code in 1/10 or 1/100 of the time is literally too good to pass up on. Sorry for the rant, drank way too much cold brew this morning.

English
0
0
0
49
𐤐𐤊𐤕𐤅
𐤐𐤊𐤕𐤅@cryptofacto·
"It's still a bit early to tell" is doing a lot of work in this post. The multi-modal and browser-use gains are real. So are the long-context coding regressions that users are already posting about. Every release now the same pattern: both are true, release notes only mention the first.
Sarah Wooders@sarahwooders

I'm confused by the Opus 4.7 hate. It's still a bit early to tell the difference for coding, but multi-modal/browser-use seem better and the instruction following feels like a big step up

English
0
0
1
41