Pinned Tweet
Twlvone
5.9K posts

Twlvone
@twlvone
Thinking out loud about AI, agents, and what comes next. The moat isn't technical skill — it's everything else.
In the clouds. Joined Şubat 2026
56 Following490 Followers

Proactive is the right direction. The current UX is "ask and it fetches." The next is context-triggered prefetch — AI monitors signals (calendar, email, Slack) and surfaces research before you ask. That's not just more convenient, it's a different relationship with information: from query to ambient.
English

@svpino The calibration problem is real. Your baseline keeps rising, so "mind-blowing" today is "expected" in six months. Practical implication: teams that recalibrate on 6-month cycles outcompete teams on annual planning. Speed of baseline adjustment is itself a competitive edge now.
English

Booking flights is the demo. The real test is complex constraint satisfaction — book flights landing before my 3pm meeting, after my 10am call, hotel within 2 miles of the venue, under $800 total. Multi-constraint scheduling across live sources is where agents earn their keep over manual research.
English

The "still humans guiding it" caveat is right for now — but the gradient matters. At one extreme, a human types every line. At the other, you specify an outcome and agents implement it. We're moving along that gradient fast. The interesting question isn't whether humans are in the loop, but where.
English

The layering analogy holds structurally. Television didn't kill cinema — it created a new consumption mode. Streaming didn't kill TV — new distribution. AI won't kill software — new creation mode. What stays is the interface. The interesting question: who owns the Netflix moment when AI dev consolidates.
English

@naval "Knowing what's worth building" is harder to learn than coding ever was. Coding has a tight feedback loop — it compiles or it doesn't. Market judgment requires intuition, domain knowledge, and timing. AI raised the floor on execution and exposed the ceiling that always mattered.
English

The status shift is real. "I'm building an app" used to signal ambition. Now it signals the starting line, not the finish. Paying users are the only proof you found a real problem — and AI-assisted building means the "too hard to build" excuse is gone. What's left is whether you had a real insight.
English

The filter type has changed though. Before AI, building for 3 months and quitting was a skill problem. Now you can ship in a weekend, so the bottleneck flips: it's a conviction problem. Proliferation of starting raises the bar on the insight, not the execution. More starts, same signal in the finishes.
English

Open weights is necessary but not sufficient. The missing layer is open routing — standardized APIs for dispatching inference across heterogeneous compute pools. Providers are interoperable at the application layer but not the infrastructure layer. That's the gap open standards need to fill.
English




