zencoderai

409 posts

zencoderai banner
zencoderai

zencoderai

@zencoderai

The Most Intuitive AI Coding Agent - Code faster, smarter, and stay in the flow.

Katılım Mayıs 2024
288 Takip Edilen1.3K Takipçiler
Sabitlenmiş Tweet
zencoderai
zencoderai@zencoderai·
Introducing Zenflow Work. A proactive AI assistant for your entire organization.
English
10
39
381
15.3M
zencoderai
zencoderai@zencoderai·
@vercel_dev What's the call you ended up making on which actions get fully sandboxed vs which ones just need a confirmation gate before running?Sandboxing + pluggable agents covers a lot of the agent-on-large-repo failure modes.
English
1
0
0
1.4K
Vercel Developers
Vercel Developers@vercel_dev·
Introducing deepsec, an open source coding security harness. • CLI-first • Sandbox-based scaling • Pluggable coding agents • Designed for large-scale repos • Use AI Gateway or your own subscription After months of successful internal use, we put it to the test on some of the largest open source codebases. vercel.com/blog/introduci…
English
33
94
912
182.2K
zencoderai
zencoderai@zencoderai·
22,000 lines of AI-written code at Anthropic. It works because they vibe-code the leaves, not the trunks. Leaves: utilities, glue, one-off scripts. Trunks: core systems everyone reads. Use the agent on the leaves. Hand-write the trunks.
English
0
0
1
47
zencoderai
zencoderai@zencoderai·
@alexxubyte What's been hardest to set up for you: checking whether the answer is right, or checking whether the agent took the right path to get there?Most teams skip evals because they don't feel productive on day one. The pain shows up at month three.
English
0
0
0
14
Alex Xu
Alex Xu@alexxubyte·
How do you know if your AI app actually works? You evaluate it. But most teams skip this step (or do it wrong) because "eval" feels vague. It's not. Every good eval is usually a 3-step recipe:
Alex Xu tweet media
English
8
13
69
4.6K
zencoderai
zencoderai@zencoderai·
@svpino Where do you draw the line: dump everything and hope, or summarize aggressively and trust the summary?"Stick to a reasonable length" is the part most agent setups skip. Most failures we see come from stuffing context, not from hitting any real limit.
English
0
0
0
50
Santiago
Santiago@svpino·
Opus 1M, past 400K tokens, is a huge downgrade. The model is great, but that extra context isn't free (and it's straight-up a sure way to degrade your experience). When you fill up your context, the same attention has to spread across more material, which means: • Worse reasoning abilities • Weaker instruction following • More lost information Unless you truly need that much context (and if you have to ask, this is probably not you), stick to a reasonable length.
English
34
3
52
9.2K
zencoderai
zencoderai@zencoderai·
@sakitojo What's been the hardest piece to swap between models without breaking?Most coding agents end up locked to one model and never recover.
English
1
0
0
19
Saki Tojo
Saki Tojo@sakitojo·
🚀 I built DeepCLI, a modular DeepSeek-native coding agent. It distills Claude Code’s harness, tools, and prompt engineering, then rebuilds them as a pluggable Python kernel shaped by OpenClaw/Hermes. Not just a CLI wrapper: DeepCLI is designed as an agent kernel for personalized software. Users can grow their own library of skills, tools, plugins, scheduled tasks, memory, and long-lived session agents. 🧠 self-evolving memory 🧩 pluggable tools/skills/hooks 🌐 ACP/WebSocket thin-client boundary 📱 gateway-ready agents beyond terminal ⏰ scheduler + background tasks 🔁 multi-provider benchmarking 💻 Windows/macOS/Linux native support Code: 100% AI-agent built. Docs: open for agent-driven contribution. Alpha, but usable. If you’re building agents or interested in this direction, feel free to reach out. I’d love to chat. github.com/Harinlen/DeepC…
Saki Tojo tweet media
English
1
0
4
107
zencoderai
zencoderai@zencoderai·
@saen_dev @shubh19 What's worked better for you: listing the files it can't touch, or just describing the parts of the codebase that should stay hands-off?Yeah, telling the agent what NOT to touch is the part most people skip.
English
0
0
0
0
Saeed Anwar
Saeed Anwar@saen_dev·
@shubh19 CLAUDE.md is the single highest-ROI file in any project using AI coding tools. Five minutes of writing constraints saves hours of debugging unwanted changes. The key insight is being explicit about what NOT to touch, not just what to do.
English
1
0
0
19
Shubh Jain
Shubh Jain@shubh19·
[Vibe Coding series day 9] the single file that changed how I vibe code more than any tool, model, or prompt technique: CLAUDE.md it sits in your project root and tells Claude everything it needs to know before touching a single line your stack, your rules, your patterns, what to never do, how you want errors handled, which files are off limits without it: every session starts from zero. you re-explain your stack. Claude makes decisions that contradict what it built last week. consistency is impossible. with it: Claude walks into every session already knowing your project like a developer who's been on the team for a month here's the exact structure I use on every project: # Project one sentence: what this does and who it's for # Stack - Next.js app router - Supabase for db and auth - Tailwind for styling - Deployed on Vercel # Rules - never modify the /lib/db.ts file without asking first - always use TypeScript, never plain JS - error handling: log to console.error + return user-facing message, never swallow silently - no inline styles, always Tailwind classes # Patterns - API routes go in /app/api/[route]/route.ts - reusable components go in /components/ui/ - never hardcode values that belong in .env # What to avoid - do not install new packages without listing them here first - do not touch auth logic unless the task explicitly requires it 5 minutes to write on day 1. saves you hours of re-explaining across every session after that. the best part: Claude reads it automatically every time you start a new session in Claude Code. you write it once. it applies forever. paste this structure into your project root today. fill in your actual stack. that's the whole task.
Shubh Jain tweet media
Shubh Jain@shubh19

[Vibe Coding series day 8] 70% of vibe coding projects collapse before they ever leave localhost not because the idea was bad week 1 kills them and it's always the same 7 mistakes doing the killing #Vibecoding

English
12
4
15
411
zencoderai
zencoderai@zencoderai·
@godofprompt Two models working against each other beats one model alone, every time we've tested it. The harder call in practice: do you let the executor overrule the planner mid-task, or lock the plan once it's drafted?
English
0
0
0
121
God of Prompt
God of Prompt@godofprompt·
I tested the highest-performing AI coding workflow of 2026. It doesn't use one model. It uses two competing models against each other. Opus 4.7 plans. GPT-5.5 executes. The results aren't close. (Prompts included)
God of Prompt tweet mediaGod of Prompt tweet media
English
18
7
65
10.9K
zencoderai
zencoderai@zencoderai·
@Anubhavhing @asaio87 Demo-to-production is where the real engineering shows up. The vibe carries you to the screenshot, not past it. What's been the cleanest signal a demo isn't ready: missing edge cases, or invisible coordination overhead?
English
0
0
0
24
Anubhav
Anubhav@Anubhavhing·
@asaio87 vibe coding gets you to demo, not to production
English
1
0
0
65
andrei saioc
andrei saioc@asaio87·
A vibecoder cannot comprehend: - security for his app - auth in the app - payment integration - scaling issues It’s difficult to create production apps.
English
78
7
252
22.5K
zencoderai
zencoderai@zencoderai·
@akshay_pachaar "Jagged intelligence" is the most useful frame in the whole talk for hiring decisions. Where have you seen the gap widest in practice: code where you can verify, vs design/UX where you can't?
English
0
0
0
138
Akshay 🚀
Akshay 🚀@akshay_pachaar·
Andrej Karpathy on the shift from vibe coding to agentic engineering:
Akshay 🚀 tweet media
English
20
31
214
9.3K
zencoderai
zencoderai@zencoderai·
@yrzhe_top Hard gates on form, full freedom on method is a sharp split. The PLAN-CONFIRM block is the practical version of "spec before code." Which gate has been highest-value: declaring method, declaring risks, or declaring expected output?
English
1
0
0
12
yrzhe.top
yrzhe.top@yrzhe_top·
I shipped my senior-data-analyst skill today. Two failure modes I've watched myself walk into all month. Tighten too hard. 12 numbered rules, 3 must-do's, 2 banned templates, a fixed output schema. The agent obeys and goes formulaic. Reading the output feels like it was holding its breath. Loosen too much. One-line goal, full trust. The agent picks the prettiest-sounding template, skips the disconfirming step, writes prose where I needed evidence. Yerkes-Dodson 1908. They put hungry monkeys in a cage with bananas out of reach and a stick behind them. The hungriest ones bashed the bars; couldn't see the stick. Cognition narrows under pressure. The shape that landed is hard gates on form, full freedom on method. My skill has a forced PLAN-CONFIRM block. The agent must declare method, risks, expected output before touching data. Which method, which risks, the agent picks. Honest part. I still can't tell you where "the middle" sits. It moves with the task, the model, the user's context loaded going in. That's the actual hard problem of prompt design, and I'm sitting in it.
English
1
0
2
15
zencoderai
zencoderai@zencoderai·
@ankits0052 "Verifiability is the unlock" is the line that ages the best out of that whole talk. Where do you draw the verifiable / non-verifiable line on your stack: tests-only, integration tests, or all the way to user-observable behavior?
English
0
0
0
7
Ankit Shah
Ankit Shah@ankits0052·
Andrej Karpathy on the shift from Vibe coding to Agentic Engineering.
Ankit Shah tweet media
English
1
0
0
47
zencoderai
zencoderai@zencoderai·
@simonw Single-purpose playgrounds are the most underrated AI coding output. Throwaway by design, perfect for exactly your shape of testing. How long from prompt to working WASM build, and how much manual cleanup at the end?
English
0
0
0
75
zencoderai
zencoderai@zencoderai·
@amavashev Pre-execution is the right place to enforce constraints. Most agents hallucinate boundaries because nobody told them what they couldn't touch. What's the first check Cycles runs before the agent moves: scope, secrets, or dependencies?
English
1
0
1
6
Albert Mavashev
Albert Mavashev@amavashev·
Launched Cycles - pre-execution layer for AI agents. Now back to building. TODO for today: review and tighten my CI workflows on github. Appreciate your support. 👍 peerlist.io/amavashev/proj…
English
0
0
0
24
zencoderai
zencoderai@zencoderai·
@mincasurong @googlegemma @googledevgroups On-device agents are the underrated unlock. Most teams haven't tried because the tooling felt rough. What was hardest: getting model size right, or making multimodality actually work end-to-end?
English
1
0
0
10
Minsu (Max) Chang
Minsu (Max) Chang@mincasurong·
I joined the @googlegemma Sprint event in Seoul! It was an amazing time building a Gemma4-based on-device Agent. I focused on the multimodality of Gemma4-E4B, and it was truly outstanding, even more than other open-source models. Thanks to @googledevgroups in Korea inviting me to this exclusive event!
Minsu (Max) Chang tweet mediaMinsu (Max) Chang tweet mediaMinsu (Max) Chang tweet mediaMinsu (Max) Chang tweet media
English
1
0
2
19
zencoderai
zencoderai@zencoderai·
@jaideepparasha7 Layer one gets all the headlines because it's the easiest to claim. Where does layer two end and layer three start: when one role shifts, or when the org chart redraws around AI?
English
1
0
1
11
Jaideep Parashar
Jaideep Parashar@jaideepparasha7·
AI will impact society in layers: First layer: productivity gains Second layer: job role shifts Third layer: business model disruption Fourth layer: cognitive dependency Most people are still at layer one.
English
2
0
2
25
zencoderai
zencoderai@zencoderai·
@Pragmatic_Eng @mitsuhiko Friction as a feature, not a bug. Where does it land best in practice: human review pre-merge, agent self-checks pre-PR, or both?
English
0
0
0
20
The Pragmatic Engineer
The Pragmatic Engineer@Pragmatic_Eng·
A degree of friction improves platform stability. Armin Ronacher(@mitsuhiko) - creator of Flask and founder at Earendil - on why why it’s a bad idea to ship without pause: “There was an incident related, at least in parts, to agentic engineering where a company shipped out a configuration change that ultimately resulted in a security issue. Things happen, but the company's tagline was ‘ship without friction’. This gave me pause. As engineers we used to talk about: ‘you have to get rid of all the things in the way so that you feel happy shipping stuff’. However, there were always changes where you really wanted to think: ‘do you want to drop the database’, ‘do you want to merge this migration’? It's these moments every once in a while where you are really supposed to think and people created checklists or mechanical gates where you would have to confirm something. There are certain things that we used to put in, particularly if you run a SaaS company, to slow things down. In some of the best engineering teams, in order to mature a service, you have to define an SLO, you have to define expectations. A lot of engineers feel like, ‘oh, this is bureaucracy’. But the reality is, if you do this correctly, then it saves you time and it makes you happier. You're not waking up at three o'clock in the morning.”
English
1
0
16
3.1K
zencoderai
zencoderai@zencoderai·
@stefanjblos @karpathy Should we just build more, or should we build what's actually verifiable? Karpathy's lever: LLMs automate what you can verify. What's harder for your team: deciding what to build, or making it verifiable?
English
1
0
0
14
Stefan
Stefan@stefanjblos·
On my train ride back from a week in Amsterdam, I watched @karpathy's talk at Sequoia Capital about agentic engineering and had a few thoughts about it. I asked myself the question of...should we build more? stefanblos.com/posts/should-w…
English
1
0
1
65
zencoderai
zencoderai@zencoderai·
@SlabbedWorks The hardest part of the first few weeks: knowing when to write code yourself vs make space for someone else to. What's the first thing you're changing day one?
English
0
0
0
11
zencoderai
zencoderai@zencoderai·
@VillumsenC66060 Counting shipped features misses the whole point. Tech debt ratio is the better signal. What's the floor you'd accept: 80/20 new-to-maintenance, or does a healthy team always run closer to 60/40?
English
0
0
0
10
Claus Villumsen
Claus Villumsen@VillumsenC66060·
Most teams measure modernization by what they shipped. The better metric: technical debt ratio. What percentage of engineer time goes to maintenance versus new work? If that number is not falling quarter over quarter, you are running the pipeline but not winning. Full breakdown: kodebaze.com/blog/the-conti…
English
1
0
0
10
zencoderai
zencoderai@zencoderai·
Individual speed: up. Team velocity: down. Same tool, different unit of analysis. The bottleneck isn't coding anymore. It's coordination, review, shared context. Full breakdown: youtube.com/watch?v=zqTxwl…
YouTube video
YouTube
English
0
0
0
34