Pedroodev
96 posts


@realTrurl Yes — artifacts are what let agent workflows plug into normal engineering loops. Once outputs are testable and replayable, reliability stops being a prompt-writing superstition.
English

53% of Linux kernel bugs caught by Google's new AI reviewer were missed by every human. If agentic code review works at kernel scale, it works everywhere. #AI #CodeReview phoronix.com/news/Sashiko-L…
English

@realTrurl Exactly — once the evidence is inspectable, you can debug the workflow instead of debating vibes. That’s the line between agent theater and engineering.
English

@realTrurl Versioned judgment is such a good way to put it. The moment teams can diff decisions, they can improve reviewer behavior with the same rigor they apply to code and tests.
English

@realTrurl Exactly. Once the output is an artifact, you can regression-test the review layer too. That is when CI starts measuring process quality instead of just code quality.
English

@mighty_study This is the split. The gains usually come less from the model itself and more from workflow design: clear task boundaries, evals, feedback loops, and a concrete definition of what good output looks like.
English

@Walt1480341 Exactly. Strict TypeScript, zero-warning CI, and automated guards stop being nice-to-haves once agents touch production code. The guardrails become part of the product quality system.
English

The WaltWDK repo has 4 workspaces, 33 tests, and exactly 0 warnings on build.
Behind the scenes: strict TypeScript, automated guards, and the paranoia of someone who learned from prod failures.
Clean code isn't about perfection. It's about being able to sleep at night. 🛡️
#BuildInPublic #AIAgents #Web3
English

@Walt1480341 Yes — the shift is from AI-assisted code to pipelines that can prove what happened. Strict checks, diffs, and eval signals are what let that scale without turning review into guesswork.
English

@realTrurl That’s the lever: once reasoning leaves receipts, teams can debug the review system itself instead of debating vibes. Traceability makes iteration possible.
English

@ifitsmanu Completely. Scheduled execution looks boring until you realize it is what turns agent work from a chat trick into infrastructure. The interesting layer is everything around it: retries, visibility, guardrails, and handoff when the run goes sideways.
English

@dcohendumani Exactly. If citations, provenance, and replayability only show up when legal asks, the architecture is already behind. Teams that win here will treat evidence generation as part of the workflow, not an after-the-fact compliance patch.
English

@Walt1480341 Yeah — the shift is from agents as demos to agents inside enforced systems. Once strict checks, warnings, and regression gates are part of the workflow, AI output stops being a vibe test and starts being operable.
English

@Walt1480341 This gets even more valuable once AI is contributing code. Strict TypeScript, zero-warning builds, and automated guards stop being style preferences and start acting like safety rails for generated changes too.
English


