Siddhant Khare

9.3K posts

Siddhant Khare banner
Siddhant Khare

Siddhant Khare

@Siddhant_K_code

23 | Devtools ❤️ | Engineer @ona_hq | Maintainer @OpenFGA, github1s 💚 | Building agent infra & tools | Author of Agentic Engineering Guide | opinions are mine

India Katılım Temmuz 2018
1.6K Takip Edilen5.6K Takipçiler
Sabitlenmiş Tweet
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
I wrote a book for engineering leaders. The Agentic Engineering Guide. 10 parts, 33 chapters. Context engineering, authorization, cost control, sandboxing, evaluation, governance, adoption playbooks. Not a prompt tutorial. The decisions you face when shipping agents to production this quarter. Early version. Open source. Rough edges included. Spot a mistake, open a PR. agents.siddhantkhare.com
English
4
37
304
23.4K
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
@svpino you will see a new fork on Hackernews in next ~6 hours
English
0
0
3
584
Mihai Maruseac
Mihai Maruseac@mihaimaruseac·
Is AI killing open source? No. Lazy people that submit slop and don't care about the project, only about their ego, do.
English
3
0
12
615
eric
eric@ericlim·
excited to share that I’m joining @openai as part of the GTM team. I’ll be building with openai’s customers to bring the latest research to all of us: users. AGI is coming.
English
2
0
6
515
Malte Ubl
Malte Ubl@cramforce·
These challenges are very real. Like, very real! But the answer isn't to wait for the tools to get better, because the problem isn't the tool, it is how we work. This is not everything, but one aspect is that you gotta be absolutely ruthless about throwing away stuff. Sunk cost fallacy was always bad. But now it got 10x worse with brains that trigger "wow, that demo is so far along, we might as well ship it" where they should trigger "This was 30 minutes of work. That investment should play absolutely zero role in the decision whether to proceed"
David Cramer@zeeg

im fully convinced that LLMs are not an actual net productivity boost (today) they remove the barrier to get started, but they create increasingly complex software which does not appear to be maintainable so far, in my situations, they appear to slow down long term velocity

English
11
6
141
20.6K
David Cramer
David Cramer@zeeg·
im fully convinced that LLMs are not an actual net productivity boost (today) they remove the barrier to get started, but they create increasingly complex software which does not appear to be maintainable so far, in my situations, they appear to slow down long term velocity
English
467
227
3.5K
657.6K
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
Siddhant Khare@Siddhant_K_code

Heads up for teams using @vercel + GitHub 👇 There’s a subtle auth edge case where if your GitHub account gets connected to a different Vercel org, deployments in the original org can start failing due to broken OAuth linkage. Symptoms: - GitHub repo access randomly failing - Deploy hooks not triggering - Permission errors despite correct org membership Root cause appears to be stale/overridden OAuth association between GitHub <-> Vercel org. Fix: Re-auth your GitHub account in Vercel auth settings Would love to see clearer multi-org OAuth isolation here, this can become a real auth mess in larger teams

English
0
0
0
220
Soham Datta
Soham Datta@tech_savvy_guy_·
hey @rauchg - there seems to be a flaw in the deployment protection in @vercel i am part of an organization in both GitHub and Vercel with my organization email added to both platforms. Now, when I author commits to my repo with my organization email, it blocks the deployment
Soham Datta tweet media
English
2
0
12
11.4K
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
@icedcoffeecoder it should work now, can you try it again with hard refresh (w/ clear cache) or in incognito window
English
0
0
0
46
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
This might be interesting for teams running agents on the same codebase every day. Most agents start every session from scratch. No memory of yesterday. No knowledge of what other agents learned. No awareness of past mistakes. A team of ten engineers running five agent sessions a day generates fifty sessions of institutional knowledge daily. And throws all of it away. Three types of memory that change this: 1. Session memory. The conversation history within a single run. Simple, but it grows with every step. By turn 20, you're sending 200K tokens per turn. The cost grows quadratically, not linearly. 2. Persistent memory. Survives across sessions. When an agent finishes, it saves what it did, what it learned, what went wrong. Next time, it loads those summaries instead of rediscovering everything. The simplest version is an AGENTS.md file. The sophisticated version uses a vector database. 3. Shared memory. One agent's knowledge available to others. The code review agent discovers a tricky initialization sequence. The code generation agent working on the same module should know about it. Without shared memory, every agent is a new hire on their first day. The most valuable form: learning from mistakes. When a human corrects agent output, that correction is signal. Store it. Retrieve it next time. An agent that repeats the same mistake twice is a tool problem. An agent that repeats it once is a memory problem. I wrote about this in Chapter 19 of the Agentic Engineering Guide
Siddhant Khare tweet media
English
11
17
119
6.9K
Payal 🌻
Payal 🌻@icedcoffeecoder·
@Siddhant_K_code Hey Siddhant! I wanted to read this but seems like the link is broken for me 😐
English
1
0
0
14
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
Siddhant Khare@Siddhant_K_code

Agent tooling is moving fast. The part that has not caught up yet is how we think about permissions across teams. When one engineer uses an agent, permissions are simple. You know what the agent can reach. You trust your own judgment. When three teams use agents across a shared codebase, the question changes. It is no longer "can this agent access this file." It is "which team's agent should have access to which service, under what conditions, and who decides." That is an organizational question, not a technical one. The way I think about it: teams own services. Engineers belong to teams. Agents act on behalf of engineers. These are relationships. And relationships are how permissions should work. "Agent belongs to team:payments. team:payments owns service:checkout. Therefore agent can write to service:checkout." A new engineer's agent inherits the boundaries of their team. A cross-team refactoring agent gets temporary, scoped access that expires when the task ends. An overnight agent gets the narrowest access the task requires. An orchestrator can share read access with sub-agents, but not write. The permissions mirror the org. When someone changes teams, the agent's access changes with them. When a service changes ownership, the permissions follow. No policy files to update manually. No role explosion. This is not a new idea. Google's Zanzibar has modeled permissions as relationships at massive scale for years. The insight is that the same pattern applies naturally to agents, because agents operate within the same org structures humans do. I wrote a chapter on this in the Agentic Engineering Guide. It covers the model, the code, and how it plays out across teams.

QME
0
0
1
148
Larsen Cundric
Larsen Cundric@larsencc·
Who is solving Agent Auth guys?
English
21
0
23
6.9K
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
strace for AI agents 👀
Siddhant Khare tweet media
English
3
2
15
1.1K
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
I wrote the full thing. A week ago, I discussed the gaps in agent observability in a thread. Session-level performance, context changes, signal-to-noise, production traces. A lot of you had the same frustrations. So I wrote it up properly. We have better observability for a Node.js service than for an AI agent that just rewrote half a codebase.
Siddhant Khare tweet media
English
23
40
413
18.8K
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
@Kernoga_Siarhei Yeah both are important at their places - x.com/Siddhant_K_cod…
Siddhant Khare@Siddhant_K_code

Agent tooling is moving fast. The part that has not caught up yet is how we think about permissions across teams. When one engineer uses an agent, permissions are simple. You know what the agent can reach. You trust your own judgment. When three teams use agents across a shared codebase, the question changes. It is no longer "can this agent access this file." It is "which team's agent should have access to which service, under what conditions, and who decides." That is an organizational question, not a technical one. The way I think about it: teams own services. Engineers belong to teams. Agents act on behalf of engineers. These are relationships. And relationships are how permissions should work. "Agent belongs to team:payments. team:payments owns service:checkout. Therefore agent can write to service:checkout." A new engineer's agent inherits the boundaries of their team. A cross-team refactoring agent gets temporary, scoped access that expires when the task ends. An overnight agent gets the narrowest access the task requires. An orchestrator can share read access with sub-agents, but not write. The permissions mirror the org. When someone changes teams, the agent's access changes with them. When a service changes ownership, the permissions follow. No policy files to update manually. No role explosion. This is not a new idea. Google's Zanzibar has modeled permissions as relationships at massive scale for years. The insight is that the same pattern applies naturally to agents, because agents operate within the same org structures humans do. I wrote a chapter on this in the Agentic Engineering Guide. It covers the model, the code, and how it plays out across teams.

English
0
0
0
90
Siarhei Kernoga
Siarhei Kernoga@Kernoga_Siarhei·
@Siddhant_K_code Agent observability is useful. But observability explains actions after they happen. Governance decides whether they should happen at all.
English
1
0
0
109
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
@codevlogger @0xkanth Agreed. That's the gap. The infra doesn't support it yet because nobody has standardized what an agent trace even looks like. We have OpenTelemetry for services. We have nothing equivalent for agents. Someone needs to build that layer.
English
0
0
0
26
Siddhant Khare
Siddhant Khare@Siddhant_K_code·
yes!! The reasoning trace is the closest thing we have to "why". But it shouldn't depend on the model vendor. The observability layer should capture the decision context (what was in the window, what tools were available, what the agent considered) regardless of whether the model exposes chain-of-thought. That's an infra thing, not a model feature.
English
1
0
0
27
Guna Sekhar Venkata Chennaiah Chakka
@Siddhant_K_code @0xkanth At sometimes i feel in the every tool call in needs to comes with some message that why it's making that tool call. Which is helpful for better debugging and understanding of the system. Unfortunately this is possible only with gpt reasoning models where as claude supports this🌋
English
1
0
0
25