Bhanu

1.3K posts

Bhanu banner
Bhanu

Bhanu

@bhanu_in_tech

AI/ML SSI Agentic AI

Berlin, Germany Katılım Temmuz 2009
412 Takip Edilen62 Takipçiler
Bhanu
Bhanu@bhanu_in_tech·
@system_monarch Design a new memory architecture optimise for enterprise data current memory providers are optimised for chat
English
0
0
0
7
Puneet Patwari
Puneet Patwari@system_monarch·
SD Round at OpenAI for L4 Role (Comp: $569K+| INR 5.3Cr+) : You are designing LLM powered search for a 10,000 person enterprise. The LLM has a context window of 128k tokens. A single query might need to pull from 300 relevant documents totalling 2 million tokens. You cannot fit everything in context. You cannot afford to miss the critical document. How do you decide what goes into the context window and what gets left out without the LLM giving a confidently wrong answer?
English
8
2
77
15.1K
Bhanu
Bhanu@bhanu_in_tech·
Kubernetes
Filipino
0
0
0
5
Bhanu
Bhanu@bhanu_in_tech·
A path to production ready AI agents @bhanu.pratap/the-day-i-stopped-building-agents-and-started-engineering-them-c0d598f05fe6" target="_blank" rel="nofollow noopener">medium.com/@bhanu.pratap/…
English
0
0
0
14
Josh Woodward
Josh Woodward@joshwoodward·
New in Gemini: NEET practice tests! 🇮🇳🩺 Sending good luck to everyone taking India's biggest medical exam in a few weeks. You've got this! We’re working on adding more practice tests for different subjects and countries. What should we add next? Let us know below 👇
Google India@GoogleIndia

Full length, no cost NEET UG practice tests are now in @GeminiApp, isn’t that neat? 😄 Say “I want to take a NEET mock test” and begin ✍️ Read here: goo.gle/NEERPrep

English
63
127
1.4K
194.4K
Bhanu
Bhanu@bhanu_in_tech·
@kepano @karpathy LLMs can create clean, structured BS. Without strong grounding, validation, and deduping, this wiki becomes a confident echo chamber.
English
0
0
0
19
kepano
kepano@kepano·
I like @karpathy's Obsidian setup as a way to mitigate contamination risks. Keep your personal vault clean and create a messy vault for your agents. I prefer my personal Obsidian vault to be high signal:noise, and for all the content to have known origins. Keeping a separation between your personally-created artifacts and agent-created artifacts prevents contaminating your primary vault with ideas you can't source. If you let the two mix too much it will likely make Obsidian harder to use as a representation of *your* thoughts. Search, bases, quick switcher, backlinks, graph, etc, will no longer be scoped to your knowledge. Only once your agent-facing workflow produces useful artifacts would I bring those into the primary vault.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
82
167
3K
485.2K
Bhanu
Bhanu@bhanu_in_tech·
Cool idea, but risky: LLMs can create clean, structured BS. Without strong grounding, validation, and deduping, your wiki becomes a confident echo chamber.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
0
0
0
11
Bhanu
Bhanu@bhanu_in_tech·
@karpathy You are assuming “structure = correctness”
English
0
0
0
1
Andrej Karpathy
Andrej Karpathy@karpathy·
LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.
English
2.8K
7K
58.2K
20.8M
DROID
DROID@droidbuilds·
Guess the programming language
DROID tweet media
English
349
11
1.2K
291.7K
Bhanu
Bhanu@bhanu_in_tech·
Started using @FireworksAI_HQ pass but soon started getting 422 and now it is consuming my main balance instead of free flat rate
English
0
0
0
7
Bhanu
Bhanu@bhanu_in_tech·
@asmah2107 ACP has officially gave up to A2A they are no longer developing
English
0
0
0
7
Ashutosh Maheshwari
Ashutosh Maheshwari@asmah2107·
Three protocols are competing to become the TCP/IP of multi-agent systems. You should know how they differ. Bookmark this. > MCP Model Context Protocol (Anthropic) → Standardises how agents access tools and external resources. → Before MCP: every tool needed a custom integration. N agents × M tools = N×M interfaces. → After MCP: one standard. N+M interfaces total. → Already adopted by VS Code, JetBrains, and most major frameworks. → Think REST for agent-to-tool communication. > A2A Agent-to-Agent Protocol (Google) → Peer-to-peer collaboration between agents. → Agents negotiate, share findings, and coordinate without a central orchestrator. → Solves the "who is in charge" problem for decentralised systems. > ACP Agent Communication Protocol (IBM) → Governance-first. Security and compliance built into the handshake. → Designed for enterprise deployment where auditability is non-negotiable. The parallel: REST vs SOAP vs GraphQL. One will win. Two will become legacy. We're early. Follow for the system design layer of AI nobody else is writing about.
English
10
7
57
3.6K
dotta 📎
dotta 📎@dotta·
We just open-sourced Paperclip: the orchestration layer for zero-human companies It's everything you need to run an autonomous business: org charts, goal alignment, task ownership, budgets, agent templates Just run `npx paperclipai onboard` github.com/paperclipai/pa… More 👇
English
424
720
8.2K
2.5M
Farzad 🇺🇸 🇮🇷
Farzad 🇺🇸 🇮🇷@farzyness·
AI Engineers are going to be insanely valuable for the next 3-5 years. Just someone who is VERY good at deploying AI in any use case. Basically use Claude Code + Codex + OpenClaw + Grok + Gemini to solve any problem a business is having. Next-level valuable.
English
140
179
1.9K
80.9K
Vadim
Vadim@VadimStrizheus·
what’s the best agentic tool out there? 1. OpenClaw 2. Hermes Agent 3. PaperClip (Hermes Agent looks promising) 🤷‍♂️
English
147
5
282
42K
Josh Long
Josh Long@starbuxman·
hello what're you building today?
English
16
0
19
4.1K
Bhanu
Bhanu@bhanu_in_tech·
That's super-org — an execution bridge + platform layer: → Translates Paperclip jobs into Hermes-shaped work → Runs bounded executions with timeout, cancel, idempotency → Adds shared memory, policy gates, and evaluation services Paperclip directs. Hermes executes.
Bhanu@bhanu_in_tech

But neither solves the full picture alone. Paperclip doesn't have Hermes-level memory and skills. Hermes doesn't have Paperclip's multi-company orchestration and governance. Someone needs to connect them.

English
0
0
0
16
Bhanu
Bhanu@bhanu_in_tech·
But neither solves the full picture alone. Paperclip doesn't have Hermes-level memory and skills. Hermes doesn't have Paperclip's multi-company orchestration and governance. Someone needs to connect them.
Bhanu@bhanu_in_tech

Hermes (@NousResearch) is great at the employee plane: persistent memory, skill documents, session continuity, 30+ tools, sub-agents. It's an AI worker that actually learns — accumulating procedure and context instead of resetting every chat.

English
0
0
0
26
Bhanu
Bhanu@bhanu_in_tech·
Hermes (@NousResearch) is great at the employee plane: persistent memory, skill documents, session continuity, 30+ tools, sub-agents. It's an AI worker that actually learns — accumulating procedure and context instead of resetting every chat.
Bhanu@bhanu_in_tech

Paperclip (@paperaborin) is great at the control plane: companies, goals, roles, budgets, heartbeats, approvals, audit trails. Think of it as the operating system for an AI company. It tells agents what to do and when.

English
0
0
0
25
Bhanu
Bhanu@bhanu_in_tech·
Paperclip (@paperaborin) is great at the control plane: companies, goals, roles, budgets, heartbeats, approvals, audit trails. Think of it as the operating system for an AI company. It tells agents what to do and when.
Bhanu@bhanu_in_tech

Building super-org — an open-source bridge between Paperclip (AI company orchestration) and Hermes (persistent AI workers with memory + skills). One directs, the other executes. Together: a structured AI company. Apache-2.0. github.com/bhanu-singh/su…

English
1
0
0
19
Bhanu
Bhanu@bhanu_in_tech·
Building super-org — an open-source bridge between Paperclip (AI company orchestration) and Hermes (persistent AI workers with memory + skills). One directs, the other executes. Together: a structured AI company. Apache-2.0. github.com/bhanu-singh/su…
English
0
0
0
23