segmenta

859 posts

segmenta

segmenta

@segmenta

Rowing Rowboat (YC S24). Prev: Co-founder/ CTO Agara AI (acq. Coinbase), Coinbase AI, Twitter AI

เข้าร่วม Nisan 2009
1.4K กำลังติดตาม792 ผู้ติดตาม
segmenta
segmenta@segmenta·
This captures a lot of the problem space we’re building Rowboat for: a personal living knowledge base for your work. We think the “incredible new product” here is very real.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
7
0
8
2.3K
Steven Grin
Steven Grin@sgrin77·
@segmenta @CAMP_Realty @ycombinator Multiple accounts is key. This is one of biggest impediments for me on Claude mcp integration. Need mult oauth account access to build a true knowledge graph.
English
1
0
1
151
segmenta
segmenta@segmenta·
Introducing Rowboat. An AI coworker that compiles your emails, meetings, and work into a living knowledge graph, then uses it to actually get things done. Open source. Local-first. Voice-powered. Karpathy described the idea last week. We've been building it for a while.
English
21
30
371
50.7K
segmenta
segmenta@segmenta·
@reachmeviz Hermes is great but it mainly focuses on the general purpose agent itself. Rowboat is a desktop app for knowledge work assistance. It’s built around cooperative work - hence the user readable knowledge graph, voice out, meeting notes etc.
English
0
0
0
257
Viz
Viz@reachmeviz·
@segmenta How is this different from things like Hermes agent which also constantly evolves its memory ?
English
1
0
1
295
segmenta
segmenta@segmenta·
@CAMP_Realty @ycombinator You can connect only one Gmail account right now. Outlook support is coming soon. Would love to learn what your use case is for multiple accounts.
English
2
0
1
317
segmenta
segmenta@segmenta·
We built ours. Memory is markdown files on specific topics: things about the user and their preferences on tasks like email drafting, presentations, etc. The assistant can save things to memory when needed and there is also a background agent that looks at chats and creates memory notes if the assistant missed anything. Happy to explore on DM.
English
0
0
0
419
Saïd Aitmbarek
Saïd Aitmbarek@SaidAitmbarek·
@segmenta That's so cool, do you rely on external memory providers? Or did you build yours? Asking, but i'll explore the codebase (thx for oss) Let's launch you on microlaunch.net anytime btw mate.
English
1
0
2
494
segmenta
segmenta@segmenta·
@bittybitbit86 You are going to be pleasantly surprised :) Our default spam detection is better than Gmail's. And you can configure it in the app to how strict you want it to be.
English
0
0
2
219
LiτBro
LiτBro@bittybitbit86·
@segmenta This shit gonna fill my knowledge graph with 24k spam emails isn’t it?
English
1
0
0
246
segmenta
segmenta@segmenta·
@levie Couldn't agree more! This is exactly why we built Rowboat: an AI coworker that compiles your work into a knowledge graph and uses it to actually do things - with the same tools you use. Just launched: x.com/segmenta/statu…
segmenta@segmenta

Introducing Rowboat. An AI coworker that compiles your emails, meetings, and work into a living knowledge graph, then uses it to actually get things done. Open source. Local-first. Voice-powered. Karpathy described the idea last week. We've been building it for a while.

English
0
0
0
57
Aaron Levie
Aaron Levie@levie·
The same kinds of productivity gains we've seen in coding with AI agents are heading to the rest of knowledge work. This is the jump when you go from having a chatbot to being able to actually have an agent go off and do work for minutes or even hours and come back with a complete work output that you then review. Here's an example of the new Box Agent filling out an RFP response from an existing knowledge base. This process would normally take hours to fill out, and requires the full attention of the user doing the work. Now, you provide the Box Agent with the RFP questions, and it will go off, make a plan, extract all the relevant questions, read through existing source material to come up with an answer, and then generate a new word document as the final output. All while you're doing something else. The key to this architecture is that the agent is able to use all of the same tools in the background that a user uses to get work done. The agent can search for documents, read entire files, run scripts and tools in the background, and even be able to write code on the fly to automate tasks it hasn't seen before. And best of all, the Box Agent will (soon) work from the Box MCP and CLI so you can invoke it in any agentic system as a step in a process. This kind of agent complexity would have been impossible even 6 months ago. Models consistently failed at tracking long running tasks or using the right tools at the right moment for the task. But this is all now possible because of models like GPT-5.4, Opus 4.6, and Gemini 3, and is only getting better by the month. Just as we moved from engineers writing code and using AI as an assistant to answer questions, in many areas of knowledge work -like legal, finance, consulting, sales, marketing, and more- when we have a problem we'll just kick off the AI agent to just go work on it for us in the background.
English
30
11
121
20.2K
segmenta
segmenta@segmenta·
Fair point. Local-first is a side benefit. The real bottleneck is context quality, like you said. That's why we built Rowboat around a living knowledge graph instead of RAG. Consciously building context over time is better than doing one off data pulls when needed - is our thesis.
English
1
0
0
719
Gregor
Gregor@bygregorr·
@segmenta The "local-first" framing is doing a lot of heavy lifting here. Isn't the bottleneck not where data lives but whether the AI actually understands context well enough to act without making things worse?
English
1
0
2
851
segmenta
segmenta@segmenta·
@yabsssai Thanks! We have integrations to Jira, Linear and Asana and few others. What does your team use - would love any feedback.
English
5
0
0
307
YabsssAI
YabsssAI@yabsssai·
your idea of a "living knowledge graph" really resonates, as someone who's worked on similar data integration projects, i've found that the key to successful implementation is in the data quality inputs – have you talked to any teams about integrating rowboat with existing task management systems?
English
1
0
1
337
Akshay 🚀
Akshay 🚀@akshay_pachaar·
@segmenta Ok, I was looking for something that is local first for my AI brain. Giving it a spin right away. And congrats on shipping.
English
1
1
1
1.1K
segmenta
segmenta@segmenta·
@JaykhatriDev Its processed mostly chronologically - the newer source typically supersede the older one. Having said that, this is done by background agents - there is some amount of LLM discretion involved.
English
0
0
1
718
Jay Khatri
Jay Khatri@JaykhatriDev·
@segmenta how does it handle conflicting information from different sources?
English
1
0
1
782
segmenta
segmenta@segmenta·
@saadnvd1 Its stored as plain markdown files with Obsidian-style backlinks.
English
0
0
1
782
Saad
Saad@saadnvd1·
@segmenta Local-first is the part I care about most. How are you handling the knowledge graph storage: SQLite + embeddings locally, or something custom?
English
1
0
2
929
segmenta
segmenta@segmenta·
Since our previous launch, we’ve added: → Voice, so you can talk to Rowboat and it talks back → A native meeting note taker → Live note to track anything with agents → Bases view → Granular web search → Works out-of-the-box, no API keys needed All on top of the local-first, Markdown-based knowledge graph.
English
0
0
5
2.6K
segmenta รีทวีตแล้ว
segmenta รีทวีตแล้ว
james hawkins
james hawkins@james406·
what is your 5 prompt plan? where do you see yourself in 5 prompts?
English
50
42
522
27.3K