Matt

421 posts

Matt banner
Matt

Matt

@mattdotfun

building @ @worldnetwork . i ramble a bit. believer in the tek.

San Francisco, CA Tham gia Mart 2016
478 Đang theo dõi203 Người theo dõi
Matt
Matt@mattdotfun·
I’m extremely happy I took advantage of the annual max coding plan promo earlier this year
0xSero@0xSero

Droid - GLM-5.1 I asked it to install Ghidra Pointed at compiled typescript for a popular product "Reverse engineer this and provide me code that compiles to it then make it so I can use the app from my own dev build" 20 million tokens 1 shot later Full application, compiles down to the same thing, usable, can rip out any BS Making a bet, ZAI will have the first non-western #1 model in almost all benchmarks soon. Don't discount the value of training data, people who use ZAI are especially from the west have a higher technical cieling. Model is excellent. 432$ a quarter well spent.

English
0
0
0
35
Matt
Matt@mattdotfun·
@sebgoddijn @tryramp @AnthropicAI thank you for this amazing rundown of Glass. I’m tinkering with something similar to help at work too and this is great guidance
English
0
0
1
872
Seb Goddijn
Seb Goddijn@sebgoddijn·
We hit 99% AI adoption at @tryramp but it wasn't enough. Most people were still stuck in a chat window while power users ran laps around them, and teaching everyone how to use the terminal wasn’t going to scale. So we built Glass: our own AI productivity suite on @AnthropicAI's Claude Agent SDK with 700 daily active users in one month. We don't believe in lowering the ceiling. We believe in raising the floor.
Seb Goddijn@sebgoddijn

x.com/i/article/2042…

English
19
24
467
156.1K
Matt
Matt@mattdotfun·
@agrimsingh @gabrielchua @karpathy try @DuffelHQ instead of amadeus - amadeus is killing their self serve API i think. also love how this is taking you on a world tour - $ per hour on biz class is maxxed 🔥🔥🔥
English
1
0
1
1.9K
agrim singh
agrim singh@agrimsingh·
i took @karpathy's autoresearch loop and pointed it at business class flights because i wanna fly cheap but in class. (at least in theory - i'm a poor founder flying coach) here's autofare - autoresearch but for always flying in lie-flat beds made with codex + gpt 5.4 mini @gabrielchua @reach_vb @romainhuet @OpenAIDevs
English
59
39
919
173.6K
Matt đã retweet
Sarah Wooders
Sarah Wooders@sarahwooders·
The new Anthropic managed agents API is basically the Letta API that we've had since a year ago, but closed source and with provider lock-in. They even have read-only memory blocks and memory block sharing -- something which was unique to the Letta agents for a long time. Funny enough, we actually don't think this is the direction agents are going to go. Having API interfaces for memory blocks and tools is certainly convenient - you can spin up stateful agents as API services with just a few lines of code. But its also limiting: LLMs today are extremely adept at computer-use, and representing their memories in this way limits the action space of agents and their ability to learn. It's important to remember that just because something comes out of a frontier lab, doesn't mean its the "right" answer long-term. The Letta API ~1 year ago was somewhat of an antipattern in a sea of agent framework libraries offered by every lab. But now, stateful agent APIs are becoming the new norm - especially as providers try to lock in memory/state into their platforms to increase switching costs (which is exactly why we believe memory should live outside of model providers) If you want to see what the future is going to look like, follow @Letta_AI
Sarah Wooders tweet mediaSarah Wooders tweet media
Claude@claudeai

Introducing Claude Managed Agents: everything you need to build and deploy agents at scale. It pairs an agent harness tuned for performance with production infrastructure, so you can go from prototype to launch in days. Now in public beta on the Claude Platform.

English
46
50
592
116.3K
Factory
Factory@FactoryAI·
Today we're releasing the Factory desktop app. A native interface for autonomous AI agents that work across every part of your software business.
English
108
75
930
242.5K
Matt đã retweet
Arcee.ai
Arcee.ai@arcee_ai·
Today we're releasing Trinity-Large-Thinking. Available now on the Arcee API, with open weights on Hugging Face under Apache 2.0. We built it for developers and enterprises that want models they can inspect, post-train, host, distill, and own.
English
100
244
2.1K
682.8K
Matt đã retweet
0xSero
0xSero@0xSero·
This didn't receive the attention it deserved. They pre-trained this model completely peer 2 peer, no data-centers. Everything was done over a permissionless network, I have tried the model, it's honestly not a good LLM but that's beyond the point. We NEED this, we NEED an alternative. - Download OpenCode - Download Pi - Pay for OpenSource - Share your AI sessions - Learn to do RL We can't be at the mercy of ANY lab. arxiv.org/abs/2603.08163
0xSero tweet media
English
45
113
1.1K
48.7K
Matt
Matt@mattdotfun·
@clairevo lots of our community members have been trying hard to figure things out. would love to have a chat!
English
0
0
0
8
Matt
Matt@mattdotfun·
@clairevo my family runs a baby/kids clothing brand in singapore and i’ve started seeding our blog with OpenClaw content!
English
1
0
0
368
Matt
Matt@mattdotfun·
@clairevo am i adding an unnecessary step by still creating a ticket in linear so that i can track things in my team?
English
0
0
1
104
claire vo 🖤
claire vo 🖤@clairevo·
"PR >> PRD" Yep. the handoff era is over. but it's not just the roles collapsing. it's the tools. Every PM tool was built for a world where humans did the coordination. tickets docs roadmaps presentations all of that was scaffolding for work AI now does faster and cheaper. slapping AI on top doesn't fix it. The foundation is already out of date. I build @chatprd every day knowing i have to replace its core before something else does: claude code, another startup, something i haven't imagined yet. Radical humility and endless paranoia are the only product strategies that make sense right now. So sure. the PRD is dead. But I'll kill it before you do.
Bilgin Ibryam@bibryam

PR >> PRD. The handoff era is over. → When opening a PR is faster than writing a PRD, AI changes how product gets built. The old roles start to collapse.

English
21
6
165
36.9K
Matt
Matt@mattdotfun·
@clairevo curious - why double wield cursor and codex?
English
2
0
0
103
claire vo 🖤
claire vo 🖤@clairevo·
POV you text me these days and ask what my AI stack is
claire vo 🖤 tweet media
English
24
7
103
8.4K
Ivan Leo
Ivan Leo@ivanleomk·
I built a personal OS in @GoogleAIStudio with just a few prompts - build your own apps with their own icons and see them on an mac/ios view. Everything is backed up by a firebase database too with our full stack deployments :)
English
7
3
43
2.5K
Matt
Matt@mattdotfun·
@mil000 it is actually fantastic - you should try. the speed allows for really really rapid iterations
English
0
0
5
1.3K
Matt
Matt@mattdotfun·
Friday is for exploring. Today was the day I merged my first PR from @SlackHQ using @linear and @cursor_ai agents. Felt like a lightbulb moment - going to spend the weekend thinking about how this can become more common in our team.
English
1
0
2
70
Hila Shmuel
Hila Shmuel@HilaShmuel·
Meet Cabinet: Paper Clip + KB. for quite some time I've been thinking how LLMs are missing the knowledge base - where I can dump CSVs, PDFs, and most important - inline web app. running on Claude Code with agents with heartbeats and jobs runcabinet.com
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
72
35
954
250.7K
Matt
Matt@mattdotfun·
got asked today on which was my go-to tool. @cursor_ai is my go to. over codex, or claude code.
English
0
0
0
45