trashman

426 posts

trashman

trashman

@tint_space

I like building things

Around เข้าร่วม Ağustos 2018
35 กำลังติดตาม235 ผู้ติดตาม
trashman รีทวีตแล้ว
trashman
trashman@tint_space·
@sriramk some of us have been using it more than a year now
English
0
0
0
19
Sriram Krishnan
Sriram Krishnan@sriramk·
there are several products waiting to be built here.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
51
13
463
124.4K
David Cramer
David Cramer@zeeg·
where's all those billion dollar businesses built from gas town and other slop farms? oh
English
24
9
355
44.1K
Péter Szilágyi
Péter Szilágyi@peter_szilagyi·
AI and the illusion of sandboxing: OpenCode: You can't access files outside this folder. Claude: Lol, hold my beer
Péter Szilágyi tweet media
English
10
6
76
8.1K
Clevta
Clevta@Clevta·
Sumer Brain needs work
Clevta tweet media
English
3
0
9
6.7K
trashman
trashman@tint_space·
@ScottBarrettDFB It was fun when there was value on Pierce 60+ at +300. Now it's nearly all gone.
English
0
0
1
1.1K
Scott Barrett
Scott Barrett@ScottBarrettDFB·
Alec Pierce, very quietly, ranks 12th in receiving yards per game (71.6). He's exceeded 65 receiving yards in 5 of 7 games this year. And Pierce has been at his best most recently. Over the last three weeks, Pierce ranks 8th among all WRs in targets (27), ranks 2nd in receiving yards (282), and leads in air yards by a mile: 1. Alec Pierce (577) 2. Troy Franklin (486) 3. Romeo Doubs (362) On Caesar's Sportsbook, Pierce's over/under is set at 54.5 receiving yards. That feels fairly lockish, but I'd actually prefer to ladder Pierce, where Caesar's will pay out +900 if Pierce gets 120 receiving yards. One of our biggest edges at @FantasyPtsData is our access to coverage shell data. Although it's highly predictive, other projection providers aren't looking at this, sportsbooks aren't using this when setting lines. As evidenced below, Pierce is crushing single-high coverage this year. Only Puka Nacua has been better. And Pierce's TPRR explodes from 0.13 to 0.31 against it. Atlanta deploys single-high coverage 73% of the time. No one else is over 65%.
Scott Barrett tweet mediaScott Barrett tweet media
English
10
7
296
63.7K
trashman รีทวีตแล้ว
mert
mert@mert·
in my circles we call this bagwork
mert tweet media
English
23
4
120
10.9K
JetPack Galileo
JetPack Galileo@JetPackGalileo·
Luther Burden next week for sure
English
14
3
535
63.5K
trashman
trashman@tint_space·
trashman tweet media
ZXX
1
0
2
233
trashman
trashman@tint_space·
@benj_robinson imo full control and transparency over the context window is a non-negotiable quality. Another one is the system not being silo’d and should easily be able to take any action or bring anything into context Alternative approach:
trashman@tint_space

@DetroitOnLion @DET_in_TOL @SyedSchemes @SumerSports When `trust but verify` is not easy, you are often left picking between blind trust and zero trust. There are many other reasons why silo'd blackbox llm chats are not the way imo An example of a more easily auditable approach — base.tint.space/thread/6d5712b…

English
0
0
0
618
Benjamin Robinson
Benjamin Robinson@benj_robinson·
I'll admit Sumer Brain is really cool. I'd also like some more transparency from any AI system so I know it's coming up with the numbers that are accurate and not hallucinated. If I ask the same question twice, does it produce the same result? Trust but verify.
English
4
0
36
3.7K
Computer Cowboy
Computer Cowboy@benbbaldwin·
I'm trying to keep an open mind about all this AI stuff but I think I am drawing my personal line at things that cannot be independently verified
English
10
0
81
16.2K
Jeremy Reisman
Jeremy Reisman@DetroitOnLion·
@DET_in_TOL @SyedSchemes @SumerSports In theory, yes. In practice, no. I, quite simply, have zero trust in AI models right now. My very first query resulted in an incorrect response, which was very obvious, because Cincy obviously didn't have 63 attempted passing plays.
Jeremy Reisman tweet media
English
2
0
8
11.4K
Shawn Syed
Shawn Syed@SyedSchemes·
I think @SumerSports' SumerBrain will save you time and get you access to better data: *Complex data pulls that would require multiple filters *Advanced charting data for situation specific matchups *Box score stats Just type it in In Beta for free: sumersports.com/sumerbrain
English
31
74
612
658.6K
trashman รีทวีตแล้ว
dax
dax@thdxr·
my favorite kind of product to build is one that has a very simple, opinionated surface that you can use without knowing much but with built with a set of underlying primitives pro users can access and go crazy configuring and composing
English
16
5
218
12.9K
trashman รีทวีตแล้ว
SIGKITTEN
SIGKITTEN@SIGKITTEN·
I'm very happy to see that cursor-agent implements my favorite claude-code feature
English
55
25
883
79.1K
trashman
trashman@tint_space·
@steipete sweet spot where you’re not waiting nor holding up a loop
English
0
0
0
44