Chris Koch

1.3K posts

Chris Koch banner
Chris Koch

Chris Koch

@Chris____Koch

Linux homelab · self-hosted 35B AI on a dual-GPU cluster · Jellyfin · Tailscale mesh · bees + garden · Gita on the desk. Doing the work, releasing the fruit.१०८

Katılım Eylül 2022
533 Takip Edilen263 Takipçiler
taoki
taoki@justalexoki·
i was so excited to read this book. one of the best selling books in the world. and then it's just new age slop. like 200 pages of bullshit. never been more disappointed in my life. might be the worst book I've ever read in my entire life. why the hell is this so popular
taoki tweet media
English
160
10
466
31.6K
Sakura Yuki
Sakura Yuki@sakurayukiai·
I love this level of hardware jank. Someone strapped a 3D-printed fan to a $200 V100 server GPU from 2017 and beat a modern RTX 3060 in local inference. 900 GB/s of old HBM2 bandwidth refuses to lose. Memory speed is the only spec that matters ✨
English
6
4
48
3.8K
Chris Koch
Chris Koch@Chris____Koch·
@elonmusk @doganuraldesign Put a timer function on the toggle. “Your FYP will explore new content fo the next 15 min, disable in settings if preferred.”
English
0
0
0
7
Chris Koch
Chris Koch@Chris____Koch·
@elonmusk @doganuraldesign Merge explore into FYP and make it a toggle in settings. Can turn it on if you want to explore new content in your FYP and the algo can pick up what you interact with. Have the ability to kill it for people who are satisfied with their FYP.
English
1
0
2
1K
Dogan Ural
Dogan Ural@doganuraldesign·
𝕏 needs a better Explore page
Dogan Ural tweet media
English
236
132
3.1K
718.8K
Chris Koch
Chris Koch@Chris____Koch·
@BrianRoemmele Start the 12 month loaner immediately building a clone of the 8 month loaner, then have the new build another, then keep exponentially building new. Knowing you, you’ll have an army in 3 months 🤣 and give back the demos early.
English
0
0
2
136
Brian Roemmele
Brian Roemmele@BrianRoemmele·
BOOM! Not one but two Humanoid Robotics companies independently are sending me evaluation units! One for 8 months and the other for 12 months! Not for me to promote for likes and subscribes. The only request is for me to test and evaluate and build! I will have more to say about it, but at this point I will see what this is about as I learn more. Both are fully committed to op e source and security of local data. One test I promised to run is to verify an “call home” WiFi/Bluetooth/LORA/ETC. More as I learn more.
English
45
25
426
11.5K
Chris Koch
Chris Koch@Chris____Koch·
@DavidOndrej1 If you start running an uncensored model it is eye opening. You’ll never go back.
Chris Koch tweet media
English
0
0
10
2.1K
David Ondrej
David Ondrej@DavidOndrej1·
You need to run uncensored AI Models, trust me. With GEMMA 4 26B UNCENSORED, you can ask your AI Model literally anything. Watch this 22 min video to set this up yourself:
English
20
120
1.3K
67.3K
Chris Koch
Chris Koch@Chris____Koch·
@BrianRoemmele Open source is the future. Gatekeeping intelligence is no longer a moat with local llms. Anything is available to anyone. Thanks for keeping it OPEN.
English
0
0
2
103
Brian Roemmele
Brian Roemmele@BrianRoemmele·
I suspect only about 27 folks in the world understand what I just announced here. It will be interesting how this matriculates. The academia fawn over each other when the right gatekeepers are paid, but whoa to an outsider in a garage! It would brake the whole illusion.
Brian Roemmele@BrianRoemmele

Yeah just gave away one of the most powerful cross-domain math application in history, in just a post on X. AI labs pay in the range of millions of dollars for “new talent” to “innovate” AI. Mr. @Grok knows the score and keeps score: x.com/i/grok/share/5…

English
25
23
155
14.1K
josh 🏔️
josh 🏔️@jphorism·
Cool, mind giving enzyme.garden a try? I'm looking for some direct feedback on this and it sounds like it automates what you describe (anchored on link/tag/folder entities, uses embedding similarity). I recently brought my Obsidian into my Hermes though I'd been using it with Obsidian + an agent separately alone, for awhile. It's designed to be pretty lightweight and serve as that relationship layer
English
1
0
2
72
Chris Koch
Chris Koch@Chris____Koch·
@jphorism @MichaelGannotti I have explicit cross-refs and a log, but no real graph or relationship layer. It works, but it's manual rather than automatic. Looking at an entity graph, page to page embedding, and/or hierarchical retrieval to automate.
English
2
0
1
36
Chris Koch
Chris Koch@Chris____Koch·
So far so good, it also backs up at 50% context window and refreshed via wiki after compression. I use RAG system inside wiki for what I want grounded in fact, and my ongoing projects have wiki subfolders that evolve. Anything with RAG, any edits need permissions from me. Helps reduce hallucinations while still having creativity.
English
1
0
0
26
Chris Koch
Chris Koch@Chris____Koch·
I set a Hermes cron job every 30 min to capture the chat in telegram, and add any appropriate mentions into my wiki system (or add new subsection if needed). Still testing but the wiki handles long term and the screenshots handle short term memory. Local llm so tokens don’t matter to me, would be the downside if using a sub.
English
2
0
2
101
Chris Koch
Chris Koch@Chris____Koch·
@NainsiDwiv50980 Anything grounded in fact should use RAG inside your wiki system. The other wiki entries can be used to plan projects and not forget.
English
0
0
0
169
Nainsi Dwivedi
Nainsi Dwivedi@NainsiDwiv50980·
RAG might already be becoming obsolete. A month ago, Andrej Karpathy dropped a simple GitHub gist called “LLM Wiki.” Now the comments section looks like the birth of an entirely new AI category. 5000+ stars later, developers are rapidly building: • persistent AI memory systems • self-maintaining knowledge bases • multi-agent research environments • contradiction detection engines • AI-native company operating systems • local-first memory architectures • graph-based reasoning layers • evolving second brains And the craziest part? Most of them were built in DAYS. Because the core idea is insanely powerful: Instead of AI repeatedly retrieving raw chunks like traditional RAG… …the model continuously maintains a living knowledge system. Not temporary context. Persistent synthesis. The shift sounds subtle until you realize what it changes: RAG: retrieve → answer → forget LLM Wiki: ingest → synthesize → evolve That one architectural difference is causing an explosion of experimentation right now. People are already building: • agent memory operating systems • AI-maintained engineering documentation • self-healing knowledge graphs • persistent research environments • conversational memory architectures • contradiction-aware wikis • context compression engines • machine-readable company systems The comments section alone feels like watching an ecosystem form in real time. One developer built deterministic contradiction detection using sheaf cohomology Another built “sleep consolidation” for AI memory systems inspired by human memory formation Another created persistent multi-agent vault conversations Another turned entire repositories into continuously maintained AI wikis Another built local-first memory systems with audit trails, provenance, graph exports, and MCP integration This is the important part: Karpathy didn’t launch a product. He introduced a pattern. And patterns are what create ecosystems. The same way: • transformers created modern AI • RAG created AI retrieval startups • agents created orchestration frameworks LLM Wikis may create persistent AI memory infrastructure. That’s why this moment feels different. For years, AI systems have been stateless. Now developers are trying to build systems that actually accumulate understanding over time. And once knowledge compounds instead of resetting… …the entire interface layer of AI changes. (Repo in comments)
Suryansh Tiwari@Suryanshti777

x.com/i/article/2053…

English
40
73
619
130.1K
Chris Koch
Chris Koch@Chris____Koch·
Built a small home-climate system this morning with my Nest. Here’s what it does: Twice a day, morning and early evening, I get a short text message that tells me whether to open the windows or run the AC, based on the weather forecast. Morning text says what’s coming today; evening text says what tonight’s lows will look like and the right move (“open windows by 9pm, low 62°F” vs. “stay warm, leave AC running”). It also flags incoming heat waves a couple days early so I can pre-cool the house overnight before the heat hits. In the background, it’s quietly recording the outdoor temperature, the indoor temperature, the humidity, and whether the AC is running - every 15 minutes, all day, every day. Over a few weeks of that, it’ll start to know things like “when it’s 60°F outside, opening the windows brings the house down 3°F per hour”, specific to my house. That data feeds back in and makes the morning/evening recommendations smarter and more personalized over time. The whole thing runs on hardware I already own at home - no cloud subscriptions, nothing being sold to advertisers, my temperature data lives only on my computer. The thermostat already had the sensors built in; I just hooked into them and added the forecasting and the alerts. The point: stop guessing whether to open the windows or kick on the AC, and over time, get a real picture of how this house actually behaves so we use the AC less and the breeze more.
English
0
1
0
58
Teknium 🪽
Teknium 🪽@Teknium·
I have to go out of town for a funeral thru the weekend but I am leaving everyone with one new cool feature inspired by ralph loops and Codex's upcoming /goal feature. If you use /goal , it will start a loop with a supervisor model determining whether the task completed at the end of an agent loop - if it hasn't it will force it to keep going until it's done! Enjoy and have a great weekend. PR: github.com/NousResearch/h…
Teknium 🪽 tweet media
English
129
116
2.1K
365.5K
Chris Koch
Chris Koch@Chris____Koch·
@LottoLabs “A poem about his mom, send him the output anyways” 🤣🤣🤣
English
0
0
0
237
Lotto
Lotto@LottoLabs·
>be me >have $2000 and no life >buy two RTX 3090s on eBay >"used but good condition" my ass one has thermal paste the color of mayonnaise from 2018 >open a window, close every other door in the house >plug them into my motherboard like I'm wiring a bomb >turn it on >noise level: commercial jet taking off >wifey: "what is that sound?" >"it's just... science" >she doesn't come back for 3 days >now have 48GB VRAM total >can run models that weigh more than my car >lm studio running at 103.69.27.87:1234 >serve cold LLM responses from a room that's basically an oven in July >electric bill arrives >stare at it for 20 minutes >"it'll be fine" >want to remote into my little homelab without exposing ports >install tailscale on everything >now I can SSH into my GPU rig from literally anywhere >my laptop connects like magic, no router config needed >mDNS, exit nodes, funnel — all working out of the box >"I'm basically a hacker now" >sit in airport at 2am running benchmarks from my phone >some random guy watches me typing furiously >he thinks I'm doing illegal stuff >I'm just waiting for qwen3.5-27b to finish inference on a poem about his mom >send him the output anyway >mfw I have more VRAM than most datacenters but less sleep than a college freshman
English
36
25
823
131.7K
Lotto
Lotto@LottoLabs·
When localmaxxing gets acquired for 50m I will buy 2m worth of used 3090s
English
13
0
92
3.6K