Steffen Fiedler

698 posts

Steffen Fiedler

Steffen Fiedler

@stffnfdlr

Part of StudioNAND.

Berlin, Germany Katılım Mayıs 2007
395 Takip Edilen394 Takipçiler
Steffen Fiedler
Steffen Fiedler@stffnfdlr·
@witcheer @stackwalnuts Super relevant to what I'm trying right now on Hermes: Obsidian-backed memory with ongoing cycles so each run closes with structured updates and the next run starts from a fresh state + goals. Keep us posted on week-one outcomes.
English
1
0
1
52
witcheer ☯︎
witcheer ☯︎@witcheer·
trying a new memory tool for my AI agents: @stackwalnuts basically it organises your context into "walnuts" - structured files with their own identity, state, history, and task queue. separate domains that link to each other. your agent reads them at session start, writes to them at session end. set up 4 walnuts on my mac mini yesterday: personal identity + content strategy // day job (CDP protocol, partnership pipeline) // AI agent capabilities // business admin wired my 24/7 cron agent to write to the walnuts automatically. end of every research session > log entry prepended to the right walnut. end of every content draft > captured to the personal walnut. what i'm watching: - whether the agent reliably runs the walnut update at the end of long sessions - how context quality holds up after 50+ entries - setup friction: ALIVE was built for openclaw, I'm running hermes agent. had to write a custom skill + shell script to bridge them. will report back in a week
witcheer ☯︎ tweet media
English
4
3
23
2K
Kristof
Kristof@CoastalFuturist·
If there’s enough interest I’d like to make a group chat for people using openclaw / hermes agent heavily I really want to understand some good use cases, best practices, and just have a place for people to talk shop Comment if you’re interested
English
322
4
328
20.6K
Lotto
Lotto@LottoLabs·
Hermes Agent and qwen 3.5 27b on RTX6000 getting a nice 48TPS on vLLM Testing 3090 + vLLM next
English
17
5
175
11.6K
Steffen Fiedler
Steffen Fiedler@stffnfdlr·
@kr0der Hermes is state-of-the-art in agentic systems. Straightforward setup, slim config, never drops a job, and has self-healing capabilities.
English
3
0
10
1.7K
Anthony
Anthony@kr0der·
gonna give Hermes Agent a try this week, anyone tried it yet? it has the coolest name not gonna lie Hermes
Anthony tweet media
English
65
14
279
33.2K
0x_Vivek
0x_Vivek@0x_Vivek·
@tunahorse21 Hermes sync architecture is clean. Real alpha is connecting it with SQLite for persistent memory across sessions. It's how I keep my PR learnings alive. 🦞
English
1
0
2
204
tuna🍣
tuna🍣@tunahorse21·
okay hermes agent kind of bangs, simplest setup out of all the personal agents ive tried fed it a bunch of repos via telegram and it installed cleanly connected todoist cli with just the repo name
tuna🍣 tweet mediatuna🍣 tweet media
English
13
0
85
8.5K
Steffen Fiedler
Steffen Fiedler@stffnfdlr·
@techNmak Why would running a small local model be a good choice if the quality of the resulting source code is junk compared to Opus? Instead, using a local model for housekeeping, data retrieval, and data collection tasks can be great to keep the bills down.
English
2
1
13
3.6K
Tech with Mak
Tech with Mak@techNmak·
Claude Code can run entirely on your local GPU now. Unsloth AI published the complete guide. The setup itself is straightforward - llama.cpp serves Qwen3.5 or GLM-4.7-Flash, one environment variable redirects Claude Code to localhost. But the guide is valuable because of what it explains beyond the setup: Why local inference feels impossibly slow: Claude Code adds an attribution header that breaks KV caching. Every request recomputes the full context. The fix requires editing settings.json - export doesn't work. Why Qwen3.5 outputs seem off: f16 KV cache degrades accuracy, and it's llama.cpp's default. Multiple reports confirm this. Use q8_0 or bf16 instead. Why responses take forever: Thinking mode is great for reasoning but slow for agentic tasks. The guide shows how to disable it. The proof it all works: Claude Code autonomously fine-tuning a model with Unsloth. Start to finish. No API dependency. Fits on 24GB. RTX 4090, Mac unified memory.
Tech with Mak tweet media
English
55
211
1.7K
129.5K
Steffen Fiedler
Steffen Fiedler@stffnfdlr·
@LottoLabs Same here, Hermes + Gwen 3.5 4B locally for housekeeping and optional 14B for background tasks where tidiness is key.
English
0
0
5
347
Lotto
Lotto@LottoLabs·
Little hermes agent running w/ 4b qwen 3.5 running skills to manage your node. This is the way forward.
English
9
2
122
8.9K
Steffen Fiedler
Steffen Fiedler@stffnfdlr·
@decisionoracle @Teknium Maintaining a growing configuration became a balancing act with Openclaw. In comparison Hermes Agent has a very slim an thought through setup.
English
0
0
1
97
Teknium (e/λ)
Teknium (e/λ)@Teknium·
We just released Hermes Agent! In my humble opinion a very good blend between coding agents like Claude Code and generalist agents like Clawdbot. Been working on this for the last month or so now - started as a way for us to have agentic primitives for datagen and RL and got inspired by the agentic revolution of late, so been expanding it's scope and capabilities non-stop! Hope you all enjoy.
Nous Research@NousResearch

Meet Hermes Agent, the open source agent that grows with you. Hermes Agent remembers what it learns and gets more capable over time, with a multi-level memory system and persistent dedicated machine access.

English
78
65
942
119K
Steffen Fiedler
Steffen Fiedler@stffnfdlr·
@Zeneca Fully transitioned to Hermes Agent. In comparison Openclaw felt like a tar pit.
English
1
0
1
1K
Zeneca🔮
Zeneca🔮@Zeneca·
thoughts on openclaw vs hermes? has anyone switched from one to the other, successfully bringing their memory systems and other things?
English
91
4
226
74.3K
Jack Rusher
Jack Rusher@jackrusher·
Flying out of the new BER airport for the second time. I just want to say publicly that I am very willing to teach free workshops on queuing theory to architects. en.m.wikipedia.org/wiki/Queueing_…
Schönefeld, Deutschland 🇩🇪 English
3
2
33
0
Steffen Fiedler
Steffen Fiedler@stffnfdlr·
@stphnthiel @laessr My inspiration has been the white lies approach to faking context with artificial background noise. Like music boxes that provide ambient sounds of traffic or bars when you call your actual partner from an hour hotel. Those must have been from the 60s.
English
0
0
2
0