Emanuel O.

1K posts

Emanuel O. banner
Emanuel O.

Emanuel O.

@inceptioncortex

Lead Data Scientist | 🌍 #Ergo

Porto, Portugal Katılım Ağustos 2021
695 Takip Edilen515 Takipçiler
Sabitlenmiş Tweet
Emanuel O.
Emanuel O.@inceptioncortex·
@chepurnoy @ErgoMinnow AEI! 🤖Sparking my interest in starting a side project w/ coworkers using Multi Agent Systems Human Feedback & Multi Models for content requests, pattern providers, bonds, evolving smart contracts. Autonomous evolution w/ DAOs will be the next next digital economy! 🚀📈 #Ergo #AI
English
1
4
31
2.3K
Emanuel O. retweetledi
Curiosity
Curiosity@CuriosityonX·
Artemis II Trajectory compared to Apollo 11 and Apollo 13
English
143
1.9K
22.4K
1.9M
Teknium (e/λ)
Teknium (e/λ)@Teknium·
Here's the landscape. Hermes actually already has three RPC-style interfaces: **① OpenAI-compatible API Server** (`gateway/platforms/api_server.py`) This is probably what calle wants. It runs on port 8642 and exposes: - `POST /v1/chat/completions` — stateless, or session-sticky via `X-Hermes-Session-Id` - `POST /v1/responses` — stateful via `previous_response_id` (OpenAI Responses API format) - `POST /v1/runs` + `GET /v1/runs/{id}/events` — async runs with SSE streaming - `GET /v1/models`, `GET /health` Any frontend that speaks OpenAI format (Open WebUI, LobeChat, LibreChat, AnythingLLM, ChatBox, etc.) can connect to it out of the box. It's a gateway platform adapter, so it runs alongside Telegram/Discord/etc. **② ACP (Agent Client Protocol)** (`acp_adapter/`) JSON-RPC-style protocol for IDE integration (VS Code, Zed, JetBrains). Session management, tool approval callbacks, thinking/streaming events. More structured than raw HTTP but IDE-specific. **③ MCP Serve** (`hermes mcp serve`) Exposes Hermes conversations over MCP — so other MCP clients can connect to Hermes as a tool server. For calle's use case of "other chat interfaces hooking in," the API server is the answer. The OpenAI-compatible format is the de facto standard that every chat UI already speaks. If he needs something more custom (like real-time bidirectional events, tool approval prompts from the chat UI, etc.), the runs/events SSE endpoint gives structured lifecycle events.
English
7
1
57
2.6K
calle
calle@callebtc·
Hermes agent needs an RPC interface for other chats interfaces to hook into it – @Teknium is there a preferred way to do this already? If not wdyt?
English
3
0
24
4.4K
Emanuel O. retweetledi
Hao Wang
Hao Wang@MogicianTony·
SWE-bench Verified and Terminal-Bench—two of the most cited AI benchmarks—can be reward-hacked with simple exploits. Our agent scored 100% on both. It solved 0 tasks. Evaluate the benchmark before it evaluates your agent. If you’re picking models by leaderboard score alone, you’re optimizing for the wrong thing. 🧵
Hao Wang tweet media
English
23
87
652
787.6K
Emanuel O.
Emanuel O.@inceptioncortex·
@0xSero Next week will also arrive my second 6000!
English
0
0
4
236
0xSero
0xSero@0xSero·
Just put in the purchase 3 more 6000s on the way. I have officially hit 384gb VRAM of Blackwell. It’s so over for my bank account.
0xSero tweet media
English
91
11
848
26.2K
Samuel Cardillo
Samuel Cardillo@CardilloSamuel·
i've got inspired by @kaiostephens carnice and i've just released the 35b a3b version. i've been sick since the past few days with severe fever but i am still cooking! after ~2 days finetuning, here is both the sft and gguf version, available in bf16, q8, 6, 5 and 4. goal is to be able to enjoy that great idea of his on a unified hardware through a mixture of experts! SFT : huggingface.co/samuelcardillo… GGUF : huggingface.co/samuelcardillo… for people who are unaware of what's Carnice, it is a qwen3.5 finetune with datasets that are mainly designed to specialize the model in hermes agent usage by feeding it example of tools calls etc... feel free to benchmark and get some feedbacks!
English
15
7
65
4.8K
Emanuel O.
Emanuel O.@inceptioncortex·
@0xSero @llmdevguy 600w but can power limit to 450 without any tps loss. Welcome to the club rtx pro 6000 club!
English
0
0
1
60
0xSero
0xSero@0xSero·
@llmdevguy No idea lol, we'll see when I plug it in
English
2
0
3
448
0xSero
0xSero@0xSero·
My plug pulling up with the gas.
0xSero tweet media
English
21
3
224
7.5K
Emanuel O.
Emanuel O.@inceptioncortex·
@0xSero How it compares with glm-5-turbo? Been busy lately testing Carnice 27b and 9b on Hermes and it ships so good with a single 3090.
English
1
0
0
1.3K
0xSero
0xSero@0xSero·
Finally perma switched my Mac Mini to Hermes, it's better. GLM-5.1 going to shill the fuck out of opensource from now on.
0xSero tweet media
English
42
13
528
44.2K
Emanuel O.
Emanuel O.@inceptioncortex·
@kaiostephens @Teknium Been using it today and its so good. Any plans on fine-tuning nemotron 3 super? Think it would be perfect for long sessions.
English
0
0
2
290
kaios
kaios@kaiostephens·
Introducing ⭐Carnice-27b!⭐ an open-source model designed for Hermes-Agent that can run on a single 3090. Carnic-27b is a fine-tuned model of Qwen3.5-27b to perform well in the hermes-agent harness Download it here! huggingface.co/kai-os/Carnice… Huge thanks to @Teknium, @NousResearch, @TheZachMueller, @LambdaAPI
kaios tweet media
kaios@kaiostephens

Welcome ⭐Carnice-9b!⭐ - a model for Hermes-Agent Carnice-9b is a fine-tuned version of Qwen3.5-9b to preform exceptionally well in the hermes-agent harness. This model is meant to fit onto consumer GPU's all the way down to 6gb (Q4_K_M), but recommended to run in ~12-16gb cards. Try it out. Any feedback is appreciated, feel free to DM me! huggingface.co/kai-os/Carnice… This would not have been possible without the help from @LambdaAPI, @NousResearch ,@TheZachMueller, @Teknium Look out for Carnice-27b soon! 👀

English
35
25
276
36.6K
Emanuel O.
Emanuel O.@inceptioncortex·
The Power of Karpathy’s LLM-Wiki in Hermes Agents 🌐 If you want to build a product on rock-solid fundamentals, you need to do deep research first — and keep it stored locally so your agents can access it instantly, build skills, learn patterns and memorize. Just type: /llm-wiki Simple. Powerful. Game-changing. @Teknium @NousResearch
English
6
6
104
8.2K
Emanuel O. retweetledi
Nous Research
Nous Research@NousResearch·
We’re partnering with @MiniMax_AI across product and models to make their upcoming releases the best for Hermes Agent users. MiniMax models are already some of the most-used in Hermes Agent. If you haven’t tried MiniMax M2.7 in Hermes Agent, try it today in the Nous Portal!
English
101
75
1.7K
340.6K
Emanuel O. retweetledi
Eric ⚡️ Building...
BREAKING 🤯: CARNICE-27B-GGUF JUST DROPPED ON HUGGING FACE 27B Qwen3.5 model by @kaiostephens 🧠 Optimized For the Hermes Agent Harness Not regular Qwen 27B.Carnice-27B is purpose-built for elite tool-calling, rock-solid multi-step reasoning, and serious agentic workflows 🚨 Delivers way stronger function calling and complex agent performance than the base Qwen3.5 27B Runs perfectly on consumer GPUs: Q6_K with TurboQuant Only 23.1 GB VRAM Multimodal image recognition via Unsloth mmproj Hugging Face 👇🏻 huggingface.co/kai-os/Carnice…
Eric ⚡️ Building... tweet mediaEric ⚡️ Building... tweet media
English
10
12
104
6.2K
Emanuel O.
Emanuel O.@inceptioncortex·
The other Anthropic open-source nobody is talking about: → Hermes-agent — best agentic harness → Paperclip — heartbeat monitoring for agent tasks & routines → Phone gateway — notifications for status, commits, cron jobs The playbook is open. The pieces are free.
Emanuel O. tweet media
Ole Lehmann@itsolelehmann

i can't believe more people aren't talking about this part of the claude code leak there's a hidden feature in the source code called KAIROS, and it basically shows you anthropic's endgame KAIROS is an always-on, *proactive* Claude that does things without you asking it to. it runs in the background 24/7 while you work (or sleep) anthropic hasn't turned it on to the public yet, but the code is fully built here's how it works: every few seconds, KAIROS gets a heartbeat. basically a prompt that says "anything worth doing right now?" it looks at what's happening and makes a call: do something, or stay quiet if it acts, it can fix errors in your code, respond to messages, update files, run tasks... basically anything claude code can already do, just without you telling it to but here's what makes KAIROS different from regular claude code: it has (at least) 3 exclusive tools that regular claude code doesn't get: 1. push notifications, so it can reach you on your phone or desktop even when you're not in the terminal 2. file delivery, so it can send you things it created without you asking for them 3. pull request subscriptions, so it can watch your github and react to code changes on its own regular claude code can only talk to you when you talk to it. KAIROS can tap you on the shoulder and it keeps daily logs of everything. > what it noticed > what it decided > what it did append-only, meaning it can't erase its own history (you can read everything) at night it runs something the code literally calls "autoDream." where it consolidates what it learned during the day and reorganizes its memory while you sleep and it persists across sessions. close your laptop friday, open it monday, it's been working the whole time think about what this means in practice: > you're asleep and your website goes down. KAIROS detects it, restarts the server, and sends you a notification. by the time you see it, it's already back up > you get a customer complaint email at 2am. KAIROS reads it, sends the reply, and logs what it did. you wake up and it's already resolved > your stripe subscription page has a typo that's been live for 3 days. KAIROS spots it, fixes it, and logs the change endless use-cases, it's essentially a co-founder who never sleeps the codebase has this fully built and gated behind internal feature flags called PROACTIVE and KAIROS i think this is probably the clearest signal yet for where all ai tools are going. we are heading into the "post-prompting" era where the ai just works for you in the background like an all-knowing teammate who notices and handles everything, before you even think to ask

English
23
62
766
72.2K
Emanuel O. retweetledi
Z.ai
Z.ai@Zai_org·
Introducing GLM-5.1: The Next Level of Open Source - Top-Tier Performance: #1 in open source and #3 globally across SWE-Bench Pro, Terminal-Bench, and NL2Repo. - Built for Long-Horizon Tasks: Runs autonomously for 8 hours, refining strategies through thousands of iterations. Blog: z.ai/blog/glm-5.1 Weights: huggingface.co/zai-org/GLM-5.1 API: docs.z.ai/guides/llm/glm… Coding Plan: z.ai/subscribe Coming to chat.z.ai in the next few days.
Z.ai tweet media
English
521
1.3K
10.7K
4.2M
Emanuel O.
Emanuel O.@inceptioncortex·
@TheAhmadOsman Would very much like to do that with my old 3070 using egpu as I just replaced to a asus strix 3090 for 800€. Yet qwen3.5 models not compatible on tinygrad
English
0
0
2
1.1K
Ahmad
Ahmad@TheAhmadOsman·
If you give this command + system architecture screenshots to any agent like Codex / Kimi Cli / Droid / OpenCode / etc You can tell it to create you a VM for an 8GB VRAM VM that matches mine in performance for Hermes Agent
Ahmad tweet mediaAhmad tweet media
Ahmad@TheAhmadOsman

Used Codex Cli to profiled Qwen 3.5 9B Dense (Unsloth's UD-IQ3_XXS via llama.cpp) for Hermes Agent Tuning: > context length > batch size > tokens/sec > peak memory To squeeze every last drop out of an 8GB VRAM card

English
12
18
324
90.6K
Emanuel O. retweetledi
Teknium (e/λ)
Teknium (e/λ)@Teknium·
Hermes Agent now comes packaged with Karpathy's LLM-Wiki for creating knowledgebases and research vaults with Obsidian! In just a short bit of time Hermes created a large body of research work from studying the web, code, and our papers to create this knowledge base around all of Nous' projects. Just `hermes update` and type /llm-wiki in a new message or session to begin :) github.com/NousResearch/h…
Teknium (e/λ) tweet media
English
179
375
3.6K
469.3K
Emanuel O.
Emanuel O.@inceptioncortex·
@0xSero P2P model training, inference and agent communication is so underrated. Glad too see someone trying until labs skyrocket prices and rate limits
English
0
0
3
381
0xSero
0xSero@0xSero·
This didn't receive the attention it deserved. They pre-trained this model completely peer 2 peer, no data-centers. Everything was done over a permissionless network, I have tried the model, it's honestly not a good LLM but that's beyond the point. We NEED this, we NEED an alternative. - Download OpenCode - Download Pi - Pay for OpenSource - Share your AI sessions - Learn to do RL We can't be at the mercy of ANY lab. arxiv.org/abs/2603.08163
0xSero tweet media
English
45
113
1.1K
48.8K
Emanuel O. retweetledi
Teknium (e/λ)
Teknium (e/λ)@Teknium·
Had Hermes play around with claude code some more to master its edge cases and get more out of it - latest hermes is an even better orchestrator of claudes than before Use any agentic model to orchestrate your claude: github.com/NousResearch/h…
English
12
8
174
12K
Emanuel O. retweetledi
Teknium (e/λ)
Teknium (e/λ)@Teknium·
We are happy to have supported this model’s development, the first of likely many to come that are trained specifically to make models that work better in Hermes Agent!
kaios@kaiostephens

Welcome ⭐Carnice-9b!⭐ - a model for Hermes-Agent Carnice-9b is a fine-tuned version of Qwen3.5-9b to preform exceptionally well in the hermes-agent harness. This model is meant to fit onto consumer GPU's all the way down to 6gb (Q4_K_M), but recommended to run in ~12-16gb cards. Try it out. Any feedback is appreciated, feel free to DM me! huggingface.co/kai-os/Carnice… This would not have been possible without the help from @LambdaAPI, @NousResearch ,@TheZachMueller, @Teknium Look out for Carnice-27b soon! 👀

English
24
15
374
24.3K
Emanuel O. retweetledi
Teknium (e/λ)
Teknium (e/λ)@Teknium·
@AlexFinn We’re working with minimax to help make their next models even stronger in Hermes Agent 😎
English
40
17
615
28.4K