Sudo su

8.4K posts

Sudo su banner
Sudo su

Sudo su

@sudoingX

GPU/local LLM. more RAM and OSS... everywhere

Bangkok, Thailand Katılım Ağustos 2022
967 Takip Edilen28.9K Takipçiler
Nous Research
Nous Research@NousResearch·
Join the Nous Research team for another Hermes Agent Jam in our Discord This one will be an interactive session, so come prepared to discuss ideas and show off your projects!
Nous Research tweet media
English
37
32
365
566.7K
Sudo su
Sudo su@sudoingX·
two stories at the top of the X timeline right now. > "OpenClaw faces skepticism as users switch to Hermes Agent" at 721 posts, 16h trending. > "Nous Research adds seamless computer control to Hermes Agent" at 403 posts, 15h trending. while i've been saying this for months, the timeline caught up. users are walking from the framework i've been calling a babysitting trap. and the harness i've been recommending is shipping NEW capability while the competitor faces "skepticism." this is what i mean every time i say harness matters more than the model. the model is open. the harness decides whether you ship or wait for approval prompts like a person waiting for plane at train station. bloated tool users, the door is open. one tool, hermes agent, ships your work autonomously. computer control just landed. the throne is still not crowded. your cognition deserves a better tool.
Sudo su tweet media
English
12
5
80
4.7K
Sudo su
Sudo su@sudoingX·
@dee_hw 8x 5090s landing in actual business hands is how local AI goes from "interesting take" to default infra. sweet build, those racks look clean.
English
1
0
8
247
Dee
Dee@dee_hw·
@sudoingX yeah i think businesses running their own local ai will be the norm within the next 12-24 months. we set up these boxes for our business and they are the sweetest thing ever.
Dee@dee_hw

On-Premise Business AI Center After my posts on the 2-GPU and 4-GPU builds, people reached out asking how to build an 8-GPU box for their businesses. Why? - Protect their IP - Protect customer data - Save on inference costs - Train their own models Here's how to build one: 🧵

English
1
0
10
613
Sudo su
Sudo su@sudoingX·
i hear it often. what's the point of local AI when cloud models are so far ahead. and yes, frontier closed-source models are ahead. that's true. before you assume i'm trying to convince you otherwise, hold the question one minute. if you handle client financials, do you want that data sitting in anthropic's next training set. if you have a private case open and need a second brain to think through it, are you comfortable hitting openai api with the details. if you're working on something new that could matter, do you want that thinking captured upstream before you ship it. some tasks stay between me and my machine. not everything has to be cloud-driven. i don't write what i think you might like. i write what's already happening at the frontier of this age of acceleration. and there is no way i'm comfortable reasoning on top of my private data with frontier corporate models. i don't want my data to be the next training dataset. what i want is to mine frontier intelligence instead of being mined. that can happen if you orchestrate well, organize the work so frontier models solve the actual problem without knowing the core context. a 27b dense on a single 3090 is intelligence enough to play orchestrator, hold your context, route the public-facing problem out, integrate what comes back. nothing unveils until you actually try running this. words on screen don't load weights into vram. flip the lens. start running local ai. you will find things, for the first time, where every word stays between you and your machine. that is what independent thinking actually feels like. your context is your moat. don't lease it.
English
25
11
194
7.4K
Sudo su
Sudo su@sudoingX·
my dgx spark just woke up and chose violence. step-3.5-flash-REAP-121B loaded. 121 billion parameters, 11B active per token, running locally on unified memory. It asked using hermes agent if i want to install dependencies and fire the test suite. i said yes from my phone while stuck in bangkok traffic. by the time i got home it was already executing code autonomously. this model is becoming my new favorite on spark. fast enough for agentic loops. smart enough for real work. if you have a spark and you're not running step-3.5 REAP, you're leaving performance on the table.
Sudo su tweet media
English
15
2
91
9.2K
chase
chase@therealchaseeb·
Went to this thanks to an invite from @themoe and the rest of the @SuperteamTH team. FINALLY there a huge push for AI and tie ins to blockchain in Thailand. Met the prior minister of finance who is helping support the initiative w/ talks of an innovation center. 🇹🇭
Pichapen@pichapen

See you today at AI DEMO DAY! Updating today’s agenda | May 12, 2026 14:30 - The Living Fair Opens: Check-in via AI Passport and start your mission across the Showcase and Booth Zones. 15:00 - Opening Remarks by Abhisit Vejjajiva @Abhisit_DP former Prime Minister, Leader of Democrat Party and MP 15:10 | Forum 1: Built in Bangkok — Why Builders Are Choosing This City (Typoon LLM, Cysmiq, OnlyFounders) 16:00 | BKK Showcase Batch #1 16:40 | Forum 2: Scaling the Engine — Capital, Infrastructure, and Bangkok’s AI Future (Tiwa York, @KornGoThailand ) 17:20 | The Hero Workshop: Live Build - Vibe Coding Session by Claw Collective 18:20 | BKK Showcase Batch #2 19:00 | Closing Remarks by Korn Chatikavanij (Former Finance Minister, Deputy Leader of Democrat Party and MP)

English
7
3
38
5.6K
Sudo su
Sudo su@sudoingX·
hermes agent from bangkok traffic. the agent moves faster than i do.
Sudo su tweet media
English
1
0
45
4.5K
Sudo su
Sudo su@sudoingX·
notice who's loudest saying local AI is not there yet. it's almost never independent builders running models on their own hardware. it's two groups specifically. > 1. AI wrappers whose entire business model depends on you needing their cloud API layer. if local catches up, their pricing collapses. > 2. corporate salesmen on fat paychecks selling closed source contracts to enterprises. their commission depends on you believing local cannot do serious work. both groups have a financial incentive to keep telling you it's not there yet. neither runs the models they're talking about on a single 3090. the question is not whether local AI is there. it's who you trust to tell you the truth. someone paid to keep you on cloud, or someone running it on their desk and posting receipts. receipts are right there on the timeline. update your take.
Sudo su@sudoingX

look anon, those of you who kept saying local AI is not there yet, who said open source can't compete, who said you need cloud APIs to get anything serious done, look at this gameplay for one minute. every pixel on this screen was written by one model, in one shot, on a single rtx 3090 with 24gb of vram. the model is qwen 3.6 27b dense q4. the harness is hermes agent. the hardware is a single consumer card you can buy used for 900 dollars. the prompt is open source on github. every claim verifiable, on your own desk. if your local AI take is from 2024, update it. the consumer tier is shipping work that was supposed to need 8 gpus and an api key. open source moved the floor while the rest of the field was busy explaining why it cannot. 24gb tier owners are eating ramen with half boiled egg and double chocolate.

English
21
8
148
8K
Sudo su
Sudo su@sudoingX·
zoom in on this. when you pair this model with hermes agent equipped with all the tools and browser automation, it does not just write code and hope. it loads the page itself in a real browser, watches the console, finds issues, patches them, reloads, watches again. autonomous debug loop with no human in the middle. this is what i mean every time i say harness matters more than the model. same model on a worse harness sits there generating bad code that never reaches a browser to verify itself. same model on hermes agent ships working software. the model thinks. the harness lets the model see, act, and verify.
Sudo su tweet media
English
5
2
15
1.4K
Sudo su
Sudo su@sudoingX·
look anon, those of you who kept saying local AI is not there yet, who said open source can't compete, who said you need cloud APIs to get anything serious done, look at this gameplay for one minute. every pixel on this screen was written by one model, in one shot, on a single rtx 3090 with 24gb of vram. the model is qwen 3.6 27b dense q4. the harness is hermes agent. the hardware is a single consumer card you can buy used for 900 dollars. the prompt is open source on github. every claim verifiable, on your own desk. if your local AI take is from 2024, update it. the consumer tier is shipping work that was supposed to need 8 gpus and an api key. open source moved the floor while the rest of the field was busy explaining why it cannot. 24gb tier owners are eating ramen with half boiled egg and double chocolate.
Sudo su@sudoingX

update: qwen 3.6 27b dense q4 just one shotted octopus invaders game on a single 3090. hermes agent drove the whole thing, ~41 tok/s gen 21gb vram at full 262k context, thinking mode on. one prompt in and the canonical multi-file space shooter benchmark out, the same exact prompt i ran on qwen 3.5 27b dense back in march on the same card. 3.5 needed one external scope bug fix before the game would even load on first play. 3.6 needed nothing. 11 of 11 files written, 2411 lines of code, zero steering interventions, zero external fixes, playable on first load. 16 minutes 41 seconds wall clock from prompt to playable. consumer tier king on a single 3090 is locked tonight, and the silicon underneath my desk did not change between march and now. the open source ecosystem just moved the floor. watch it ship itself, the full 16 minutes 41 seconds sped to 3 minutes 45, no human touched the keyboard between the first prompt and the final frame.

English
22
19
237
23.2K
Sudo su
Sudo su@sudoingX·
12 tmux sessions are easy. anyone can spin them up. what slows me down is the context switching between them. if you are seriously building what you build while running paid work and life and relationships in parallel, the bottleneck stops being execution and becomes context maintenance. i don't wake up asking myself what work i should do first these days. i wake up asking which part of the context has not been updated in the last 12 hours. that is where the work actually is. there has to be a better mental model than the one i have.
English
12
0
47
2.6K
Sudo su
Sudo su@sudoingX·
@Scobleizer this is the way. messenger of intellect pervades everywhere
English
1
0
6
353
Sudo su
Sudo su@sudoingX·
@Miguel07Code exactly that anon. one tool, all under your roof. nothing else needed.
English
0
0
2
661
Sudo su
Sudo su@sudoingX·
let me say this out loud here: there is absolutely zero reason to use openclaw in may 2026. a general agent exists. hermes agent does coding, video editing, marketing design, research, browser automation, terminal work. one tool, all under your roof.
Joel - coffee/acc@JoelDeTeves

@sudoingX What is the reason for using Openclaw at this point? I have had zero issues with Hermes.

English
48
24
440
46.5K
Sudo su
Sudo su@sudoingX·
if you do not benchmark you do not build. you narrate.
English
2
0
38
2.4K
Sudo su retweetledi
Sudo su
Sudo su@sudoingX·
when bro find out qwen 3.6 27B dense runs on a single rtx 3090
Vasko@RoliumGens

@sudoingX wait qwen 3.6 27B runs on a single 3090 ?

English
30
26
538
32.1K