Sudo su

6.6K posts

Sudo su banner
Sudo su

Sudo su

@sudoingX

GPU/local LLM and more RAM.

Bangkok, Thailand Присоединился Ağustos 2022
782 Подписки13.5K Подписчики
Закреплённый твит
Sudo su
Sudo su@sudoingX·
let me get you started in local AI and bring you to the edge. if you have a GPU or thinking about diving into the local LLM rabbit hole, first thing you do before any setup is join x/LocalLLaMA. this is the community that will help you at every step. post your issue and we will direct you, debug with you, and save you hours of work. once you're in, follow these three: @TheAhmadOsman the oracle. this is where you consume the latest edges in infrastructure and AI. if something dropped you hear it from him first. his content alone will keep you ahead of most. @0xsero one man army when it comes to model compression, novel quantization research, new tools and tricks that make your local setup better. you will learn, experiment, and discover things you didn't know existed. @Teknium maker of Hermes Agent, the agent i use every day from @NousResearch. from Teknium you don't just stay at the frontier, you get your hands on the tools before everyone else. this is where things are headed. if you follow me follow these three and join the community. you will be ahead of most people in this space. if you run into wrong configs, stuck debugging hardware, or can't get a model to load, post there so we can help. get started with local AI now. not only understand the stack but own your cognition. don't pay openai fees on top of giving them your prompts, your research, and your most valuable thinking to be monitored and metered. buy a GPU and build your own token factory.
Sudo su tweet media
English
42
50
659
32.5K
Sudo su
Sudo su@sudoingX·
@r0ck3t23 the gap between "AI will destroy us" and "I ran a 9B model on a $300 GPU and it built a game" is the entire problem with this conversation. builders know what this is. commentators don't.
English
0
0
0
6
Dustin
Dustin@r0ck3t23·
Jensen Huang just told every AI leader in the room to grow up. Stop scaring the public with science fiction. Start communicating like the weight of civilization is on your shoulders. Because it is. Huang: “AI is not a biological being. It is not alien. It is not conscious. It is computer software.” That single statement dismantles half the panic surrounding this industry. The mainstream conversation is dominated by people projecting human malice onto math. Alien consciousness onto code. Existential dread onto a software architecture we built, we trained, and we can read. Huang: “We say things like, ‘We don’t understand it at all.’ It is not true. We understand a lot of things about this technology.” When builders tell the public they don’t understand their own creation, the public hears threat. The state responds with control. That is already happening. Palihapitiya asked Huang what he would have told Anthropic during their regulatory clash with the Department of Defense. Huang didn’t attack the technology. He attacked the communication. Huang: “The desire to warn people about the capability of the technology is really terrific. We just have to make sure that we understand that the world has a spectrum, and that warning is good, scaring is less good because this technology is too important to us.” Warning shows risks, mitigation, why upside overwhelms downside. Scaring says we might be building something that destroys us and we can’t stop it. One builds trust. The other invites regulation written in panic. Huang: “To say things that are quite extreme, quite catastrophic, that there’s no evidence of it happening, could be more damaging than people think.” Projecting catastrophe without evidence is not caution. It is sabotage. When your technology is embedded in national defense, the financial system, and healthcare infrastructure, your words carry structural weight. If the architects act terrified of their own product, the response is predictable. Governments step in. They restrict. They seize control of something they don’t understand because the builders told them to be afraid. Huang: “There was a time when nobody listened to us, but now because technology is so important in the social fabric, such an important industry, so important to national security, our words do matter.” Most tech founders have not internalized this. You are no longer a startup founder disrupting an industry. You are running infrastructure that nations depend on. Your statements move policy. Your framing shapes legislation. Your tone determines whether governments treat you as partner or threat. Huang: “We have to be much more circumspect, we have to be more moderate, we have to be more balanced, we have to be far more thoughtful.” Huang did not ask for silence. He asked for precision. The leaders who cannot tell the difference will not be leading for long.
English
76
47
273
24.7K
Sudo su
Sudo su@sudoingX·
@signulll this is exactly what running local models teaches you. you stop writing code and start evaluating it. the model outputs, you judge. the faster you can spot what's off the faster you ship.
English
0
0
0
19
signüll
signüll@signulll·
with ai increasingly writing more & more code, engineers shift from makers to critics. taste, judgment, & the ability to recognize when something is wrong without being able to immediately articulate why is what compounds now more than ever before. i.e. the terminal skill is aesthetic discernment applied to large systems, which was always rare as hell & is now the only scarce thing.
English
21
4
66
3.9K
Sudo su
Sudo su@sudoingX·
@xeraphims so openai already proved the calc tool approach works. they just over optimized the trigger. the tool itself was the right call. now the open source local stack needs the same tool without the reward hacking.
English
0
0
1
28
Sudo su
Sudo su@sudoingX·
thinking out loud. every model gets math wrong. 7B, 9B, 70B. doesn't matter. pattern matching is not computation. hermes agent has code_execution which spins up a full python sandbox with RPC over unix sockets. powerful but heavy. a 9B isn't going to navigate that reliably for basic arithmetic. what if there was a lightweight calc tool built in. model hits a math question, calls the tool, gets the exact answer computed on your hardware. no interpreter overhead. sandboxed. simple enough schema that a 9B can call it every time. the accuracy problem stops being a model problem and becomes an infrastructure problem. and infrastructure is solvable. @Teknium would this belong in hermes agent or is code_execution enough?
English
11
0
45
1.4K
Sudo su
Sudo su@sudoingX·
@yaboilyrical how do i get my hands on one of those. shipping to bangkok is worth it for hermes merch.
English
0
0
0
9
Sudo su
Sudo su@sudoingX·
@uttertard the language doesn't matter much. the key is the schema the model sees. one field, expression in, answer out. whether the backend is JS, python, or raw C the model just needs to output "847 * 293" and get the right number back.
English
1
0
0
82
uttertard
uttertard@uttertard·
@sudoingX If you want to skip python would a calculator built in javascript and passed as a skill make sense?
English
1
0
0
102
Sudo su
Sudo su@sudoingX·
@startupideaspod that's a lot of duct tape for a problem hermes agent solved at the framework level. persistent memory, session search, daily context. no manual setup. you deserve better tools
English
0
0
0
17
The Startup Ideas Podcast (SIP) 🧃
"Why does my OpenClaw forget everything?" Because nothing was saved in the first place. Here's the 3-layer memory fix: memory.md: - Your agent's long-term brain. - High-level learnings, preferences, insights. - If this file doesn't exist yet, tell your agent to create it. Daily memory folder: - Granular logs created every day. - More detailed than memory.md. - This is where session-level context lives. Compaction flush: - Before your agent summarizes and compresses a long session, force it to write everything to memory first. - Otherwise context gets lost when the window fills up. Then add a 30-minute auto-save heartbeat: - Check if today's memory file exists - Create it if missing - Log a summary of the current session Fix your memory system before you touch anything else. That's where it clicks.
GREG ISENBERG@gregisenberg

THE ULTIMATE GUIDE TO OPENCLAW (1hr free masterclass) 1. fix memory so it compounds add MEMORY.md + daily logs. instruct it to promote important learnings into MEMORY.md because this is what makes it improve over time 2. set up personalization early identity.md, user.md, soul.md. write these properly or everything feels generic. this is what makes it sound like you and understand your world 3. structure your workspace properly most setups break because the foundation is messy. folders, files, and roles need to be clean or everything downstream degrades 4. create a troubleshooting baseline make a separate claude/chatgpt project just for openclaw. download the openclaw docs (context7) and load them in. when things break, it checks docs instead of guessing this alone fixes most issues!! 5. configure models and fallbacks set primary model to GPT 5.4 and add fallbacks across providers. this is what keeps tasks running instead of failing mid-way 6. turn repeat work into skills install summarize skill early. anything you do 2–3 times → turn into a skill. this is how it starts executing real workflows 7. connect tools with clear rules add browser + search (brave api). use managed browser for automation. use chrome relay only when login is neededthis avoids flaky behavior 8. use heartbeat to keep it alive add rules to check memory + cron healthif jobs are stale, force-run themthis prevents silent failures 9. use cron to schedule real work set daily and weekly tasksreports, follow-ups, content workflowsthis is where it starts acting without you 10. lock down security properly move secrets to a separate env file outside workspace. set strict permissions (folder 700, file 600). use allowlists for telegram access. don’t expose your gateway publicly 11. understand what openclaw actually is it’s a system that remembers, acts, and improves. basically, closer to an employee than a tool this ep of @startupideaspod is now out w/ @moritzkremb it's literally a full 1hr free course to take you from from “i installed openclaw”to “this thing is actually working for me” most people are one step away from openclaw working they installed it, they tried it and it didn’t click this ep will make it click all free, no advertisers, i just want to see you build your ideas with ideas with this ultimate guide to openclaw watch

English
7
8
77
7.9K
Sudo su
Sudo su@sudoingX·
@schinsly for sure, on a capable model you can ask and it handles it. the gap shows up when you're running 7B-14B on consumer hardware. those models call tools reliably but can't generate correct python consistently. that's who this is for.
English
0
0
1
31
Schinsly
Schinsly@schinsly·
@sudoingX i guess yeah it hasnt been done before but like i could oneshot that by just asking my agent if it was something i needed
English
2
0
1
33
Sudo su
Sudo su@sudoingX·
@DasMarky99 exactly. one tool, one field, model outputs the expression, hardware computes the answer. that's the whole idea.
English
0
0
0
86
Matu
Matu@DasMarky99·
@sudoingX Wouldn't be enough to expose a new "calc" function to the llm ?
English
1
0
0
96
Sudo su
Sudo su@sudoingX·
you're right, the pieces exist. the question is whether a 9B can use them reliably. code_execution needs the model to generate valid python with correct syntax, imports, and print statements. a calc tool with a one field schema just needs the model to output "847 * 293". the tool computes the result. same math, completely different reliability at 7B-14B scale.
English
2
0
2
159
Schinsly
Schinsly@schinsly·
@sudoingX this isn't super novel imo. the agent can literally just do math in console manually, follow a skill, or call a cli.
English
1
0
0
199
Sudo su
Sudo su@sudoingX·
@drewsky1 can't say which yet. but nothing was removed. only added.
English
0
0
0
13
Sudo su
Sudo su@sudoingX·
What have I done. holly shit this is magic a literal magic.
English
5
0
51
4.1K
Sudo su
Sudo su@sudoingX·
@ArbitorofOZ the greater good is making it so accessible that nobody needs permission to use it. open source everything. let the magic spread.
English
1
1
8
202
TradeVet
TradeVet@ArbitorofOZ·
@sudoingX I feel this so hard. Each of your replies to others also what I’m going through. Looking forward to reading about the good works you will accomplish with the magic you have discovered. My only question, do you feel the urge to apply it for the greater good yet?
English
1
0
0
224
Sudo su
Sudo su@sudoingX·
they are keeping me so compute constrained. I would have double gpus, double ram but hardware prices said no. for now.
English
4
0
23
1.1K
Sudo su
Sudo su@sudoingX·
@DasMarky99 some things stay in a little longer.
English
0
0
3
407
DustRoll
DustRoll@dustrollapp·
@sudoingX this is literally psyop magic
English
1
0
1
533
Sudo su
Sudo su@sudoingX·
reward the right behavior long enough and anything learns. RL is just the universe's update rule. good morning btw, lets fucking go..🔥
English
0
0
4
469
Sudo su
Sudo su@sudoingX·
every morning i wake up and ask myself what can't be solved with reinforcement learning. still waiting for an answer.
English
2
0
18
969