

Sudo su
6.6K posts





this guy has 29 models on huggingface at page 2 ranking. no lab behind him. no sponsorship. $2,000 from his own pocket on GPU rentals. he compressed GLM-4.7 to run on a MacBook and quantized Nemotron Super the week it dropped. all public. all free. nvidia is a trillion dollar company with hundreds of teams but they are not the ones quantizing models middle of the night and pushing them out before sunrise. if nvidia stopped tomorrow their employees stop working. people like @0xSero would not. that is the difference between a paycheck and a mission. @NVIDIAAI you talk about making AI accessible. the people actually doing it are right here. 29 models deep burning their own compute with no ask except more hardware to keep going. you do not need to build another program. just look at who is already building for you. one GPU to this man would produce more public value than a hundred internal sprints. i am not asking for charity. i am asking you to invest in someone who already proved it.





Putting out a wish to the universe. I need more compute, if I can get more I will make sure every machine from a small phone to a bootstrapped RTX 3090 node can run frontier intelligence fast with minimal intelligence loss. I have hit page 2 of huggingface, released 3 model family compressions and got GLM-4.7 on a MacBook huggingface.co/0xsero My beast just isn’t enough and I already spent 2k usd on renting GPUs on top of credits provided by Prime intellect and Hotaisle. ——— If you believe in what I do help me get this to Nvidia, maybe they will bless me with the pewter to keep making local AI more accessible 🙏

@sudoingX Are you open to taking donations on the GitHub?


this is what 12 gigs of VRAM built in 2026. a 9 billion parameter model running on a 5 year old RTX 3060 wrote a full space shooter from a single prompt. blank screen on first try. i came back with a bug list and the same model on the same card fixed every issue across 11 files without touching a single line myself. enemies still looked wrong so i pushed another iteration and now the game has pixel art octopi, particle effects, screen shake, projectile physics and a combo system. all running locally on a card that was designed to play fortnite. three iterations. zero cloud. zero API calls. every token generated on hardware sitting under my desk. the model reads its own code, finds what's broken, patches it, validates syntax and restarts the server. i just describe what's wrong and it handles the rest. people are paying monthly subscriptions to type into a browser tab and wait for a server farm to respond. meanwhile a GPU you can find used on ebay is running a full autonomous hermes agent framework with 31 tools, 128K context window and thinking mode generating at 29 tokens per second nonstop. the game still needs work. level upgrades don't trigger and boss fights need tuning. but the fact that i'm iterating on gameplay balance instead of debugging whether the code runs at all tells you where this is headed. every iteration the game gets better on the same hardware. same 12 gigs. same 9 billion parameters. same RTX 3060 from 5 years ago your GPU is not a gaming card anymore. it's a local AI lab that never sends your data anywhere.






Are we a merch company or a publishing company?



THE ULTIMATE GUIDE TO OPENCLAW (1hr free masterclass) 1. fix memory so it compounds add MEMORY.md + daily logs. instruct it to promote important learnings into MEMORY.md because this is what makes it improve over time 2. set up personalization early identity.md, user.md, soul.md. write these properly or everything feels generic. this is what makes it sound like you and understand your world 3. structure your workspace properly most setups break because the foundation is messy. folders, files, and roles need to be clean or everything downstream degrades 4. create a troubleshooting baseline make a separate claude/chatgpt project just for openclaw. download the openclaw docs (context7) and load them in. when things break, it checks docs instead of guessing this alone fixes most issues!! 5. configure models and fallbacks set primary model to GPT 5.4 and add fallbacks across providers. this is what keeps tasks running instead of failing mid-way 6. turn repeat work into skills install summarize skill early. anything you do 2–3 times → turn into a skill. this is how it starts executing real workflows 7. connect tools with clear rules add browser + search (brave api). use managed browser for automation. use chrome relay only when login is neededthis avoids flaky behavior 8. use heartbeat to keep it alive add rules to check memory + cron healthif jobs are stale, force-run themthis prevents silent failures 9. use cron to schedule real work set daily and weekly tasksreports, follow-ups, content workflowsthis is where it starts acting without you 10. lock down security properly move secrets to a separate env file outside workspace. set strict permissions (folder 700, file 600). use allowlists for telegram access. don’t expose your gateway publicly 11. understand what openclaw actually is it’s a system that remembers, acts, and improves. basically, closer to an employee than a tool this ep of @startupideaspod is now out w/ @moritzkremb it's literally a full 1hr free course to take you from from “i installed openclaw”to “this thing is actually working for me” most people are one step away from openclaw working they installed it, they tried it and it didn’t click this ep will make it click all free, no advertisers, i just want to see you build your ideas with ideas with this ultimate guide to openclaw watch

