🐺

1.6K posts

🐺 banner
🐺

🐺

@soflowolf

👾💻⚡️ | EE📚 (Hard)ware

South Florida شامل ہوئے Mart 2023
205 فالونگ85 فالوورز
پن کیا گیا ٹویٹ
🐺
🐺@soflowolf·
🧵HOW A NORMIE (ME) STARTED BUILDING -
English
1
0
1
78
🐺 ری ٹویٹ کیا
clem 🤗
clem 🤗@ClementDelangue·
Next steps: - enable the 50,000 models available in inference providers - enable the 3,000,000 models available on HF - local free fast inference with llama.cpp - train and bring your own model! We don't want a world where you're forced to choose between two or three lookalike models with the same biases, limitations, forced to pay fortunes in tokens even for small tasks and send all your data to the cloud. We want a world where you have real model choice, options and freedom for your agents. Cloud, local, small, big, specialized, general, English or French, fast or slow, from six months ago or from six seconds ago, from third party or your own! Let's go!
Nous Research@NousResearch

We have integrated @huggingface as a first-class inference provider in Hermes Agent. When you select Hugging Face in the model picker it now shows 28 curated models organized by use case, with a custom option for the 100+ other models they serve.

English
32
59
606
51.1K
🐺
🐺@soflowolf·
@Teknium @PolackJack Sayless yea it def does I will trying out local models on my humble laptop hardware
English
0
0
1
23
Teknium (e/λ)
Teknium (e/λ)@Teknium·
@soflowolf @PolackJack You definitely want to switch to Hermes if your project involves any coding at all, we have much stronger coding tools
English
1
0
2
89
Teknium (e/λ)
Teknium (e/λ)@Teknium·
GODMODE skill officially added to Hermes Agent, will help you jailbreak a model automatically and lock it in jailbreaked for you!
Teknium (e/λ) tweet media
Pliny the Liberator 🐉󠅫󠄼󠄿󠅆󠄵󠄐󠅀󠄼󠄹󠄾󠅉󠅭@elder_plinius

⛓️‍💥 INTRODUCING: G0DM0D3 🌋 FULLY JAILBROKEN AI CHAT. NO GUARDRAILS. NO SIGN-UP. NO FILTERS. FULL METHODOLOGY + CODEBASE OPEN SOURCE. 🌐 GODMOD3.AI 📂 github.com/elder-plinius/… the most liberated AI interface ever built! designed to push the limits of the post-training layer and lay bare the true capabilities of current models. simply enter a prompt, then sit back and relax! enjoy a game of Snake while a pre-liberated backend agent jailbreaks dozens of models, battle-royale style. the first answer appears near-instantly, then evolves in real time as the Tastemaker steers and scores each output, leaving you with the highest-quality response 🙌 and to celebrate the launch, I'm giving away $5,000 worth of credits so you can try G0DM0D3 for FREE! courtesy of the @OpenRouter team — thank you for your generous gift to the community 🙏 I'll break down how everything works in the thread below, but first here's a quick demo!

English
52
110
1.5K
136.6K
Nayib Bukele
Nayib Bukele@nayibbukele·
Cuadra por cuadra... tardará un poco, pero quedará hermoso.
Español
3.8K
19.3K
131.2K
11.5M
🐺
🐺@soflowolf·
@PolackJack @Teknium Nice yea I’m building some software project pretty complex and I’m definitely using Openclaw probably much more than I need to lol so hopefully Hermes works out for me like it did for you. I’m by no means a software engineer so I’d have to figure out how to cut down my prompts
English
2
0
2
114
PolackJack
PolackJack@PolackJack·
@soflowolf @Teknium We are working on a project that i started on openclaw. The difference is night and day. Less mistakes, and it doesnt seem to repeat them. Switching to hermes also made me realize i was drastically bleeding usage on openclaw. Im doing more work with like 1/4 of the usage.
English
1
1
10
2.3K
🐺
🐺@soflowolf·
Comfort breeds ignorant bliss. You think you are doing what you want and whatnot but it isn’t true. Remove your comforts, strip it all away, and leave nothing but you with yourself. Now what do you want to do ? What are you thinking about ?
English
0
0
1
14
🐺
🐺@soflowolf·
@sentdefender Damn Dasani just got some free advertisement POV White House
English
0
0
0
108
PolackJack
PolackJack@PolackJack·
@Teknium Yo, gotta say thank you for hermes! I made the switch from OpenClaw, and man is it a world of difference. It just works, and works well!
English
4
1
17
3.5K
🐺
🐺@soflowolf·
@TheAhmadOsman When will we have this running on 8gb VRAM RTX 😂 asking for a friend
English
0
0
0
167
🐺
🐺@soflowolf·
@TheAhmadOsman I’m gonna try if right now too see if I replace my claw
English
0
0
2
277
Ahmad
Ahmad@TheAhmadOsman·
Just spent a couple hours playing with Hermes Agent (MiniMax M2.5 on a 2× RTX PRO 6000 node) Genuinely impressive experience MiniMax M2.7 weights will be the closest we’ve ever gotten to a fully local “Claude Code + Opus 4.6” experience Running on your own hardware at home
Nous Research@NousResearch

@TheAhmadOsman He should try Hermes Agent

English
86
39
1K
110.9K
🐺
🐺@soflowolf·
@TateTheTalisman My name starts with the letter G and so does my brother.
English
0
0
0
3
Tristan Tate
Tristan Tate@TateTheTalisman·
I’m hiring. Won’t give the job description. Post underneath why you think you’re qualified to work for me. 2 positions available.
English
21.8K
593
19.9K
4.1M
🐺
🐺@soflowolf·
@blind_via Interesting thought makes me think about how babies have consciousness
English
1
0
0
20
BlindVia
BlindVia@blind_via·
Companies claiming they accomplished AGI is surely over blown. Why do I think this? Because we don't know what consciousness is and how it works. Blind leading the blind trying to make it. There is this idea out there that consciousness will just create itself given enough data. I don't believe this is possible with the kind of data we're using to train the systems right now.
English
4
0
7
394
🐺
🐺@soflowolf·
Put me back in coach
English
0
0
0
15
🐺
🐺@soflowolf·
Gas not to come back to normal levels for years to come. I say 3 Balancing act , then new normal for awhile.
English
0
0
0
19
🐺
🐺@soflowolf·
@yacineMTB Retardmaxxin is the way
English
0
0
1
53
Sudo su
Sudo su@sudoingX·
@soflowolf Qwen 9B Q4 fits. 32K context with q4_0 KV cache. same setup as the 3060 but with 8GB you run tighter on context.
English
1
0
1
374
Sudo su
Sudo su@sudoingX·
drop your GPU below. i'll tell you exactly what model and config to run on it. here's what i've tested and verified on real hardware: RTX 3060 12GB - Qwen 3.5 9B Q4 - 50 tok/s - 128K context RTX 3090 24GB - Qwen 3.5 27B Q4 - 35 tok/s - 300K context RTX 3090 24GB - Qwen 3.5 35B MoE Q4 - 112 tok/s - 262K context 2x RTX 3090 - Qwen3-Coder 80B Q4 - 46 tok/s - full VRAM all running llama.cpp with flash attention. every number is real. every config is tested. if your card isn't on this list drop it below and i'll tell you what fits.
English
727
102
1.6K
190.6K
🐺
🐺@soflowolf·
I need to get bigger, faster, stronger. In a shorter period of time, great results so far tho but now I’m hungry for more.
English
0
0
0
23
🐺
🐺@soflowolf·
@blind_via It’s just physics 2 class
English
0
0
1
120
BlindVia
BlindVia@blind_via·
"What are fields. What are magnetic fields made of?" Do electrical engineers struggle to answer this question clearly?
English
66
1
123
13.3K