Major FAFO
2.4K posts

Major FAFO retweetledi
Major FAFO retweetledi

Claude Code rate limited me so hard I bought a $5,000 NVIDIA DGX Spark.
Arriving tomorrow. A personal AI supercomputer.
Anthropic cut off OpenClaw users.
Slashed Claude Opus 4.6 rate limits.
Told $200/month Max plan customers to use less.
Then gave us a credit as an apology.
This is what happens when AI companies have too much power over your workflow.
One update and your entire stack breaks.
Local models are the only infrastructure no one can throttle.
No rate limits. No 529 errors.
No surprise policy changes.
Tomorrow I'm testing the DGX Spark live on stream.
Running local models through real vibe coding workflows.
The goal is simple.
Never depend on a single provider again.

English

@eracoon @YourAnonOne beat me to it, I would have farted my way out of this
English

@nikitabier @nypost they probably were right, bloody bastards may be on to something after all
English

@nypost Can I say something without everyone getting mad
English

Notorious Gen. Soleimani's sultry grandniece led lavish lifestyle touring US hotspots, as her mom promoted Iranian regime trib.al/y38evjw

English

@loktar00 llama.cpp is great with its webUI, only needs a HF model downloader and a simple UI to set loading parameters
English

@LottoLabs Check with nvlink and report back please, should keep the speed similar while doubling the VRAM
English

I just bought another 3090 don’t listen to this guy
Lotto@LottoLabs
2x used 5060ti might be better than 1x3090
English

Qwen3.5 27B vs Gemma4 31B | Canvas Creativity Test
Why HTML Canvas? Two reasons:
1. It's unforgiving, one small mistake and the whole thing breaks
2. We kept prompts short to test real creativity, not instruction following
4 rounds:
- Analog Clock
- Hyperspace Tunnel
- Growing Tree
- Black Hole
Both nailed the clock, but the other three is where it gets interesting.
Looking forward to Qwen3.6 open-weight release!
English

@julien_c Qwen3.5-27b as a good all-a rounder, reasons well don't go below Q4, uses about 350W on 3090, Gemma4-31B-it is fine but trained for benchmaxing, it memorised the answers, then fails if you change variables
English


@hallerite you're right, running gemma 4 on llama.cpp CPU better speed on CPU than GPU (6 vs 1.5 tok/s) in VRAM restricted scenario. I remember long time ago @ggerganov said all you need is CPU for inference, I think he saw it early

English















