Major FAFO

2.4K posts

Major FAFO banner
Major FAFO

Major FAFO

@MajorFAFO

Texas, USA Katılım Mart 2017
402 Takip Edilen147 Takipçiler
Elon Musk
Elon Musk@elonmusk·
✝️ Happy Easter! ✝️
English
10.5K
12.1K
162.1K
7.2M
Major FAFO
Major FAFO@MajorFAFO·
27b still King an Gemma 4 31b a Queen, test them both, start with ggml or unsloth quants, some other releases may not be so accurate. I think the best way to use 2x3090's is to use the NvLink connector, if you find one. pci-e will kill your bandwidth between them, risers: massacre.
English
0
0
0
37
Lotto
Lotto@LottoLabs·
Im doing the typical gamer setup to local inference pipeline rn Learn from my upcoming posts What are pcie lanes, what’s x16/x8x/4, how slow is 2x3090 on prosumer mobos, all coming up. Just to see if qwen 27b is still the king.
English
13
1
68
4.6K
Major FAFO retweetledi
Donald J. Trump
Donald J. Trump@realDonaldTrump·
I wish everyone, including the haters and losers, a very happy Easter!
Palm Beach, FL 🇺🇸 English
4.9K
60.4K
195.3K
0
Major FAFO
Major FAFO@MajorFAFO·
Stop normalizing saying "Happy Easter" sounds pretty harmless to me. Better to say "Peaceful Easter" - this is a powerful message. Thank you for your attention to this matter!
English
0
0
0
32
Secretary of War Pete Hegseth
The tomb is empty. The promise is fulfilled. Through His sacrifice, we are redeemed. We stand firm in faith, courage, and truth—Happy Easter.
Secretary of War Pete Hegseth tweet media
English
8.8K
12.4K
65.8K
783K
BridgeMind
BridgeMind@bridgemindai·
Claude Code rate limited me so hard I bought a $5,000 NVIDIA DGX Spark. Arriving tomorrow. A personal AI supercomputer. Anthropic cut off OpenClaw users. Slashed Claude Opus 4.6 rate limits. Told $200/month Max plan customers to use less. Then gave us a credit as an apology. This is what happens when AI companies have too much power over your workflow. One update and your entire stack breaks. Local models are the only infrastructure no one can throttle. No rate limits. No 529 errors. No surprise policy changes. Tomorrow I'm testing the DGX Spark live on stream. Running local models through real vibe coding workflows. The goal is simple. Never depend on a single provider again.
BridgeMind tweet media
English
281
66
1.5K
100.9K
Anonymous
Anonymous@YourAnonOne·
An isolated astronaut in weightless can remain stuck forever if he cannot reach a surface to propel himself
English
999
273
9.2K
1.7M
Lotto
Lotto@LottoLabs·
Bros pcie4 x4 is bad
Português
7
0
18
1.5K
Nikita Bier
Nikita Bier@nikitabier·
@nypost Can I say something without everyone getting mad
English
648
74
5.6K
397.3K
New York Post
New York Post@nypost·
Notorious Gen. Soleimani's sultry grandniece led lavish lifestyle touring US hotspots, as her mom promoted Iranian regime trib.al/y38evjw
New York Post tweet media
English
1.1K
3.1K
14.3K
3M
Major FAFO
Major FAFO@MajorFAFO·
@loktar00 llama.cpp is great with its webUI, only needs a HF model downloader and a simple UI to set loading parameters
English
0
0
0
34
Loktar 🇺🇸
Loktar 🇺🇸@loktar00·
HuggingFace building a local model provisioner on top of llama.cpp.... this is going to make Ollama completely irrelevant. Harbor already does this but having HF behind it means model compatibility day one.
English
2
0
28
1.6K
Major FAFO
Major FAFO@MajorFAFO·
@LottoLabs Check with nvlink and report back please, should keep the speed similar while doubling the VRAM
English
0
0
0
11
Ahmad
Ahmad@TheAhmadOsman·
- Gemma 4 26B-A4B > Qwen 3.5 35B-A3B - Qwen 3.5 27B > Gemma 4 31B
CY
46
33
728
54.1K
Maziyar PANAHI
Maziyar PANAHI@MaziyarPanahi·
Gemma 4 watches raw video. Understands the scene. Then prompts SAM 3 to segment and RF-DETR to track. One AI directing two others. Fighter jets. Crowds. Aerial defense footage. All three models running locally on a MacBook. No cloud. What scene should I point this at next?
English
72
119
2K
212K
stevibe
stevibe@stevibe·
Qwen3.5 27B vs Gemma4 31B | Canvas Creativity Test Why HTML Canvas? Two reasons: 1. It's unforgiving, one small mistake and the whole thing breaks 2. We kept prompts short to test real creativity, not instruction following 4 rounds: - Analog Clock - Hyperspace Tunnel - Growing Tree - Black Hole Both nailed the clock, but the other three is where it gets interesting. Looking forward to Qwen3.6 open-weight release!
English
63
101
1.4K
159.6K
Sudo su
Sudo su@sudoingX·
wooo grok is different now
English
9
0
27
6.9K
Major FAFO
Major FAFO@MajorFAFO·
@julien_c Qwen3.5-27b as a good all-a rounder, reasons well don't go below Q4, uses about 350W on 3090, Gemma4-31B-it is fine but trained for benchmaxing, it memorised the answers, then fails if you change variables
English
0
0
8
2K
Major FAFO
Major FAFO@MajorFAFO·
@hallerite you're right, running gemma 4 on llama.cpp CPU better speed on CPU than GPU (6 vs 1.5 tok/s) in VRAM restricted scenario. I remember long time ago @ggerganov said all you need is CPU for inference, I think he saw it early
Major FAFO tweet media
English
2
1
37
10.5K
hallerite
hallerite@hallerite·
you can tell that gemma4 was not trained on nvidia gpus
English
30
9
923
175.9K