Major FAFO

2.4K posts

Major FAFO banner
Major FAFO

Major FAFO

@MajorFAFO

Texas, USA เข้าร่วม Mart 2017
402 กำลังติดตาม148 ผู้ติดตาม
Elon Musk
Elon Musk@elonmusk·
✝️ Happy Easter! ✝️
English
18.8K
32.8K
475.9K
42M
Major FAFO
Major FAFO@MajorFAFO·
27b still King an Gemma 4 31b a Queen, test them both, start with ggml or unsloth quants, some other releases may not be so accurate. I think the best way to use 2x3090's is to use the NvLink connector, if you find one. pci-e will kill your bandwidth between them, risers: massacre.
English
0
0
0
62
Lotto
Lotto@LottoLabs·
Im doing the typical gamer setup to local inference pipeline rn Learn from my upcoming posts What are pcie lanes, what’s x16/x8x/4, how slow is 2x3090 on prosumer mobos, all coming up. Just to see if qwen 27b is still the king.
English
13
1
81
7.2K
Major FAFO รีทวีตแล้ว
Donald J. Trump
Donald J. Trump@realDonaldTrump·
I wish everyone, including the haters and losers, a very happy Easter!
Palm Beach, FL 🇺🇸 English
5K
61.7K
207.7K
0
Major FAFO
Major FAFO@MajorFAFO·
Stop normalizing saying "Happy Easter" sounds pretty harmless to me. Better to say "Peaceful Easter" - this is a powerful message. Thank you for your attention to this matter!
English
0
0
0
32
Secretary of War Pete Hegseth
The tomb is empty. The promise is fulfilled. Through His sacrifice, we are redeemed. We stand firm in faith, courage, and truth—Happy Easter.
Secretary of War Pete Hegseth tweet media
English
12K
17.2K
94.7K
1.3M
Major FAFO รีทวีตแล้ว
Secretary Marco Rubio
Secretary Marco Rubio@SecRubio·
He is Risen.
English
10.1K
38.6K
205.3K
4.2M
BridgeMind
BridgeMind@bridgemindai·
Claude Code rate limited me so hard I bought a $5,000 NVIDIA DGX Spark. Arriving tomorrow. A personal AI supercomputer. Anthropic cut off OpenClaw users. Slashed Claude Opus 4.6 rate limits. Told $200/month Max plan customers to use less. Then gave us a credit as an apology. This is what happens when AI companies have too much power over your workflow. One update and your entire stack breaks. Local models are the only infrastructure no one can throttle. No rate limits. No 529 errors. No surprise policy changes. Tomorrow I'm testing the DGX Spark live on stream. Running local models through real vibe coding workflows. The goal is simple. Never depend on a single provider again.
BridgeMind tweet media
English
358
94
2.1K
313.4K
Lotto
Lotto@LottoLabs·
Bros pcie4 x4 is bad
Português
9
0
22
2.2K
Nikita Bier
Nikita Bier@nikitabier·
@nypost Can I say something without everyone getting mad
English
654
77
5.8K
586.4K
New York Post
New York Post@nypost·
Notorious Gen. Soleimani's sultry grandniece led lavish lifestyle touring US hotspots, as her mom promoted Iranian regime trib.al/y38evjw
New York Post tweet media
English
1.2K
3.3K
15K
3.3M
Major FAFO
Major FAFO@MajorFAFO·
@loktar00 llama.cpp is great with its webUI, only needs a HF model downloader and a simple UI to set loading parameters
English
0
0
0
36
Loktar 🇺🇸
Loktar 🇺🇸@loktar00·
HuggingFace building a local model provisioner on top of llama.cpp.... this is going to make Ollama completely irrelevant. Harbor already does this but having HF behind it means model compatibility day one.
English
2
0
28
1.7K
Major FAFO
Major FAFO@MajorFAFO·
@LottoLabs Check with nvlink and report back please, should keep the speed similar while doubling the VRAM
English
0
0
0
11
Ahmad
Ahmad@TheAhmadOsman·
- Gemma 4 26B-A4B > Qwen 3.5 35B-A3B - Qwen 3.5 27B > Gemma 4 31B
CY
47
34
741
55.9K
Maziyar PANAHI
Maziyar PANAHI@MaziyarPanahi·
Gemma 4 watches raw video. Understands the scene. Then prompts SAM 3 to segment and RF-DETR to track. One AI directing two others. Fighter jets. Crowds. Aerial defense footage. All three models running locally on a MacBook. No cloud. What scene should I point this at next?
English
80
152
2.5K
270.8K
stevibe
stevibe@stevibe·
Qwen3.5 27B vs Gemma4 31B | Canvas Creativity Test Why HTML Canvas? Two reasons: 1. It's unforgiving, one small mistake and the whole thing breaks 2. We kept prompts short to test real creativity, not instruction following 4 rounds: - Analog Clock - Hyperspace Tunnel - Growing Tree - Black Hole Both nailed the clock, but the other three is where it gets interesting. Looking forward to Qwen3.6 open-weight release!
English
63
104
1.5K
164.6K
Sudo su
Sudo su@sudoingX·
wooo grok is different now
English
9
0
28
7.1K
Major FAFO
Major FAFO@MajorFAFO·
@julien_c Qwen3.5-27b as a good all-a rounder, reasons well don't go below Q4, uses about 350W on 3090, Gemma4-31B-it is fine but trained for benchmaxing, it memorised the answers, then fails if you change variables
English
0
0
8
2K
Julien Chaumond
Julien Chaumond@julien_c·
so…. Qwen3.5 or Gemma 4?
Indonesia
204
19
876
193.2K
Major FAFO
Major FAFO@MajorFAFO·
@hallerite you're right, running gemma 4 on llama.cpp CPU better speed on CPU than GPU (6 vs 1.5 tok/s) in VRAM restricted scenario. I remember long time ago @ggerganov said all you need is CPU for inference, I think he saw it early
Major FAFO tweet media
English
2
1
37
10.5K
hallerite
hallerite@hallerite·
you can tell that gemma4 was not trained on nvidia gpus
English
30
9
924
176.3K