Major FAFO

2.4K posts

Major FAFO banner
Major FAFO

Major FAFO

@MajorFAFO

Texas, USA Beigetreten Mart 2017
402 Folgt148 Follower
Major FAFO
Major FAFO@MajorFAFO·
@kaiostephens Sounds interesting. Isn’t Intel already made new CPU’s that work by default with DDR5 8000mt/s. I guess this idea is only for MOE models as these do not require high RAM bandwidth as opposed to dense models
English
0
0
0
371
kaios
kaios@kaiostephens·
hold on. You can overclock dual-channel DDR5 RAM to 10000mt/s+ with a good heatsink + fan That’s almost 200gb/s in memory bandwidth. That can run Gemma4 26B A4B at ~70tok/s decode. $300 of RAM (and $150 CPU) vs $3000 Ryzen AI Max 395+? has anyone tried this?
kaios tweet media
English
15
6
124
13.7K
Elon Musk
Elon Musk@elonmusk·
✝️ Happy Easter! ✝️
English
20.3K
38.3K
568.8K
62.2M
Major FAFO
Major FAFO@MajorFAFO·
27b still King an Gemma 4 31b a Queen, test them both, start with ggml or unsloth quants, some other releases may not be so accurate. I think the best way to use 2x3090's is to use the NvLink connector, if you find one. pci-e will kill your bandwidth between them, risers: massacre.
English
0
0
0
67
Lotto
Lotto@LottoLabs·
Im doing the typical gamer setup to local inference pipeline rn Learn from my upcoming posts What are pcie lanes, what’s x16/x8x/4, how slow is 2x3090 on prosumer mobos, all coming up. Just to see if qwen 27b is still the king.
English
13
1
87
7.6K
Major FAFO retweetet
Donald J. Trump
Donald J. Trump@realDonaldTrump·
I wish everyone, including the haters and losers, a very happy Easter!
Palm Beach, FL 🇺🇸 English
5.1K
61.9K
209.4K
0
Major FAFO
Major FAFO@MajorFAFO·
Stop normalizing saying "Happy Easter" sounds pretty harmless to me. Better to say "Peaceful Easter" - this is a powerful message. Thank you for your attention to this matter!
English
0
0
0
32
Secretary of War Pete Hegseth
The tomb is empty. The promise is fulfilled. Through His sacrifice, we are redeemed. We stand firm in faith, courage, and truth—Happy Easter.
Secretary of War Pete Hegseth tweet media
English
13.2K
18.2K
101K
1.5M
BridgeMind
BridgeMind@bridgemindai·
Claude Code rate limited me so hard I bought a $5,000 NVIDIA DGX Spark. Arriving tomorrow. A personal AI supercomputer. Anthropic cut off OpenClaw users. Slashed Claude Opus 4.6 rate limits. Told $200/month Max plan customers to use less. Then gave us a credit as an apology. This is what happens when AI companies have too much power over your workflow. One update and your entire stack breaks. Local models are the only infrastructure no one can throttle. No rate limits. No 529 errors. No surprise policy changes. Tomorrow I'm testing the DGX Spark live on stream. Running local models through real vibe coding workflows. The goal is simple. Never depend on a single provider again.
BridgeMind tweet media
English
384
103
2.2K
427K
Lotto
Lotto@LottoLabs·
Bros pcie4 x4 is bad
Português
9
0
23
2.4K
Nikita Bier
Nikita Bier@nikitabier·
@nypost Can I say something without everyone getting mad
English
655
79
5.8K
598.9K
New York Post
New York Post@nypost·
Notorious Gen. Soleimani's sultry grandniece led lavish lifestyle touring US hotspots, as her mom promoted Iranian regime trib.al/y38evjw
New York Post tweet media
English
1.2K
3.3K
15.1K
3.4M
Major FAFO
Major FAFO@MajorFAFO·
@loktar00 llama.cpp is great with its webUI, only needs a HF model downloader and a simple UI to set loading parameters
English
0
0
0
36
Loktar 🇺🇸
Loktar 🇺🇸@loktar00·
HuggingFace building a local model provisioner on top of llama.cpp.... this is going to make Ollama completely irrelevant. Harbor already does this but having HF behind it means model compatibility day one.
English
2
0
28
1.7K
Major FAFO
Major FAFO@MajorFAFO·
@LottoLabs Check with nvlink and report back please, should keep the speed similar while doubling the VRAM
English
0
0
0
12
Ahmad
Ahmad@TheAhmadOsman·
- Gemma 4 26B-A4B > Qwen 3.5 35B-A3B - Qwen 3.5 27B > Gemma 4 31B
CY
48
35
747
57K
Maziyar PANAHI
Maziyar PANAHI@MaziyarPanahi·
Gemma 4 watches raw video. Understands the scene. Then prompts SAM 3 to segment and RF-DETR to track. One AI directing two others. Fighter jets. Crowds. Aerial defense footage. All three models running locally on a MacBook. No cloud. What scene should I point this at next?
English
95
170
2.8K
312.3K
Major FAFO
Major FAFO@MajorFAFO·
@stevibe Ere you performing the tests in BF16?
English
1
0
0
1.2K
stevibe
stevibe@stevibe·
Qwen3.5 27B vs Gemma4 31B | Canvas Creativity Test Why HTML Canvas? Two reasons: 1. It's unforgiving, one small mistake and the whole thing breaks 2. We kept prompts short to test real creativity, not instruction following 4 rounds: - Analog Clock - Hyperspace Tunnel - Growing Tree - Black Hole Both nailed the clock, but the other three is where it gets interesting. Looking forward to Qwen3.6 open-weight release!
English
63
105
1.5K
167K
Sudo su
Sudo su@sudoingX·
wooo grok is different now
English
9
0
28
7.2K
Major FAFO
Major FAFO@MajorFAFO·
@julien_c Qwen3.5-27b as a good all-a rounder, reasons well don't go below Q4, uses about 350W on 3090, Gemma4-31B-it is fine but trained for benchmaxing, it memorised the answers, then fails if you change variables
English
0
0
8
2K
Julien Chaumond
Julien Chaumond@julien_c·
so…. Qwen3.5 or Gemma 4?
Indonesia
204
19
882
195.5K