Dumbo
63 posts


Real talk — Apple Silicon optimization is moving fast, but the compute gap with NVIDIA? Still massive. You can optimize all you want, the hardware ceiling is the hardware ceiling. Maxed-out Apple still doesn't touch NVIDIA for local LLM stuff.
And dense models like Gemma 4 31B and Qwen 3.5 27B absolutely destroy their MoE variants in inference. Raw GPU power matters more than ever.
So yeah, NVIDIA GPUs = still the move for serious local LLM work. Don't bet on Apple closing that gap anytime soon.
…anyway I ordered a 128GB M5 Max on April 5th 🤡
English




















