void
4.3K posts


Introducing 𝐆𝐞𝐦𝐦𝐚 𝟒 𝟑𝟏𝐁 𝐓𝐮𝐫𝐛𝐨 ⚡️ It runs on a 𝘴𝘪𝘯𝘨𝘭𝘦 RTX 5090, at 51 tok/s (single) and 1244 tok/s (batched). And prefills up to 15359 tok/s. It's 𝟔𝟖% 𝐬𝐦𝐚𝐥𝐥𝐞𝐫 in GPU memory and ~𝟐.𝟓𝐱 𝐟𝐚𝐬𝐭𝐞𝐫 than the base model, and retains nearly 𝐢𝐝𝐞𝐧𝐭𝐢𝐜𝐚𝐥 𝐪𝐮𝐚𝐥𝐢𝐭𝐲 on benchmarks (1-3% loss). Turbo is a derivative of the NVFP4 quant that NVIDIA released a few days ago. It fully leverages NVIDIA Blackwell FP4 tensor cores for ~𝟐× 𝐡𝐢𝐠𝐡𝐞𝐫 𝐜𝐨𝐧𝐜𝐮𝐫𝐫𝐞𝐧𝐭 𝐭𝐡𝐫𝐨𝐮𝐠𝐡𝐩𝐮𝐭 𝐭𝐡𝐚𝐧 𝐨𝐭𝐡𝐞𝐫 𝐪𝐮𝐚𝐧𝐭𝐬. I'm using it for hard classification tasks — on internal benchmarks it showed 𝐒𝐨𝐧𝐧𝐞𝐭-𝟒.𝟓-𝐥𝐞𝐯𝐞𝐥 𝐢𝐧𝐭𝐞𝐥𝐥𝐢𝐠𝐞𝐧𝐜𝐞 (scored well above Haiku 4.5), at a 600𝘵𝘩 of the cost. A single RTX 5090 scales up to 18 req/s at 1000in/20out 🥵. Model card and benchmark in comments 👇 I'd love to hear your use cases.

Introducing Claude Managed Agents: everything you need to build and deploy agents at scale. It pairs an agent harness tuned for performance with production infrastructure, so you can go from prototype to launch in days. Now in public beta on the Claude Platform.

how to run claude code with gemma 4 completely free (beginner's guide): this guide shows you how to use claude code completely free with gemma 4, no subscriptions &no api keys. just your laptop + 15 mins setup. this lets you run open-source models (like google’s gemma) locally, meaning: — no costs — full privacy what you need before starting, make sure you have: vs code installed — node.js (version 18+) — stable internet (for one-time model download) _____________ step 1: install ollama (the engine) ollama is what runs ai models locally on your machine. → mac: go to ollama.com/download click download for mac, open the file and install like any normal app. no terminal needed. → windows: go to ollama.com/download, click download and install → linux: curl -fsSL ollama.com/install.sh | sh check it worked: ollama --version _____________ step 2: download gemma 4 this is the ai model you’ll run locally, pick based on your system: → low-end (8gb ram): ollama pull gemma4:e2b → recommended (16gb ram): ollama pull gemma4:e4b → high-end (32gb ram): ollama pull gemma4:26b ⚠️ it’s a big download (7gb–18gb), so give it time. after download is completed, verify with the command: ollama list _____________ step 3: install claude code in VS code or any other IDE this is your interface. — open vs code — press ctrl + shift + x — search claude code install the one by anthropic after install → you’ll see a ⚡ icon in sidebar _____________ step 4: connect claude code to ollama by default, claude connects to the cloud. we’re redirecting it to your local machine. so do this: — press ctrl + shift + p — search: open user settings (json) — then paste this inside: "claude-code.env": { "ANTHROPIC_BASE_URL": "http://localhost:11434", "ANTHROPIC_API_KEY": "", "ANTHROPIC_AUTH_TOKEN": "ollama" } what this does: — it routes everything to your local ollama server. — nothing leaves your device. _____________ step 5: run everything 1. start ollama with this command: ollama serve leave this running. 2. open claude code in vs code click ⚡ icon 3. select your model type: gemma4:e4b (or whichever you downloaded) you’re done _____________ you now have: — claude code running — powered by gemma 4 — fully local completely free try: “explain this file” “write a function” “refactor this code” _____________ common issues (quick fixes) “unable to connect” run: ollama serve asked to sign in your json config is wrong check for missing commas/brackets very slow responses your model is too big switch to: gemma4:e2b model not found run: ollama list copy exact name quick recap you just built: a free claude setup powered by local ai no api costs Follow for more AI contents like this!!!





Esse é o tipo de coisa que hater de Mac só vai sacar o dia que tiver um, são vários detalhes premium que você só sente usando no dia a dia e aí pensa “hmmm agora entendi pq falam tão bem”


New in Claude Code: auto mode. Instead of approving every file write and bash command, or skipping permissions entirely, auto mode lets Claude make permission decisions on your behalf. Safeguards check each action before it runs.

🚨 SAM ALTMAN: “We see a future where intelligence is a utility, like electricity or water, and people buy it from us on a meter.”





⚠️ ATENÇÃO MINEIROS Sabe o custo da reconstrução das cidades que foram arrebentadas ontem pelas chuvas? Vai sair do bolso de vocês. Agora, sabe quanto custa mitigação de danos? 15 vezes menos. Isso mesmo, estudos mostram que mitigar os danos reduzem em até 15 vezes os gastos com reconstrução. @RomeuZema não só não investiu em mitigação de danos como CORTOU VERBAS, mas sabe o que ele fez com o próprio salário né? Parem de ser idiotas, façam campanhas, se esforcem para impedir esse comedor de banana com casca de continuar destruído o estado de Minas. Meu pai era de Minas, eu adoro a cultura do estado, é uma pena ver a situação que esse incompetente está deixando um povo tão bem visto.











