Olivier Chafik
877 posts

Olivier Chafik
@ochafik
Work @ Anthropic on MCP (views expressed = my own), ex-Google, past contrib. to OpenSCAD & llama.cpp; he/him 🏳️🌈 @ochafik.bsky.social @[email protected]








Your work tools are now interactive in Claude. Draft Slack messages, visualize ideas as Figma diagrams, or build and see Asana timelines.





Your work tools are now interactive in Claude. Draft Slack messages, visualize ideas as Figma diagrams, or build and see Asana timelines.







You really can just do things! Use *any* Hugging Face space as a MCP server along with your Local Models! 🔥 Here in we use Qwen 3 30B A3B with @ggml_org llama.cpp and @huggingface tiny agents to create images via FLUX powered by ZeroGPU ⚡ It's quite a bit crazy to see local models be capable of so much and just be able to understand/ infer from tool description! There's a lot of potential here in automating video generation workflows, content curation and a lot more.. Bonus: you can plug any other Inference Provider if you don't want to run locally too! npx @ huggingface/tiny-agents run [TASK] oh, and we provide both typescript and python client! 🐐

biggest takeaway of all is Qwen 3 30B A3B is slept on and you should be playing around with MCP if you haven't already!

Wanna disable thinking in llama.cpp? Try the new `--reasoning-budget 0` flag github.com/ggml-org/llama… Should work w/ Qwen3, QwQ, DeepSeek R1 distills, Command R7B; please report any issues! (Upcoming per-request behaviour discussed on github.com/ggml-org/llama… @ngxson) #llamacpp




