

Wolfbane
10.6K posts

@zzddfge
informático de nacimiento, escritor vocacional y abrumado por el Big Data.




Ollama is now updated to run the fastest on Apple silicon, powered by MLX, Apple's machine learning framework. This change unlocks much faster performance to accelerate demanding work on macOS: - Personal assistants like OpenClaw - Coding agents like Claude Code, OpenCode, or Codex







Sabéis que habitualmente soy del team optimista. Huyo de visiones apocalípticas y normalmente conservo bien la calma bajo presión. Pues empiezo a estar sinceramente preocupada.


Introducing TurboQuant: Our new compression algorithm that reduces LLM key-value cache memory by at least 6x and delivers up to 8x speedup, all with zero accuracy loss, redefining AI efficiency. Read the blog to learn how it achieves these results: goo.gle/4bsq2qI



