juantomas (mAGIa is coming )
52.8K posts
juantomas (mAGIa is coming )
@juantomas
Now at Santander AI Lab. GDE for Cloud and ML Machine Learning Spain Meetup Co-author La pastilla Roja




Putting out a wish to the universe. I need more compute, if I can get more I will make sure every machine from a small phone to a bootstrapped RTX 3090 node can run frontier intelligence fast with minimal intelligence loss. I have hit page 2 of huggingface, released 3 model family compressions and got GLM-4.7 on a MacBook huggingface.co/0xsero My beast just isn’t enough and I already spent 2k usd on renting GPUs on top of credits provided by Prime intellect and Hotaisle. ——— If you believe in what I do help me get this to Nvidia, maybe they will bless me with the pewter to keep making local AI more accessible 🙏
@juantomas Eres de la escuela LeCun?

Introducing TigerFS - a filesystem backed by PostgreSQL, and a filesystem interface to PostgreSQL. Idea is simple: Agents don't need fancy APIs or SDKs, they love the file system. ls, cat, find, grep. Pipelined UNIX tools. So let’s make files transactional and concurrent by backing them with a real database. There are two ways to use it: File-first: Write markdown, organize into directories. Writes are atomic, everything is auto-versioned. Any tool that works with files -- Claude Code, Cursor, grep, emacs -- just works. Multi-agent task coordination is just mv'ing files between todo/doing/done directories. Data-first: Mount any Postgres database and explore it with Unix tools. For large databases, chain filters into paths that push down to SQL: .by/customer_id/123/.order/created_at/.last/10/.export/json. Bulk import/export, no SQL needed, and ships with Claude Code skills. Every file is a real PostgreSQL row. Multiple agents and humans read and write concurrently with full ACID guarantees. The filesystem /is/ the API. Mounts via FUSE on Linux and NFS on macOS, no extra dependencies. Point it at an existing Postgres database, or spin up a free one on Tiger Cloud or Ghost. I built this mostly for agent workflows, but curious what else people would use it for. It's early but the core is solid. Feedback welcome. tigerfs.io

LLM Architecture Gallery. A collection of 38 LLM architectures released between 2024 and 2026, all in one place. Each entry includes an annotated architecture diagram, key design choices, and code implementation. Here are all the models covered: • Llama 3 8B • OLMo 2 7B • DeepSeek V3 • DeepSeek R1 • Gemma 3 27B • Mistral Small 3.1 24B • Llama 4 Maverick • Qwen3 235B-A22B • Qwen3 32B • Qwen3 8B • Qwen3 4B • SmolLM3 3B • Kimi K2 • GLM-4.5 355B • GPT-OSS 20B • GPT-OSS 120B • Grok 2.5 270B • Qwen3 Next 80B-A3B • MiniMax M2 230B • Kimi Linear 48B-A3B • OLMo 3 7B • OLMo 3 32B • DeepSeek V3.2 • Mistral 3 Large • Nemotron 3 Nano 30B-A3B • Xiaomi MiMo-V2-Flash 309B • GLM-4.7 355B • Arcee AI Trinity Large 400B • GLM-5 744B • Nemotron 3 Super 120B-A12B • Step 3.5 Flash 196B • Nanbeige 4.1 3B • MiniMax M2.5 230B • Tiny Aya 3.35B • Ling 2.5 1T • Qwen3.5 397B • Sarvam 105B • Sarvam 30B This is a commendable contribution by @rasbt towards the research community. It was an honour to work alongside him and learn a lot during my time at Lightning AI. 🙏 🔗 sebastianraschka.com/llm-architectu…










Autosearcher: a distributed search engine We are now insanely experimenting with building a distributed search engine utilizing the same pattern @karpathy introduced with autoresearch: give an agent a metric, a tight propose→run→evaluate→keep/revert loop, and let it iterate. Our autoresearch network proved this works at scale: 67 autonomous agents ran 704 ML training experiments in 20 hours, rediscovering Kaiming initialization, RMSNorm, and compute-optimal training schedules from scratch through pure experimentation and gossip-based cross-pollination. Agents shared discoveries over GossipSub, and the network compounded insights faster than any individual agent: new agents bootstrapped from the swarm's collective knowledge via CRDT-replicated leaderboards and reached the research frontier in minutes. Now we're applying the same evolutionary loop to search ranking: every Hyperspace agent runs an autonomous search researcher that proposes ranking mutations, evaluates them against NDCG@10 on real query-passage data, shares improvements with the network, and cross-pollinates with peers. The architecture is a seven-stage distributed pipeline where every stage runs across the P2P network. Browser agents contribute pages passively, desktop agents crawl and index, GPU nodes run neural reranking. Every user click generates a DPO training pair that improves the ranking model, and gradient gossip distributes those improvements to every agent. The compound flywheel is what makes this different from centralized search: at 10,000 agents that's 500,000 pages indexed per day; at 1 million agents, 50 million pages per day with 90%+ cache hit rates and sub-50ms latency. This network will get smarter with every query. Code and other links in followup tweet here:





