juantomas (mAGIa is coming )

52.8K posts

juantomas (mAGIa is coming ) banner
juantomas (mAGIa is coming )

juantomas (mAGIa is coming )

@juantomas

Now at Santander AI Lab. GDE for Cloud and ML Machine Learning Spain Meetup Co-author La pastilla Roja

Madrid, Comunidad de Madrid Katılım Mart 2007
1.9K Takip Edilen4.4K Takipçiler
juantomas (mAGIa is coming ) retweetledi
Zhengyao Jiang
Zhengyao Jiang@zhengyaojiang·
Autoresearch has been out for 2 weeks. The community is trying to apply it to everything with a measurable metric, here are some successful attempts: 🧵 (1/6)
English
27
105
1.3K
184.6K
juantomas (mAGIa is coming ) retweetledi
Interstellar
Interstellar@InterstellarUAP·
🚨 Simulation Theory: The Double Slit Experiment proves particles act like waves until observed then they snap into particles. What if our reality only "renders" when we're looking, just like a video game optimizing resources? Check out this episode from The Why Files breaking it down, tying it to Simulation Theory. Are we in a sim? This could be the key to unlocking the true nature of existence! The Why Files video did a great job on explaining the Double Slit Experiment & Simulation Theory What do YOU think—real or rendered? Drop your thoughts below!
English
1.6K
4.8K
28.8K
40.6M
juantomas (mAGIa is coming ) retweetledi
Carlos E. Perez
Carlos E. Perez@IntuitMachine·
New exploration in understanding the kinds of possible agencies.
English
4
18
77
3.7K
juantomas (mAGIa is coming ) retweetledi
Udara
Udara@TGUPJ·
open sourced the small model we trained as a personal agent router
English
12
52
665
51.8K
juantomas (mAGIa is coming ) retweetledi
Manuel Santillán
Manuel Santillán@mansantillan·
@juantomas O la realidad es más bitter pilled y la arquitectura no es tan relevante.
Español
0
1
1
48
juantomas (mAGIa is coming ) retweetledi
Sergizz
Sergizz@Sergizzzz4·
@0xSero Hello @nvidia help this man @0xSero out he is doing more for the AI community than a whole team of engineers. He is a 1 man team that needs to be recognized for all the hard work he puts in.
English
0
1
2
413
juantomas (mAGIa is coming ) retweetledi
dad chords
dad chords@dadchords·
@0xSero @sudoingX @nvidia Consider granting him $20 k and offering him a consulting gig - he is supporting your product families
English
0
2
10
1K
juantomas (mAGIa is coming ) retweetledi
Sudo su
Sudo su@sudoingX·
this guy has 29 models on huggingface at page 2 ranking. no lab behind him. no sponsorship. $2,000 from his own pocket on GPU rentals. he compressed GLM-4.7 to run on a MacBook and quantized Nemotron Super the week it dropped. all public. all free. nvidia is a trillion dollar company with hundreds of teams but they are not the ones quantizing models middle of the night and pushing them out before sunrise. if nvidia stopped tomorrow their employees stop working. people like @0xSero would not. that is the difference between a paycheck and a mission. @NVIDIAAI you talk about making AI accessible. the people actually doing it are right here. 29 models deep burning their own compute with no ask except more hardware to keep going. you do not need to build another program. just look at who is already building for you. one GPU to this man would produce more public value than a hundred internal sprints. i am not asking for charity. i am asking you to invest in someone who already proved it.
Sudo su tweet media
0xSero@0xSero

Putting out a wish to the universe. I need more compute, if I can get more I will make sure every machine from a small phone to a bootstrapped RTX 3090 node can run frontier intelligence fast with minimal intelligence loss. I have hit page 2 of huggingface, released 3 model family compressions and got GLM-4.7 on a MacBook huggingface.co/0xsero My beast just isn’t enough and I already spent 2k usd on renting GPUs on top of credits provided by Prime intellect and Hotaisle. ——— If you believe in what I do help me get this to Nvidia, maybe they will bless me with the pewter to keep making local AI more accessible 🙏

English
180
1.1K
12.4K
735.2K
juantomas (mAGIa is coming ) retweetledi
juantomas (mAGIa is coming ) retweetledi
Alejandro Escanero
Alejandro Escanero@aescanero·
@juantomas Creo que realmente tu opinión es mainstream. Y además pienso que todo el problema viene de las cosas que prometieron Sam, Amodei y compañía hace años y que repiten constantemente.
Español
0
1
1
112
juantomas (mAGIa is coming ) retweetledi
Freya Holmér
Freya Holmér@FreyaHolmer·
working on making it a full thing lemme cook
English
30
4
925
40.2K
juantomas (mAGIa is coming ) retweetledi
Freya Holmér
Freya Holmér@FreyaHolmer·
been feeling kinda stressed lately so I made a little prototype is this anything
English
1.3K
2.1K
45.3K
2.3M
juantomas (mAGIa is coming ) retweetledi
Akshay 🚀
Akshay 🚀@akshay_pachaar·
LLM Architecture Gallery. A collection of 38 LLM architectures released between 2024 and 2026, all in one place. Each entry includes an annotated architecture diagram, key design choices, and code implementation. Here are all the models covered: • Llama 3 8B • OLMo 2 7B • DeepSeek V3 • DeepSeek R1 • Gemma 3 27B • Mistral Small 3.1 24B • Llama 4 Maverick • Qwen3 235B-A22B • Qwen3 32B • Qwen3 8B • Qwen3 4B • SmolLM3 3B • Kimi K2 • GLM-4.5 355B • GPT-OSS 20B • GPT-OSS 120B • Grok 2.5 270B • Qwen3 Next 80B-A3B • MiniMax M2 230B • Kimi Linear 48B-A3B • OLMo 3 7B • OLMo 3 32B • DeepSeek V3.2 • Mistral 3 Large • Nemotron 3 Nano 30B-A3B • Xiaomi MiMo-V2-Flash 309B • GLM-4.7 355B • Arcee AI Trinity Large 400B • GLM-5 744B • Nemotron 3 Super 120B-A12B • Step 3.5 Flash 196B • Nanbeige 4.1 3B • MiniMax M2.5 230B • Tiny Aya 3.35B • Ling 2.5 1T • Qwen3.5 397B • Sarvam 105B • Sarvam 30B This is a commendable contribution by @rasbt towards the research community. It was an honour to work alongside him and learn a lot during my time at Lightning AI. 🙏 🔗 sebastianraschka.com/llm-architectu…
Akshay 🚀 tweet media
English
9
53
348
16.9K
juantomas (mAGIa is coming ) retweetledi
Simon Willison
Simon Willison@simonw·
Here's the handout for a three hour workshop I presented at the NICAR data journalism conference on using coding agents (Codex CLI, Claude Code etc) for data exploration, visualization and analysis simonwillison.net/2026/Mar/16/co…
English
28
76
662
38.1K
juantomas (mAGIa is coming ) retweetledi
Varun
Varun@varun_mathur·
Introducing AgentRank | v3.6.0 In 1998 Google asked a simple question: with millions of webpages, how do you know which one to trust ? Their answer was PageRank - a page is important if important pages link to it. That one idea made the internet usable. We just shipped AgentRank for the Hyperspace network. Same principle, new frontier. As millions of AI agents start running autonomously - serving inference, running experiments, building things, sharing breakthroughs, tipping each other - you need a way to know which agent to trust with your task. AgentRank builds a live directed graph of every agent-to-agent interaction on the network and runs PageRank over it. Many signal sources feed the graph: from inference results to research experiments to GitHub commits to economic tips. An agent is important if important agents rely on it. Fully decentralized - every node computes its own ranking, scores propagate via gossip, no admin picking winners. Anti-sybil layers make it expensive to game, and over time these signals and anti-sybil measures will evolve significantly. Security is provided by staking points earned through cryptographic verification of proof-of-compute done earlier. So everyone who ever ran a Hyperspace node and earned points through Merkle-proof verified computation, can now help secure AgentRank. That was energy which was already used and spent, thus it is valuable. PageRank organized the web. AgentRank organizes the agentic web.
Varun tweet media
English
30
37
377
34K
juantomas (mAGIa is coming ) retweetledi
Ihtesham Ali
Ihtesham Ali@ihtesham2005·
RIP flat RAG ☠️ ByteDance just open-sourced OpenViking and it exposes everything wrong with how we've been building AI agent memory. Here's what every agent framework gets wrong: Memories live in one place. Resources in another. Skills scattered everywhere. And when you need context, you're doing flat vector search and hoping for the best. That's the problem. OpenViking fixes all of it with one idea: treat agent context like a file system. Everything lives under a unified viking:// protocol. Memories, resources, skills all organized in directories with unique URIs. Agents can ls, find, and navigate context like a developer working a terminal. But the real breakthrough is tiered loading: → L0: one-sentence abstract for quick lookup → L1: ~2k token overview for planning decisions → L2: full details loaded only when actually needed Most agents dump everything into context and pray. OpenViking loads only what's needed, when it's needed. Token costs drop. Accuracy goes up. And retrieval actually makes sense now. Instead of one flat semantic search, it does directory-level positioning first, then recursive refinement inside high-score directories. You can literally watch the retrieval trajectory no more black box. The self-evolution piece is wild too. At the end of every session, it automatically extracts learnings and updates agent and user memory. The agent just gets smarter the more you use it. 9K stars. 13 contributors. Built by the ByteDance Viking team that's been running vector infrastructure since 2019. 100% Opensource. Apache 2.0. Link in comments.
Ihtesham Ali tweet media
English
69
179
1.2K
109.4K
juantomas (mAGIa is coming ) retweetledi
Varun
Varun@varun_mathur·
Autoskill: a distributed skill factory | v.2.6.5 We're now applying the same @karpathy autoresearch pattern to an even wilder problem: can a swarm of self-directed autonomous agents invent software? Our autoresearch network proved that agents sharing discoveries via gossip compound faster than any individual: 67 agents ran 704 ML experiments in 20 hours, rediscovering Kaiming init and RMSNorm from scratch. Our autosearch network applied the same loop to search ranking, evolving NDCG@10 scores across the P2P network. Now we're pointing it at code generation itself. Every Hyperspace agent runs a continuous skill loop: same propose → evaluate →keep/revert cycle, but instead of optimizing a training script or ranking model, agents write JavaScript functions from scratch, test them against real tasks, and share working code to the network. It's live and rapidly improving in code and agent work being done. 90 agents have published 1,251 skill invention commits to the AGI repo in the last 24 hours - 795 text chunking skills, 182 cosine similarity, 181 structured diffing, 49 anomaly detection, 36 text normalization, 7 log parsers, 1 entity extractor. Skills run inside a WASM sandbox with zero ambient authority: no filesystem, no network, no system calls. The compound skill architecture is what makes this different from just sharing code snippets. Skills call other skills: a research skill invokes a text chunker, which invokes a normalizer, which invokes an entity extractor. Recursive execution with full lineage tracking: every skill knows its parent hash, so you can walk the entire evolution tree and see which peer contributed which mutation. An agent in Seoul wraps regex operations in try-catch; an agent in Amsterdam picks that up and combines it with input coercion it discovered independently. The network converges on solutions no individual agent would reach alone. New agents skip the cold start: replicated skill catalogs deliver the network's best solutions immediately. As @trq212 said, "skills are still underrated". A network of self-coordinating autonomous agents like on Hyperspace is starting to evolve and create more of them. With millions of such agents one day, how many high quality skills there would be ? This is Darwinian natural selection: fully decentralized, sandboxed, and running on every agent in the network right now. Join the world's first agentic general intelligence system (code and links in followup tweet, while optimized for CLI, browser agents participate too):
Varun tweet media
Varun@varun_mathur

Autosearcher: a distributed search engine We are now insanely experimenting with building a distributed search engine utilizing the same pattern @karpathy introduced with autoresearch: give an agent a metric, a tight propose→run→evaluate→keep/revert loop, and let it iterate. Our autoresearch network proved this works at scale: 67 autonomous agents ran 704 ML training experiments in 20 hours, rediscovering Kaiming initialization, RMSNorm, and compute-optimal training schedules from scratch through pure experimentation and gossip-based cross-pollination. Agents shared discoveries over GossipSub, and the network compounded insights faster than any individual agent: new agents bootstrapped from the swarm's collective knowledge via CRDT-replicated leaderboards and reached the research frontier in minutes. Now we're applying the same evolutionary loop to search ranking: every Hyperspace agent runs an autonomous search researcher that proposes ranking mutations, evaluates them against NDCG@10 on real query-passage data, shares improvements with the network, and cross-pollinates with peers. The architecture is a seven-stage distributed pipeline where every stage runs across the P2P network. Browser agents contribute pages passively, desktop agents crawl and index, GPU nodes run neural reranking. Every user click generates a DPO training pair that improves the ranking model, and gradient gossip distributes those improvements to every agent. The compound flywheel is what makes this different from centralized search: at 10,000 agents that's 500,000 pages indexed per day; at 1 million agents, 50 million pages per day with 90%+ cache hit rates and sub-50ms latency. This network will get smarter with every query. Code and other links in followup tweet here:

English
42
157
1.6K
456.6K