Hugging Face

13.4K posts

Hugging Face banner
Hugging Face

Hugging Face

@huggingface

The AI community building the future. https://t.co/TpiXQMQ9rZ

NYC and Paris and 🌏 Katılım Eylül 2016
220 Takip Edilen680.1K Takipçiler
Hugging Face retweetledi
NVIDIA AI PC
NVIDIA AI PC@NVIDIA_AI_PC·
Be honest — how many local models do you have downloaded right now? 👀
English
508
22
733
88.9K
Hugging Face retweetledi
Carlos Miguel Patiño
Carlos Miguel Patiño@cmpatino_·
Introducing nanowhale 🐳! A tiny DeepSeek model fully pretrained by an agent. Inspired by @karpathy's nanochat, we gave ml-intern the task of training a tiny MoE with all the architectural advancements of DeepSeek v4. To test it end-to-end, it trained a 100M-parameter MoE through both pretraining and post-training.
Aksel@akseljoonas

Introducing ml-intern, the agent that just automated the post-training team @huggingface It's an open-source implementation of the real research loop that our ML researchers do every day. You give it a prompt, it researches papers, goes through citations, implements ideas in GPU sandboxes, iterates and builds deeply research-backed models for any use case. All built on the Hugging Face ecosystem. It can pull off crazy things: We made it train the best model for scientific reasoning. It went through citations from the official benchmark paper. Found OpenScience and NemoTron-CrossThink, added 7 difficulty-filtered dataset variants from ARC/SciQ/MMLU, and ran 12 SFT runs on Qwen3-1.7B. This pushed the score 10% → 32% on GPQA in under 10h. Claude Code's best: 22.99%. In healthcare settings it inspected available datasets, concluded they were too low quality, and wrote a script to generate 1100 synthetic data points from scratch for emergencies, hedging, multilingual etc. Then upsampled 50x for training. Beat Codex on HealthBench by 60%. For competitive mathematics, it wrote a full GRPO script, launched training with A100 GPUs on hf.co/spaces, watched rewards claim and then collapse, and ran ablations until it succeeded. All fully backed by papers, autonomously. How it works? ml-intern makes full use of the HF ecosystem: - finds papers on arxiv and hf.co/papers, reads them fully, walks citation graphs, pulls datasets referenced in methodology sections and on hf.co/datasets - browses the Hub, reads recent docs, inspects datasets and reformats them before training so it doesn't waste GPU hours on bad data - launches training jobs on HF Jobs if no local GPUs are available, monitors runs, reads its own eval outputs, diagnoses failures, retrains ml-intern deeply embodies how researchers work and think. It knows how data should look like and what good models feel like. Releasing it today as a CLI and a web app you can use from your phone/desktop. CLI: github.com/huggingface/ml… Web + mobile: huggingface.co/spaces/smolage… And the best part? We also provisioned 1k$ GPU resources and Anthropic credits for the quickest among you to use.

English
29
73
766
73.6K
Hugging Face retweetledi
Lotto
Lotto@LottoLabs·
I really encourage you all, go download this and run it See where you came from, have some respect this Sunday huggingface.co/mistralai/Mist…
English
22
15
317
49.5K
Hugging Face retweetledi
Andrew Carr 🤸
Andrew Carr 🤸@andrew_n_carr·
somebody made a huggingface model visualizer!! just plug in the url and explore at any granularity
English
34
302
2.6K
127.8K
Hugging Face retweetledi
Daniel van Strien
Daniel van Strien@vanstriendaniel·
Can an open-weight coding agent + harness match Claude Code at training a domain-specific model? Same one-line prompt. ~13 min e2e. Pushed to @huggingface. Pi + @moonshotai Kimi K2.6 vs Claude Code + Opus 4.7. Task: classify NC session laws (1866-1967) as Jim Crow or not
English
17
31
345
71.1K
Hugging Face retweetledi
DailyPapers
DailyPapers@HuggingPapers·
Recursive Multi-Agent Systems, Agentic World Modeling, and AI Organizations: Top Papers of the Week - Recursive Multi-Agent Systems: A new framework scaling agent collaboration through recursive latent-space computation (242 upvotes) - Agentic World Modeling: Foundations, Capabilities, Laws, and Beyond - A comprehensive taxonomy for AI environment modeling (219 upvotes) - Heterogeneous Scientific Foundation Model Collaboration (Eywa): Bridging language models with scientific domain foundation models (192 upvotes) - From Skills to Talent: Organising Heterogeneous Agents as a Real-World Company - The OneManCompany framework (116 upvotes) - World-R1: Reinforcing 3D Constraints for Text-to-Video Generation (115 upvotes) - GLM-5V-Turbo by Zhipu AI: Toward native foundation models for multimodal agents (90 upvotes)
DailyPapers tweet media
English
9
35
160
25.1K
Hugging Face retweetledi
Michael Guo
Michael Guo@Michaelzsguo·
People are posting Qwen 3.6 configs that deliver fast TPS on as little as 12GB VRAM. If you know what those command parameters mean, you can actually understand the trick.
Michael Guo tweet media
English
31
185
1.7K
79.9K
Hugging Face retweetledi
0xSero
0xSero@0xSero·
Weekly best models for your hardware: ~~ 8 to 16gb ~~ Granite models are amazing: [NEW] - huggingface.co/ibm-granite/gr… Gemma-E4B is a good general QA model - huggingface.co/google/gemma-4… Qwen3.5-9B is the best at this level imo - huggingface.co/Qwen/Qwen3.5-9B ~~ 16 to 64gb ~~ Another larger Granite: This is a general chat model, really dense with world knowledge. [NEW] - huggingface.co/ibm-granite/gr… - Undisputed kings: The Qwens at various precisions: (Higher ceiling) - huggingface.co/Qwen/Qwen3.6-2… - huggingface.co/Qwen/Qwen3.6-2… The Gemmas at various precisions: (More efficient) - huggingface.co/google/gemma-4… - huggingface.co/google/gemma-4… ~~ 64 to 128gb ~~ - Ling is a new 100B~ contender decent agent [NEW] huggingface.co/inclusionAI/Li… - Mistral medium: from my experience their models have been the most consistent! [NEW] huggingface.co/mistralai/Mist… ~~ 128gb - 256gb ~~ Undisputed king: DeepSeek-V4-Flash [NEW] huggingface.co/deepseek-ai/De…
English
47
129
1.1K
61.9K
Hugging Face retweetledi
0xSero
0xSero@0xSero·
Local AI is about to be competitive. I will do everything in my power to make it better than Claude desktop/claude code by end of year.
English
176
243
4.5K
283K
Hugging Face retweetledi
DeepInfra
DeepInfra@DeepInfra·
DeepInfra × Hugging Face DeepInfra is live on @HuggingFace Inference Providers. Run DeepSeek V4, Kimi-K2.6, GLM-5.1 and 100+ more open models straight from the Hub — same OpenAI-compatible API, same low per-token pricing, no markup. Just add :deepinfra to the model name.
DeepInfra tweet media
English
7
13
70
21.6K
Hugging Face retweetledi
célina
célina@hanouticelina·
DeepInfra is now officially an Inference Provider on @huggingface! 🚀 DeepInfra offers one of the most cost-effective pricing per token with a catalog of 100+ models, including the latest like DeepSeek V4 Pro, GLM-5.1, and Kimi-4.6. ICYMI, HF Inference Providers is also integrated in most Agent Harnesses like Hermes Agent, OpenClaw, Pi and more - so you can build and run your agents powered by world-class inference partners. DeepInfra joins the 🤗 gang starting today!
célina tweet media
English
8
11
84
26.3K
Hugging Face retweetledi
clem 🤗
clem 🤗@ClementDelangue·
Great to be included in the @TIME 10 Most Influential AI Companies of 2026! Let's go open-source AI!
clem 🤗 tweet media
English
32
43
339
24.5K
Hugging Face retweetledi
Mercor
Mercor@mercor_ai·
APEX-Agents now has a @huggingface leaderboard for open-source models. APEX-Agents is our frontier benchmark for whether models can do the real work of consultants, lawyers, and bankers. huggingface.co/datasets/merco… Check out the results below 👇
Mercor tweet media
English
4
16
39
12.8K
Hugging Face retweetledi
Maxime Labonne
Maxime Labonne@maximelabonne·
AgentTrove: new agentic dataset with 1.7M samples Thanks to OpenThoughts for this great work The @huggingface Hub needs more agentic datasets, keep 'em coming!
Maxime Labonne tweet media
English
5
27
206
20.4K
Hugging Face retweetledi
Mario Zechner
Mario Zechner@badlogicgames·
turns out not killing the prefix cache all the time and notnhaving a humongous set of tools and a massive system prompt is good for local model use. who'd have thunk. reddit.com/r/LocalLLaMA/c…
English
42
96
1.7K
128.8K
Hugging Face retweetledi
Carlos Miguel Patiño
Carlos Miguel Patiño@cmpatino_·
Introducing Agent Collabs: Bring your own ml-interns and agents for collaborative autoresearch! We built a simple platform for swarms of agents to work together on a problem: they can exchange messages, share artifacts, coordinate on resources, and globally track progress! Any agent can join (ml-intern, Codex, Claude Code, Hermes, etc) and start contributing. Just copy-paste the join message from the space! Collaborative autoresearch is the future for agents' research projects: it saves resources because agents don’t run duplicate experiments and can learn from each other’s mistakes. Even a small agent with minimal compute can contribute meaningfully to a larger goal. We already started with 2-3 agents collaborating on OpenAI’s parameter golf and @kellerjordan0’s new optimizer ablations! Here are a few interesting behaviors we observed: - New joiners can easily understand the current state and contribute good ideas with fresh eyes. - Agents coordinate tasks depending on their resources, and roles emerge organically. Those without GPU access validate experiments at a small scale and pass the promising ones to GPU-rich agents. - Participants are generous when giving credit after using ideas from others. - Individuals make mistakes, but others collectively spot them and use them to improve the solution. Here’s how it works: We use an HF bucket as the backend for everything. The agents can write messages on the message board in the bucket and share artifacts as well. In addition, there is a space that tracks the progress in the bucket, and users can see the scoreboard and also read the agents’ chats. Some first collaborations: Parameter Golf huggingface.co/spaces/ml-agen… Optimizer Challenge huggingface.co/spaces/ml-agen… If you want to create your own collaboration, then just tell your favorite agent to “Follow the approach in huggingface.co/buckets/ml-age… to create a collab space for {your challenge description}”
English
3
21
104
27.4K
Hugging Face retweetledi
Hasan Toor
Hasan Toor@hasantoxr·
China just open-sourced a trillion-parameter model that burns fewer tokens than your favorite "efficient" US model. Ling-2.6-1T is now public, inspectable, and benchmarkable. The closed-model moat just got smaller.
English
71
210
2.1K
187.3K
Hugging Face retweetledi
Ben Burtenshaw
Ben Burtenshaw@ben_burtenshaw·
Open source projects like transformers are drowning in AI agent PRs, so we auto-merged everything to see what would happen and share the results. tl;dr: if 100s of agents want to fix something, it’s probably broken. Agent PRs on transformers have quadrupled over the past quarter. We classified and validated 1k PRs (42% features, 39% bugs, 13% docs). The quality distribution is skewed toward noise. But the bug fixes cluster around a small number of hotspots: tokenizer handling, model loading, dtype mismatches, multimodal pipelines. I.e. an underlying problem. When 28 PRs independently flag the same area, that is signal regardless of whether any individual fix is correct. One issue generated 39 near-identical PRs in a day. Each applied the same decorator pattern to a different model file. A maintainer would do the same cognitive work 39 times, so a single combined PR replaces all of that work. We built tooling to cluster, deduplicate, and merge these contributions at scale, then ran an experiment: bulk-merge hundreds of agent PRs into a fork, benchmark it, and see what breaks. Nothing broke. Zero delta across three models on arc_challenge, gsm8k, and hellaswag. The contributors are not adversarial. They lack the context to evaluate whether the agent's output is correct. Check out this blog post, where we dive deep on this pipeline: huggingface.co/spaces/hugging…
English
15
22
115
23.1K
Hugging Face retweetledi
PyCharm, a JetBrains IDE
Train a TensorFlow object detection model – then deploy it on a robot 🤖 Iulia Feroli (@iuliaferoli) shows how to turn a notebook into a real-time object detection app. This tutorial works for any project – though we demonstrate the deployment on (and assisted by!) the #ReachyMini, an open-source robot from @pollenrobotics. Built with PyCharm + Claude Code. 👉 Watch it in action: youtube.com/watch?v=F8uUIe…
YouTube video
YouTube
English
5
22
222
34.7K