Paco Guzmán

410 posts

Paco Guzmán banner
Paco Guzmán

Paco Guzmán

@guzmanhe

Researcher in Language Technologies

San Francisco, CA Katılım Mart 2009
168 Takip Edilen343 Takipçiler
Anish Athalye
Anish Athalye@anishathalye·
.@CleanlabAI has just been acquired by @joinHandshake! This “recruiting marketplace” has silently grown in just 1 year to be a dominant player in human data for AI. With Cleanlab's deep roots in research, Handshake is doubling down on building out its AI research org to strengthen the data foundations for frontier AI.
English
3
2
10
1K
Jonas Mueller
Jonas Mueller@jomulr·
News: @CleanlabAI has been acquired by @joinHandshake!! We're excited to ensure frontier AI systems are trained on the highest-quality data, accelerating our mission toward AI that's actually reliable. Data is only good as the experts producing it and the best work at Handshake
Jonas Mueller tweet media
English
5
4
13
5.5K
Markus Freitag
Markus Freitag@markuseful·
This release represents a huge team effort. So proud of this team! It's a joy to work alongside such brilliant people, and I can't wait to see how the research community uses @TranslateGemma in the future! #AI #Gemma3
Google AI Developers@googleaidevs

🗣 Introducing TranslateGemma, our new collection of open translation models built on Gemma 3. The model is available in 4B, 12B, and 27B parameter sizes, and furthers communication across languages, no matter what device you own. blog.google/innovation-and…

English
2
3
15
2.1K
Paco Guzmán
Paco Guzmán@guzmanhe·
As AI accelerates, it’s worth taking time to ask what’s actually shaping these systems. The new Handshake Research Hub brings together research and essays that explore how human expertise drives progress in AI, examining how we measure human input, codify expertise, and align AI with human judgment. Explore the work joinhandshake.com/research/
English
0
0
0
87
David Ifeoluwa Adelani 🇳🇬
Hallelujah! I’m excited to share that I’ve been selected as a 2025 AI2050 Early Career Fellow by @Schmidtsciences This year’s fellows represent 42 institutions across eight countries, working to ensure AI benefits humankind. Learn more at: lnkd.in/eZA5FHci
Schmidt Sciences@schmidtsciences

We're excited to welcome 28 new AI2050 Fellows! This 4th cohort of researchers are pursuing projects that include building AI scientists, designing trustworthy models, and improving biological and medical research, among other areas. buff.ly/riGLyyj

English
42
15
304
17.2K
Paco Guzmán retweetledi
Garrett Lord
Garrett Lord@GarrettLord·
🚀 GPT‑5 is here: deep reasoning + lightning speed. The leap toward AGI just got real. Recap + takeaways in thread.
English
22
20
79
23.9K
Paco Guzmán
Paco Guzmán@guzmanhe·
Who is coming to ACL? Let's meet!
English
1
1
4
505
Paco Guzmán
Paco Guzmán@guzmanhe·
I’ve recently joined Handshake. We’re hiring!
English
0
0
13
763
Paco Guzmán retweetledi
Handshake
Handshake@joinHandshake·
Introducing Handshake AI—the most ambitious chapter in our story. We leverage the scale of the largest early career network to source, train, and manage domain experts who test and challenge frontier models to failure for the top AI labs.
English
19
13
130
326K
Paco Guzmán retweetledi
Arena.ai
Arena.ai@arena·
BREAKING: Meta's Llama 4 Maverick just hit #2 overall - becoming the 4th org to break 1400+ on Arena!🔥 Highlights: - #1 open model, surpassing DeepSeek - Tied #1 in Hard Prompts, Coding, Math, Creative Writing - Huge leap over Llama 3 405B: 1268 → 1417 - #5 under style control Huge congrats to @AIatMeta — and another big win for open-source! 👏 More analysis below⬇️
Arena.ai tweet media
AI at Meta@AIatMeta

Today is the start of a new era of natively multimodal AI innovation. Today, we’re introducing the first Llama 4 models: Llama 4 Scout and Llama 4 Maverick — our most advanced models yet and the best in their class for multimodality. Llama 4 Scout • 17B-active-parameter model with 16 experts. • Industry-leading context window of 10M tokens. • Outperforms Gemma 3, Gemini 2.0 Flash-Lite and Mistral 3.1 across a broad range of widely accepted benchmarks. Llama 4 Maverick • 17B-active-parameter model with 128 experts. • Best-in-class image grounding with the ability to align user prompts with relevant visual concepts and anchor model responses to regions in the image. • Outperforms GPT-4o and Gemini 2.0 Flash across a broad range of widely accepted benchmarks. • Achieves comparable results to DeepSeek v3 on reasoning and coding — at half the active parameters. • Unparalleled performance-to-cost ratio with a chat version scoring ELO of 1417 on LMArena. These models are our best yet thanks to distillation from Llama 4 Behemoth, our most powerful model yet. Llama 4 Behemoth is still in training and is currently seeing results that outperform GPT-4.5, Claude Sonnet 3.7, and Gemini 2.0 Pro on STEM-focused benchmarks. We’re excited to share more details about it even while it’s still in flight. Read more about the first Llama 4 models, including training and benchmarks ➡️ go.fb.me/gmjohs Download Llama 4 ➡️ go.fb.me/bwwhe9

English
76
368
2.2K
526K
Paco Guzmán retweetledi
Ahmad Al-Dahle
Ahmad Al-Dahle@Ahmad_Al_Dahle·
Introducing our first set of Llama 4 models! We’ve been hard at work doing a complete re-design of the Llama series. I’m so excited to share it with the world today and mark another major milestone for the Llama herd as we release the *first* open source models in the Llama 4 collection 🦙. Here are some highlights: 📌 The Llama series have been re-designed to use state of the art mixture-of-experts (MoE) architecture and natively trained with multimodality. We’re dropping Llama 4 Scout & Llama 4 Maverick, and previewing Llama 4 Behemoth. 📌 Llama 4 Scout is highest performing small model with 17B activated parameters with 16 experts. It’s crazy fast, natively multimodal, and very smart. It achieves an industry leading 10M+ token context window and can also run on a single GPU! 📌 Llama 4 Maverick is the best multimodal model in its class, beating GPT-4o and Gemini 2.0 Flash across a broad range of widely reported benchmarks, while achieving comparable results to the new DeepSeek v3 on reasoning and coding – at less than half the active parameters. It offers a best-in-class performance to cost ratio with an experimental chat version scoring ELO of 1417 on LMArena. It can also run on a single host! 📌 Previewing Llama 4 Behemoth, our most powerful model yet and among the world’s smartest LLMs. Llama 4 Behemoth outperforms GPT4.5, Claude Sonnet 3.7, and Gemini 2.0 Pro on several STEM benchmarks. Llama 4 Behemoth is still training, and we’re excited to share more details about it even while it’s still in flight. A big thanks to all of our launch partners (full list in blog) for helping us bring Llama 4 to developers everywhere including @huggingface, @togethercompute, @SnowflakeDB, @ollama, @databricks and many others👏 This is just the start, we have more models coming and the team is really cooking – look out for Llama 4 Reasoning 😉 A few weeks ago, we celebrated Llama being downloaded over 1 billion times. Llama 4 demonstrates our long-term commitment to open source AI, the entire open source AI community, and our unwavering belief that open systems will produce the best small, mid-size and soon frontier models. Llama would be nothing without the global open source AI community & we are so ready to begin this next chapter with you. 🦙 Read more about the release here: llama.com, and try it in our products today.
Ahmad Al-Dahle tweet media
English
316
894
5.7K
1.1M
Paco Guzmán retweetledi
Paco Guzmán
Paco Guzmán@guzmanhe·
@ArmenAgha Congratulations Armen! Looking forward to what you build
English
0
0
1
164
Armen Aghajanyan
Armen Aghajanyan@ArmenAgha·
Say hello to our new company Perceptron AI. Foundation models transformed the digital realm, now it’s time for the physical world. We’re building the first foundational models designed for real-time, multi-modal intelligence across the real world. perceptron.inc
English
37
58
663
105.6K
Paco Guzmán retweetledi
Kushal Lakhotia
Kushal Lakhotia@hikushalhere·
Quantized Llama 3.2 1B/3B models are here! Blazing fast CPU inference at ~50 tokens/sec for 1B & ~20 tokens/sec for 3B while being competitive on quality with the respective bf16 versions. Very proud of the team. Can't wait to see what developers build with the foundation models.
AI at Meta@AIatMeta

We want to make it easier for more people to build with Llama — so today we’re releasing new quantized versions of Llama 3.2 1B & 3B that deliver up to 2-4x increases in inference speed and, on average, 56% reduction in model size, and 41% reduction in memory footprint. Details on our new quantized Llama 3.2 on-device models ➡️ ai.meta.com/blog/meta-llam… While quantized models have existed in the community before, these approaches often came at a tradeoff between performance and accuracy. To solve this, we Quantization-Aware Training with LoRA adaptors as opposed to only post-processing. As a result, our new models offer a reduced memory footprint, faster on-device inference, accuracy and portability — while maintaining quality and safety for developers to deploy on resource-constrained devices. The new models can be downloaded now from Meta and on @huggingface.

English
0
1
1
246
Paco Guzmán retweetledi
Ahmad Al-Dahle
Ahmad Al-Dahle@Ahmad_Al_Dahle·
On device and small models are a really important part of the Llama herd so we are introducing quantized versions with significantly increased speed. These models have a 2-3x increased speedup – that is fastI Add a link if you want to share what you are building with Llama! ai.meta.com/blog/meta-llam…
Ahmad Al-Dahle tweet media
English
5
17
108
74.9K