Biao Zhang

217 posts

Biao Zhang

Biao Zhang

@BZhangGo

Research Scientist @ Google. Past: PostDoc at UoE. PhD in NLP/MT @edinburghnlp. All opinions are my own.

Katılım Mart 2018
328 Takip Edilen727 Takipçiler
Biao Zhang retweetledi
Biao Zhang retweetledi
Philipp Schmid
Philipp Schmid@_philschmid·
We just release 2 new open-weight Gemma models. FunctionGemma and T5Gemma optimized for on-device agentic actions and multimodal applications. FunctionGemma 🤖 270M parameter built for on-device tool use. 📂 32K token context window. 📱 85% accuracy on mobile system call identification. 🧠 Trained on 6 trillion tokens. T5Gemma 2 🖼️ Multimodal encoder-decoder architecture handling both text and image inputs. 🌐 128K context window across over 140 languages. 📏 Available in three sizes: 270M, 1B, and 4B parameters. 👁️ Normalizes images to 896x896 resolution, encoded into 256 tokens each.
Philipp Schmid tweet mediaPhilipp Schmid tweet mediaPhilipp Schmid tweet mediaPhilipp Schmid tweet media
English
13
48
544
31.3K
Biao Zhang retweetledi
Olivier Lacombe
Olivier Lacombe@o_lacombe·
Meet T5Gemma 2, the next evolution of Google's encoder-decoder family! 🚀 Building on Gemma 3, these models bring major upgrades to efficiency and capability: 🖼️ Multimodal: Understands images + text out of the box. 📚 128K Context: Handles massive datasets with long-context support. 🌍 140+ Languages: Massive multilingual training. ⚡ Efficient Architecture: New tied embeddings & merged attention for faster inference and smaller footprints (270M, 1B, and 4B sizes). Check out the pre-trained checkpoints on Kaggle and Hugging Face now! 🛠️✨
English
1
2
4
308
Biao Zhang
Biao Zhang@BZhangGo·
Can't wait to see what you build!
English
1
0
0
67
Biao Zhang
Biao Zhang@BZhangGo·
That foundation inspired T5Gemma, our first recipe for adapting modern strong decoder-only models into encoder-decoder models🔗 arxiv.org/abs/2504.06225 Now we extend it to the multimodal and long-context regime with T5Gemma 2!
English
1
0
1
153
Biao Zhang retweetledi
Google AI Developers
Google AI Developers@googleaidevs·
Introducing T5Gemma 2, the next generation of encoder-decoder models, built on the powerful capabilities of Gemma 3. Key innovations and upgraded capabilities include: + Multimodality + Extended long context + Support of 140+ languages out of the box + Architectural improvements for efficiency + And more blog.google/technology/dev…
English
39
231
1.8K
279.9K
Biao Zhang retweetledi
Omar Sanseviero
Omar Sanseviero@osanseviero·
Introducing T5Gemma 2, the next generation of encoder-decoder models 🚀 Built on top of Gemma 3, we were able to build compact models at sizes of 270m-270m, 1B-1B, and 4B-4B sizes. While most models today are decoder-only, T5Gemma 2 is the first (I'm aware of) multimodal, long-context, and heavily multilingual (140 languages) encoder-decoder model out there. We hope this model enables the model research community as well as the community of devs ready to explore with new architectures. Blog: blog.google/technology/dev… Models: huggingface.co/collections/go… Paper: arxiv.org/abs/2512.14856
Omar Sanseviero tweet media
English
75
201
1.5K
247K
Biao Zhang retweetledi
Sundar Pichai
Sundar Pichai@sundarpichai·
Introducing EmbeddingGemma, our newest open model that can run completely on-device. It's the top model under 500M parameters on the MTEB benchmark and comparable to models nearly 2x its size – enabling state-of-the-art embeddings for search, retrieval + more.
English
199
518
7.5K
534.7K
Biao Zhang retweetledi
Philipp Schmid
Philipp Schmid@_philschmid·
Introducing EmbeddingGemma, our new open embedding model for on-device AI applications. - Highest ranking open model under 500M on the MTEB benchmark. - Runs on less than 200MB of RAM with quantization. - Dynamic output dimensions from 768 down to 128. - Input context length of 2048 tokens. - Trained on over 100 languages. - Based on Gemma 3 270M. Start building today with @GoogleDeepMind Embedding Gemma and @huggingface Sentence Transformers, @ollama, Llama.cpp, MLX, @lmstudio, @weaviate_io, @googlecloud Vertex AI, @AMD, @baseten, @Cloudflare, @nvidia, and more.
Philipp Schmid tweet media
English
25
58
461
38.6K
Biao Zhang retweetledi
Omar Sanseviero
Omar Sanseviero@osanseviero·
Introducing EmbeddingGemma🎉 🔥With only 308M params, this is the top open model under 500M 🌏Trained on 100+ languages 🪆Flexible embeddings (768 to 128 dims) with Matryoshka 🤗Works with your favorite open tools 🤏Runs with as little as 200MB developers.googleblog.com/en/introducing…
Omar Sanseviero tweet media
English
27
152
1.2K
83.6K
Biao Zhang retweetledi
Google AI Developers
Google AI Developers@googleaidevs·
Introducing EmbeddingGemma: our new open, state-of-the-art embedding model designed for on-device AI 📱
English
35
117
980
153.6K
Biao Zhang
Biao Zhang@BZhangGo·
From encoder-decoder to world-class embeddings! 🚀 Super excited to introduce EmbeddingGemma, our new open embedding model. It leverages the strong encoder from its sibling, T5Gemma, to achieve new SOTA performance for models under 500M! Dive in 👇
Google DeepMind@GoogleDeepMind

EmbeddingGemma is our new best-in-class open embedding model designed for on-device AI. 📱 At just 308M parameters, it delivers state-of-the-art performance while being small and efficient enough to run anywhere - even without an internet connection.

English
0
0
9
794
Biao Zhang retweetledi
Markus Freitag
Markus Freitag@markuseful·
Our Google Translate team is bringing a strong presence to #ACL2025 in Vienna this week! 🇦🇹 My group is excited to present several of our latest papers. 👇 Don't miss them!
English
1
5
53
3.2K
Biao Zhang
Biao Zhang@BZhangGo·
@alvations Thanks for your interests, Liling! However, we are unable to release training-related codes :-(
English
1
0
1
26
Liling Tan
Liling Tan@alvations·
@BZhangGo are you folks planning to release the code that does the conversion/initialization of the enc-decoder model with the decoder-only weights?
English
1
0
0
64
Biao Zhang retweetledi
Omar Sanseviero
Omar Sanseviero@osanseviero·
Introducing T5Gemma: the next generation of encoder-decoder/T5 models! 🔧Decoder models adapted to be encoder-decoder 🔥32 models with different combinations 🤗Available in Hugging Face and Kaggle developers.googleblog.com/en/t5gemma
Omar Sanseviero tweet media
English
21
132
777
85.1K