Thomas Mesnard

166 posts

Thomas Mesnard

Thomas Mesnard

@Mesnard_Thomas

Research Scientist @Meta Superintelligence Labs. ex-@GoogleDeepMind - Gemma. PhD @IP_Paris_ | @Mila_Quebec | MSc MVA | @ENS_ULM

Paris 가입일 Ekim 2015
310 팔로잉528 팔로워
고정된 트윗
Thomas Mesnard
Thomas Mesnard@Mesnard_Thomas·
🚀 After Gemma 3 1B, we’re going tiny: Gemma 270M — fast, on-device, low-power & privacy-first. A new vision: smaller models with strong instruction-following & finetuning, ideal for low-latency edge apps & automation. Congrats to everyone involved! huggingface.co/google/gemma-3…
Omar Sanseviero@osanseviero

Introducing Gemma 3 270M 🔥 🤏A tiny model! Just 270 million parameters 🧠 Very strong instruction following 🤖 Fine-tune in just a few minutes, with a large vocabulary to serve as a high-quality foundation developers.googleblog.com/en/introducing…

English
1
3
18
1.6K
Thomas Mesnard 리트윗함
Sundar Pichai
Sundar Pichai@sundarpichai·
An exciting milestone for AI in science: Our C2S-Scale 27B foundation model, built with @Yale and based on Gemma, generated a novel hypothesis about cancer cellular behavior, which scientists experimentally validated in living cells.  With more preclinical and clinical tests, this discovery may reveal a promising new pathway for developing therapies to fight cancer.
English
545
3.3K
21.8K
6.9M
Thomas Mesnard 리트윗함
Sundar Pichai
Sundar Pichai@sundarpichai·
Introducing EmbeddingGemma, our newest open model that can run completely on-device. It's the top model under 500M parameters on the MTEB benchmark and comparable to models nearly 2x its size – enabling state-of-the-art embeddings for search, retrieval + more.
English
199
520
7.5K
534.8K
Thomas Mesnard 리트윗함
Omar Sanseviero
Omar Sanseviero@osanseviero·
Introducing EmbeddingGemma🎉 🔥With only 308M params, this is the top open model under 500M 🌏Trained on 100+ languages 🪆Flexible embeddings (768 to 128 dims) with Matryoshka 🤗Works with your favorite open tools 🤏Runs with as little as 200MB developers.googleblog.com/en/introducing…
Omar Sanseviero tweet media
English
27
153
1.2K
83.7K
Thomas Mesnard 리트윗함
Thomas Mesnard 리트윗함
Andreas Steiner
Andreas Steiner@AndreasPSteiner·
🚀🚀PaliGemma 2 is our updated and improved PaliGemma release using the Gemma 2 models and providing new pre-trained checkpoints for the full cross product of {224px,448px,896px} resolutions and {3B,10B,28B} model sizes. 1/7
Andreas Steiner tweet media
English
4
52
258
61.8K
Thomas Mesnard 리트윗함
Éloi Zablocki
Éloi Zablocki@EloiZablocki·
📢 Exciting opportunity alert! We (valeo.ai) just posted our annual research internship openings in computer vision & ML. Check out the openings and the great achievements by our past interns here: valeoai.github.io/interns/
valeo.ai@valeoai

🌟 Calling all MSc students passionate about computer vision and ML! We’re offering research internships about diffusion models, multi-modal transformers, continual learning, & more. 4 exciting openings await! 🔗 Learn more: valeoai.github.io/interns/ RT to spread the word! 🙌

English
0
4
17
2.8K
Thomas Mesnard 리트윗함
Tris Warkentin
Tris Warkentin@triswarkentin·
Gemma 2 just got even better! 🚀 New Japanese-tuned 2B model AND a $150K Kaggle competition to build Gemma models for every language. Great to have @sundarpichai here to share the excitement! Read more: goo.gle/Gemma4Japan #GemmaDeveloperDay
Tris Warkentin tweet media
English
44
74
471
141.6K
Thomas Mesnard 리트윗함
Markus Zimmermann
Markus Zimmermann@zimmskal·
Imagine a 27B LLM can beat a 405B model in writing quality code by investing a few milliseconds in static code repair. Now stop imagining and take a look at this chart 🌈 Just for Go, we have the following stats: - Increases score +22.9% across 45 applicable models - +26.2% response files compiled (avg. 17 files, 150 tasks total) - mistral-tiny has +71% in score: beats mistral-small and mistral-medium - Gemma 2 27B has +16% in score: beats GPT4o and Llama 3.1 405B Proof that the approach of doing code fixes over a static analysis should be the default for every code response and coding assistant. Makes code instantly more useful.
Markus Zimmermann tweet media
English
4
12
56
11.4K