Robert Dadashi

207 posts

Robert Dadashi

Robert Dadashi

@robdadashi

Gemma TL @GoogleDeepMind

Paris, France Katılım Eylül 2014
478 Takip Edilen2.1K Takipçiler
Sabitlenmiş Tweet
Robert Dadashi
Robert Dadashi@robdadashi·
Gemma is hiring! 🚀🚀🚀 Please apply if you want to work on pre-training and/or designing models at the intersection of multiple modalities (text, audio, vision, video, actions,) Location: Paris Apply here: job-boards.greenhouse.io/deepmind/jobs/…
English
6
39
405
36.6K
Robert Dadashi retweetledi
Sundar Pichai
Sundar Pichai@sundarpichai·
An exciting milestone for AI in science: Our C2S-Scale 27B foundation model, built with @Yale and based on Gemma, generated a novel hypothesis about cancer cellular behavior, which scientists experimentally validated in living cells.  With more preclinical and clinical tests, this discovery may reveal a promising new pathway for developing therapies to fight cancer.
English
545
3.3K
21.8K
6.9M
Robert Dadashi retweetledi
Prince Canuma
Prince Canuma@Prince_Canuma·
MLX-VLM v0.2.0 is here 🔥🙌🏽 Or should I rename it to MLX-O since Omni is taken? We now support Omni models starting with the latest Gemma3n from @GoogleDeepMind and more coming soon. Thanks to @fleetwood___ , @pcuenq, @osanseviero, @reach_vb and team for helping make this possible! Get started today: > pip install -U mlx-vlm
Prince Canuma tweet media
Prince Canuma@Prince_Canuma

🚀 Gemma 3n is HERE with day-0 MLX support! Thrilled to have collaborated with @GoogleDeepMind and @huggingface to bring the MLX community instant access to this groundbreaking model. Why Gemma 3n changes everything: 📹 True Multimodal: text, audio, image & video ⚡ Runs on just a few GB of RAM 🏆 First <10B model to hit 1300+ on @lmarena_ai Massive thanks to @osanseviero, @reach_vb, @fleetwood___, @pcuenq and Ryan for making this possible! 🙏 Get started in seconds: > pip install -U mlx-vlm Try it now and share what you build! 🔥 youtu.be/8-8R2UvUBrc?si…

English
1
7
26
8.6K
Robert Dadashi retweetledi
Sundar Pichai
Sundar Pichai@sundarpichai·
Gemini 2.5 Pro + 2.5 Flash are now stable and generally available. Plus, get a preview of Gemini 2.5 Flash-Lite, our fastest + most cost-efficient 2.5 model yet. 🔦 Exciting steps as we expand our 2.5 series of hybrid reasoning models that deliver amazing performance at the Pareto frontier of cost and speed. 🚀
Sundar Pichai tweet media
English
253
446
4.2K
1M
Robert Dadashi retweetledi
clem 🤗
clem 🤗@ClementDelangue·
Everyone is talking about how we need more AI data centers (especially the ones who would mostly benefit from them) but why is no one talking about on-device AI? Running AI on your device: - Free - Faster & takes advantage of existing hardware - 100% privacy and control (you don’t send your data to an API)
clem 🤗 tweet media
English
187
201
1.6K
230.1K
Robert Dadashi retweetledi
Neil Zeghidour
Neil Zeghidour@neilzegh·
Unmute is our new cascaded voice assistant: fast, accurate, and flexible. It doesn't have the full-duplex and zero latency of Moshi, but you can change the voice with a 10s sample and plug any LLM. A good playground for testing custom voice AIs.
kyutai@kyutai_labs

Talk to unmute.sh 🔊, the most modular voice AI around. Empower any text LLM with voice, instantly, by wrapping it with our new speech-to-text and text-to-speech. Any personality, any voice. Interruptible, smart turn-taking. We’ll open-source everything within the next few weeks.

English
2
8
65
5.1K
Robert Dadashi
Robert Dadashi@robdadashi·
Gemma 3n is out! 🚀🚀🚀 The frontier models from a year ago can now run locally on a phone! Lots of innovations (e.g. matformers, mix’n’match, per layer embeddings) to make this model mobile first. And we finally have audio/video as an input for Gemma models! 1/2
English
3
9
41
3.2K
Robert Dadashi retweetledi
Tris Warkentin
Tris Warkentin@triswarkentin·
This is my favorite demo of Gemma 3n -- multimodal live video understanding and intelligence, locally on your phone 🤯! This was only possible with the peak of foundation models at I/O last year -- the Astra demo -- the progress of small models is incredible
Google AI Developers@googleaidevs

Key features include: -Expanded multimodal understanding with video and audio input, alongside text and images -Developer-friendly sizes: 4B and 2B (and many in between!) -Optimized on-device efficiency for 1.5x faster response on mobile compared to Gemma 3 4B

English
1
4
22
2.6K
Robert Dadashi retweetledi
Johan Ferret
Johan Ferret@johanferret·
We just released Gemma 3n, a mobile-first & multimodal LLM that works with as little as 2Gb RAM. Feels crazy to interact with a model whose training I contributed to, hosted on my *own* phone (see screenshot!) 🤯 It packs so much for its size, give it a try (how to in thread)!
Johan Ferret tweet media
English
2
5
40
3.1K
Robert Dadashi retweetledi
Aditya Kusupati
Aditya Kusupati@adityakusupati·
Pocket powerhouse admist I/O awesomeness! Gemma 3n E4B & E2B are insane models, optimized for on-device while rivaling frontier models. It's a 🪆Matryoshka Transformer (MatFormer)🪆: Natively elastic b/w 4B & 2B pareto-optimally! ⭐️: free models with ZERO training cost! 🧵👇
GIF
English
9
38
296
84K
Robert Dadashi retweetledi
Alexandre Ramé
Alexandre Ramé@ramealexandre·
Releasing Gemma 3n, our new open-weight model processing audio, images and text (with improved multilingual capabilities), optimized for on-device usage with MatFormer architecture (enabling adaptive compute) and reaching 1283 on Chatbot Arena. Read more: developers.googleblog.com/en/introducing….
Alexandre Ramé tweet media
English
2
20
80
7.9K
Robert Dadashi retweetledi
Glenn Cameron Jr
Glenn Cameron Jr@GlennCameronjr·
I've been reading about Gemma 3n for months. It sounded great, but my mind was blown when I started seeing the demos. 🤯 Check out this quick demo:
English
1
5
19
1.4K
Robert Dadashi retweetledi
Philipp Schmid
Philipp Schmid@_philschmid·
Gemini Nano meets Gemma! Gemma 3n the next generation of Gemini Nano is expanding to multimodality for edge devices! ✨ Gemma 3n will be an open, offline first model to run and build agents from browsers to on-device! 🚀 Gemma 3n will: 🔤 👀 🖼️ Understand text, images and audio and generate text 2️⃣come in 2 sizes 2B & 4B with Per-Layer Embedding caching and a MatFormer architecture 📏 has a 32K token context (1x image 256 tokens, 1s audio 6.25 tokens) ⚡ Optimized edge with for 1.52x faster response on mobile compared to Gemma 3 4B 🧠 Offline agentic use and built for Privacy, no connection required 🔜 Today in AI Studio, Soon everywhere
Philipp Schmid tweet media
English
6
47
333
17.5K
Robert Dadashi retweetledi
Pier Giuseppe Sessa
Pier Giuseppe Sessa@piergsessa·
Gemini Diffusion is out! Very excited to have worked on the post-training of such a state-of-the-art text diffusion model. Incredible performance at lightspeed⚡️ Congrats to everyone involved!!
Google DeepMind@GoogleDeepMind

We’ve developed Gemini Diffusion: our state-of-the-art text diffusion model. Instead of predicting text directly, it learns to generate outputs by refining noise, step-by-step. This helps it excel at coding and math, where it can iterate over solutions quickly. #GoogleIO

English
0
9
37
2.9K