Basil Mustafa

209 posts

Basil Mustafa

Basil Mustafa

@_basilM

ML research @ google deepmind Zürich vision things in gemini opinions are mine but registered copyright of my mother

Zürich, Schweiz Katılım Kasım 2019
128 Takip Edilen1.7K Takipçiler
Basil Mustafa
Basil Mustafa@_basilM·
🔥
Arena.ai@arena

BREAKING: Gemini 2.5 Pro is now #1 on the Arena leaderboard - the largest score jump ever (+40 pts vs Grok-3/GPT-4.5)! 🏆 Tested under codename "nebula"🌌, Gemini 2.5 Pro ranked #1🥇 across ALL categories and UNIQUELY #1 in Math, Creative Writing, Instruction Following, Longer Query, and Multi-Turn! Massive congrats to @GoogleDeepMind for this incredible Arena milestone! 🙌 More highlights in thread👇

ART
0
1
24
2K
Basil Mustafa retweetledi
Kevin Patrick Murphy
Kevin Patrick Murphy@sirbayes·
Gemma 3 is best in class for a VLM that runs on 1 GPU. Should make RL fine tuning feasible. Also Academic researchers can apply for Google Cloud credits (worth $10,000 per award) to accelerate their Gemma 3-based research.
Zoubin Ghahramani@ZoubinGhahrama1

Introducing Gemma 3: our open model that runs on single GPU/TPU! SoTA AI for its size; multimodal; supporting 140 languages; long context of 128k; fast and high-performance with quantization; open weights; function calling. What more do you want? @google blog.google/technology/dev…

English
8
42
305
40.4K
Basil Mustafa retweetledi
Fabian Mentzer
Fabian Mentzer@mentzer_f·
Chilling more
Fabian Mentzer tweet media
English
10
4
63
4.7K
Basil Mustafa retweetledi
Michael Tschannen
Michael Tschannen@mtschannen·
Have you ever wondered how to train an autoregressive generative transformer on text and raw pixels, without a pretrained visual tokenizer (e.g. VQ-VAE)? We have been pondering this during summer and developed a new model: JetFormer 🌊🤖 arxiv.org/abs/2411.19722 A thread 👇 1/
Michael Tschannen tweet media
English
15
142
838
248.6K
Basil Mustafa
Basil Mustafa@_basilM·
I cannot recommend this enough. Paul is absolutely legendary, the whole team in London is fantastic and the vibes are impeccable. You will learn so much and accomplish awesome things. Apply!!! 🚀♊
Paul Michel@pmichelX

Interested in working on Gemini pre-training? I'm hiring a research scientist to work on pre-training data @GoogleDeepMind in London: boards.greenhouse.io/deepmind/jobs/… I am unfortunately not at #NeurIPS2024 but feel free to reach out to ask questions or see the team at the booth there!

English
0
1
9
904
Basil Mustafa retweetledi
Oriol Vinyals
Oriol Vinyals@OriolVinyalsML·
Gemini 2.0 Flash ⚡️ has arrived! 2.0 Flash > 1.5 Pro (again!) 📈 Interacts with a browser 🤖 Native image generation 🖼️ and much more! Try it out aistudio.google.com/prompts/new_ch… As a preview of what is possible, wishing you all a Drastic Holiday powered by 2.0!
English
12
59
354
41.2K
Basil Mustafa
Basil Mustafa@_basilM·
🔥💪🏾♊😤
Jeff Dean@JeffDean

Gemini 1.5 Model Family: Technical Report updates now published In the report we present the latest models of the Gemini family – Gemini 1.5 Pro and Gemini 1.5 Flash, two highly compute-efficient multimodal models capable of recalling and reasoning over fine-grained information from millions of tokens of context, including multiple long documents and hours of video and audio. Our latest report details notable improvements in Gemini 1.5 Pro within the last four months. Our May release demonstrates significant improvement in math, coding, and multimodal benchmarks compared to our initial release in February. Furthermore, the 1.5 Pro Model is now stronger than 1.0 Ultra. The latest Gemini 1.5 Pro is now our most capable model for text and vision understanding tasks, surpassing 1.0 Ultra on 16 of 19 text benchmarks and 18 of 21 of the vision understanding benchmarks. The table below highlights the improvement in average benchmark performance for different categories in 1.5 Pro since Feb, and also shows the strength of the model relative to the 1.0 Pro and 1.0 Ultra models. The 1.5 Flash model also compares very well against the 1.0 Pro and 1.0 Ultra models. One clear example of this can be seen on MMLU On MMLU we find that 1.5 Pro surpasses 1.0 Ultra in the regular 5-shot setting scoring 85.9% versus 83.7%. However with additional inference compute, via majority voting on top of multiple language model samples, we can get a performance of 91.7% versus Ultra’s 90.0%, which extends the known performance ceiling of this task. @OriolVinyalsML and I are very proud of the whole Gemini team, and it’s fantastic to see this progress and to share these highlights from our Gemini Model Family. Read the updated report here: goo.gle/GeminiV1-5

ART
0
0
13
617
Basil Mustafa retweetledi
Pietro Schirano
Pietro Schirano@skirano·
I was so impressed with the Astra demo at Google I/O yesterday that I decided to build my own version using Gemini 1.5 Pro Flash. It's so fast and really good. ⚡️ It was even able to detect the gate! Content is streamed directly from my camera. Voice via @elevenlabs
English
48
78
663
107.7K
Basil Mustafa retweetledi
Jacob Austin
Jacob Austin@jacobaustin132·
This is something I've worked on for a while! You can save the activations of one LLM call and reuse them for a follow-up that overlaps with the first. This means asking a question about a big codebase can take 30 seconds the first time and 1s after that!
Jaana Dogan ヤナ ドガン@rakyll

Gemini’s context caching is one of the most exciting releases that came out it of Google I/O. ai.google.dev/gemini-api/doc…

English
14
47
434
104.4K