Mohammad Saffar

173 posts

Mohammad Saffar

Mohammad Saffar

@msaffar3

Research Scientist @googledeepmind, Gemini multi modal | past: @reveimage, Google brain

Mountain View, CA Katılım Temmuz 2016
432 Takip Edilen886 Takipçiler
Sabitlenmiş Tweet
Mohammad Saffar
Mohammad Saffar@msaffar3·
After a fantastic year at Reve AI, I’ve rejoined Google DeepMind to continue working on VEO. I was deeply involved in its early days, but the rapid progress from VEO 1 to VEO 3 in just one year has truly amazed me. It’s a testament to what can happen when you combine compute, brilliant minds, and a healthy dose of the "bitter lesson."
Mohammad Saffar tweet media
English
17
2
203
24.3K
Mohammad Saffar
Mohammad Saffar@msaffar3·
@Angaisb_ It is rarely about AR vs diffusion and almost always about data and bitter lessons.
English
0
0
5
329
Angel 🌼
Angel 🌼@Angaisb_·
Midjourney should have gone full AR and left diffusion behind They had the data, the compute and the talent yet somehow they still managed to become irrelevant. This isn't any better than older Midjourney models Sad to watch a company I genuinely liked fade out in real time
Angel 🌼 tweet media
Mark Kretschmann@mark_k

The long-awaited testing phase for @Midjourney V8 has officially begun, marking a massive leap forward for the generative art platform. This latest iteration promises a significant boost in efficiency, operating at five times the speed of its predecessors while maintaining a much tighter grip on complex prompt instructions. High-resolution creators will find the native 2K modes particularly useful for professional workflows. The update also brings more reliable text rendering and enhanced "sref" styling, allowing for a level of aesthetic consistency that was previously difficult to achieve. Personalization is a major focus of this release, with improved moodboard performance to help users fine-tune their unique visual language. It is an impressive step toward making AI-assisted design both faster and more intuitive.

English
16
8
141
22.6K
Mohammad Saffar
Mohammad Saffar@msaffar3·
@zhaisf Nice idea! In a causal setup the BOS becomes a vector of all zeros out of attn which could have some implications for post-norm.
English
0
0
2
1.4K
Shuangfei Zhai
Shuangfei Zhai@zhaisf·
Say hi to Exclusive Self Attention (XSA), a (nearly) free improvement to Transformers for LM. Observation: for y = attn(q, k, v), yᵢ and vᵢ tend to have a very high cosine similarity Fix: exclude vᵢ from yᵢ via zᵢ = yᵢ - (yᵢᵀvᵢ)vᵢ/‖vᵢ‖² Result: better training/val loss across model sizes; increasing gains as sequence length grows. See more: arxiv.org/abs/2603.09078
Shuangfei Zhai tweet media
English
27
61
759
136.2K
Mohammad Saffar
Mohammad Saffar@msaffar3·
@tkipf Magic happens when we let models do their jon and learn from data
English
0
0
7
131
Stefano Ermon
Stefano Ermon@StefanoErmon·
Mercury 2 is live 🚀🚀 The world’s first reasoning diffusion LLM, delivering 5x faster performance than leading speed-optimized LLMs. Watching the team turn years of research into a real product never gets old, and I’m incredibly proud of what we’ve built. We’re just getting started on what diffusion can do for language.
English
320
587
4.2K
977.7K
Behnam Neyshabur
Behnam Neyshabur@bneyshabur·
4) Building a company to build a technology to accelerate science Now I'm starting something new—focused on core bottlenecks that could unlock step-change acceleration across science and technology. It's ambitious. I'm about to learn a lot and be humbled. I see so many similarities between my experience in this new journey now and my weeks-long backpacking trips in the Alaskan wilderness with no guide and no trail. If this excites you and you want to learn more, reach out!
English
30
4
423
32.2K
Behnam Neyshabur
Behnam Neyshabur@bneyshabur·
I've left Anthropic to start something new. 🧵
Behnam Neyshabur tweet media
English
156
62
2.9K
398K
Mohammad Saffar
Mohammad Saffar@msaffar3·
Good example of why image generation needs to be pretty smart, we are in the beyond aesthetics era.
Oliver Wang@oliver_wang2

@sama Really impressive model, huge congrats to everyone who worked on it at OpenAI! However, the calendar is wrong, I fixed it for you in Nano Banana Pro 😀

English
0
0
4
2.8K
rohan anil
rohan anil@_arohan_·
Nano banana is impressive and scary! I have been able to use it for lot of creative things like making game sprites. Mass market imagine generation with highly efficient inference tpus backing it.
English
3
1
51
4.1K
Mohammad Saffar
Mohammad Saffar@msaffar3·
This is a really nice and insightful work that embraces removing hard coded assumptions.
Xinchen Yan@skywalkeryxc

For years, RAW pixel space pretraining has been sidelined: too compute-expensive. Our new @GoogleDeepMind paper 📜 dives into the scaling trends of raw pixel models to answer the question “how far are we from scaling up next-pixel prediction?” arxiv.org/pdf/2511.08704 Forecast: Raw next-pixel modeling will reach competitive ImageNet classification (>80% top1 accuracy) and generation metrics (90 Fr’echet Distance) in five years! Threads 👇

English
0
0
2
429
theseriousadult
theseriousadult@gallabytes·
I think we're going to need some new benchmarks
theseriousadult tweet media
English
18
25
663
27.6K
Mostafa Dehghani
Mostafa Dehghani@m__dehghani·
btw, you can bring your graph back to reality. You are welcome.
Mostafa Dehghani tweet media
English
30
52
1.1K
174.4K