Our paper "Do Thought Streams Matter?" got featured in a YouTube video 👀
Quick watch, simple breakdown, and some cool insights on Gemini video reasoning.
Watch here: youtube.com/watch?v=8HN19n…
Can video AI really reason about what it sees, or does it just sound confident?
We explored this in our new paper on Gemini vision-language models for video scene understanding.
Some interesting results came out of it 👀
This matters because a lot of people assume more reasoning = better answers.
But for video models, the story may be more nuanced.
Our paper looks at this closely and helps make sense of how reasoning works in real video understanding settings.
We just released a new benchmark looking inside the "black box" of Gemini 2.5 reasoning for video understanding.
Does "thinking more" always lead to better results?
The answer is more nuanced than you’d think 💭