Benoit Vandevivere

828 posts

Benoit Vandevivere

Benoit Vandevivere

@benvdv

Entrepreneur. 🇪🇺 eu/acc. @euacchq Speaker and advisor on disruption, innovation. Strong believer in innovation for positive change and impact.

Spain Katılım Aralık 2007
1.1K Takip Edilen935 Takipçiler
Babak Hassibi
Babak Hassibi@BabakHassibi·
Today I feel very proud and am honored to introduce PrismML. This company grew out of years of research at Caltech and a simple conviction: the future of AI will not be defined only by ever-growing models. It will be defined by intelligence density - how much useful intelligence we can deliver per unit of compute, memory, and energy. At PrismML, we seek to build the most concentrated form of intelligence. Our first proof point is the 1-bit Bonsai family: models that are small, fast, and efficient enough to run locally, while remaining competitive with full-precision models in their class. We see this not as an endpoint, but as the beginning of a new paradigm for AI, one that expands where intelligence can exist: on-device, at the edge, in the cloud, and in entirely new products and systems. We are excited to begin sharing that vision.
PrismML@PrismML

Today, we are emerging from stealth and launching PrismML, an AI lab with Caltech origins that is centered on building the most concentrated form of intelligence. At PrismML, we believe that the next major leaps in AI will be driven by order-of-magnitude improvements in intelligence density, not just sheer parameter count. Our first proof point is the 1-bit Bonsai 8B, a 1-bit weight model that fits into 1.15 GBs of memory and delivers over 10x the intelligence density of its full-precision counterparts. It is 14x smaller, 8x faster, and 5x more energy efficient on edge hardware while remaining competitive with other models in its parameter-class. We are open-sourcing the model under Apache 2.0 license, along with Bonsai 4B and 1.7B models. When advanced models become small, fast, and efficient enough to run locally, the design space for AI changes immediately. We believe in a future of on-device agents, real-time robotics, offline intelligence and entirely new products that were previously impossible. We are excited to share our vision with you and keep working in the future to push the frontier of intelligence to the edge.

English
36
26
424
36.7K
Benoit Vandevivere
Benoit Vandevivere@benvdv·
@aminkarbasi @GoogleResearch Congrats. It’s very interesting. Do you think this could enable current SOTAs to run on current Macbooks , even iPhones? Like opening the gate to on-device AI with great models..
English
0
0
1
765
Amin Karbasi
Amin Karbasi@aminkarbasi·
I left @GoogleResearch almost two years ago, so it makes me genuinely happy to see our work on polar quantization (my last project), which eventually led to extreme compression, being recognized there. It is a nice reminder that good fundamental work tends to find its place with time.
Google Research@GoogleResearch

Introducing TurboQuant: Our new compression algorithm that reduces LLM key-value cache memory by at least 6x and delivers up to 8x speedup, all with zero accuracy loss, redefining AI efficiency. Read the blog to learn how it achieves these results: goo.gle/4bsq2qI

English
46
104
2.6K
203.8K
Benoit Vandevivere retweetledi
Alessandro Palombo
Alessandro Palombo@thealepalombo·
These are all fantastic ideas. In general, there is no EU Inc. without accountability on performance. There should be an embedded methodology for tracking multiple performance factors…KPIs and, in case of failure, the possibility for remedies to kick in immediately. Without this approach, the EU Inc. will be a “nice to have” piece of theoretical harmonization that unfortunately won’t succeed in practice
English
1
2
8
1.3K
Benoit Vandevivere retweetledi
eu/acc
eu/acc@euacchq·
Re: EU Inc — the role of courts There are two issues Europe needs to address, for EU Inc and beyond: 🇪🇺 National courts are extremely slow (hundreds of days, often years). There is no real or rapid consistency in how EU law is applied across the 27 national court systems. 1/4
English
11
14
180
16.5K
Benoit Vandevivere retweetledi
eu/acc
eu/acc@euacchq·
We want to congratulate the @EU_Commission , @EUCssrMcGrath , @EZaharievaEU @vonderleyen - and the Europeans who spoke up - for this draft EU inc proposal, as it's a good starting point. Pending further review, we'll highlight now some of the interesting items 1/7
Ursula von der Leyen@vonderleyen

With EU Inc., we are making it drastically easier to start and grow a business all across Europe ↓ twitter.com/i/broadcasts/1…

English
3
19
140
10.2K
Benoit Vandevivere
Benoit Vandevivere@benvdv·
We did. And we did say it’s important to set KPIs on speed and cost: caps. Regulation or directive, in any case we need a results-oriented company law where registry fees are not 10x or 100x the fees of other successful countries Not 10x slower. We need speed and efficiency
eu/acc@euacchq

What we said to the EU Commission when building EU Inc is simple. Look at the World’s best : fastest, most affordable, most digital company law. Make it as fast or faster. Same price. Make Europe number one. In 2 days we’ll find out how ambitious is the @EU_Commission

English
0
0
2
259
Benoit Vandevivere
Benoit Vandevivere@benvdv·
@NXT4EU Yes. I hope so. Seems there’s reason for hope. In any case, we’ll keep pushing :)
English
0
0
2
56
NXT EU
NXT EU@NXT4EU·
@benvdv Cool! Hopefully it will be unleashed in all its glory
English
1
0
15
661
Benoit Vandevivere
Benoit Vandevivere@benvdv·
March 18th. 🗓️ That is the date the EU Commission is scheduled to unveil the "EU Inc" (28th regime) draft law. 🇪🇺 After contributing so much effort to shaping this, I’m excited for the ecosystem to finally see the text coming out soon. Stay tuned.
English
1
1
28
2.4K
Groq Inc
Groq Inc@GroqInc·
Groq has entered into a non-exclusive licensing agreement with Nvidia for Groq’s inference technology. GroqCloud will continue to operate without interruption. Learn more here: groq.com/newsroom/groq-…
English
129
284
2.4K
1.4M
Christopher Manning
Christopher Manning@chrmanning·
Re: “Google has cannibalized a big chunk of 3.0 Pro use cases”: Being willing to cannibalize yourself is the only way to avoid The Innovator’s Dilemma (en.wikipedia.org/wiki/The_Innov…) – and @Google is in the lucky position of having the industry-best economics for enabling low prices.
Delip Rao e/σ@deliprao

I had early access to Gemini 3 Flash (ty @GoogleDeepMind) and it shocked my vibe test as I walked in with 2.5 Pro/Flash expectations. Looking at the evals now it all makes sense. The latent story here, from my POV, is Google has cannibalized a big chunk of 3.0 Pro use cases (besides smoking competition). The fact that Google pushed this out shortly after 3.0 makes me think they already know future 3.x Pro will have stellar performance. That’s something to look forward to. Flash is now the best agentic model hands down (tau2, mcp atlas, swe verified) for its price point. The lower score on HLE and GPQA diamond over Pro means it is not as knowledgeable as Pro, which makes sense. The choice is clear: Flash 3.0 should be the de facto agentic model unless you are in a knowledge heavy domain. But I suspect, even there, with sufficient context management you can get good value out of Flash 3.0. Gemini LLMs have been a black swan for a big chunk of 2025. I doubt any outsider could’ve predicted total Pareto frontier domination by the Gemini franchise by EOY. Congratulations to the Google/DeepMind teams for this exciting program execution!

English
14
14
195
39.9K
Demis Hassabis
Demis Hassabis@demishassabis·
For a fast model, Gemini 3 Flash offers incredible performance, allowing us to provide frontier intelligence to everyone globally. Try the 'fast' mode from the model picker in the @GeminiApp - it’s shockingly speedy AND smart. Best pound-for-pound model out there ⚡️⚡️⚡️
Demis Hassabis tweet media
English
119
144
2K
264.2K
Benoit Vandevivere
Benoit Vandevivere@benvdv·
@mikeknoop Interesting. Somehow reasoning capabilities could be said to develop overfitting on the fly on the new data they are tested on.. Given their no prior training, they’re good at (over) fitting a new set of data..
English
0
0
0
336
JFPuget 🇺🇦🇨🇦🇬🇱
One ingredient of our solution is the Tiny Recursive Model of @jm_alexia . During the competition we got a score of 10% on the semi private dataset of arc agi2, and 10.41% on the public eval dataset. I further trained TRM for 10 more days using the same recipe as in our solution. There is some volatility, but the best pass@2 I got on the public eval dataset was 18.05%. This is within Kaggle arc prize compute limits.
JFPuget 🇺🇦🇨🇦🇬🇱@JFPuget

We also appear on the ARC AGI2 leaderboard. Not best score, but clearly on the Pareto frontier with a much lower cost than best scores.

English
6
13
152
42.8K
Remi Cadene
Remi Cadene@RemiCadene·
Read this to learn about Deep Learning development through the years (AI + Robotics) 😇
Pierre Sermanet@psermanet

I’m really happy to share that we’re launching UMA. Together with @RemiCadene, @alibert_s, @therobotstudio, and an exceptional founding team, we’re building general-purpose mobile and humanoid robots. If you want to be part of this adventure, reach out at uma.bot Throughout my career, I have been obsessed with scalable learning and data acquisition methods that require little to no labels. Back in 2005 with @ylecun, we were self-supervising our “deep” 2-layer network to do long range vision using short range stereo information, this was running live onboard our robot. However, because our deep model was so slow, the robot would crash constantly, so I designed a decoupled fast & far architecture for robust navigation, allowing fast control to coexist with slow long horizon thinking, much like systems 1 & 2 in modern humanoids. My PhD was focused on making deep learning work for computer vision, including unsupervised feature learning with @koraykv, writing and open-sourcing a C++ deep learning library with @soumithchintala, and open-sourcing one of the first deep learning vision systems. I came back towards robotics at @Google Brain and @GoogleDeepMind, where I pushed for entirely label-free methods on real robots. In 2017, @coreylynch and I managed to make our robot imitate human motion by co-training self-supervision across sim and real domains jointly, without any labels. With @imkelvinxu and @svlevine , we showed that unsupervised visual reward learning could be used for RL in the real world. In 2020, Corey and I developed the first manipulation VLA, which was trained with very few language labels thanks to self-supervision on play data (playing is an efficient way to demonstrate and practice a broad set of skills and is essential for human development). I was never satisfied with the status quo of top-down data collection, where researchers decide a few tasks to collect data on. Instead, I believed that we should let the data speak: tasks should be automatically discovered bottom-up (scalable and general) from cheap and continuous data collection, with a sprinkle of more expensive data and labels. In 2022, I explored long-horizon reasoning for robotics using scalable automatic labeling augmentations for VQA tasks and studied the economics of different data collection schemes. Most recently, I developed approaches to scalably discover laws of robotics from real data (images, hospital reports, sci-fi literature) in a broad and bottom-up fashion, which improved robot behavior over top-down approaches like Asimov’s laws. All these experiences nourished my vision for UMA as Chief Scientist, I’m incredibly excited to put everything together and so grateful I get to contribute to this incredible moment in human history. Picture: Yann supporting UMA as an advisor and investor, with the team in Paris a couple weeks ago.

English
2
4
72
10.7K