François-Régis Chaumartin

311 posts

François-Régis Chaumartin banner
François-Régis Chaumartin

François-Régis Chaumartin

@FR_Chaumartin

Entrepreneur, #AI, #LLM, #NLP. Ingénieur en informatique, docteur en linguistique. Fondateur de Proxem, acheté par Dassault Systèmes en 2020. Opinions are mine.

Paris, France Katılım Nisan 2013
293 Takip Edilen198 Takipçiler
François-Régis Chaumartin retweetledi
Anthropic
Anthropic@AnthropicAI·
We invited Claude users to share how they use AI, what they dream it could make possible, and what they fear it might do. Nearly 81,000 people responded in one week—the largest qualitative study of its kind. Read more: anthropic.com/features/81k-i…
English
481
918
6.3K
2.6M
François-Régis Chaumartin
François-Régis Chaumartin@FR_Chaumartin·
2026 est l'année de bascule de l'IA au travail
Andrej Karpathy@karpathy

It is hard to communicate how much programming has changed due to AI in the last 2 months: not gradually and over time in the "progress as usual" way, but specifically this last December. There are a number of asterisks but imo coding agents basically didn’t work before December and basically work since - the models have significantly higher quality, long-term coherence and tenacity and they can power through large and long tasks, well past enough that it is extremely disruptive to the default programming workflow. Just to give an example, over the weekend I was building a local video analysis dashboard for the cameras of my home so I wrote: “Here is the local IP and username/password of my DGX Spark. Log in, set up ssh keys, set up vLLM, download and bench Qwen3-VL, set up a server endpoint to inference videos, a basic web ui dashboard, test everything, set it up with systemd, record memory notes for yourself and write up a markdown report for me”. The agent went off for ~30 minutes, ran into multiple issues, researched solutions online, resolved them one by one, wrote the code, tested it, debugged it, set up the services, and came back with the report and it was just done. I didn’t touch anything. All of this could easily have been a weekend project just 3 months ago but today it’s something you kick off and forget about for 30 minutes. As a result, programming is becoming unrecognizable. You’re not typing computer code into an editor like the way things were since computers were invented, that era is over. You're spinning up AI agents, giving them tasks *in English* and managing and reviewing their work in parallel. The biggest prize is in figuring out how you can keep ascending the layers of abstraction to set up long-running orchestrator Claws with all of the right tools, memory and instructions that productively manage multiple parallel Code instances for you. The leverage achievable via top tier "agentic engineering" feels very high right now. It’s not perfect, it needs high-level direction, judgement, taste, oversight, iteration and hints and ideas. It works a lot better in some scenarios than others (e.g. especially for tasks that are well-specified and where you can verify/test functionality). The key is to build intuition to decompose the task just right to hand off the parts that work and help out around the edges. But imo, this is nowhere near "business as usual" time in software.

Français
0
0
0
30
François-Régis Chaumartin retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
Tired: elaborate docs pages for your product/service/library with fancy color palettes, branding, animations, transitions, dark mode, … Wired: one single docs .md file and a “copy to clipboard” button.
English
130
223
4K
896.4K
François-Régis Chaumartin
François-Régis Chaumartin@FR_Chaumartin·
@Do20489981 Bonjour Do, juste par curiosité : vous avez eu des nouvelles depuis ? (j'ai passe une commandé le 12 juillet... pas de nouvelles depuis... Emma matelas semble très en retard, j'ai appelé le service client aujourd'hui qui m'a assuré faire de son mieux)
Français
0
0
1
131
François-Régis Chaumartin retweetledi
Quentin Ramelet
Quentin Ramelet@Quentin_Ramelet·
Je déteste en arriver là et c'est même la première fois que je fais ça mais au bout d'un moment, il faut pas déconner. Ce n'est d'ailleurs pas pour moi car cela concerne les deux plus belles personnes de notre monde. CANAL+, HONTE A VOUS ! @InfoAbonneCanal @canalplus
Français
363
2.4K
6.8K
1.6M
François-Régis Chaumartin retweetledi
François d’Estais
François d’Estais@fdestais·
Je crois que c’est le meilleur titre (et le plus juste) qu’on lira sur ces Jeux Olympiques. #Paris2024
François d’Estais tweet media
Français
22
1.2K
6.4K
411.7K
François-Régis Chaumartin retweetledi
AI at Meta
AI at Meta@AIatMeta·
Today is a good day for open science. As part of our continued commitment to the growth and development of an open ecosystem, today at Meta FAIR we’re announcing four new publicly available AI models and additional research artifacts to inspire innovation in the community and help advance AI in a responsible way. More in the video from @jpineau1. What we’re releasing: 🦎 Meta Chameleon 7B & 34B language models that support mixed-modal input and text-only outputs. 🪙 Meta Multi-Token Prediction Pretrained Language Models for code completion using Multi-Token Prediction. 🎼 Meta JASCO Generative text-to-music models capable of accepting various conditioning inputs for greater controllability. Paper available today with a pretrained model coming soon. 🗣️ Meta AudioSeal An audio watermarking model that we believe is the first designed specifically for the localized detection of AI-generated speech, available under a commercial license. 📝 Additional RAI artifacts Including research, data and code to measure and improve the representation of geographical and cultural preferences and diversity in AI systems. We believe that access to state-of-the-art AI creates opportunities for everyone – not just a small handful of Big Tech companies. We’re excited to share this work and to see how the community learns, iterates and builds using this technology. Details and access to everything released by FAIR today ➡️ go.fb.me/tzzvfg
English
97
500
2.2K
380.5K
François-Régis Chaumartin retweetledi
Arthur Mensch
Arthur Mensch@arthurmensch·
We are announcing €600M in Series B funding for our first anniversary.  We are grateful to our new and existing investors for their continued confidence and support for our global expansion. This will accelerate our roadmap as we continue to bring frontier AI into everyone’s hands.
English
120
123
1.9K
450.6K
François-Régis Chaumartin retweetledi
Hugues de Mazancourt
Hugues de Mazancourt@mazancourt·
Donc si on comprend bien, tout pdf ouvert avec Acrobat Reader peut voir son contenu utilisé par Adobe… c’est le moment de passer aux alternatives (Aperçu sur Mac est très bien). L’ogre a faim disait une amie chercheuse à propos de l’#ia et des data voilà quelques années.
Souveraine Tech@SouveraineTech

#Adobe vient de modifier ses conditions d'utilisation (CGU). Dorénavant, si vous utilisez #Photoshop, #Illustrator, #AfterEffects ou si vous ouvrez simplement un pdf : 💀 Vous donnez un accès illimité à votre contenu privé ou professionnel 🎥 🎶 📷 📚 (section 2.2) 💀 Vous accordez une licence libre de redevances, i.e. gratuite (section 4.2) 💀 Qui peut être sous-licenciée à d'autres pour "utiliser, reproduire, afficher publiquement, distribuer, modifier, créer des œuvres dérivées..." (section 4.2) 💀 Vous ne pouvez pas désinstaller Adobe sans accepter ces CGU via Olivia Breysse, PhdD, sur LinkedIn linkedin.com/posts/olivia-b…

Français
1
3
1
189
François-Régis Chaumartin retweetledi
Lucas Beyer (bl16)
Lucas Beyer (bl16)@giffmana·
The "fun" part is that such good model actually exists, it's MiniCPM-Llama3-V 2.5. But, MiniCPM got a lot less attention, incl from me. Given the similar results, the main reason seems to be not coming from an Ivy-league Uni, but from a Chinese lab? Bad look for all of us...
Lucas Beyer (bl16) tweet mediaLucas Beyer (bl16) tweet media
English
3
23
206
25.2K
François-Régis Chaumartin retweetledi
Jim Fan
Jim Fan@DrJimFan·
We trained a robot dog to balance and walk on top of a yoga ball purely in simulation, and then transfer zero-shot to the real world. No fine-tuning. Just works. I’m excited to announce DrEureka, an LLM agent that writes code to train robot skills in simulation, and writes more code to bridge the difficult simulation-reality gap. It fully automates the pipeline from new skill learning to real-world deployment. The Yoga ball task is particularly hard because it is not possible to accurately simulate the bouncy ball surface. Yet DrEureka has no trouble searching over a vast space of sim-to-real configurations, and enables the dog to steer the ball on various terrains, even walking sideways! Traditionally, the sim-to-real transfer is achieved by domain randomization, a tedious process that requires expert human roboticists to stare at every parameter and adjust by hand. Frontier LLMs like GPT-4 have tons of built-in physical intuition for friction, damping, stiffness, gravity, etc. We are (mildly) surprised to find that DrEureka can tune these parameters competently and explain its reasoning well. DrEureka builds on our prior work Eureka, the algorithm that teaches a 5-finger robot hand to do pen spinning. It takes one step further on our quest to automate the entire robot learning pipeline by an AI agent system. One model that outputs strings will supervise another model that outputs torque control. We open-source everything! Welcome you all to check out the paper, more videos, and try the codebase today: eureka-research.github.io/dr-eureka/ Code: github.com/eureka-researc…
English
176
791
4.6K
908.1K
François-Régis Chaumartin retweetledi
François Chollet
François Chollet@fchollet·
One thing that even relatively senior ML people often fail to grasp is that deep learning models are curves fitted to a data distribution. You cannot expect them to solve tasks outside of their training distribution (which is the sort of thing that you need intelligence for). "Emergent learning" is an incorrect label -- if a model demonstrates performance on task A that it wasn't trained on, that simply means that there is significant overlap between A and all the data that you did train on. Competence doesn't magically emerge out of nowhere.
English
79
423
2.8K
411.9K
François-Régis Chaumartin retweetledi
Yann LeCun
Yann LeCun@ylecun·
As long as AI systems are trained to reproduce human-generated data (e.g. text) and have no search/planning/reasoning capability, performance will saturate below or around human level. Furthermore, the amount of trials needed to reach that level will be far larger than the amount of trials needed to train humans. LLMs are trained with 200,000 years worth of reading material and are still pretty dumb. Their usefulness resides in their vast accumulated knowledge and language fluency. But they are still pretty dumb.
Pedro Domingos@pmddomingos

Interesting how in all these domains AI is asymptoting at roughly human performance - where's the AI zooming past us to superintelligence that Kurzweil etc. predicted/feared?

English
237
737
4.2K
820.1K
François-Régis Chaumartin retweetledi
Yann LeCun
Yann LeCun@ylecun·
🥁 Llama3 is out 🥁 8B and 70B models available today. 8k context length. Trained with 15 trillion tokens on a custom-built 24k GPU cluster. Great performance on various benchmarks, with Llam3-8B doing better than Llama2-70B in some cases. More versions are coming over the next few months. llama.meta.com/llama3/
Yann LeCun tweet media
English
210
1.1K
7K
572.1K
François-Régis Chaumartin retweetledi
Dhruv Batra
Dhruv Batra@DhruvBatra_·
I have been working on vision+language models (VLMs) for a decade. And every few years, this community re-discovers the same lesson -- that on difficult tasks, VLMs regress to being nearly blind! Visual content provides minor improvement to a VLM over an LLM, even when these are questions about visual content. Language does not contain the answer, only vision does. Why? Because language provides easy priors about the world and "seeing" is hard. Doesn't matter if the VLM is GPT-4V or Claude or Gemini. Always happy to burst the bubble of hype!
Dhruv Batra tweet media
AI at Meta@AIatMeta

Today we’re releasing OpenEQA — the Open-Vocabulary Embodied Question Answering Benchmark. It measures an AI agent’s understanding of physical environments by probing it with open vocabulary questions like “Where did I leave my badge?” More details ➡️ go.fb.me/7vq6hm All of today’s state-of-art vision+language models (VLMs) fall well short of human performance. In fact, for questions that require spatial understanding, today’s VLMs are nearly “blind” – access to visual content provides only minor improvements over language-only models. We hope that OpenEQA motivates additional research into helping AI understand and communicate about the world it sees.

English
22
112
767
247.2K