Jean Ibarz

7.2K posts

Jean Ibarz banner
Jean Ibarz

Jean Ibarz

@_ibarz

https://t.co/DbyKd5baXF https://t.co/6GQLeVtmHm https://t.co/mMoG2Vc9pQ

France Katılım Mayıs 2014
415 Takip Edilen132 Takipçiler
Jean Ibarz
Jean Ibarz@_ibarz·
Qwen 3.6 35B A3B is great. Today might be the takeoff of local models ! ❤️
English
0
0
0
34
0dteezy
0dteezy@0dteezy·
@sudoingX Aint no way this is fitting on a 3090 with that context
English
1
0
0
249
Sudo su
Sudo su@sudoingX·
85-100 tok/s on the 3090 with qwen 3.6 already? that's in line with what 3.5 MoE was doing. drop your full flags and context length you tested at, i'm pulling 3.6 on the 5090 24gb and will run the same config for a direct comparison. if anyone else is running qwen 3.6 on a 3090 or any consumer card drop your tok/s, quant, and flags below. building the community benchmark sheet before i publish my own numbers
Jacob Verdoorn@VerdoornJacob

@sudoingX 3090 getting 85-100 t/s on cpp server with new qwen3.6 35b a3b ud q4 k m 262k context

English
43
2
206
15.5K
Jean Ibarz
Jean Ibarz@_ibarz·
@stevibe @ivanfioravanti I tried quantized model from unsloth: didn't work at all. I than tried quantized model from abiray: works great on chat conversation and via cline through LM Studio server.
English
0
0
3
173
stevibe
stevibe@stevibe·
Qwen3.6 35B-A3B: smarter, but forgot how to use tools? Running 6 Bench Packs on BenchLocal across 3 open-source Qwen models. ✅ ReasonMath: 92 vs 85 vs 86 — 3.6 wins ✅ InstructFollow: 97 / 97 / 97 — tied ❌ ToolCall: 83 vs 97 vs 100 — 3.6 tanks Qwen3.5 27B still the tool-calling champ. 3.6 clearly leveled up reasoning, but tool use took a hit. DataExtract live now. BugFind + StructOutput next.
stevibe tweet mediastevibe tweet mediastevibe tweet mediastevibe tweet media
English
31
26
310
22.4K
Jean Ibarz retweetledi
Qwen
Qwen@Alibaba_Qwen·
⚡ Meet Qwen3.6-35B-A3B:Now Open-Source!🚀🚀 A sparse MoE model, 35B total params, 3B active. Apache 2.0 license. 🔥 Agentic coding on par with models 10x its active size 📷 Strong multimodal perception and reasoning ability 🧠 Multimodal thinking + non-thinking modes Efficient. Powerful. Versatile. Try it now👇 Blog:qwen.ai/blog?id=qwen3.… Qwen Studio:chat.qwen.ai HuggingFace:huggingface.co/Qwen/Qwen3.6-3… ModelScope:modelscope.cn/models/Qwen/Qw… API(‘Qwen3.6-Flash’ on Model Studio):Coming soon~ Stay tuned
Qwen tweet media
English
372
1.3K
9.2K
1.6M
Jean Ibarz
Jean Ibarz@_ibarz·
On produit la même richesse, en y consacrant beaucoup moins de temps de vie humaine. Oui, les informaticiens dont je fais partie contribuent à la suppression de leur propre emploi, et c’est un bénéfice pour l’humanité. Il n’y a aucun intérêt à faire un travail qui peut être automatisé et réalisé par une machine à moindre coût. Si vous voulez payer des gens pour effectuer un travail qui ne produit rien d’utile pour la société, libre à vous de le faire, mais il n’y a aucun intérêt à les obliger à travailler.
Français
0
0
0
252
François Ruffin
François Ruffin@Francois_Ruffin·
Les informaticiens seront-ils remplacés par l'IA ? L'IA Claude répond : « Ils ont construit l'outil qui les remplace. Collectivement, les informaticiens ont construit quelque chose qui se retourne contre eux. » Voir en intégralité sur ma chaîne YouTube : youtu.be/OSAHTpRaKtw Sur une idée de Bernie Sanders.
YouTube video
YouTube
Français
101
51
207
99.1K
Jean Ibarz
Jean Ibarz@_ibarz·
@AlexXplore La délinquance et l'injustice galopante dans notre pays ne crée aucune richesse, mais elle crée au moins des emplois.
Français
0
0
1
170
Alex Xplore
Alex Xplore@AlexXplore·
🇫🇷 Les actes de malveillance, incluant vols de cuivre, vandalisme et sabotages, ont causé plus de 800 000 minutes de retard et impacté près de 40 000 trains en 2024. 📡 La SNCF a une flotte de 200 drones fabriqués par la start-up française Delair pour surveiller ses 28 000 km de voies ferrées 24h/24, notamment la nuit avec des caméras infrarouges. 👀 Ces drones permettent de détecter les personnes en flagrant délit et ont déjà contribué au démantèlement de plusieurs réseaux criminels. 🛡️ L’objectif affiché est de « dissuader l’acte quel qu’il soit » grâce à une surveillance dissuasive et une intervention rapide de la police ferroviaire. 💰 Cette contre-attaque high-tech s’inscrit dans un budget annuel de 100 millions d’euros dédié à la protection du réseau ferré français. youtube.com/watch?v=4J2-fE…
YouTube video
YouTube
Français
26
243
1K
80.7K
Jean Ibarz retweetledi
Ivan Burazin
Ivan Burazin@ivanburazin·
Dylan Patel says GPUs are no longer the biggest bottleneck. According to @dylan522p, now CPUs are the constraint. In the early AI era, CPUs were the laggers. You used them for storage, checkpointing, pre-processing, etc. (pretty light workloads) The models weren't agentic and couldn't go step by step. Just string in and string out (simple inference) Then OpenAI launched O1 preview in September '24, and RL training loops have since tightened every month. - initially it was checking model output with regex - then running classifiers - followed by code unit tests + compilation - and finally agentic flows calling databases & scientific simulations The model outputs to an environment, gets verified, and trains on it. Coding agent revenue went from a couple billion to north of $10B in roughly 6 months. Something like Codex 5.4 can work agentically on its own for 6-7 hrs straight - doing all sorts of calls (databases, cron servers, scraping) That requires insane CPU capabilities. And over the last two quarters, the entire cloud market ran out of CPUs. - GitHub has been really unstable lately - Amazon's CPU server installations 3x'd year over year - Microsoft sold all of its spare CPUs to Anthropic & OpenAI Earlier, it was 100 megawatts of GPUs served by 1 megawatt of CPUs. Now that ratio is getting much closer for both RL training and agentic inference. There's simply no capacity anywhere, and it's causing massive instability.
English
62
106
1.1K
297.6K
Jean Ibarz
Jean Ibarz@_ibarz·
@floor_per_area @AlecStapp Per article XIV, page 203 of the second amendment, structures connected via underground passage are to be counted as one building. Source: trust me, but also maybe don’t.
English
0
0
3
688
Chris Goldammer
Chris Goldammer@floor_per_area·
@AlecStapp In high demand: Duplex houses of 149sq meter each, possibly connected by a tunnel.
English
5
4
1.2K
74.7K
Alec Stapp
Alec Stapp@AlecStapp·
Government regulation in France: Above a certain size, building new homes requires a licensed architect. Outcome:
Alec Stapp tweet media
English
168
1.3K
16.4K
2.1M
Jean Ibarz retweetledi
Julian Ibarz
Julian Ibarz@julianibarz·
Big news from Europe for Tesla!
Sawyer Merritt@SawyerMerritt

NEWS: Dutch regulators (RDW), which just approved @Tesla FSD (Supervised) in the Netherlands, have just issued an official statement: "Due to the continuous strict monitoring of the driver in the vehicle, the system is safer than other driver assistance systems. We have thoroughly researched and checked this system, more than a year and a half. The RDW has issued a type approval for Tesla's driver's assistance system, FSD Supervised. This driver's assistance system has been extensively researched and tested on our test track and on public roads for more than a half years. Safety is paramount for the RDW. The proper use of this driver's system makes a positive contribution to road safety." This approval from the RDW clears the path for approval in other European countries. Tesla owners in the Netherlands will be receiving FSD (Supervised) on their cars shortly. Amazing day!

English
4
5
187
6.3K
JO
JO@appealstoheaven·
@_ibarz @thsottiaux That makes sense. Especially the training point. In my experience when I was a pro subscriber they had an option in settings to say you don’t want your data used for training. Is it trustworthy? I’m not sure
English
1
0
1
10
JO
JO@appealstoheaven·
@_ibarz @thsottiaux So can you not just use the Pro plan if the business plan isn’t sufficient? Are businesses not allowed to use the plans individuals use? Set up an “IT” user at your company or something and use that to sub to the pro plan
English
1
0
0
13
Tibo
Tibo@thsottiaux·
I get so many personal thank you emails for something and it’s not even a feature. All we needed apparently was a new codex plan. Can we rest now?
Tibo tweet media
English
187
18
1.6K
112.2K
David
David@davicorn·
@thsottiaux Would love some clarification on what multiplier the $200/mo plan is, to make a better decision :) Is it also 5x and then 20x like the competitor?
English
2
0
1
779
CyberSecDev
CyberSecDev@cybersecdev_·
@heeney_luke Same here , after the receipt of the “Updates to codex credit pricing “ mail, burnt through my 5hrs limit in 20mins
English
1
0
4
1.6K
Luke Heeney
Luke Heeney@heeney_luke·
did codex just change their 5 hour limits? I am suddenly burning through it in an hour after never coming close before. Gah, usage limits are why I use it over Claude!
English
116
16
857
95.3K
Dwayne
Dwayne@CtrlAltDwayne·
@heeney_luke What plan are you on? I'm on the $200 Pro plan and haven't noticed any change in session limits. I've been using xhigh reasoning for the past couple of hours and have 93% remaining on my 5 hour limit.
English
1
0
4
1.3K
Jean Ibarz
Jean Ibarz@_ibarz·
@heeney_luke They change the rates to match API pricing, that's why. Now business plans have 4x less quotas than ChatGPT Plus plan
English
0
0
1
705
Therealsteal
Therealsteal@Therealsteal01·
@thsottiaux Why did you guys completely nuke my 5hr limit window? 2 tasks now consume 40% of usage??? Insane
English
3
0
5
518
Jean Ibarz
Jean Ibarz@_ibarz·
@thsottiaux I just noticed why we hit limits in no time with business plans: you changed pricing to align with token pricing, now since 2 April business plans have much less quotas than chatgpt plus users, that's a very low blow. Not thank you👎
English
0
0
6
253