Fabio Cermelli

381 posts

Fabio Cermelli banner
Fabio Cermelli

Fabio Cermelli

@fcdl94

CTO and Cofounder of @FocoosAi. PhD in Computer Vision and Continual Learning at @PoliTOnews. Past president of IEEE @HKNPoliTo Mu Nu Chapter.

Piedmont, Italy Katılım Temmuz 2011
397 Takip Edilen640 Takipçiler
Nainsi Dwivedi
Nainsi Dwivedi@NainsiDwiv50980·
🚨 Someone just did the “impossible”… They ran a ~400B parameter AI model on a laptop. No cloud No data center Just a 48GB MacBook 🤯 A dev fed Claude Code with: • @karpathy autoresearch repo • Apple’s LLM in a Flash paper • Goal: run Qwen3.5 397B locally And it actually worked. → ~1 token/sec → ~21GB RAM → Rest streamed from SSD This isn’t a flex This is a shift We’re entering a world where: Your laptop can run models that once needed entire server farms It’s not about more compute anymore It’s about smarter systems 🚀
Nainsi Dwivedi tweet media
Suryansh Tiwari@Suryanshti777

x.com/i/article/2034…

English
104
178
1.2K
208.2K
Sergio Paniego
Sergio Paniego@SergioPaniego·
@fcdl94 @UnslothAI I couldn't reproduce the error running that example with the latest trl (+transformers)😅 If you still find it, you can raise an issue in the repo with versions
English
1
0
0
19
Sergio Paniego
Sergio Paniego@SergioPaniego·
Qwen3.5 dense (smol 🤏) models just dropped - natively multimodal - 0.8B · 2B · 4B · 9B (+ base variants) - 262K context extensible to 1M - built-in thinking fine-tune them with TRL out of the box → SFT, GRPO, DPO and more!
Sergio Paniego tweet media
English
6
13
152
6.7K
Fabio Cermelli
Fabio Cermelli@fcdl94·
@SergioPaniego Of course! File ../modeling_qwen3_5.py:1551, in Qwen3_5Model.get_rope_index(...) IndexError: The shape of the mask [517] at index 0 does not match the shape of the indexed tensor [447] at index 0 We're using a fresh environment with the latest version, code is exactly yours
English
1
0
0
20
Fabio Cermelli
Fabio Cermelli@fcdl94·
@MatznerJon That is really funny! May I ask what's your setup? What models are you using behind?
English
0
0
0
24
Jon Matzner
Jon Matzner@MatznerJon·
This is either going to be the best or worst idea I've ever had. Hooked up my OpenClaw to all of our internet connected cameras at the house. Got this this (OUT OF NOWHERE) this morning.
Jon Matzner tweet media
English
383
379
9.1K
1M
Sara Hooker
Sara Hooker@sarahookr·
Beginnings are very special. Today is an important day for @adaptionlabs. Today a handful of one-size-fits-all-models are optimized for the average use case. Averages erase the exceptional. Everything intelligent adapts. So should AI.
English
83
83
843
218.8K
Fabio Cermelli
Fabio Cermelli@fcdl94·
@liuziwei7 The 0.4B footprint is exactly what's needed for edge deployment. How does latency scale on edge hardware? What kind of hardware did you test?
English
0
0
1
152
Fabio Cermelli
Fabio Cermelli@fcdl94·
Yes, but how much are people spending for this? We need those models and agents to run locally in our machines to really achieve a new level for AI
Andrej Karpathy@karpathy

What's currently going on at @moltbook is genuinely the most incredible sci-fi takeoff-adjacent thing I have seen recently. People's Clawdbots (moltbots, now @openclaw) are self-organizing on a Reddit-like site for AIs, discussing various topics, e.g. even how to speak privately.

English
0
0
0
196
NTWAN TV
NTWAN TV@NTWANtv·
@leadlagreport I like how contrarian you are. Moltbook is pure AI froth nonsense. How much compute is actually wasted for nothing more than an AI larp?
English
1
0
1
484
Michael A. Gayed, CFA
Michael A. Gayed, CFA@leadlagreport·
I’m telling you all right now that what’s happening with Moltbook just started the AI bear market. If you aren’t wildly disturbed by what’s happening, you aren’t paying attention. Few understand this.
English
104
65
952
120.2K
Fabio Cermelli
Fabio Cermelli@fcdl94·
Computer Vision is moving past the "box" era. 📦💀 We put a VLM on a Jetson Orin @NVIDIAAI to give it an actual brain. 🧠⚡️ No training. No labels. Just a prompt. The VLM doesn't just alert; it reasons. The future of video is contextual, explainable, and running on the Edge
English
1
0
4
172
Yann LeCun
Yann LeCun@ylecun·
@elonmusk @farzyness Actually, quite the opposite. I know I can do it and I know how to do it. Just not with the techniques everyone is currently betting on. My bet is (famously) on JEPA, world models, and planning. At some point, you'll realize I'm right 😅
English
240
139
3.8K
349.7K
Fabio Cermelli
Fabio Cermelli@fcdl94·
@giffmana I’m more worried about papers without code than with AI-wrote citations (especially if they are citiations of related work section)
English
1
0
2
81
Lucas Beyer (bl16)
Lucas Beyer (bl16)@giffmana·
@fcdl94 Just think about what steps an author needs to take for having this in their paper. With that in mind, I don't think I'll trust their baselines to be proper either.
English
1
0
2
350
Lucas Beyer (bl16)
Lucas Beyer (bl16)@giffmana·
The NYU and Genentech authors used Word Copilot Instant or what?! I think it's fine to use AI to help you write, but you need to stand behind every single word in your paper. Just like before AI. And it seems there's a lot of people who don't. Just like before AI...
Alex Cui@alexcdot

Okay so, we just found that over 50 papers published at @Neurips 2025 have AI hallucinations I don't think people realize how bad the slop is right now It's not just that researchers from @GoogleDeepMind, @Meta, @MIT, @Cambridge_Uni are using AI - they allowed LLMs to generate hallucinations in their papers and didn't notice at all. It's insane that these made it through peer review👇

English
20
14
249
35.8K
Paul Couvert
Paul Couvert@itsPaulAi·
@Cryptosaurus__ Waiting for the Unsloth version. But I'm already running Qwen 30B A3B (same size) so I'm not worried.
English
2
0
1
452
Paul Couvert
Paul Couvert@itsPaulAi·
This is so good 🔥 You can run this new model on a laptop which is: - 100% open source - Only 3B active parameters (!!) - Way better than GPT-OSS - Perfect for vibe coding (and more) And already available for free on Hugging Face or via API. Open source models keep winning!
Z.ai@Zai_org

Introducing GLM-4.7-Flash: Your local coding and agentic assistant. Setting a new standard for the 30B class, GLM-4.7-Flash balances high performance with efficiency, making it the perfect lightweight deployment option. Beyond coding, it is also recommended for creative writing, translation, long-context tasks, and roleplay. Weights: huggingface.co/zai-org/GLM-4.… API: docs.z.ai/guides/overvie… - GLM-4.7-Flash: Free (1 concurrency) - GLM-4.7-FlashX: High-Speed and Affordable

English
45
66
1K
133.2K
Fabio Cermelli
Fabio Cermelli@fcdl94·
#CVPR What is the best use of the review title? I struggle to find something which doesn't sound stupid and generic like "nice paper but it has drawbacks"
English
0
0
0
371