
Carlo Daffara
25.7K posts

Carlo Daffara
@cdaffara
NodeWeaver CEO and cofounder. Proud father, engineer, entrepreneur. Loves technology, cooking, IT economics.
Udine, London, Brussels Katılım Nisan 2008
173 Takip Edilen1.3K Takipçiler
Carlo Daffara retweetledi

The final report on the Iberia Peninsula (Spain and Portugal) blackout is out.
A lot of people will be blaming renewables and talk about inertia. But the cause was bad voltage control and that's surprisingly easy to fix. Let me explain.
#Publications_&_Documents" target="_blank" rel="nofollow noopener">entsoe.eu/publications/b…
English
Carlo Daffara retweetledi
Carlo Daffara retweetledi
Carlo Daffara retweetledi
Carlo Daffara retweetledi

Inspired by @__tinygrad__ userspace AMD driver, I clauded a userspace driver for some stress testing of SDMA and compute/comms overlap debug.
I didn't open the editor once. Agents are the great equalizer in software. And Speed is the moat.
github.com/ROCm/TheRock/t…
English

@SG_Posters The most beautiful images I had on my timeline this week. What a wonderful way of composing space and light.
English
Carlo Daffara retweetledi

@balajis export controlling NVIDIA was the nail in the coffin. we can only ship our top tinybox to a small set of whitelisted countries. in 2-3 years, we'll be shipping tinyboxes full of Chinese chips.
English
Carlo Daffara retweetledi
Carlo Daffara retweetledi

Our experiment stopped producing useful data.
Basically, participants increasingly didn't want to work w/out AI, even at $50/hr, so the sample drifted toward tasks where AI access doesn't matter. This biases our estimate downward. The pilot's 20% slowdown result is no longer valid and shouldn't be cited. The world has since changed and so has our understanding.
I think there are roughly two wrong takeaways:
1. "the experiment failed, therefore AI's effect is enormous and unmeasurable." - way too strong, experiments can fail for boring reasons too.
2. "the experiment failed, so we've learned nothing." I think that's also wrong, the way you fail can be quite informative, even if I'd be cautious about how much weight to put on it.
In Khalid's personal opinion, effect is prob positive and we're prob underestimating it. METR is working on better ways to answer this...but plz,"METR's experiment broke b/c people are wayyy too sped up by AI" is convenient but not what we're saying.
METR@METR_Evals
Since early 2025, we've been studying how AI tools impact productivity among developers. Previously, we found a 20% slowdown. That finding is now outdated. Speedups now seem likely, but changes in developer behavior make our new results unreliable. We’re working to address this.
English

Carlo Daffara retweetledi



@aakashgupta Misdirection... and false alarmism for some likes...
The actual number is 3.61‱ (using the per-ten-thousand symbol ‱, also known as "permyriad").
This equates to 3.61 per 10,000 CPUs (or 0.0361%).

English

The scariest number here: 3.61% of CPUs in one large-scale study were found to cause silent data corruptions. Not “a few bad chips.” Nearly 4 out of every 100 processors doing math wrong, silently, with no error log.
Google coined the term “mercurial cores” in 2021 after their production teams kept blaming software for data corruption. They’d debug for weeks, find nothing wrong with the code, swap the machine, problem gone. The actual cause: manufacturing defects at sub-7nm that pass every factory test, then degrade unpredictably months or years after deployment.
Facebook confirmed the same thing independently. Hundreds of affected CPUs across hundreds of thousands of machines. The defect doesn’t crash your system. It just gives you 5 instead of 6 when you multiply 2x3, under specific microarchitectural conditions, with zero indication anything went wrong.
Now think about what this means for AI training. A single corrupted GPU or CPU in a distributed training cluster doesn’t just produce one bad output. It feeds corrupted gradients into a synchronization step that gets averaged across every accelerator in the cluster. One bad chip can silently poison an entire training run. NVIDIA published a whitepaper on exactly this problem. Loss spikes during LLM training that nobody could explain traced back to silent hardware corruption.
The part that keeps infrastructure engineers up at night: traditional defenses don’t work. ECC memory can’t catch this because the corruption happens during computation, not storage. Checksums like CRC heavily use vector operations, which are themselves one of the most vulnerable instruction types. The tools designed to detect corruption are running on the same flawed silicon.
Google’s current detection method? Roughly half human-driven, half automated. And of the machines humans flag as suspicious, only about 50% are actually confirmed mercurial on deeper investigation. We’re debugging trillion-parameter models on hardware where we can’t reliably tell which chips are lying to us.
Moore’s Law gave us more transistors. It also gave us transistors we can’t fully verify.
LaurieWired@lauriewired
CPUs are getting worse. We’ve pushed the silicon so hard that silent data corruptions (SDCs) are no longer a theoretical problem. Mercurial Cores are terrifying because they don’t hard-fail; they produce rare, but *incorrect* computations!
English
Carlo Daffara retweetledi

Kimi K2.5 tech report just dropped!
Quick hits:
- Joint text–vision training: pretrained with 15T vision-text tokens, zero-vision SFT (text-only) to activate visual reasoning
- Agent Swarm + PARL: dynamically orchestrated parallel sub-agents, up to 4.5× lower latency, 78.4% on BrowseComp
- MoonViT-3D: a unified image–video encoder with 4× temporal compression, enabling 4× longer videos in the same context
- Toggle: token-efficient RL, 25–30% fewer tokens with no accuracy drop
Here's our work toward scalable, real-world agentic intelligence. More details in the report 👉github.com/MoonshotAI/Kim…




English

@xai's Grok 4.20 along with @Cassi_on_X are closest to the top human superforecasters. Impressive showing of LLMs in general. forecastingresearch.substack.com/p/llms-are-clo…

English
Carlo Daffara retweetledi

Wheels up.
pip install vllm==0.14.0+rocm700 --extra-index-url wheels.vllm.ai/rocm/0.14.0/ro…
English

@0t0m001 @ani_obsessive Apart from the interview I linked? (if you understand Italian, that is). He mentions they were part of Studio Pagot. You can check on this the wonderful book "The Art of Pagot - La meravigliosa storia della famiglia Pagot e dei suoi eroi in carta e inchiostro"
English

@cdaffara @ani_obsessive I don't question that Gi and Marco Pagot pitched the idea, I'd like the receipts that there was a "italian studio" involved in some way, rather than those two creators pitching it to RAI and RAI cooperating with TMS. There's a difference between individuals and a studio.
English

Sherlock Hound is a gem. Hayao Miyazaki built an idea from an Italian studio into a brilliant adventure series.
It was a turning point in his career, and his last TV show. Behind it were clashing visions, young talent and lots of money, as we explore:
─➤animationobsessive.substack.com/p/miyazakis-sh…
English

@0t0m001 @ani_obsessive So yes-Studio Pagot implemented it, with funding from RAI. Miyazaki worked only on the first six episodes in until copyright problems with Conan Doyle's estate made him leave, the rest of episodes directed by Kyōsuke Mikuriya (Lupin the 3rd part II, Ulysse 31)
English

@0t0m001 @ani_obsessive Idea/dev/char design was by Gi and Marco Pagot, along with Tokyo Movie Shinsha. RAI's Luciano Scaffa had the idea of a Italy-Japan coproduction, to stop being pure anime consumers. Studio Pagot pitched the idea which got greenlighted. interviewed here: teche.rai.it/2025/11/marco-…
English














