Bruno Neri

3.4K posts

Bruno Neri banner
Bruno Neri

Bruno Neri

@neribr

Technical Leader - Artificial Intelligence and Machine Learning Enthusiast - Senior Software Engineer @altenitalia

Rome, Lazio Katılım Ağustos 2011
2.7K Takip Edilen1.1K Takipçiler
Sabitlenmiş Tweet
Bruno Neri
Bruno Neri@neribr·
If you wanna change the world, start off by making your bed. If you make your bed every morning, you have accomplished the first task of the day and it will encourage you to do another task ... #leadership youtu.be/pxBQLFLei70
YouTube video
YouTube
English
0
0
7
0
Bruno Neri retweetledi
Simone Scardapane
Simone Scardapane@s_scardapane·
*Alice goes to Japan!* The book is available for purchase from Amazon.co.jp! As always, price is almost all printing (except a ☕). 🙃 amazon.co.jp/-/en/微分可能な不思議の国のアリスの冒険-ニューラルネットワーク設計入門書-第1巻-Simone-Scardapane/dp/B0D9QHS5NG/
Simone Scardapane tweet media
English
0
5
33
1.3K
Bruno Neri retweetledi
Petar Veličković
Petar Veličković@PetarV_93·
This paper kicked off our team's studies into the intricate relationship LLMs have with confidence. Now landed in @NatMachIntell 🚀 Give it a read, esp if you enjoy CogSci-style analyses of LLMs 🧠 Thoroughly impressed by Dharsh's leadership on this work! More outputs soon 👀
Petar Veličković tweet media
English
1
10
43
5.2K
Bruno Neri retweetledi
Simone Scardapane
Simone Scardapane@s_scardapane·
*Learning without training: The implicit dynamics of in-context learning* by Dherin et al. An interesting work that tries to explains in-context learning by "contextual" low-rank updates on the MLP components. arxiv.org/abs/2507.16003
Simone Scardapane tweet media
English
0
23
162
9.4K
Bruno Neri retweetledi
Thomas Kipf
Thomas Kipf@tkipf·
There's finally a community implementation of Neural Assets (in PyTorch)! Go check it out 👇 Neural Assets was one of the first (and maybe even the first scalable?) solution(s) to the long-standing problem of multi-entity consistency in visual generative models. One of the most fun projects I had the chance to work on (with the amazing @Dazitu_616).
Mingtian@MingtianZhang

Just released an unofficial PyTorch reimplementation of Neural Assets arxiv.org/pdf/2406.09292. Check it out if you want to build on top of this github.com/Wenlin-Chen/ne…

English
2
32
265
25.8K
Bruno Neri
Bruno Neri@neribr·
If you want to improve your skills about flow matching and diffusion model take a look at this amazing course by @peholderrieth
Peter Holderrieth@peholderrieth

🚀MIT Flow Matching and Diffusion Lecture 2026 Released (diffusion.csail.mit.edu)! We just released our new MIT 2026 course on flow matching and diffusion models! We teach the full stack of modern AI image, video, protein generators - theory and practice. We include: 📺 Videos: Step-by-step derivations. 📝 Notes: Mathematically self-contained lecture notes 💻 Coding: Hands-on exercises for every component We fully improved last years’ iteration and added new topics: latent spaces, diffusion transformers, building language models with discrete diffusion models. Everything is available here: diffusion.csail.mit.edu A huge thanks to Tommi Jaakkola for his support in making this class possible and Ashay Athalye (MIT SOUL) for the incredible production! Was fun to do this with @RShprints! #MachineLearning #GenerativeAI #MIT #DiffusionModels #AI

English
0
0
3
160
Bruno Neri
Bruno Neri@neribr·
"Did you ever play ball, Max?" "No, never have." "But I make it a little more fun to watch, you see." RIP Robert Duvall
Bruno Neri tweet media
English
0
0
0
55
Bruno Neri retweetledi
Ravid Shwartz Ziv
Ravid Shwartz Ziv@ziv_ravid·
Our paper "Attention Sinks and Compression Valleys in LLMs are Two Sides of the Same Coin" has been accepted to ICLR 🎉 This work connects two puzzling phenomena in LLMs that were previously studied separately: attention sinks and compression valleys, showing they're fundamentally linked through massive activations. We propose a three-stage theory of how LLMs structure computation across depth, which also helps explain why intermediate layers work best for embedding tasks while generation needs full-depth processing. Great collaboration with @arroyo_alvr, Enrique Q, @fedzbar, @epomqo, @mmbronstein, and @ylecun.
Alvaro Arroyo@arroyo_alvr

🚨 How do attention sinks relate to information flow in LLMs? We show how massive activations create attention sinks and compression valleys, revealing a three-stage theory of information flow in LLMs. 🧵 w/ Enrique* @fedzbar @epomqo @mmbronstein @ylecun @ziv_ravid

English
10
15
102
9.5K
Bruno Neri retweetledi
Mathelirium
Mathelirium@mathelirium·
The Elusive Concept of Time. Your instincts treat time like a background meter the whole universe shares. Relativity does not take time away, it forces you to earn it operationally. Events are points. Motion is a curve through them. A clock is not a metaphor, it is a worldline with a number attached to it. Two observers can disagree on which distant events are simultaneous, and nothing contradictory happens, because causal structure is still pinned down by light cones and invariants. Even in weak gravity the rule shows up. A clock deeper in a gravitational potential ticks more slowly relative to one far away. Near a compact object the difference becomes hard to ignore. Add rotation and spacetime itself picks up a twist. That twist is frame dragging, not a new force, just geometry telling you that time and angle are coupled in a rotating spacetime. In the animation You are watching a geometry lesson disguised as a black hole scene. The fabric is a visualization of the field shaping clock-rates and the paths light can take. The ripples are driven by local proper time, so their phase visibly slows as you approach the horizon. The accretion disk is lensed through Kerr ray tracing, and its brightness is pushed by redshift and beaming so the approaching side can flare while the receding side dims. Beacon points at different radii pulse at different rates, so you can see time dilation without any labels. The bead ring is a redshift tracer, with intensity scaled by a g³ proxy so deeper emission arrives weaker and shifted. The math breakdown Start with what a clock actually measures. Proper time τ is the accumulated time along an observer’s worldline. In special relativity, the invariant interval is ds² = c² dt² − dx² − dy² − dz² Along a timelike path, dτ = (1/c) √(ds²) = √( dt² − (1/c²)(dx²+dy²+dz²) ) If the observer moves with speed v, so dx²+dy²+dy²+dz² = v² dt², then dτ = dt √(1 − v²/c²) That is time dilation as geometry. The moving clock accumulates less τ between the same pair of events. Now add gravity. General relativity replaces the flat interval with a metric gᵤᵥ that depends on position: ds² = gᵤᵥ dxᵘ dxᵛ For a stationary clock in Schwarzschild geometry (mass M), the time component is g_tt = −(1 − 2GM/(rc²)) If the clock sits at fixed r (no spatial motion), ds² = g_tt c² dt², so dτ = dt √(1 − 2GM/(rc²)) Closer to the mass means a smaller factor, so the clock ticks more slowly relative to a clock far away. That is the rule used to drive the fabric phase in the animation. Now connect time to light. A gravitational field shifts photon frequency. Between an emitter at rₑ and an observer at rₒ, f_obs / f_emit = √( (1 − 2GM/(rₑ c²)) / (1 − 2GM/(rₒ c²)) ) For a far-away observer rₒ → ∞, f_obs / f_emit = √(1 − 2GM/(rₑ c²)) Deeper emission arrives redshifted. Lower frequency. Lower energy per photon. In the render, the disk intensity uses a Kerr-derived redshift factor g (clipped for stability). The bead ring uses a simple radiative proxy I_obs ∝ g³ I_emit to make that effect visible. Finally, why rotation looks like a twist. A rotating black hole is Kerr geometry. The key structural change is a nonzero g_tφ term, which couples time to angle. That coupling is frame dragging in equations. Near the hole, being stationary is not the same notion everywhere, because the local inertial frames are being pulled around the spin axis. So the moral stays clean. Time is not a universal substance flowing everywhere at one rate. It is what clocks accumulate along worldlines. Light cones constrain what can influence what. Invariants are what everyone agrees on. The rest is operational detail that only feels universal because our daily corner of the universe is slow and mild. #GeneralRelativity #Gravity #FrameDragging #BlackHoles #Spacetime
English
107
641
3.1K
149K
Bruno Neri
Bruno Neri@neribr·
At St. Peter Basilica
Bruno Neri tweet mediaBruno Neri tweet mediaBruno Neri tweet media
English
0
0
0
61
Bruno Neri retweetledi
Petar Veličković
Petar Veličković@PetarV_93·
for the first time, eeml treads mountainous terrain 🏔️ while being very close to the seaside 🌊 not to mention it's the best AI summer school in EE & beyond! come join us, it'll be amazing! @clarelyle @dorothychou @ninamiolane @sedielem @apsarathchandar @re_rayne @fedzbar
EEML@EEMLcommunity

📢 EEML2026 🇲🇪 is now accepting applications! Check the website for instructions and see the amazing speakers confirmed so far (links in thread). Deadline for applications: March 31, 2026! Join us in beautiful Cetinje! 🎉🇲🇪⛰️

English
0
9
45
8.6K