Itay Evron

528 posts

Itay Evron banner
Itay Evron

Itay Evron

@itayevron

PhD; Research Scientist @Meta (opinions are my own)

Katılım Temmuz 2018
397 Takip Edilen821 Takipçiler
Sabitlenmiş Tweet
Itay Evron
Itay Evron@itayevron·
Why does L1 regularization induce sparse models? Many illustrate this using the least squares problem with a norm constraint. The least squares level sets are drawn next to the different unit "circles". I prepared a cool animation which I believe makes it even clearer 🙂
GIF
English
11
226
1.3K
0
Théo Vincent
Théo Vincent@Theo_Vincent_·
While reviewing for ICML/RLC, I was surprised to find submissions citing the same work twice with slightly different BibTeXs🤔 It is always nice to see authors giving double credit to the works they rely on!👌 ...or it is an artifact of an LLM helper😅
Théo Vincent tweet media
English
5
0
16
2.1K
Itay Evron retweetledi
Adi Simhi
Adi Simhi@AdiSimhi·
How does an LLM’s past influence its future?🤔 In our new paper with @FazlBarez,@mtutek,@boknilev, Shay Cohen, we show that conversational history creates a "geometric trap" in the latent space, confining the model’s trajectory➡️making old habits e.g. hallucinations hard to break
Adi Simhi tweet media
English
1
15
99
10.6K
Itay Evron retweetledi
Boris Hanin
Boris Hanin@BorisHanin·
🚨 2026 @Princeton ML Theory Summer School Mini-courses by: - Subhabrata Sen @subhabratasen90 - Lenaic Chizat @LenaicChizat - Sinho Chewi - Elliot Paquette @poseypaquet - Elad Hazan @HazanPrinceton - Surya Ganguli @SuryaGanguli August 3 - 14, 2026 Apply by March 31. Link 👇 Sponsors: @NSF, @PrincetonAInews, @EPrinceton @JaneStreetGroup, @DARPA, @PrincetonPLI, Princeton NAM, Princeton AI2, Princeton PACM Some amazing speakers from previous years: @Andrea__M, @TheodorMisiakie, @KrzakalaF, @_brloureiro, @rakhlin, @DimaKrotov, @CPehlevan, @SoledadVillar5, @SebastienBubeck, @tengyuma
English
3
37
421
43.4K
Tsoof Bar Or
Tsoof Bar Or@tsoofbaror·
כבר הרבה זמן שהמטריקה היחידה שאני עוקב אחריה באדיקות היא זו. המודלים הופכים לחכמים יותר, זה נכון - ובמקביל לחרטטנים יותר.
Tsoof Bar Or tweet media
Amit Mandelbaum@Amit_Mandelbaum

עדיין בשוק כמה ג'מיני מחרטט למוות ביחס למודלים אחרים אפילו במודלים הכי חזקים שלו ואפילו כשאני ליטרלי מבקש מקור לכל דבר שהוא כותב.. זה לא קורה בקלוד או ב ChatGPT. ג'מיני פשוט ממציא, בלי הפסקה.

עברית
10
1
79
12.5K
Itay Evron
Itay Evron@itayevron·
Please rewrite your proof, fully rigorously. - I can’t honestly rewrite that previous document into a fully rigorous proof because the core mechanism I relied on does not hold in the way the draft claimed.
English
1
0
5
781
Itay Evron retweetledi
Lorenzo Perini
Lorenzo Perini@LorenzoPerini95·
1/6 🧵 Calibration is hard. Multicalibration—fixing errors across every possible subgroup—is usually impossible at scale. Until now. Introducing MCGrad: A production-ready multicalibration library from Meta, accepted at KDD 2026. 🚀 github.com/facebookincuba…
English
1
1
6
916
Itay Evron retweetledi
Maya Bechler-Speicher
Maya Bechler-Speicher@mayabechlerspei·
A glimpse into the research I’ve been leading over the past year at Meta 🥹 . So many organizations own rich graphs that remain largely underutilized. GraphBFF shows how to build feasible, powerful Graph Foundation Models from these graphs, end to end, from data curation and modeling choices to production. We rely on real data, and solve real problems, no toy setups, just what it actually takes to make a Graph Foundation Model work in practice. This has been a life-changing experience for me, taking something from an idea all the way to a deployed GFM that is now having real impact at Meta. The preprint is now available on arxiv.
Maya Bechler-Speicher tweet media
English
7
9
86
4.3K
Itay Evron
Itay Evron@itayevron·
@miniapeur Higher-Order Learning Dynamics on Cellular Complexes M Alain, Terrence Tao, Yoshua Bengio Annals of Mathematics, 2026
English
1
0
1
40
Itay Evron
Itay Evron@itayevron·
An LLM hallucinating a citation with your name on it is just the universe telling you to write that paper
English
1
0
12
481
Itay Evron retweetledi
Guy Bar-Shalom
Guy Bar-Shalom@GuyBarSh·
📌 [1/4] A Graph Meta-Network for Learning on Kolmogorov-Arnold Networks We introduce a weight-space model for KANs, where learning happens directly over the KANs' 1D functions. This work was done during my Meta internship. openreview.net/pdf?id=ONpyYav…
English
1
3
7
565
Itay Evron
Itay Evron@itayevron·
@boknilev @HebAcademy ועל כך כתוב בספרי זוטא: תמהני עליך יונתן שאמרת דבר זה
עברית
0
0
1
26
Aryeh Kontorovich
Aryeh Kontorovich@aryehazan·
I'm submitting 400 icml papers wish me luck
English
6
0
133
14.2K
Francesco Orabona
Francesco Orabona@bremen79·
ICLR decisions are being rolled out by ID number
English
1
0
41
8.8K
Tsoof Bar Or
Tsoof Bar Or@tsoofbaror·
הAI יחליף את כולנו? לא יודע מה איתכם, ככל שהAI ״מחליף אותי״ יותר אני עובד יותר שעות.
עברית
19
1
188
14.7K
Itay Evron retweetledi
RL Theory Virtual Seminars
We are back. For the first presentation of the year, Uri will present his recent work on agnostic reinforcement learning. See you next Tuesday!
RL Theory Virtual Seminars tweet media
English
0
3
9
1.1K
Itay Evron
Itay Evron@itayevron·
(This is, of course, a simplified statement which refers to the *worst case*)
English
0
0
0
96
Itay Evron
Itay Evron@itayevron·
One of my papers I'm especially fond of, now accepted to ALT2026. 🥳 A question kept me busy for a few years: Do continual linear models under random task orderings converge more slowly in high dimension? By reducing this problem to stepwise-optimal SGD, we show they do not! pic.twitter.com/iiFzqlHb1M
Itay Evron@itayevron

In continual learning of linear models random task orderings diminish forgetting even in high dimensions! Better Rates for Random Task Orderings in Continual Linear Models Evron*, @ranlevinstein*, @MatanSchliserm1*, Sherman*, Koren, @soudry_daniel, Srebro arxiv.org/abs/2504.04579

English
2
0
4
324
Itay Evron retweetledi
Shai Shalev-Shwartz
Shai Shalev-Shwartz@shai_s_shwartz·
In memory of my father, an educator who loved the Mishna. One teaching he cherished describes four kinds of students sitting before the sages as "tools". Adapted to modern terms, these are: A sponge, a funnel, a sifter, and a colander.
Shai Shalev-Shwartz tweet media
English
1
5
31
1.7K
Itay Evron retweetledi
Maya Bechler-Speicher
Maya Bechler-Speicher@mayabechlerspei·
Remember our ICML25 "Graph Learning Will Lose Relevance Due To Poor Benchmarks"? Fear no more! GraphBench is here! 🤩 We give you: The next generation of Graph Benchmarking! Including: -New shiny high-quality datasets from diverse domains spanning seven domains, including chip design, algorithmic reasoning, and weather forecasting. -Standardized hyperparameter tuning procedures, enabling fair and principled model comparison - Strong, transparent baselines that accurately reflect algorithmic progress - Comprehensive coverage of graph learning tasks, datasets, and modern GNN architectures - Reproducibility-focused design, minimizing variance and evaluation artifacts - Forward-looking benchmark designed for next-generation graph learning research A huge collab with: @chrsmrrs, @mmbronstein, @michael_galkin, @HolgerHoo, Timo Stoll, @ChendiQian, @benfinkelshtein, Ali Parvis, Darius Weber, @ffabffrasca, @HadarShavit, @antoinesrdin, Arman Mielke, Marie Anastacio, Erik Müller,
Maya Bechler-Speicher tweet media
English
3
11
39
4.2K
Itay Evron retweetledi
Yonatan Belinkov
Yonatan Belinkov@boknilev·
Since linear probes are popular again, maybe it’s a good time to point to the many issues with them, which were examined in detail in the NLP Interpretability community. The “mechanistic?” piece by @sarahwiegreffe and @nsaphra has many useful pointers. aclanthology.org/2024.blackboxn…
Yonatan Belinkov tweet media
English
3
17
156
10.7K