Luca Ambrogioni

10.5K posts

Luca Ambrogioni banner
Luca Ambrogioni

Luca Ambrogioni

@LucaAmb

Ass. prof. of Machine Learning. PI of Generative Memory Lab (@DondersInst). Generative diffusion and statistical physics. AI realist.

Nijmegen, Nederland Katılım Temmuz 2011
2.6K Takip Edilen6.9K Takipçiler
Sabitlenmiş Tweet
Luca Ambrogioni
Luca Ambrogioni@LucaAmb·
1/2) How do patterns form in diffusion models? Out-of-equilibrium phase transitions! Symmetry breaks → low-frequency modes destabilize → large-scale structure emerges. Preprint: How Out-of-Equilibrium Phase Transitions Seed Pattern Formation in Trained Diffusion Models The peper offers a statistical field theory analysis of this process!
Luca Ambrogioni tweet media
English
2
63
355
23.2K
Luca Ambrogioni
Luca Ambrogioni@LucaAmb·
@arthur_spirling @itaisher Especially if since since arxiv is too understaffed to be capable of litigating anything at thus scale. Any small decision is going yo be irreversible
English
0
0
1
21
Arthur Spirling
Arthur Spirling@arthur_spirling·
@itaisher Yes and I think many posters are perhaps not very familiar with what’s it like to try to litigate edge cases. They think of the slam dunks (pure and obvious AI slop) but these things are often subtle and need considerable care.
English
2
1
2
64
Itai Sher
Itai Sher@itaisher·
Disagreement about the arXiv policy is a small instance of a common difference in focus between — punishing wrongdoing and — protecting people from harsh or unfair application of punitive rules.
English
1
0
5
198
Luca Ambrogioni
Luca Ambrogioni@LucaAmb·
@LorenzoDiPi I am positive about hep but I think it still he's as many issues as any other academic field
English
0
0
0
6
Lorenzo Di Pietro
Lorenzo Di Pietro@LorenzoDiPi·
@LucaAmb Depends who you listen to. I don’t find scarcity of interesting things to work on
English
1
0
1
12
Luca Ambrogioni
Luca Ambrogioni@LucaAmb·
I am quite convinced that, under these arxive guidelines, every single major PI in the field will be banned within a few years
English
17
1
62
25K
Mathieu
Mathieu@miniapeur·
Although I generally feel something should be done about arXiv, I can understand the backlash given the severity of the proposed punishment. Could we at least institute a points-based system, like for a driving licence? A direct one-year ban, plus losing the ability to share unpublished work on arXiv, seems too harsh. That said, failing to notice that a reference is incorrect is different from submitting a paper that contains an AI prompt, which could literally be detected by reading the said paper.
Christopher D. Long 🇺🇦🏳️‍🌈🌹@octonion

The backlash against arXiv is a bit odd. All they're asking is that you read your papers before submitting them.

English
9
3
45
12.6K
Luca Dellanna
Luca Dellanna@DellAnnaLuca·
@LucaAmb My humble suggestion: arXiv asks researchers to choose a label for their paper, with these two options: - I take responsibility for it or - I take no responsibility, I didn't even re-read it, lol
English
1
0
0
72
Luca Ambrogioni
Luca Ambrogioni@LucaAmb·
I feel some people fails to distinguish the fact of something being wrong, and something deserving maximum punishment
Nirmalya Kajuri@Kaju_Nut

@WKCosmo @LucaAmb Yeah I have no sympathy either. The least they can do is take responsibility for ensuring the paper complies with the guidelines.

English
3
1
10
713
Luca Ambrogioni retweetledi
Stanislav Fort
Stanislav Fort@stanislavfort·
How powerful is gradient descent exactly? For a small CNN on CIFAR-10 I've looked at the typical loss change due to a random step of the same length as a gradient step starting at the same weights. The gradient step is literally a 185 sigma event => ~impossible~ at random ✅
English
7
27
288
0
Luca Ambrogioni retweetledi
Itai Sher
Itai Sher@itaisher·
I think this is the best post I have seen on arXiv’s new policy
Ben Blaiszik@BenBlaiszik

There are important distinctions getting blurred in the conversation about arXiv and generative AI. arXiv’s written policy does not prohibit LLM use (see attached). In fact, it directly acknowledges that authors may use these tools, while emphasizing that authors remain fully responsible for the entire submission. But, the recent X thread on enforcement raises deeper questions and seems to mingle evidence of LLM usage with more severe misconduct. What counts as evidence of misconduct versus evidence of imperfect editing using tools that are explicitly allowed? If a manuscript contains hallucinated references, fabricated claims, invented data, or misleading content, that is a serious problem. Such authors have failed in their responsibility, regardless of whether the source was an LLM or not and deserve penalties. But if a paper contains a leftover drafting comment from an LLM - or a stray human comment judged as similar to an LLM, e.g., to a postdoc "Please add the newest data here in a short paragraph" - that also seems to meet criteria for the proposed bans. I feel that treating every visible AI artifact as evidence that the entire paper is untrustworthy risks overshooting the balance between allowing AI slop to run wild and curating impactful research. This is a very tough balance to define, and ArXiv is ground zero for many of these changes and they have done a noble job trying to figure things out. But, circling back to the beginning, this blurring is especially confusing as official arXiv policy says that generative AI use is allowed. If that is the policy, then the presence of some sign of AI-assisted drafting should not be considered the core violation. In my opinion, a more durable standard would separate things like minor drafting artiacts, or sloppy manuscript preparation from the much worse problems of hallucinated references, or worst of all - fabricated claims, figures, or data. I want to conclude by saying thank you to arXiv for their brilliant work as a whole; they are a foundational capability that has allowed AI and science across the board to flourish. But, there is work to be done on this policy. Also, can we please write this into an official arXiv policy instead of trying to infer from Prof. Dietterich's X post alone? :)

English
0
2
6
907
Luca Ambrogioni retweetledi
Peter Richtarik
Peter Richtarik@peter_richtarik·
@predict_addict @dakovalev1 @srchvrs @LucaAmb I know arXiv faces issues. And I said something needs to be done. I just do not think this is the best way to handle the this all. I am 100% convinced the intention is good. The implementation looks a bit too draconian though. Time will tell.
English
2
1
1
151
Luca Ambrogioni
Luca Ambrogioni@LucaAmb·
I guess that the difference here is that I think it's a complete social disaster if enforced, with a large fraction of PIs in mant fields ending up being hit in a few years I apologize for being too 'aggressive', I just find this policy very bothering and I think that even people who make mistakes, especially if unintentional, deserve our empathy. We live in rapidly changing times and it's hard to adapt. Most people are doing their best, in their own imperfect way
English
0
0
0
23
Nirmalya Kajuri
Nirmalya Kajuri@Kaju_Nut·
@LucaAmb @WKCosmo I don’t think the current policy is perfect. I wrote a post arguing lifetime punishments are too harsh
English
1
0
1
33
Will Kinney
Will Kinney@WKCosmo·
This is really remarkable. Citations to nonexistent papers are not "typos", any more than fabricated data or synthetic figures are.
Ted Pavlic (he/him/his)@TedPavlic

@WKCosmo You're saying that you already do this? You are willing to vouch that no paper you've been co-author on has any typos (because you would have caught them in your complete read despite only being involved in some small aspect of the stats or methodology)?

English
12
26
523
12.4K
Luca Ambrogioni retweetledi
Ravid Shwartz Ziv
Ravid Shwartz Ziv@ziv_ravid·
A few more thoughts: I agree with the concern that Arxiv will become full of millions of AI-slop papers written by agents. **But** this is not the answer. Not only because banning an author for life is not fair (it might that you write only some sections and ofcourse that you can't check all the references one by one), but also because this is not the real problem. AI is writing our papers at this point, and the only question is at what level. If I ask the model to rephrase a sentence and it keeps, " Sure, I will do it for you", it doesn't mean anything about the quality of the paper. The same as we didn't punish papers with grammar errors in the past. Moreover, I can create entire fake papers that will pass this bar from one prompt, and soon, we will have systems that will catch these mistakes automatically. At the end, we need to come up with a semi-automatic system that will check for correctness without the question of whether it was generated by AI (the question of whether we want humans to be responsible for each paper is a different story)
Ravid Shwartz Ziv@ziv_ravid

We must think about how to handle the fact that LLMs can generate papers without any human intervention, but this is not the way. Also, I want to see the arXiv ban senior PIs who upload 40 papers a year.

English
6
4
26
3K
Luca Ambrogioni
Luca Ambrogioni@LucaAmb·
@Kaju_Nut @WKCosmo For that case, my only issue is that it's unfair to give equal responsibility to all coathors I understand that collaborations are smaller in theoretical physics, but it's not the case in many other fields A single fabricated citation could ban half of CERN :p
English
0
0
0
16
Nirmalya Kajuri
Nirmalya Kajuri@Kaju_Nut·
@LucaAmb @WKCosmo You know I don't like the idea of lifetime punishment, and also I don't see all offenses as equally serious. But one year ban for hallucinated references for example? Totally fair.
English
2
0
2
60
Nirmalya Kajuri
Nirmalya Kajuri@Kaju_Nut·
@LucaAmb @WKCosmo I agree that leftover prompt and say, AI-generated data are not at the same level of misconduct.
English
2
0
4
92
Luca Dellanna
Luca Dellanna@DellAnnaLuca·
@LucaAmb @s_scardapane That seems like a feature rather than a bug? That sloppy researchers get seen a bit less than non-sloppy ones, all other things equal, while still having the means to make their research public?
English
1
0
0
9
Nirmalya Kajuri
Nirmalya Kajuri@Kaju_Nut·
@WKCosmo @LucaAmb Yeah I have no sympathy either. The least they can do is take responsibility for ensuring the paper complies with the guidelines.
English
1
0
1
775
Mariano
Mariano@cociclo·
@LucaAmb Authors are now going to need to read what they upload! There is probably something in the universal declaration of human rights that precludes that, right??
English
1
0
0
24
Luca Ambrogioni
Luca Ambrogioni@LucaAmb·
@WKCosmo @Kaju_Nut Yes, lifetime ban. I am sure you are 100% perfect Will and no mistake ever slips through you. I wish I were like you, but I must admit that it's possible that an LLM prompt will end up in a paper a deeply cared about and work hard on. Have mercy or us mare mortals
English
0
0
0
20
Will Kinney
Will Kinney@WKCosmo·
@LucaAmb @Kaju_Nut PIs of medium to large labs who have are so uninvolved in the actual research that they don't even do a basic check on their own group's papers before slapping their name on them are a scourge, and I hope there should be more consequences for them, not protection.
GIF
English
3
0
10
106