SubstProf weakmath (@[email protected])

3K posts

SubstProf weakmath (@weakmath@algebra.chat) banner
SubstProf weakmath (@weakmath@algebra.chat)

SubstProf weakmath (@[email protected])

@weakmath

Mathematician, computer scientist and humanist looking for truth. For this journey, benevolent and scientifically interested people are always welcome.

Hildesheim/Göttingen/Berlin Inscrit le Mayıs 2018
513 Abonnements215 Abonnés
SubstProf weakmath (@weakmath@algebra.chat)
On average one out of three ML researchers I talk to knows about JL. One out of 10 can actually say something about the proof. One out of 20 can tell me a correct definition of VC dimension. Maybe I talk to the wrong people at conferences...
Delip Rao e/σ@deliprao

The Google turboquant paper is making ML folks in this decade discover JL lemma and interact with math folks (which is cool). It appears more cool and mysterious if you do not read ML papers from the 90s and early 2000s :)

English
0
0
0
24
SubstProf weakmath (@[email protected]) retweeté
Alex Clemmer 🔥🔥🔥😅🔥🔥🔥
The first time you hear about the JL lemma, it will seem too good to be true. And it is, kind of, I'll explain. The idea is: if you have points in large d-dimensional space, a RANDOM projection to much smaller k-dim subspace will be "nearly optimal" "in the general case." Or, more specifically: with high probability, the pairwise distances between points are preserved, given a couple other requirements around d and k. So why don't we just use random projections instead of carefully-constructed ones all the time? This is the most common misunderstanding of the JL lemma, and the one thing to really understand about it: in many (most?) datasets that are meaningful to humans, you actually CAN do better with something like maybe PCA. If your dataset is pathological, e.g., the points all lie on a plane even though it's technically in 3 dimensions, then clearly some planes you project onto will be better than others. The JL lemma does not apply to 2 and 3 dimensions, but you can imagine this would be true in large numbers of dimensions too. (See screenshot 1, i hope you like it because i made it myself lol.) If you know just those facts, you will be pretty well-prepared to answer most questions about its use. Most of the papers Delip mentions do presuppose that you know this. At least when I was a student, I found this to be non-obvious.
Alex Clemmer 🔥🔥🔥😅🔥🔥🔥 tweet media
Delip Rao e/σ@deliprao

The Google turboquant paper is making ML folks in this decade discover JL lemma and interact with math folks (which is cool). It appears more cool and mysterious if you do not read ML papers from the 90s and early 2000s :)

English
8
54
636
80.6K
SubstProf weakmath (@[email protected]) retweeté
TGDK
TGDK@tgdkjournal·
TGDK announces a Special Issue on "Neuro-Symbolic Modeling for Human-Centric AI" with submissions due on June 30, 2026. We welcome research, survey and resource articles at the intersection of graph-structured data / knowledge and human-centric AI. See: url-shortener.me/A3Q5
English
0
1
3
125
SubstProf weakmath (@[email protected]) retweeté
Valeriy M., PhD, MBA, CQF
Valeriy M., PhD, MBA, CQF@predict_addict·
When KAN was released, I said it was one of the truly transformative technologies Predictably, the “sceptics” arrived. Mostly math-devoid Kagglers, plus a few loud voices confusing leaderboard tricks with scientific progress. Fast-forward less than two years KAN has now surpassed the only genuinely transformative time-series models of the last two decades N-BEATS and NHIT. Let that sink in. And yes — before someone asks — KAN has also smashed transformers in this domain. Not by vibes. By results. By structure. By empirical evidence. Paper for those who actually read: [arxiv.org/pdf/2601.18837](arxiv.org/pdf/2601.18837) Moral of the story: A Kaggle badge is not a certificate of mathematical understanding. It is not proof of scientific judgment. And it certainly isn’t a crystal ball for identifying foundational ideas. Real progress doesn’t come from leaderboard folklore. It comes from theory, structure, and results that survive contact with reality History keeps repeating this lesson. Some people still haven’t learned it.
Valeriy M., PhD, MBA, CQF tweet media
English
2
17
205
15.6K
SubstProf weakmath (@[email protected]) retweeté
Peter Richtarik
Peter Richtarik@peter_richtarik·
I am an AC for ICLR 2026. One of the papers in my batch was just withdrawn. The authors wrote a brief response, explaining why the reviewers failed at their job. I agree with most of their comments. The authors gave up. They are fed up. Just like many of us. I understand. We pretend the emperor has clothes, but he is naked. Here is the final part of their withdrawal notice. I took the liberty to make it public, to highlight that what we are doing with AI conference reviews these last few years is, basically, madness. --- Comment: We thank the reviewers for their time. However, upon reading the reviews for our paper, it became immediately apparent that the four "reject" ratings are not based on good-faith academic disagreement, but on a critical failure to read the submitted paper. The reviews are rife with demonstrably false claims that are directly contradicted by the text. The core justifications for rejection rely on asserting that key components are "missing" when they are explicitly detailed in the manuscript. Some specific examples are (and many are even fake claims). Claim: Harder tasks like GSM8K are missing. Fact: GSM8K results are in many tables, like Table 2 (Section 4.2) and Appendix G. Claim: The method does not use per-layer ranks. Fact: This is the entire point of our method. The reviewer clearly mistook our method for the baselines. (Section 2, Table 1). Claim: The GP kernel is not specified. Fact: It is specified in Appendix E (Table 6). Claim: There is no ablation of the method's three stages. Fact: Section 4.4 ("Ablation Study") and Appendix J are dedicated to this. Reviewers have a fundamental responsibility to read and evaluate the work they are assigned. The nature of these errors is so fundamental, so systemic in overlooking explicit content, that it goes far beyond what "limited time" or "oversight" can explain. This work has gone through several rounds of revision over the last year. In earlier submissions, the paper usually received borderline or weak-accept scores. Numerous signs strongly suggest that some reviewers are relying entirely on AI tools to automatically generate peer reviews, rather than fulfilling their fundamental responsibility of personally reading and evaluating manuscripts. We strongly protest this. This is a gross disrespect to the authors. It is a flagrant desecration of the reviewer's sacred duty. It fundamentally undermines the integrity of the entire peer-review process. Given that the reviews are not based on the actual content of our paper, we have decided to withdraw the submission. We leave this comment so that future readers of the OpenReview page are aware that the items described as "missing" are already present in the submitted manuscript. These negative reviews for this submission are factually unsound and do not reflect the content of the paper. We cannot and will not accept an assessment that is not based on the work we actually submitted.
English
33
205
1.5K
149.5K
Klaus Schmid
Klaus Schmid@1KlausSchmid·
@weakmath Come on, this sounds so negative! Be more positive! You got your baggage 😉
English
1
0
0
22
SubstProf weakmath (@[email protected]) retweeté
SubstProf weakmath (@weakmath@algebra.chat)
What a positive view on the future :-) So far, it seems everybody is puzzled when I tell them about it. In particular, almost all computer scientists are completely unaware ... arxiv.org/abs/1801.07985
Gappy (Giuseppe Paleologo)@__paleologo

In the near future, basic concentration of measure will be taught in undergraduate probability. And concentration of probability for matrices in masters-level courses. arxiv.org/abs/1501.01571

English
0
0
0
45