Thien Le

22 posts

Thien Le

Thien Le

@tle_96

PhD Researcher at MIT/CSAIL. Theoretical aspects of machine learning and computational biology

Cambridge, MA Katılım Ekim 2022
91 Takip Edilen72 Takipçiler
Thien Le
Thien Le@tle_96·
I'm presenting a poster at @NeurIPSConf DiffCoALG workshop today throughout the day in Upper Level Room 25ABC. I will present our initial theoretical findings on how choosing the right learning architectures can help with model distillation, under a linear rep hypothesis!
English
0
0
2
80
Thien Le retweetledi
Jeremy Bernstein
Jeremy Bernstein@jxbz·
It's been wild to see our work on Muon and the anthology start to get scaled up by the big labs. After @Kimi_Moonshot released Moonlight, people have asked whether Muon is compatible with muP. I wanted to write up an explainer, as there is something deeper going on here! (1/8)
Jeremy Bernstein tweet media
English
9
78
441
68.8K
Thien Le retweetledi
Hannah Lawrence
Hannah Lawrence@HLawrenceCS·
Are you canonicalizing your data? Depending on the group, you might be unavoidably introducing discontinuity...but there's a fix! Come to our ICML poster, #511, on Wednesday 1:30 - 3:00 PM to hear more! Joint work with Nadav Dym and Jonathan Siegel arxiv.org/pdf/2402.16077
Hannah Lawrence tweet media
English
1
10
64
4.4K
Thien Le
Thien Le@tle_96·
Finally, we derive a heuristic to our algorithm and test it on citation networks for both node classification transferability experiments (train on small graphs, test on large graphs) and positional encoding computations (compute PE on subsampled graphs, then zero pad).(4/4)
English
0
0
0
79
Thien Le
Thien Le@tle_96·
Interestingly, by using connections to spectral clustering geometry, we show that if the graphon is well-fitted to a mixture model (e.g., stochastic-block), then the necessary number of vertices to sample can be as small as the (finite!) number of components in the mixture.(3/4)
English
1
0
0
85
Thien Le
Thien Le@tle_96·
How do you robustly subsample vertices of large graphs that are changing in size? By marrying the theory of spectral clustering and graph limit! Check out our poster #36 in Halle B on Friday, 10:45-12:45, presented by the amazing Luana Ruiz! 🔗:openreview.net/pdf?id=l3qtSNs… 🧵(1/4)
Thien Le tweet media
English
1
0
2
189
Thien Le
Thien Le@tle_96·
We also derive an exponential lower bound for group averaging shallow real MLPs and a superpolynomial lower bound for frame-averaging them. Along the way, we also introduce techniques that may help bridge traditional learning theory and its equivariance counterpart. (5/5)
English
0
0
0
92
Thien Le
Thien Le@tle_96·
In contrast, for real-valued functions, we demonstrate a large family of shallow GNNs that are almost orthogonal in Gaussian space, thus requiring exponentially many correlational SQ queries (subsuming noisy gradient descent computations) to distinguish. (4/5)
English
1
0
1
95
Thien Le retweetledi
Josh Robinson
Josh Robinson@Josh_d_robinson·
Looking forward to mentoring a project at this great summer school! Students interested in GNNs / geometric deep learning should consider applying 😀
LOGML Summer School@LogmlSchool

🌟Applications open - LOGML 2024🌟 👥 Mentor-led projects, expert talks, tutorials, socials, and a networking night 📝 Application form: logml.ai 📅 Apply by 26 Jan 2024 📝 📩 Questions? logml.committee@gmail.com #LOGML #Geometry #MachineLearning #SummerSchool

English
0
10
28
5K
Thien Le retweetledi
Yufei Zhao
Yufei Zhao@yufeizhao·
yufeizhao.com/blog/2023/12/2… I’m delighted to showcase and celebrate a selection of recent papers by the talented combinatorialists at MIT, especially those by students and postdocs. These papers reflect the culmination of their hard work, dedication, and innovative problem solving.
English
3
15
100
13.3K
Thien Le retweetledi
Nikos Karalias
Nikos Karalias@AspectStalence·
I'll be presenting our work on optimal approximation algorithms for combinatorial problems with graph neural nets at the Mathematics of Modern Machine Learning (M3L) workshop. Saturday 4pm, room 242! Arxiv: arxiv.org/abs/2310.00526
English
1
19
102
18.7K
Thien Le retweetledi
Hannah Lawrence
Hannah Lawrence@HLawrenceCS·
Excited to present this work on Wednesday! Also, if you’re at NeurIPS and want to chat about equivariant learning, AI for science, or anything else, don’t hesitate to reach out 😄
Quentin Garrido@garridoq_

We will be presenting our work at the poster session on Wednesday from 5 to 7 p.m. CST, so come check it out if you're interested in AI4Science and/or Self-Supervised learning 😉 Paper link: openreview.net/forum?id=ZULq9…

English
2
9
64
10.6K
Thien Le
Thien Le@tle_96·
In our paper, we leverage a new tool from the theory of graph limit, called graphop and action convergence (Backhausz and Szegedy, 2022) to give structural conditions under which parameters of GNNs on a sequence of graphs discretized from the same limit object can transfer. 4/4
English
0
0
0
69
Thien Le
Thien Le@tle_96·
This observation has been formalized for dense graphs via different random graph models and graph limit notions (e.g. graphons). On the other hand, little is known about size-transferability of sparse graph sequences - whose limit spectrum may exhibit pathological behaviors. 3/4
English
1
0
0
70
Thien Le
Thien Le@tle_96·
Excited to present a joint work with Stefanie Jegelka at NeurIPS 2023 Thursday morning poster session. Our paper (openreview.net/pdf?id=kDQwoss…) studies size transferability of graph neural networks with special attention to sparse (even bounded-degree) graph sequences. 1/4
Thien Le tweet media
English
1
4
18
1K