Wu Lin

102 posts

Wu Lin

Wu Lin

@LinYorker

Postdoctoral fellow at @VectorInst. ML PhD at UBC. Mathematical and computational structures for ML. Geometric and algebraic methods.

Toronto, Canada 参加日 Ağustos 2020
40 フォロー中283 フォロワー
固定されたツイート
Wu Lin
Wu Lin@LinYorker·
#ICML2024 Can We Remove the Square-Root in Adaptive Methods? arxiv.org/abs/2402.03496 Root-free (RF) methods are better on CNNs and competitive on Transformers compared to root-based methods (AdamW) Removing the root makes matrix methods faster: Root-free Shampoo in BFloat16 /1
Wu Lin tweet media
English
9
16
60
12.6K
Wu Lin
Wu Lin@LinYorker·
This work builds on my ICML 2019 paper (with @MarkSchmidtUBC and @EmtiyazKhan), extending a variational Bayes-based geometric framework to modern NN optimization. It can be used to design methods for Bayesian inference, numerical optimization, and gradient-free optimization.
English
1
0
3
102
Wu Lin
Wu Lin@LinYorker·
Within an information-geometric framework, we reconnect Shampoo/SOAP with both classical quasi-Newton ideas and Gaussian whitening, and develop practical methods that naturally handle tensor-valued weights in language model pre-training. arxiv.org/abs/2509.03378 opt-ml workshop
Wu Lin tweet media
English
1
7
8
1K
Wu Lin がリツイート
Runa Eschenhagen
Runa Eschenhagen@runame_·
1/9 In practice, the Shampoo optimizer crucially relies on several heuristics. In our NeurIPS 2025 spotlight paper, we investigate the role of learning rate grafting and infrequent preconditioner updates in Shampoo by decomposing its preconditioner. arxiv.org/abs/2506.03595
Runa Eschenhagen tweet media
English
3
21
90
13.2K
Wu Lin がリツイート
Thomas Möllenhoff
Thomas Möllenhoff@tmoellenhoff·
Are you LoRA fine-tuning LLMs and looking for easy ways to get improvements in accuracy? And also Bayesian uncertainty on top for free? Then check our recent work, accepted @neurips24fitml workshop! arxiv.org/abs/2411.04421
English
7
16
62
7.4K
Wu Lin
Wu Lin@LinYorker·
Natural gradient descent: (steepest) gradient descent under a norm induced by the Fisher matrix yorkerlin.github.io/posts/2021/10/… Riemannian gradient descent (with geodesic retraction) : gradient descent in Riemannian normal coordinate
Frank Nielsen@FrnkNlsn

At Maximum Likelihood Estimator: Key property: observed Fisher information = Fisher information 2nd order Taylor expansion of likelihood: - likelihood curvature = Fisher information - radius of osculating circle=Variance of MLE for large sample size

English
1
5
38
4.1K
Wu Lin
Wu Lin@LinYorker·
Some hardcore theory people complain that "second-order" methods in DL do not have a superlinear convergence rate. At the same time, they are happy to consider SGD a first-order method with only a sublinear rate.
English
0
0
2
125
typedfemale
typedfemale@typedfemale·
i've noticed many (including myself) say "2nd second-order" in machine learning to refer to a set of optimizers (FOOF, shampoo) that... don't actually use second-order information? idk it's weird
English
10
1
47
14.2K
Wu Lin がリツイート
Elad Hazan
Elad Hazan@HazanPrinceton·
My talk on spectral transformers, given at Princeton workshop on learning in dynamical systems, is now online: youtube.com/watch?v=D_NwH5…
YouTube video
YouTube
English
0
44
256
30.7K
Wu Lin がリツイート
Frank Nielsen
Frank Nielsen@FrnkNlsn·
Lin (1991) definition of Jensen-Shannon divergence JS(p,q)= (KL(p:(p+q)/2)+KL(q:(p+q)/2))/2 is a *variational divergence* defined by JS(p,q)=min_{c} (KL(p:c)+KL(q:c))/2 where optimum value is c=(p+q)/2 Defined as information radius by Sibson (1969) 👉mdpi.com/1099-4300/23/4…
Frank Nielsen tweet media
English
1
23
119
11.6K
Wu Lin
Wu Lin@LinYorker·
We can use NGD to obtain many methods: natural evolution strategies (CMA-ES, NES), natural policy gradients, natural-gradient variational inference, exact Bayesian inference on conjugate models, Newton's method, root-free adaptive methods, and Riemannian GD on submanifolds.
English
0
1
3
504