Sebastian G. Gruber

50 posts

Sebastian G. Gruber

Sebastian G. Gruber

@SebGGruber

Postdoc in Uncertainty Quantification (Machine Learning)

Leuven Beigetreten Haziran 2022
199 Folgt71 Follower
David Pfau
David Pfau@pfau·
@FrnkNlsn Check out the now-properly-referenced preprint here, if you have a penchant for information geometry, or just want to get a snapshot of what I was thinking about in the ancient history of 2013: arxiv.org/abs/2511.08789
English
3
3
19
2.3K
David Pfau
David Pfau@pfau·
New (sort of) preprint on arXiv today: a generalized bias-variance decomposition for Bregman divergences!
David Pfau tweet media
English
4
14
129
10.6K
David Holzmüller
David Holzmüller@DHolzmueller·
@LChoshen @Eugene_Berta @BachFrancis 4) The irreducible loss (due to noise) plus the extra "grouping loss" that arises from having the same prediction f(X_1) = f(X_2) for X_1, X_2 with P(Y|X_1) != P(Y|X_2). (This means you can't separate the two probabilities with a post-hoc transformation g(f(X)) anymore.)
English
1
0
1
97
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
Looking forward to any discussions about kernels, uncertainty, bias-variance decompositions, or my other research: calibration!! See you on Wed 24 Jul 11:30 a.m. CEST at Hall C 4-9
English
4
0
1
78
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
The decomposition represents a framework for understanding generative models' generalization performance and uncertainty. Specifically, we outperform other baselines of uncertainty in predicting the answer correctness of LLMs arxiv.org/abs/2310.05833
English
1
0
2
116
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
28 years ago the bias-variance-covariance decomposition of the mean squared error was introduced. This week at @icmlconf, @BuettnerFlo and I will show that it also holds for kernel-based loss functions, which we use for image, audio, and language generation experiments...
Sebastian G. Gruber tweet media
English
1
1
7
391
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
@BlackHC In my experience, minimal changes to the optimization procedure can have major effects on the TS results. It is difficult to pinpoint the issue without toying around with the implementation.
English
1
0
1
38
Andreas Kirsch 🇺🇦
Andreas Kirsch 🇺🇦@BlackHC·
Is there anything magic I need to do to make netcal 's TemperatureScaling work? I just handwrote an LBFGS-based calibration, which seems to work, but TemperatureScaling somehow doesn't do anything at all 😐
English
1
0
1
614
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
@profgavinbrown @csmcr @RiccardoAli1 Conditionally yes. If you have parametric proper scores then you have a bijection to NLL expfam and bregman divergences. However, proper scores are also defined for non-parametric distributions. That's why we had to use functional bregman divergences in the paper
English
1
0
0
96
Gavin Brown
Gavin Brown@profgavinbrown·
@SebGGruber @csmcr @RiccardoAli1 Yes absolutely. Likewise - love your paper there too - am slowly digesting. Aren’t strictly proper rules / NLL expfam in a bijection tho? Also with Bregman divergences, that we studied?
English
1
0
1
141
Gavin Brown
Gavin Brown@profgavinbrown·
The most surprising paper I've ever published... “Bias/Variance is not the same as Approximation/Estimation”. openreview.net/forum?id=4TnFb… We figured out the precise connection between two seminal results in ML theory... somehow this was overlooked for 50+ years? @csmcr @RiccardoAli1
English
4
23
129
18.6K
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
@iScienceLuvr If you are interested for more, the MMD can also be used for other tasks, like audio and text generation. There even exists a bias-variance-covariance decomposition for it via kernel scores (=MMD + const) arxiv.org/pdf/2310.05833…
English
0
0
2
245
Tanishq Mathew Abraham, Ph.D.
Tanishq Mathew Abraham, Ph.D.@iScienceLuvr·
Rethinking FID: Towards a Better Evaluation Metric for Image Generation abs: arxiv.org/abs/2401.09603 This paper from Google Research proposes the use of CLIP MMD distance (CMMD) as an alternative to FID for text-to-image generation eval. CMMD does not make assumption about normality like FID does (which is anyway violated by the Inception embeddings and causes problems), it is an unbiased estimator, it is sample efficient, it better matches with expected trends (ex: distorting images reliably increased CMMD unlike with FID), and appears to better match human perception.
Tanishq Mathew Abraham, Ph.D. tweet media
English
4
37
244
32.1K
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
@iScienceLuvr I highly enjoy that the MMD is used more for generative models. However, as mentioned in the paper, the KID already uses the MMD of feature embeddings. So, it would be interesting to have an ablation study of the kernel choice and embedder choice. The paper only evaluates FID
English
0
0
1
339
Alan Jeffares
Alan Jeffares@Jeffaresalan·
Have we been training deep ensembles with the wrong objective? 😱 Our new #NeurIPS paper investigates why training ensembles *jointly* is almost never observed in practice and uncovers some pretty surprising behaviour… 🧵 [1/N]
Alan Jeffares tweet media
English
11
37
337
81.3K
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
@bryancsk In the first book, yes. However, in the sequels, the author goes into great detail about how technology threatens the existence of every neutral actor. Ironically, the only safe solution in the books is to isolate yourself into a bubble of no progress for eternity
English
0
0
0
72
Bryan Cheong
Bryan Cheong@bryancsk·
The Three Body Problem is an e/acc piece of science fiction because, unlike the others that narrate the dangers of new technology, it describes the civilisation-ending consequences of *delaying* the development of new tech
English
53
68
1.1K
132.6K
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
@BlackHC @ilyasut "allowing the company to be destroyed 'would be consistent with the mission'" ?!?! Does not seem like a sustainable mindset from a stochastic process perspective. You only have to slip once into the impression of 'kill the company' and it's over.
English
0
0
1
1K
Andreas Kirsch 🇺🇦
Andreas Kirsch 🇺🇦@BlackHC·
It seems @ilyasut was outvoted by the board 1:3 yesterday after removing Sam and Greg. Independent board members having no shares, stakes, or employment mean that they care much less about destroying the company than actual employees. incl e.g. Ilya... x.com/ilyasut/status…
English
2
2
38
20.2K
Sebastian G. Gruber
Sebastian G. Gruber@SebGGruber·
@tengyuma @Voyage_AI_ Well, coincidentally I am looking for a research internship on a novel calibration project, where I could use your embedders ;)
English
0
0
0
59
Tengyu Ma
Tengyu Ma@tengyuma·
@SebGGruber @Voyage_AI_ Very interesting! Thanks for sharing! please try Voyage embeddings in your next project :)
English
1
0
1
929
Tengyu Ma
Tengyu Ma@tengyuma·
📢 Introducing Voyage AI @Voyage_AI_! Founded by a talented team of leading AI researchers and me 🚀🚀. We build state-of-the-art embedding models (e.g., better than OpenAI 😜). We also offer custom models that deliver 🎯+10-20% accuracy gain in your LLM products. 🧵
Tengyu Ma tweet media
English
38
89
753
225.4K
Christoph Molnar 🦋 christophmolnar.bsky.social
Is there a good web app to demonstrate underfitting / overfitting? Like some scatterplot (dots = data) with a line (=model) going through and you have a slider for model complexity?
English
3
3
8
3K