Parthe Pandit

217 posts

Parthe Pandit

Parthe Pandit

@PartheP

Thakur Family Chair Assistant Professor @ C-MInDS, IIT Bombay

Powai, Mumbai 가입일 Şubat 2014
524 팔로잉334 팔로워
Parthe Pandit
Parthe Pandit@PartheP·
@mybmc No action taken on this blocked & broken drainage chamber on LJ Rd despite formal complaint, follow-ups. Please take immediate action. Serious health menace for residents. Local sewage not draining into municipal gutters ⇒ clogged sewage ⇒ rats, mosquitos, cockroaches...
English
1
0
0
61
Shubhendu Trivedi
Shubhendu Trivedi@_onionesque·
At the risk of sounding repetitive: Everyone should check this paper and line of work out! First, because it's cool, and second, they somehow (still not sure how) found some old paper of ours, which I really liked*, that literally no one cared about! 😜 columbia.edu/~skk2175/Paper…
Daniel Beaglehole@dbeagleholeCS

We identify that Conv Nets implement a variant of the same general mechanism of feature learning as in fully-connected networks. The covariances of the filters in CNNs again recover the average gradient outer-product (AGOP) of the model, additionally averaged over input patches.

English
1
0
14
4.6K
Arya Mazumdar
Arya Mazumdar@MountainOfMoon·
Super proud that my student Nami Matsumoto just won an NSF Graduate Research Fellowship. This is sooo well-deserved!
English
1
0
35
2.9K
Parthe Pandit
Parthe Pandit@PartheP·
@andrewgwils 2. arxiv.org/abs/2302.02605 How to train kernel models with large number of centers (or inducing points)? A new algorithm, EigenPro3, to train such models requiring only O(p) memory for p centers. Prior work such as FALKON needed O(p^2) memory, ie, infeasible for large models
English
0
0
4
172
Andrew Gordon Wilson
Andrew Gordon Wilson@andrewgwils·
...and 3) conjugate gradients! :). I look forward to reading this. I suppose "pure kernel methods" means no kernel learning? I've always wanted to combine these infinite-width methods with kernel learning... perhaps a promising next step.
Jasper@latentjasper

Really excited to share this new paper: Kernel Regression with Infinite-Width Neural Networks on Millions of Examples. We found that the recipe for success with kernels is 1) highly expressive kernels (infinitely wide deep nets) and 2) lots of data. arxiv.org/abs/2303.05420

English
5
2
26
13K
Parthe Pandit
Parthe Pandit@PartheP·
@andrewgwils You might find these papers interesting too: 1. arxiv.org/abs/2212.13881 recursively adapting kernel functions inspired by how neural networks learn features These models, Recursive Feature Machines, are SotA on tabular datasets and bridge the gap to fully connected networks
English
0
0
5
311
Raaz Dwivedi
Raaz Dwivedi@raazdwivedi·
Extremely delighted to join @Cornell_ORIE @cornell_tech @Cornell. Heartfelt thanks to my amazing advisors, @LesterMackey, Susan Murphy, Devavrat Shah, Martin Wainright, & Bin Yu, the wonderful collaborators, friends, & family for making this possible. The chapter starts Jan 1st!
Susan Murphy lab@SusanMurphylab1

Postdoc @raazdwivedi has accepted a tenure-track assistant professorship at @Cornell, Operations Research and Information Engineering @Cornell_orie, @Cornell Tech in NYC! Congratulations, Raaz!!!

English
29
11
245
40.8K
Parthe Pandit
Parthe Pandit@PartheP·
@nmallinar you can't just get a single dream of Costco. you have to get a 32 pack
English
1
0
4
140
Neil Mallinar
Neil Mallinar@nmallinar·
i dreamt of Costco again last night
English
1
0
9
603
Parthe Pandit
Parthe Pandit@PartheP·
@PreetumNakkiran i am hoping someone creates a generative model that converts hand drawings to professional figures
English
0
0
0
259
Preetum Nakkiran
Preetum Nakkiran@PreetumNakkiran·
don’t know tikz, so I have to waste time the old fashioned way (drawing unnecessary figures by hand)
English
4
1
9
5.8K
Gururaj Saileshwar
Gururaj Saileshwar@gururajS92·
Thrilled to share that our HPCA'23 paper "Scalable and Secure Row-Swap" has been selected for the Best Paper Award! This was a collaborative work led by @JeonghyunW and @Prashxnt_Nair at @UBC ! JH will present our paper at @HpcaArchConf on 27 Feb in Montreal!
Gururaj Saileshwar tweet media
English
4
0
39
3K
Parthe Pandit
Parthe Pandit@PartheP·
We now have a simple, powerful, and stable alternative to Deep Nets A challenge in classical kernel models is choosing the 'right' kernel By examining NN training, we identified the modification necessary to empower kernels - FEATURE LEARNING #NeuralNetworkFreeSince2023
Daniel Beaglehole@dbeagleholeCS

What is the nature of feature learning in deep networks? We propose that neural networks recover a statistic known as the average gradient outer product (AGOP). Github: github.com/aradha/recursi… arXiv: arxiv.org/abs/2212.13881

English
0
2
19
2K
Parthe Pandit
Parthe Pandit@PartheP·
@Yizhezhu_ How many samples do I need to generate to achieve the rate? If there are n true samples, can you generate O(sqrt(n)) samples and still achieve small W1-distance?
English
1
0
0
79
Yizhe Zhu
Yizhe Zhu@Yizhezhu_·
The reason we care about 1-Wasserstein distance is that by the Kantorovich-Rubinstein duality, it uniformly bounds the utility loss for any Lipschitz function applied to the synthetic data
English
1
0
1
259
Parthe Pandit
Parthe Pandit@PartheP·
@ysbhalgat Paragraph structure is very important for technical communication. The same text should perhaps be divided into 3-4 separate paragraphs, each with a purpose.
English
1
0
1
18
Parthe Pandit
Parthe Pandit@PartheP·
@ysbhalgat Not a good summary - Needs bullets. Currently, I need to read an entire sentence with obscure 10-word-named schemes to get to the start of the next sentence - The Capitalization Of Words In The Sentences Makes It Harder To Read Easily - Topics can be segregated based on themes
English
1
0
1
14
Shubhendu Trivedi
Shubhendu Trivedi@_onionesque·
A poster by Larry Wasserman to put on your refrigerator.
Shubhendu Trivedi tweet media
English
9
33
237
33.7K
Yash Bhalgat
Yash Bhalgat@ysbhalgat·
I just used humata.ai to explain the Union Budget recently released by the Indian Govt. I have a minimal understanding of Economics/Politics, but I think it did a great job summarizing the key points in a 58-page doc. (2/n) #UnionBudget2023
Yash Bhalgat tweet media
English
1
0
1
285
Parthe Pandit 리트윗함
Rob Nowak
Rob Nowak@rdnowak·
Rahul Parhi and I wrote a tutorial style article to explain modern theory of deep learning and neural function spaces via elementary signal/image processing concepts (Fourier transform, Radon transform, L1 regularization, sparsity) arxiv.org/abs/2301.09554
English
8
81
401
65.3K
Dimitris Papailiopoulos
Dimitris Papailiopoulos@DimitrisPapail·
Nobody calls denoising bullshit generation, but suddenly if it happens using a transformer then it is?
English
5
0
10
4.5K