Ran Levinstein

3 posts

Ran Levinstein

Ran Levinstein

@ranlevinstein

Bergabung Nisan 2020
38 Mengikuti3 Pengikut
Ran Levinstein me-retweet
Daniel Soudry
Daniel Soudry@soudry_daniel·
Accelerate your transformer model with the new Block-Sparse-Flash-Attention! github.com/Danielohayon/B… This training-free, drop-in replacement extends FlashAttention-2 with minimal code changes (CUDA Kernels Included). Paper: arxiv.org/abs/2512.07011
English
0
7
18
450
Ran Levinstein me-retweet
Matan Tsipory
Matan Tsipory@tsipory·
[1/6] Tomorrow (Thursday) at #NeurIPS: Are Greedy Task Orderings Better Than Random in Continual Linear Regression? Q: Do models learn better when consecutive tasks are similar or dissimilar? A: Our analysis suggests that they should be dissimilar! openreview.net/forum?id=8JdPq…
Matan Tsipory tweet media
English
1
2
7
821
Ran Levinstein me-retweet
Itay Evron
Itay Evron@itayevron·
[1/5] Next week at #NeurIPS *Optimal Rates in Continual Linear Regression via Increasing Regularization* In the brain, ageing naturally reduces synaptic plasticity. Our theory suggests continual learning models may benefit from a similar mechanism! openreview.net/forum?id=lDh78…
Itay Evron tweet media
English
2
11
115
5K