David Page

180 posts

David Page

David Page

@dcpage3

Machine learning researcher

Beigetreten Nisan 2018
1K Folgt2.4K Follower
Mario Zechner
Mario Zechner@badlogicgames·
Anytime someone posts "skill issues"/"ngmi", ask them to show you what they build, preferably with a link to a git repo.
English
4
1
39
4.6K
David Page
David Page@dcpage3·
The paper that introduced Batch Norm arxiv.org/abs/1502.03167 combines clear intuition with compelling experiments (14x speedup on ImageNet!!) So why has 'internal covariate shift' remained controversial to this day? Thread 👇
David Page tweet media
English
13
313
1.1K
0
David Page retweetet
Zeyuan Allen-Zhu, Sc.D.
Zeyuan Allen-Zhu, Sc.D.@ZeyuanAllenZhu·
Excited to announce our new work, a unified theory towards explaining 3 black magics in deep learning: (1) ensemble, (2) knowledge distillation, and (3) self-distillation. An accessible blog post is below.
Microsoft Research@MSFTResearch

Microsoft and CMU researchers begin to unravel 3 mysteries in deep learning related to ensemble, knowledge distillation & self-distillation. Discover how their work leads to the first theoretical proof with empirical evidence for ensemble in deep learning: aka.ms/AAavp1k

English
1
39
258
0
David Page
David Page@dcpage3·
@bozavlado @iiSeymour CTC_CRF extends flipflop to output scores for multiple (six) consecutive bases not just two. Output layer is mostly orthogonal to choice of RNN/CNN encoder so CNN improvements are very welcome! More details coming soon..
English
1
0
2
0
Vlado Boza
Vlado Boza@bozavlado·
@iiSeymour Oh the pair decoding! BTW, is there any description what CTC_CRF does? Is it flipflop? Or something even completelly different?
English
1
0
0
0
David Page retweetet
Chris Seymour
Chris Seymour@iiSeymour·
Big accuracy update coming in the next version of Bonito 🚀 v0.3.0 combines everything we have learned with structured and unstructured approaches - @dcpage3, Tim and myself are working hard on the finished touches this week - watch this space 👀
Clive G. Brown@The__Taybor

Some base-caller updates coming within 5-10 days. 98%+modal and many reads above Q20. Note the X-axis. Generally, sig +ve uplift in consensus and mutation detection. Slightly slower speed in research version.

English
7
34
80
0
Jonny
Jonny@TMVector·
@dcpage3 @nanopore Congratulations and all the best in your new role! I hope they know how fantastic a hire they've made 😎
English
1
0
1
0
David Page
David Page@dcpage3·
First day of new job @nanopore where I get to apply ML to a bunch of fun science and engineering problems. Pretty excited!
English
4
2
48
0
David Page retweetet
Alex Thiery
Alex Thiery@alexxthiery·
Preparing a short course on neural nets can be fun. Below is one of the fast Resnets by @dcpage3 on CIFAR10. Would have been nice to track a UMAP-like representation of some internal layer, but have not found a reasonably fast/stable way to do so. Any idea? @NikolayOskolkov
English
4
5
16
0
David Page
David Page@dcpage3·
Undertraining a large model is a good way to speed things up on toy problems myrtle.ai/how-to-train-y… but it was far from clear this should extend to large scale.
David Page tweet media
English
0
4
2
0
David Page
David Page@dcpage3·
Simple setup + attention to details -> sota self-supervised reps! LARS -> large batches -> no need for memory bank of -ve examples Random crops + color aug (to prevent hist cheating) -> no need for special arch Projn head for contrastive loss -> hidden reps preserve info
Ting Chen@tingchenai

Introducing SimCLR: a Simple framework for Contrastive Learning of Representations. SimCLR advances previous SOTA in self-supervised and semi-supervised learning on ImageNet by 7-10% (see next). arxiv.org/abs/2002.05709 Joint work with @skornblith @mo_norouzi @geoffreyhinton.

English
1
7
80
0
David Page retweetet
Jeremy Howard
Jeremy Howard@jeremyphoward·
@ylecun @viglovikov @timetravellertt @kaggle The problem though with "you can always add those tricks to get the numbers up" is that *very* often I see papers that don't do data aug, or don't tune hyper-params, etc, then claim their new idea helps. But then I find it's actually just a poor proxy for the things they skipped
English
2
3
47
0