UW RAIVN Lab

166 posts

UW RAIVN Lab banner
UW RAIVN Lab

UW RAIVN Lab

@RAIVNLab

The computer vision and reasoning lab in the Allen School at the University of Washington, led by Ali Farhadi and Ranjay Krishna.

Seattle, WA 가입일 Nisan 2020
142 팔로잉860 팔로워
UW RAIVN Lab 리트윗함
Ai2
Ai2@allen_ai·
Last year Molmo set SOTA on image benchmarks + pioneered image pointing. Millions of downloads later, Molmo 2 brings Molmo’s grounded multimodal capabilities to video 🎥—and leads many open models on challenging industry video benchmarks. 🧵
Ai2 tweet mediaAi2 tweet mediaAi2 tweet media
English
7
64
320
126.5K
UW RAIVN Lab 리트윗함
Prior @ AI2
Prior @ AI2@Ai2Prior·
📢Applications are open for summer'25 internships at the PRIOR (computer vision) team @allen_ai: Come join us in building large-scale models for: 📸 Open-source Vision-Language Models 💻 Multimodal Web Agents 🤖 Embodied AI + Robotics 🌎 Planet Monitoring Apply by December 11, 2024!
English
1
13
47
11.8K
UW RAIVN Lab 리트윗함
Ainaz Eftekhar
Ainaz Eftekhar@ainaz_eftekhar·
🎉 Very Excited to present our recent work on “Selective🔍 Visual Representations for Embodied-AI🤖” next week at ICLR in Vienna🇦🇹!! 📣📣Important update! Our code and pretrained models are now available through our project website 🌐: embodied-codebook.github.io🚀 👋Come to my poster, say hi, and learn more about our findings! (Poster #111, Session 8, on Friday, May 10th at 4:30 PM)
Ainaz Eftekhar@ainaz_eftekhar

Embodied-AI 🤖 models employ general-purpose vision backbones such as CLIP to encode the observation. How can we have a more task-driven visual perception for embodied-AI? We introduce a parameter-efficient approach that selectively filters visual representations for Embodied-AI tasks. Project page: embodied-codebook.github.io 🧵👇

English
1
8
53
9K
UW RAIVN Lab 리트윗함
Gabriel Ilharco
Gabriel Ilharco@gabriel_ilharco·
CLIP models have become a lot better since 2021
Gabriel Ilharco tweet media
English
9
81
624
161.3K
UW RAIVN Lab
UW RAIVN Lab@RAIVNLab·
Check out🪆MatFormer🪆co-led by @adityakusupati: it’s a simple yet powerful general-purpose architecture with flexibility and elasticity built within. It works across modalities and enables super cool things at web-scale tasks🔥🔥
Aditya Kusupati@adityakusupati

Announcing MatFormer - a nested🪆(Matryoshka) Transformer that offers elasticity across deployment constraints. MatFormer is an architecture that lets us use 100s of accurate smaller models that we never actually trained for! arxiv.org/abs/2310.07707 1/9

English
1
4
14
3.8K
UW RAIVN Lab 리트윗함
Mitchell Wortsman
Mitchell Wortsman@Mitchnw·
E) The attention logit growth instability is still present when replacing softmax with pointwise alternatives. Side note: If you're interested in learning more about replacing softmax with a pointwise alternative like relu^2/√seqlen, checkout arxiv.org/abs/2309.08586! (12/15)
Mitchell Wortsman tweet media
English
1
1
7
1.1K
UW RAIVN Lab
UW RAIVN Lab@RAIVNLab·
Exciting work with open source code to facilitate research on medium sized language models! 🎉
English
0
0
4
494
UW RAIVN Lab
UW RAIVN Lab@RAIVNLab·
If you are at #CVPR2023, come check out prompting-in-vision.github.io on Monday, June 19 from 9am - 12pm in West room 223-224. Speakers include @sarahmhpratt from RAIVN lab as well as @liuziwei7 @phillip_isola @hyojinbahng @lschmidt3 and @denny_zhou!
Kaiyang Zhou@kaiyangzhou

We're organizing a tutorial on Prompting in Vision at #CVPR2023 w/ @liuziwei7 @phillip_isola @hyojinbahng @lschmidt3 @sarahmhpratt @denny_zhou Please visit our website at prompting-in-vision.github.io to know more about this event

English
0
1
9
2K
UW RAIVN Lab 리트윗함
Aditya Kusupati
Aditya Kusupati@adityakusupati·
Introducing💃AdANNS: A Framework for Adaptive Semantic Search🕺 TL;DR: Up to 90× faster nearest neighbor retrieval and 2× lower memory cost for web-scale search. Applies to vector search at scale & improves all "retrieval" augmented models! arxiv.org/abs/2305.19435 [1/8]
Aditya Kusupati tweet media
English
5
89
471
91.1K
UW RAIVN Lab 리트윗함
Vaishaal Shankar
Vaishaal Shankar@Vaishaal·
1/9 I am excited to announce that our workshop "Towards the Next Generation of Computer Vision Datasets" will be happening at ICCV 2023 in Paris. We will feature DataComp submissions, other data-centric papers, and invited talks by experts. datacomp.ai/workshop
English
3
18
53
22.7K