Robert Pless
1.5K posts

Robert Pless
@rbpless
Professor at GWU, Developer of @traffickcam and @projectRephoto, mercenary interest in glitter and contemporary art. BlueSky: @pless.bsky.social
Washington DC เข้าร่วม Haziran 2009
787 กำลังติดตาม982 ผู้ติดตาม

Webcam showing severe flooding already in Jamaica:
skylinewebcams.com/en/webcam/jama…
(click through to see actual live feed)
English

I'm excited to share that GWU is hosting a World Bank Symposium on AI & the Future of Human Capital.
For those not steeped in the development, Human Capital == people’s skills, knowledge, and health
worldbank.org/en/events/2025…
Some flexibility in the deadline may be possible.
English
Robert Pless รีทวีตแล้ว

This step, where the executive decision maker is obliged to object to or approve of them, aligns the authority to make the decisions with the responsibility to get good reviews. Additionally, the Area Chairs see the name of the reviewer (I'm looking at you, reviewer #2). (3/4)
English

@rohitgUCF @karpathy I think andrej’s point is that even good data (like textbooks) is full of “crap” like little formatting snippets. I worry that dumb heuristics to fix things lose interesting content. Fineweb (afaik) is mostly filtering/dedup so probably not losing diversity
English

Mildly obsessed with what the "highest grade" pretraining data stream looks like for LLM training, if 100% of the focus was on quality, putting aside any quantity considerations. Guessing something textbook-like content, in markdown? Or possibly samples from a really giant model? Curious what the most powerful e.g. 1B param model trained on a dataset of 10B tokens looks like, and how far "micromodels" can be pushed.
As an example, (text)books are already often included in pretraining data mixtures but whenever I look closely the data is all messed up - weird formatting, padding, OCR bugs, Figure text weirdly interspersed with main text, etc. the bar is low. I think I've never come across a data stream that felt *perfect* in quality.
English

Question: Does anyone have a pre-computed FAISS indices for large datasets (e.g. LAION 2B, 5B) that have been used to train CLIP-like models, and would be willing to share the index or access for queries?
The one below doesn't seem to work anymore: knn5.laion.ai/knn-service
English

I'm lucky to get to work on important problems with really great researchers. @abby621 is talking about our research on image search tools to support sex trafficking investigations tomorrow (wednesday) at noon eastern time, link in the following post: bsky.app/profile/astyli…
English

Thank you for everything you did this year. The Program Chairs for the next couple of conferences are lucky to have you!
Yoshitomo Matsubara@yoshitomo_cs
Having served as Technical Chair for CVPR 2024, I was appointed to serve as Technical Chair for - CVPR 2025 - 2027 - ICCV 2025, 2027, and 2029 I am so excited to work for @CVPR & @ICCVConference and keep improving the review systems! 🔥🔥
English

As #CVPR2024 concluded yesterday, I think my technical chair role came to an end
It was my pleasure and great experience to closely work with many chairs, a senior advisor and @openreviewnet team for @CVPR !
Thank you all!
English
Robert Pless รีทวีตแล้ว

The @CVPR AI Art Gallery is now live 🤖🎨
Featuring 115+ artworks using or about computer vision😎
View them here: thecvf-art.com
#AIart #CVPR2024

English

At the AI Aspirations event in the old Newseum building. I love being in DC @GWtweets and finding ways to understand what AI research is most important.

English
Robert Pless รีทวีตแล้ว

We are thrilled to share our groundbreaking paper published today in @Nature: "Low Latency Automotive Vision with Event Cameras."
Paper: nature.com/articles/s4158…
Video: youtu.be/dwzGhMQCc4Y
Code & Dataset: github.com/uzh-rpg/dagr
Frame-based sensors such as the RGB cameras used in the automotive industry face a bandwidth–latency trade-off: higher frame rates reduce perceptual latency but increase bandwidth demands, whereas lower frame rates save bandwidth at the cost of missing vital scene dynamics due to increased perceptual latency (see Fig. 1a of the paper). Event cameras have emerged as alternative vision sensors to address this trade-off. Event cameras measure the changes in intensity asynchronously, offering high temporal resolution and sparsity, markedly reducing bandwidth and latency requirements. Despite these advantages, event-camera-based algorithms are either highly efficient but lag behind image-based ones in accuracy or sacrifice the sparsity and efficiency of events to achieve comparable results. To overcome this, we propose a hybrid event- and frame-based object detector based on Deep Asynchronous GNNs, which preserves the advantages of each modality and thus does not suffer from this trade-off. Our method exploits the high temporal resolution and sparsity of events and the rich but low temporal resolution information in standard images to generate efficient, high-rate object detections, reducing perceptual and computational latency. In doing so, it emulates the slow-fast pathways in biological neural networks and uses them to its advantage. We show that using a 20-Hz RGB camera plus an event camera achieves the same latency as a 5,000-Hz camera with the bandwidth of a 50-Hz camera, i.e., an over 100-fold bandwidth reduction, without compromising accuracy. Our approach paves the way for efficient and robust perception in edge-case scenarios by uncovering the potential of event cameras. We release the code and the dataset (DSEC-Detection) to the public.
Kudos to @DanielGehrig6, who, with this work, also received the UZH Annual Award for the Best PhD thesis!
**Reference**
Daniel Gehrig, Davide Scaramuzza
Low Latency Automotive Vision with Event Cameras
Nature, May 29, 2024.
DOI: 10.1038/s41586-024-07409-w
PDF (Open Access): nature.com/articles/s4158…
Video (Narrated): youtu.be/dwzGhMQCc4Y
Code & Datasets: github.com/uzh-rpg/dagr
@UZH_en @UZH_Science @UZHspacehub @ERC_Research @nccrrobotics

YouTube

English






