Jeongsoo Park

26 posts

Jeongsoo Park

Jeongsoo Park

@jespark0

PhD student @Cornell (Previously PhD candidate @UMichCSE)

New York, NY Beigetreten Mayıs 2022
165 Folgt146 Follower
Angehefteter Tweet
Jeongsoo Park
Jeongsoo Park@jespark0·
Can AI image detectors keep up with new fakes? Mostly, no. Existing detectors are trained using a handful of models. But there are thousands in the wild! Our work, Community Forensics, uses 4800+ generators to train detectors that generalize to new fakes. #CVPR2025 🧵 (1/5)
English
1
9
24
1.8K
Jeongsoo Park
Jeongsoo Park@jespark0·
Had a ton of fun presenting today at #CVPR2025! Thanks to everyone who came to my poster, and thank you for asking excellent questions!
Jeongsoo Park tweet mediaJeongsoo Park tweet media
English
0
0
16
545
Jeongsoo Park retweetet
Linyi Jin
Linyi Jin@jin_linyi·
Hello! If you are interested in dynamic 3D or 4D, don't miss the oral session 3A at 9 am on Saturday: @zhengqi_li will be presenting "MegaSaM" I'll be presenting "Stereo4D" and @QianqianWang5 will be presenting "CUT3R"
English
1
6
37
1.5K
Jeongsoo Park retweetet
Ayush Shrivastava
Ayush Shrivastava@ayshrv·
Excited to share our CVPR 2025 paper on cross-modal space-time correspondence! We present a method to match pixels across different modalities (RGB-Depth, RGB-Thermal, Photo-Sketch, and cross-style images) — trained entirely using unpaired data and self-supervision. Our approach learns correspondences through contrastive random walks across visual modalities. #CVPR2025 (1/6)
Ayush Shrivastava tweet media
English
1
26
120
9K
Jeongsoo Park retweetet
Yiming Dou
Yiming Dou@_YimingDou·
Ever wondered how a scene sounds👂 when you interact👋 with it? Introducing our #CVPR2025 work "Hearing Hands: Generating Sounds from Physical Interactions in 3D Scenes" -- we make 3D scene reconstructions audibly interactive! yimingdou.com/hearing_hands/
English
2
30
96
8.3K
Jeongsoo Park
Jeongsoo Park@jespark0·
Each image is labeled with detailed metadata, enabling more than just fake detection. We are excited to see what the community can build with this data! 🧵 (4/5)
Jeongsoo Park tweet media
English
1
0
0
81
Jeongsoo Park
Jeongsoo Park@jespark0·
Can AI image detectors keep up with new fakes? Mostly, no. Existing detectors are trained using a handful of models. But there are thousands in the wild! Our work, Community Forensics, uses 4800+ generators to train detectors that generalize to new fakes. #CVPR2025 🧵 (1/5)
English
1
9
24
1.8K
Jeongsoo Park retweetet
Daniel Geng
Daniel Geng@dangengdg·
Hello! If you like pretty images and videos and want a rec for CVPR oral session, you should def go to Image/Video Gen, Friday at 9am: I'll be presenting "Motion Prompting" @RyanBurgert will be presenting "Go with the Flow" and @ChangPasca1650 will be presenting "LookingGlass"
English
3
16
66
5.3K
Jeongsoo Park retweetet
Chris Rockwell
Chris Rockwell@_crockwell·
Ever wish YouTube had 3D labels? 🚀Introducing🎥DynPose-100K🎥, an Internet-scale collection of diverse videos annotated with camera pose! Applications include camera-controlled video generation🤩and learned dynamic pose estimation😯 Download: huggingface.co/datasets/nvidi…
English
2
38
177
42.9K
Jeongsoo Park retweetet
Ayush Shrivastava
Ayush Shrivastava@ayshrv·
We present Global Matching Random Walks, a simple self-supervised approach to the Tracking Any Point (TAP) problem, accepted to #ECCV2024. We train a global matching transformer to find cycle consistent tracks through video via contrastive random walks (CRW).
English
1
22
84
17K
Jeongsoo Park retweetet
Sarah Jabbour
Sarah Jabbour@SarahJabbour_·
📢Presenting 𝐃𝐄𝐏𝐈𝐂𝐓: Diffusion-Enabled Permutation Importance for Image Classification Tasks #ECCV2024 We use permutation importance to compute dataset-level explanations for image classifiers using diffusion models (without access to model parameters or training data!)
Sarah Jabbour tweet media
English
1
11
28
4.4K
Jeongsoo Park retweetet
Sarah Jabbour
Sarah Jabbour@SarahJabbour_·
This year I'm organizing ML4H Outreach program, and want to highlight our Author Mentorship program. Whether you're a mentee looking for guidance or a more experienced researcher with time to mentor, we'd love to have you be a part of this program! Deadline to apply is July 5!
ML4H@SymposiumML4H

Are you planning to submit a paper to ML4H 2024 and would like to receive mentorship from senior scientists? Are you interested in mentoring early-stage researchers? Sign up for the ML4H Submission Mentorship Program by July 5th AoE! Program details: ahli.cc/ml4h/mentorshi…

English
0
9
25
3.3K
Jeongsoo Park retweetet
Ziyang Chen
Ziyang Chen@CzyangChen·
These spectrograms look like images, but can also be played as a sound! We call these images that sound. How do we make them? Look and listen below to find out, and to see more examples!
English
1
40
164
35.7K
Jeongsoo Park retweetet
Yiming Dou
Yiming Dou@_YimingDou·
NeRF captures visual scenes in 3D👀. Can we capture their touch signals🖐️, too? In our #CVPR2024 paper Tactile-Augmented Radiance Fields (TaRF), we estimate both visual and tactile signals for a given 3D position within a scene. Website: dou-yiming.github.io/TaRF/ arXiv: arxiv.org/abs/2405.04534 Huge thanks to my collaborators Fengyu Yang, Yi Liu and advisors @andrewhowens @antoniloq !!!
English
10
21
115
37.5K
Jeongsoo Park retweetet
Daniel Geng
Daniel Geng@dangengdg·
What do you see in these images? These are called hybrid images, originally proposed by Aude Oliva et al. They change appearance depending on size or viewing distance, and are just one kind of perceptual illusion that our method, Factorized Diffusion, can make.
English
10
100
446
59.1K
Jeongsoo Park retweetet
Daniel Geng
Daniel Geng@dangengdg·
Can you make a jigsaw puzzle with two different solutions? Or an image that changes appearance when flipped? We can do that, and a lot more, by using diffusion models to generate optical illusions! Continue reading for more illusions and method details 🧵
English
16
115
614
125.6K
Saurabh Kumar
Saurabh Kumar@drummatick·
On another note, do you think any representation of an image will work? For instance, consider a one-way encryption that encrypts the image to some hash and let's assume non-collision. By all means, we should get the same performance in benchmarks as we get when we train using RGB(or JPEG in this case)?
English
1
0
1
216
Jeongsoo Park
Jeongsoo Park@jespark0·
Do we need RGB to train neural networks? We skip decoding JPEG to RGB, directly feed the encoded JPEG to ViT, and speed up train/eval by up to 39.2%/17.9% without accuracy loss! Check out our poster on Thu-PM-165 in #CVPR2023! (work w/ @jcjohnss) bit.ly/3qRwToV
English
5
70
250
39.7K