Tan Wang

14 posts

Tan Wang banner
Tan Wang

Tan Wang

@Wangt97

Ph.D. Student @NTU | 📈Google Phd Fellowship | 2022 Intern@Microsoft | 🐶2023 Summer Intern@Google

Katılım Nisan 2018
78 Takip Edilen47 Takipçiler
Tan Wang
Tan Wang@Wangt97·
@vivid_en Please reply to my email!!!! Contact so many times to your custom service but get no response
English
0
0
2
42
Linjie (Lindsey) Li
Linjie (Lindsey) Li@LINJIEFUN·
I am humbled to be re-featured as Women in Computer Vision for the BEST of CVPR section of the Computer Vision News July Magazine. It was great chatting with Ralph Anzarouth. I hope my unconventional career path can encourage more female researchers. rsipvision.com/ComputerVision…
Linjie (Lindsey) Li tweet media
English
3
5
134
19.5K
Tan Wang
Tan Wang@Wangt97·
Have a quick try using LLaVA v1.5 with a sample in our proposed VLM benchmark EqBen (ICCV2023 Oral), which seems still struggle😂😂😂… So… There are still many things to do! github.com/Wangt-CN/EqBen
Tan Wang tweet media
Rowan Cheung@rowancheung

🚨 BREAKING: GPT-4 image recognition already has a new competitor. Open-sourced and completely free to use. Introducing LLaVA: Large Language and Vision Assistant. I compared the viral parking space photo on GPT-4 Vision to LLaVa, and it worked flawlessly (see video).

Zurich, Switzerland 🇨🇭 English
0
0
1
471
Tan Wang
Tan Wang@Wangt97·
Meet us at 5th (Thu) 4:30-6pm (Room Paris Nord) for Oral session or 2:30-4:30pm (Room Foyer Sud) for Poster!
English
1
0
0
132
Tan Wang
Tan Wang@Wangt97·
Excited to share our EqBen/EqSim, a new benchmark/algorithm focusing on evaluating and improving the similarity measure of V&L foundation models, to be presented in Oral session at #ICCV2023 in Paris! Joint work w/ @linkeyun2 @LINJIEFUN CC Lin, ZY Yang, HW Zhang, ZC Liu, LJ Wang.
Tan Wang tweet media
Paris, France 🇫🇷 English
1
2
9
1.2K
Tan Wang
Tan Wang@Wangt97·
I will be in @icvss for the upcoming week presenting our DisCo (disco-dance.github.io), with interactive demo to turn static images into human dancing videos! Big thanks to @GMFarinella , @robertocipolla , @sebattiato for organizing ICVSS.
AK@_akhaliq

DisCo: Disentangled Control for Referring Human Dance Generation in Real World paper page: huggingface.co/papers/2307.00… Generative AI has made significant strides in computer vision, particularly in image/video synthesis conditioned on text descriptions. Despite the advancements, it remains challenging especially in the generation of human-centric content such as dance synthesis. Existing dance synthesis methods struggle with the gap between synthesized content and real-world dance scenarios. In this paper, we define a new problem setting: Referring Human Dance Generation, which focuses on real-world dance scenarios with three important properties: (i) Faithfulness: the synthesis should retain the appearance of both human subject foreground and background from the reference image, and precisely follow the target pose; (ii) Generalizability: the model should generalize to unseen human subjects, backgrounds, and poses; (iii) Compositionality: it should allow for composition of seen/unseen subjects, backgrounds, and poses from different sources. To address these challenges, we introduce a novel approach, DISCO, which includes a novel model architecture with disentangled control to improve the faithfulness and compositionality of dance synthesis, and an effective human attribute pre-training for better generalizability to unseen humans. Extensive qualitative and quantitative results demonstrate that DISCO can generate high-quality human dance images and videos with diverse appearances and flexible motions.

English
1
2
8
3.4K
Tan Wang
Tan Wang@Wangt97·
Thx @_akhaliq! Check out our DisCo at disco-dance.github.io.🔥🔥🔥 🧙‍♂️High Generalizability. No need human-specific fine-tuning! 💃Extensive human-related applications with disentangled control! 👨‍💻Easy-to-follow framework and totally opensource code!
AK@_akhaliq

DisCo: Disentangled Control for Referring Human Dance Generation in Real World paper page: huggingface.co/papers/2307.00… Generative AI has made significant strides in computer vision, particularly in image/video synthesis conditioned on text descriptions. Despite the advancements, it remains challenging especially in the generation of human-centric content such as dance synthesis. Existing dance synthesis methods struggle with the gap between synthesized content and real-world dance scenarios. In this paper, we define a new problem setting: Referring Human Dance Generation, which focuses on real-world dance scenarios with three important properties: (i) Faithfulness: the synthesis should retain the appearance of both human subject foreground and background from the reference image, and precisely follow the target pose; (ii) Generalizability: the model should generalize to unseen human subjects, backgrounds, and poses; (iii) Compositionality: it should allow for composition of seen/unseen subjects, backgrounds, and poses from different sources. To address these challenges, we introduce a novel approach, DISCO, which includes a novel model architecture with disentangled control to improve the faithfulness and compositionality of dance synthesis, and an effective human attribute pre-training for better generalizability to unseen humans. Extensive qualitative and quantitative results demonstrate that DISCO can generate high-quality human dance images and videos with diverse appearances and flexible motions.

English
2
6
8
3.8K
Tan Wang retweetledi
Jia-Bin Huang
Jia-Bin Huang@jbhuang0604·
How to network at a conference? Many students will be attending their first-ever *in-person* conference this year. How exciting! 🤩 Some tips on making the best of attending a conference. 🧵
English
22
302
1.5K
0
Tan Wang retweetledi
Aran Komatsuzaki
Aran Komatsuzaki@arankomatsuzaki·
Incorporating Convolution Designs into Visual Transformers CeiT matches DeiT with 3x fewer iterations by adding three modifications to the architecture. arxiv.org/abs/2103.11816
Aran Komatsuzaki tweet media
English
1
17
53
0