Hongyang Li (弘洋)

45 posts

Hongyang Li (弘洋) banner
Hongyang Li (弘洋)

Hongyang Li (弘洋)

@ARChaser_

Ph.D. student @SCUT, long-term research intern@IDEA Research, focus on 2D and 3D perception!

Katılım Eylül 2020
71 Takip Edilen49 Takipçiler
Hongyang Li (弘洋)
Hongyang Li (弘洋)@ARChaser_·
@torayeff Hi, thank you for your attention. Our paper is currently under review. We will release the code after this stage. 😉
English
0
0
1
20
Agajan Torayev
Agajan Torayev@torayeff·
@ARChaser_ The demo looks really cool! Are you planning to release the code anytime soon?
English
1
0
0
23
Hongyang Li (弘洋)
Hongyang Li (弘洋)@ARChaser_·
🎉 Online Demo Since TAPTRv3 is an online tracker, recently we have implemented a streaming inference mode, which allows us to process videos of any length on RTX3090! With the support of @Gradio and @huggingface, we have deployed the demo at huggingface.co/spaces/HYeungL…. Try it out!
JinyuanQu@JinyuanQu322

💡Introducing TAPTRv3. [1/3] TAPTRv3 focuses on the robust tracking of any point in long videos. Benefitting from Visibility-aware Long-temporal Attention (VLTA), Context-aware Cross Attention (CCA), and auto-triggered global matching, TAPTRv3 surpasses TAPTRv2 by a large margin

English
1
9
59
16.1K
Hongyang Li (弘洋)
Hongyang Li (弘洋)@ARChaser_·
The demo is temporarily unavailable due to a server malfunction. We are working urgently to fix the issue. 😥
English
1
0
0
142
Hongyang Li (弘洋) retweetledi
Lei Zhang
Lei Zhang@leizhangcs·
The API for prompt-free object detection is ready, which means you do not need to provide any prompt and DINO-X will automatically recognize, detect, and segment objects in the provided images. Feel free to check out github.com/IDEA-Research/… and try this feature.
Lei Zhang tweet media
Lei Zhang@leizhangcs

🌟 Introducing DINO-X, our groundbreaking unified vision model at IDEA Research! Paper: arxiv.org/abs/2411.14347 Blog: deepdataspace.com/blog/7?source=x Playground: deepdataspace.com/playground/din… API: cloud.deepdataspace.com/docs#/api/dino… #AI #ComputerVision #ObjectDetection #IDEAResearch Highlights:

English
1
1
6
667
Hongyang Li (弘洋)
Hongyang Li (弘洋)@ARChaser_·
💡Introducing our TAPTRv3 TAPTRv3 focuses on the robust tracking of any point in long videos, surpassing TAPTRv2 by a large margin and achieving SoTA performance. Even when compared with methods trained on internal real-world data, TAPTRv3 is still competitive.
JinyuanQu@JinyuanQu322

💡Introducing TAPTRv3. [1/3] TAPTRv3 focuses on the robust tracking of any point in long videos. Benefitting from Visibility-aware Long-temporal Attention (VLTA), Context-aware Cross Attention (CCA), and auto-triggered global matching, TAPTRv3 surpasses TAPTRv2 by a large margin

English
0
0
3
120
Hongyang Li (弘洋) retweetledi
AI at Meta
AI at Meta@AIatMeta·
New AI research from Meta – CoTracker3 Simpler and Better Point Tracking by Pseudo-Labelling Real Videos. More details ➡️ go.fb.me/xiyc63 Demo on @huggingface ➡️ go.fb.me/yzuqd0  Building on our previous work on CoTracker, this new model demonstrates impressive tracking results where points can be tracked for a long time even when they're occluded or leave the field of view. CoTracker3 achieves state-of-the-art, outperforming all recent point tracking approaches on standard benchmarks — often by a substantial margin. We've released the research paper, code and a demo on Hugging Face — along with models available under an A-NC license to support further research in this space.
English
44
318
1.8K
218.8K
Hongyang Li (弘洋)
Hongyang Li (弘洋)@ARChaser_·
#NeurIPS2024 Our TAPTRv2 is accepted by NeurIPS2024. TAPTRv2 is not only simpler and stronger than TAPTR but also unifies the object/point-level tracking framework (for future work). Code will be released soon. For more details, please refer to our paper. arxiv.org/abs/2407.16291
English
0
0
2
261
Hongyang Li (弘洋)
Hongyang Li (弘洋)@ARChaser_·
@RHeremans Hi, thank you for your attention, we are going to host our demo on HF, so we disabled the online demo temporally. We are sorry for the inconvenience.
English
0
0
0
6
Roel Heremans
Roel Heremans@RHeremans·
@ARChaser_ the demos seem not to work anymore... I installed python3.8.18 as mentioned and tried to pip install the requirements.txt file but i am getting the error telling:  ERROR: Could not find a version that satisfies the requirement MultiScaleDeformableAttention==1.0 (from versions
English
1
0
0
38
Hongyang Li (弘洋)
Hongyang Li (弘洋)@ARChaser_·
We are happy to introduce our new work 「TAPTR: Tracking Any Point with Transformers as Detection」. TAPTR is a simple and strong baseline with sota performance on many datasets while maintaining the advantage of speed. ProjectPage: taptr.github.io (demos and paper links)
English
2
1
5
565
naveen manwani
naveen manwani@NaveenManwani17·
🚨ECCV 2024 Paper Alert 🚨 ➡️Paper Title: TAPTR: Tracking Any Point with Transformers as Detection 🌟Few pointers from the paper 🎯In this paper authors have proposed a simple and strong framework for “Tracking Any Point with TRansformer (TAPTR)”. Based on the observation that point tracking bears a great resemblance to object detection and tracking, they borrowed designs from DETR-like algorithms to address the task of TAP. 🎯In the proposed framework, each tracking point is represented as a DETR query, which consists of a positional part and a content part. As in DETR, each query (its position and content feature) is naturally updated layer by layer. Its visibility is predicted by its updated content feature. 🎯Queries belonging to the same tracking point can exchange information through temporal self-attention. As all such operations are well-designed in DETR-like algorithms, the model is conceptually very simple. They also adopted some useful designs such as cost volume from optical flow models and developed simple designs to mitigate the feature drifting issue. 🎯Their framework demonstrates strong performance with state-of-the-art performance on various TAP datasets with faster inference speed. :hammer_and_wrench: Method Comparison with previous methods Inspired by detection transformer (DETR), they found that point tracking bears a great resemblance to object detection and tracking. 🎯In particular, tracking points can be essentially regarded as queries, which have been extensively studied in DETR-like algorithms. The well-studied DETR-like framework makes their TAPTR conceptually simple yet performance-wise strong. 🏢Organization: @SCUT1918 , International Digital Economy Academy (IDEA), @hkust , Dept. of CST., BNRist Center, Institute for AI, @Tsinghua_Uni 🧙Paper Authors: @ARChaser_ , Hao Zhang, @atasteoff , Zhaoyang Zeng, @Tianhe_Ren , @FengLiust , Lei Zhang 1️⃣Read the Full Paper here: arxiv.org/abs/2403.13042 2️⃣Project Page: taptr.github.io 3️⃣Code: github.com/IDEA-Research/… 🎥 Be sure to watch the attached Demo Video -Sound on 🔊🔊 Find this Valuable 💎 ? ♻️QT and teach your network something new Follow me 👣, @NaveenManwani17 , for the latest updates on Tech and AI-related news, insightful research papers, and exciting announcements. #ECCV2024
English
1
0
5
257
Hongyang Li (弘洋)
Hongyang Li (弘洋)@ARChaser_·
@JieWang_ZJUI 😂 Thank you Jie, hope you will like this tool. In fact 'pudb' could be more fancy, you can also have a try.
English
0
0
1
26