WANG Yue

20 posts

WANG Yue banner
WANG Yue

WANG Yue

@ayueei

Staff Algorithm Expert @Ali_TongyiLab, Ex-Senior Research Scientist @SFResearch (SG), Ph.D. in CS @cuhkcse

Singapore Katılım Mart 2019
45 Takip Edilen123 Takipçiler
WANG Yue
WANG Yue@ayueei·
3/3 This SOTA is a direct result of months of dedication and brilliant R&D. I’m incredibly lucky to lead such a talented group of innovators. We did it! 🏆 🚀 WE ARE HIRING! Want to build the future of GUI agents? DM me. 📜 Report: arxiv.org/abs/2512.22047 #Hiring #GUIAgents
English
0
0
0
23
WANG Yue
WANG Yue@ayueei·
2/3 What’s under the hood? Our research focuses on: 🧠 Hybrid MCP tool-use & Active User Interaction ☁️ Device-Cloud Collaboration 🔄 Large-scale Multi-turn online RL We’ve scaled from 2B to 235B, and we’re open-sourcing the 2B and 8B models! 🔓 🔗 github.com/Tongyi-MAI/MAI…
English
1
0
0
28
WANG Yue
WANG Yue@ayueei·
MAI-UI is officially live! 🚀 We’re releasing a family of foundation GUI agents (2B to 235B) setting new SOTA benchmarks in GUI grounding & mobile navigation. 🏆 #1 on ScreenSpot-Pro (73.5%) 🏆 #1 on AndroidWorld (76.7%) 🔗 tongyi-mai.github.io/MAI-UI-blog/
WANG Yue tweet mediaWANG Yue tweet media
English
1
1
1
113
WANG Yue retweetledi
Steven Hoi
Steven Hoi@stevenhoi·
Introducing 🔥BLIP-Diffusion🔥, a novel method for enabling Text-to-image Diffusion models with multimodal controllable generation/editing, powered by BLIP-2 pre-trained text-aligned subject representation. Paper: arxiv.org/abs/2305.14720 Project: dxli94.github.io/BLIP-Diffusion… (1/n)
English
3
57
191
40.6K
WANG Yue
WANG Yue@ayueei·
Excited to introduce our 🔥CodeT5+🔥, a new family of open-source code LLMs for both code understanding and generation, achieved new SoTA code generation performance on HumanEval, surpassing all the open-source code LLMs and the closed-source OpenAI code-cushman-001 model!
Steven Hoi@stevenhoi

Introducing 🔥CodeT5+🔥, a new family of open-source code LLMs for both code understanding and generation, achieved new SoTA code generation performance on HumanEval, surpassing all the open-source code LLMs. Paper: arxiv.org/pdf/2305.07922… Code: github.com/salesforce/Cod… (1/n)

English
1
3
13
806
WANG Yue retweetledi
Li Junnan
Li Junnan@LiJunnan0409·
A new member in the BLIP family: 🔥InstructBLIP🔥, a vision-language instruction tuning framework. InstructBLIP achieves SoTA zero-shot performance with various advantages over other multimodal models such as GPT-4! Github: github.com/salesforce/LAV… Paper: arxiv.org/abs/2305.06500
Li Junnan tweet media
English
15
108
428
95.5K
WANG Yue retweetledi
WANG Yue retweetledi
Steven Hoi
Steven Hoi@stevenhoi·
Happy to share that our BLIP2 paper has been accepted by #ICML2023! arxiv.org/abs/2301.12597 Congrats to our amazing team and coauthors for a very high-impact work in multimodal generative AI! Our work is open-source and publicly available in our LAVIS library.
English
2
25
151
16K
WANG Yue retweetledi
Michaël Trazzi
Michaël Trazzi@MichaelTrazzi·
Aran Komatsuzaki giving walkthroughs of the codeRL paper before the author arrives. After 10 minutes of SBFing his way into answering poster questions he revealed he was not the author and everyone lost their mind (Poster 138 #NeurIPS2022)
English
9
33
568
0
WANG Yue
WANG Yue@ayueei·
Excited to release CodeT5, a new pre-trained encoder-decoder model for code that achieves SoTA in multiple CodeXGLUE benchmarks and has been accepted by EMNLP2021! We also release code and pre-trained checkpoints (github.com/salesforce/Cod…) to spur further research.
Salesforce AI Research@SFResearch

Meet CodeT5 - the first code-aware encoder-decoder pre-trained model that achieves SoTA on 14 sub-tasks in CodeXGLUE! Learn how it’s disrupting software development. Blog: blog.einstein.ai/codet5/ Paper: arxiv.org/abs/2109.00859 GitHub: github.com/salesforce/Cod… #codeintelligence

English
0
0
1
0
WANG Yue
WANG Yue@ayueei·
Big thanks to my amazing co-authors! This is my internship work @SFResearch We present one of the first work to explore vision-language pertaining for VisDial. It is also the first unified model that supports both discriminative and generative settings without any decoders!
Steven Hoi@stevenhoi

VD-BERT: A unified Vision and Dialog Transformer with BERT for Visual Dialogue (VisDial), can rank or generate answers seamlessly. New SOTA on VisDial Challenge (1st in leaderboard). Code is coming soon. @ayueei @JotyShafiq @CaimingXiong @SFResearch arxiv.org/abs/2004.13278

English
0
1
3
0
WANG Yue retweetledi
Yifan Gao
Yifan Gao@Yifan__Gao·
Code and models will be released soon. Big thanks for my amazing coauthors! @SFResearch is a wonderful workplace and Singapore is the most beautiful city in Aisa!
English
1
2
4
0