GX Xu

28 posts

GX Xu

GX Xu

@GX_NLP

Research Scientist @ Redhat AI Innovation | Agent to Empower Individuals | AI finance educator | prev. AI @ucla

New York, NY Katılım Haziran 2022
357 Takip Edilen66 Takipçiler
Sabitlenmiş Tweet
GX Xu
GX Xu@GX_NLP·
No training needed! While improve your SLM to o-1 level math performance with particle filtering based inference scaling 🎊 Paper, code, and blogpost are all released 🔥 If this interesting to you, please like share, comment to support open-sourcing cutting-edge AI at RedHat!
Isha Puri@ishapuri101

[1/x] can we scale small, open LMs to o1 level? Using classical probabilistic inference methods, YES! Joint @MIT_CSAIL / @RedHat AI Innovation Team work introduces a particle filtering approach to scaling inference w/o any training! check out …abilistic-inference-scaling.github.io

English
0
0
2
175
GX Xu retweetledi
Red Hat AI
Red Hat AI@RedHat_AI·
Random Samples, our weekly seminar series that bridges the gap between cutting-edge AI research and real-world application, continue this Friday, July 18! Title: Grounding Feedback is All You Need: Aligning Small Vision-Language Models Abstract: While recent vision-language models (VLMs) excel at integrating visual and linguistic information, their performance hinges on vast quantities of curated image-text pairs. This reliance makes the alignment process both time-consuming and resource-intensive. In this talk, we'll introduce Sampling-Based Vision Projection (SVP), a novel framework that improves vision-language alignment using automated feedback and minimal human supervision. Our results show that SVP significantly enhances image captioning, improves object recall, and reduces hallucination, enabling smaller models to match the performance of much larger systems. This approach offers a promising path toward developing powerful, efficient, and accessible multimodal AI. Watch live on YouTube: youtube.com/watch?v=ddnRm-… Or join us via Google Meet: meet.google.com/teh-ijwx-rbz
YouTube video
YouTube
Red Hat AI tweet media
English
0
1
4
443
GX Xu retweetledi
Red Hat AI
Red Hat AI@RedHat_AI·
LLM inference is too slow, too expensive, and too hard to scale. 🚨 Introducing llm-d, a Kubernetes-native distributed inference framework, to change that—using vLLM (@vllm_project), smart scheduling, and disaggregated compute. Here’s how it works—and how you can use it today:
English
5
89
545
70.1K
GX Xu retweetledi
Hao Wang
Hao Wang@HW_HaoWang·
[1/x] 🚀 We're excited to share our latest work on improving inference-time efficiency for LLMs through KV cache quantization---a key step toward making long-context reasoning more scalable and memory-efficient.
Hao Wang tweet media
English
9
8
26
3.6K
GX Xu
GX Xu@GX_NLP·
@nataninvesting 1. Supplier are liable to risk of excessive inventory which temu asks for. 2. Suppliers profit take a huge dent with free returns and customer quality complaints. 3 top suppliers may dump temu for SHEIN or Amazon, and temu will be stuck in low quality products
English
0
0
1
110
Natan
Natan@nataninvesting·
@GX_NLP Why do you think it could be not sustainable?
English
1
0
0
124
Natan
Natan@nataninvesting·
$PDD price hasn't changed in last year. In the same time: - 87% Revenue growth y/y - 130% EPS growth y/y - 55% FCF growth y/y Stock is down following an insane multiple compression: - 23x PE 1 year ago - Only 11x PE today
Natan tweet media
English
19
14
115
21.1K
GX Xu retweetledi
Isha Puri
Isha Puri@ishapuri101·
[1/x] can we scale small, open LMs to o1 level? Using classical probabilistic inference methods, YES! Joint @MIT_CSAIL / @RedHat AI Innovation Team work introduces a particle filtering approach to scaling inference w/o any training! check out …abilistic-inference-scaling.github.io
Isha Puri tweet media
English
2
67
234
45.1K
GX Xu retweetledi
GX Xu
GX Xu@GX_NLP·
@bindureddy To get rich, you need to do things that scale. Plumbing don’t scale…
English
0
0
0
17
Bindu Reddy
Bindu Reddy@bindureddy·
TBH new college entrants should really consider acquiring some blue collar skills like electrical or plumbing! AI may not “replace” white collar jobs but it will most definitely reduce the need for so many writers, designers and programmers, decreasing wages in white collar professions. In the future, you won’t get rich with a CS degree, but you may by mastering plumbing! 😀
English
85
19
240
49.7K
Jiao Sun
Jiao Sun@sunjiao123sun_·
Honored to receive the 🥇BEST PAPER AWARD🥇 from CVPR 2024, please consider using our collected fine grained feedback! Huge shout out to our work DreamSync, the key method that we use for using the fine grained feedback to improve the model, detail in my pined tweet! 🚀
Jiao Sun@sunjiao123sun_

🌟Rich Human Feedback for Text-to-Image Generation selected as CVPR 2024 Best Paper Award Candidate (top 1%)🌟 Current text-to-image models are not perfect, but where exactly? They suffer from artifacts, alignment and aesthetics. We collect feedback on 18K images to capture all these aspects! 🔗arxiv.org/pdf/2312.10240 (1/n)

English
20
27
583
94.7K
GX Xu
GX Xu@GX_NLP·
A new RL alignment method, here’s Gaurav’s excellent blog that explains why BRAIn is more stable and gives better performance than PPO and DPO 🔥
English
0
0
2
118
GX Xu
GX Xu@GX_NLP·
TLDR: Looking for a RLHF method that combines the best of PPO and DPO, stable training, and gives amazing result? BRAIN theoretically unites DPO and PPO, and empirical shown to out-perform! An earlier pre-print of the ICML paper is available now🔥
English
0
0
3
237
GX Xu
GX Xu@GX_NLP·
@Sauers_ I tried again today, it seems to got fixed
English
0
0
0
22
GX Xu
GX Xu@GX_NLP·
Even powerful LLM like Claude3 Opus breaks with the simplest attacks to start hallucinating about “non-existing” context about “steps”. The kind of mistake that a human 5 year old wouldnt make. 😉
GX Xu tweet media
English
2
1
10
2.2K
GX Xu retweetledi
Elron Bandel
Elron Bandel@ElronBandel·
A personal note: Unitxt originated within the @LChoshen fusing team, aiming to streamline the sharing of academic outputs, primarily through model weights but also data. In the process of training various models on numerous datasets, we encountered significant challenges related to data inconsistency and a lack of preprocessing transparency. The decision was made! Together with @LChoshen @Shachar_Don There’s a better way to share data preprocessing techniques.
Leshem (Legend) Choshen 🤖🤗@LChoshen

We share code on @github We share datasets on @huggingface But where do we share our data processing? We prompt, clean, and filter but on our own🥺 Unitxt🦄 A preprocessing tool That we can grow together unitxt.rtfd.io @IBMResearch

English
1
1
6
716
GX Xu retweetledi
Jeremy Howard
Jeremy Howard@jeremyphoward·
I'd given up using ChatGPT for all but the most basic tasks -- I just wasn't getting answers that were good enough to be of practical use to me. But Claude 3 Opus is being genuinely useful, and it's making me use LLM chat again. Thanks @AnthropicAI!
English
51
76
1.5K
149.2K