Allen Chang

85 posts

Allen Chang banner
Allen Chang

Allen Chang

@AllenCChang

PhD student @upennnlp. Prev @USC

Katılım Nisan 2022
423 Takip Edilen196 Takipçiler
Sabitlenmiş Tweet
Allen Chang retweetledi
Jesse Thomason
Jesse Thomason@_jessethomason_·
For prospective PhD students, I plan to hire in this coming application cycle (Fall 2026) with a focus on robotics, speech, and signed languages.
English
1
2
10
1.6K
Allen Chang retweetledi
Yue Yang
Yue Yang@YueYangAI·
🎯 We release MolmoPoint, the best open model in GUI grounding 💻 by training on purely synthetic screenshots. We open-source all our models, data, and generation code. Plug it into your agents! Demo: huggingface.co/spaces/allenai… Model: huggingface.co/allenai/MolmoP… Data: huggingface.co/datasets/allen… Code: github.com/allenai/MolmoP…
Yue Yang tweet media
Ai2@allen_ai

Grounding lets vision-language models do more than describe—they can point to where a robot should grasp, which button to click, or which object to track across video frames. Today we're releasing MolmoPoint, a better way for models to point. 🧵

English
0
12
84
7.2K
Allen Chang retweetledi
Rulin Shao
Rulin Shao@RulinShao·
🔥Thrilled to introduce DR Tulu-8B, an open long-form Deep Research model that matches OpenAI DR 💪Yes, just 8B! 🚀 The secret? We present Reinforcement Learning with Evolving Rubrics (RLER) for long-form non-verifiable DR tasks! Our rubrics: - co-evolve with the policy model - are grounded on search knowledge 🧵
Rulin Shao tweet media
English
8
111
549
129.4K
Allen Chang retweetledi
Alex Spangher @ Neurips2025
Alex Spangher @ Neurips2025@AlexanderSpangh·
✨ Very overdue update: I'll be starting as an Assistant Professor in CS at University of Minnesota, Twin Cities, Fall 2026. I will be recruiting PhD students!! Please help me spread the word! [Thread] 1/n
Alex Spangher @ Neurips2025 tweet media
English
40
142
742
91.7K
Allen Chang retweetledi
Taylor Sorensen
Taylor Sorensen@ma_tay_·
🤖➡️📉 Post-training made LLMs better at chat and reasoning—but worse at distributional alignment, diversity, and sometimes even steering(!) We measure this with our new resource (Spectrum Suite) and introduce Spectrum Tuning (method) to bring them back into our models! 🌈 1/🧵
Taylor Sorensen tweet media
English
5
45
198
68K
Allen Chang retweetledi
Leena Mathur
Leena Mathur@lmathur_·
Future AI systems interacting with humans will need to perform social reasoning that is grounded in behavioral cues and external knowledge. We introduce Social Genome to study and advance this form of reasoning in models! New paper w/ Marian Qian, @pliang279, & @lpmorency!
Leena Mathur tweet media
English
2
11
38
6K
Allen Chang retweetledi
Tianyi Lorena Yan
Tianyi Lorena Yan@LorenaYannnnn·
When answering queries with multiple answers (e.g., listing cities of a country), how do LMs simultaneously recall knowledge and avoid repeating themselves? 🚀 Excited to share our latest work with @robinomial! We uncover a promote-then-suppress mechanism: LMs first recall all answers and then suppress previously generated ones. arxiv.org/abs/2502.20475 👇🧵
Tianyi Lorena Yan tweet media
English
4
20
110
16.5K
Allen Chang retweetledi
Tejas Srinivasan
Tejas Srinivasan@_Tejas_S_·
People are relying on AI assistance to make all kinds of decisions. *How* they incorporate AI recommendations is influenced by previous user-AI interactions and their evolving trust in the AI, which AI assistants are typically blind to. But what if they weren’t? We show that having AI assistants adapt their behavior in response to user trust levels can mitigate under- and over-reliance! Pre-print: arxiv.org/abs/2502.13321
Tejas Srinivasan tweet media
English
2
14
64
5.6K
Allen Chang retweetledi
Liam Dugan
Liam Dugan@LiamDugan_·
Last Friday I gave an hour long talk at the Penn ILST Seminar about the particular linguistic features that characterize AI text (e.g. "delve", repetitive syntax, agreeable tone) and how they affect detectability. Highly recommend giving it a listen. youtube.com/watch?v=j73X_R…
YouTube video
YouTube
English
0
5
18
1.8K
Nathan Dennler
Nathan Dennler@ndennler·
I successfully defended my dissertation (and finished all the fun paperwork to make it official)!!! My dissertation, “Physical and Social Adaptation for Assistive Robot Interactions," develops techniques to allow robots to efficiently adapt to users’ personal preferences.
Nathan Dennler tweet media
English
6
3
58
3.8K
Allen Chang retweetledi
Allen Chang retweetledi
Leena Mathur
Leena Mathur@lmathur_·
Our workshop will start in a few hours! > #ECCV2024 9/29 AM workshop > Suite 2, Allianz MiCo 🇮🇹 > Zoom info on our website (QR code below) Looking forward to the discussion today and learning from our keynote speakers! sites.google.com/andrew.cmu.edu…
Leena Mathur tweet media
Leena Mathur@lmathur_

In a few weeks at #ECCV2024, we will have the 3rd edition of the Artificial Social Intelligence Workshop! This workshop will occur on September 29 in Milan 🇮🇹, with an interactive hybrid option available, as well sites.google.com/andrew.cmu.edu…

English
0
4
13
2.2K
Allen Chang retweetledi
Ai2
Ai2@allen_ai·
Meet Molmo: a family of open, state-of-the-art multimodal AI models. Our best model outperforms proprietary systems, using 1000x less data. Molmo doesn't just understand multimodal data—it acts on it, enabling rich interactions in both the physical and virtual worlds. Try it for yourself: molmo.allenai.org
English
63
277
1.3K
515K
Allen Chang retweetledi
Tuhin Chakrabarty
Tuhin Chakrabarty@TuhinChakr·
GPT4-o1-preview from @OpenAI now gets 80.4% (compared to 14% performance of GPT4o) on the Connections game in 1 single attempt. Saw a thread on LinkedIn about similar bump on Wordle. I also attached some other models in comparison. This is very impressive given how hard the task is As someone who isn't so much about LLM scientivism, I am very confident the model was trained on these tasks. A sad and depressing trend where these models try to incorporate everything in training distribution make it super hard for researchers interested in generalization #LLM #GenAI
Tuhin Chakrabarty tweet mediaTuhin Chakrabarty tweet media
Tuhin Chakrabarty@TuhinChakr

New paper with students @BarnardCollege on testing orthogonal thinking / abstract reasoning capabilities of Large Language Models using the fascinating yet frustratingly difficult @nytimes Connections game. #NLProc #LLMs #GPT4o #Claude3opus 🧵(1/n)

English
6
3
23
6.2K
Allen Chang retweetledi
Sachin Kumar
Sachin Kumar@shocheen·
You think your model just fell out of a coconot tree 🥥? It should not always comply in the context of all it has seen in the request. Check out our paper on contextual noncompliance.
AK@_akhaliq

The Art of Saying No Contextual Noncompliance in Language Models Chat-based language models are designed to be helpful, yet they should not comply with every user request. While most existing work primarily focuses on refusal of "unsafe" queries, we posit that the scope of noncompliance should be broadened. We introduce a comprehensive taxonomy of contextual noncompliance describing when and how models should not comply with user requests. Our taxonomy spans a wide range of categories including incomplete, unsupported, indeterminate, and humanizing requests (in addition to unsafe requests). To test noncompliance capabilities of language models, we use this taxonomy to develop a new evaluation suite of 1000 noncompliance prompts. We find that most existing models show significantly high compliance rates in certain previously understudied categories with models like GPT-4 incorrectly complying with as many as 30% of requests. To address these gaps, we explore different training strategies using a synthetically-generated training set of requests and expected noncompliant responses. Our experiments demonstrate that while direct finetuning of instruction-tuned models can lead to both over-refusal and a decline in general capabilities, using parameter efficient methods like low rank adapters helps to strike a good balance between appropriate noncompliance and other capabilities.

English
3
8
57
16K
Allen Chang
Allen Chang@AllenCChang·
@_Tejas_S_ @jieyuzhao11 Ugh, sorry to hear that you had to go through this 6 times ☠️. Can't imagine what else goes on behind closed doors
English
0
0
0
394
Tejas Srinivasan
Tejas Srinivasan@_Tejas_S_·
Really heartbreaking to see Wenda and @jieyuzhao11's experiences with CBP. I also got taken to the "detention room" yesterday (sixth time 🤙), the demographic there is really telling. CBP officers love going on a power trip and barking at and bullying vulnerable PoCs.
English
15
13
144
33.8K
Allen Chang retweetledi
Tejas Srinivasan
Tejas Srinivasan@_Tejas_S_·
Our work on improving selective prediction for VLMs has been accepted to #ACL2024 Findings! Read on to learn how you can make your VLM both reliable *and* usable ✨ Paper: arxiv.org/abs/2402.15610 Code: github.com/tejas1995/ReCo…
Tejas Srinivasan@_Tejas_S_

When vision-language models are uncertain about their answers, abstaining (“I don’t know”) enhances system reliability, but at the cost of utility. We introduce ReCoVERR (arxiv.org/abs/2402.15610) to mitigate over-abstention in VLM systems without sacrificing prediction accuracy.

English
1
7
45
10.3K