Xinyue Liu

107 posts

Xinyue Liu banner
Xinyue Liu

Xinyue Liu

@irisiris_l

PhD student @sbucompsc. Prev @LTIatCMU

Stony Brook Beigetreten Temmuz 2023
264 Folgt59 Follower
Angehefteter Tweet
Xinyue Liu
Xinyue Liu@irisiris_l·
🙌 Excited to share our new paper and my first project in my PhD journey! We show finetuning on a writing task unlocks verbatim recall of copyrighted books from authors not in the finetuning data. It’s been an incredible experience working with such an amazing group of people ✨
Tuhin Chakrabarty@TuhinChakr

🚨New paper on AI & Copyright 👨‍⚖️Courts have credited LLM companies' claims that safety alignment prevents reproduction of copyrighted expression. But what if fine-tuning on a simple writing task ruins it all? Worse : Fine-tuning on a single author's books (e.g., Murakami) unlocks verbatim recall of copyrighted books from 30+ unrelated authors, sometimes as high as 90%. Joint work with @niloofar_mire (@LTIatCMU), Jane Ginsburg ( @ColumbiaLaw) and my amazing PhD student @irisiris_l (@sbucompsc ) (1/n)🧵

English
3
5
30
3.2K
Xinyue Liu retweetet
Kawin Ethayarajh
Kawin Ethayarajh@ethayarajh·
Is the Internet quietly being rewritten to serve AI agents? How do we even measure this? New paper: We find that post-ChatGPT, listings on Etsy have been systematically reshaped to influence how agents behave—without making humans worse off. We call these “mecha-nudges”.🧵
GIF
English
1
8
15
3.7K
Xinyue Liu
Xinyue Liu@irisiris_l·
🙌 Excited to share our new paper and my first project in my PhD journey! We show finetuning on a writing task unlocks verbatim recall of copyrighted books from authors not in the finetuning data. It’s been an incredible experience working with such an amazing group of people ✨
Tuhin Chakrabarty@TuhinChakr

🚨New paper on AI & Copyright 👨‍⚖️Courts have credited LLM companies' claims that safety alignment prevents reproduction of copyrighted expression. But what if fine-tuning on a simple writing task ruins it all? Worse : Fine-tuning on a single author's books (e.g., Murakami) unlocks verbatim recall of copyrighted books from 30+ unrelated authors, sometimes as high as 90%. Joint work with @niloofar_mire (@LTIatCMU), Jane Ginsburg ( @ColumbiaLaw) and my amazing PhD student @irisiris_l (@sbucompsc ) (1/n)🧵

English
3
5
30
3.2K
Michael Bommarito
Michael Bommarito@mjbommar·
after years of being gaslit by ignorant clowns or bought industry shills, it's wonderful to see the blossoming of truth - that, no shit, minimizing next token loss results in a compressed, often lossless copy of input data. who'da'thunk?
Tuhin Chakrabarty@TuhinChakr

🚨New paper on AI & Copyright 👨‍⚖️Courts have credited LLM companies' claims that safety alignment prevents reproduction of copyrighted expression. But what if fine-tuning on a simple writing task ruins it all? Worse : Fine-tuning on a single author's books (e.g., Murakami) unlocks verbatim recall of copyrighted books from 30+ unrelated authors, sometimes as high as 90%. Joint work with @niloofar_mire (@LTIatCMU), Jane Ginsburg ( @ColumbiaLaw) and my amazing PhD student @irisiris_l (@sbucompsc ) (1/n)🧵

English
2
8
42
2.5K
Xinyue Liu retweetet
Niloofar
Niloofar@niloofar_mire·
One important finding here, besides the copyright ramifications, is that emergent misalignment can occur when you fine-tune on benign-looking data as well, and there are no alarm bells for it. You cannot always predict how your fine-tuning data can have transitive effects and what harms it can cause to other domains. This was a super fun collaboration. Check out our interactive demo: cauchy221.github.io/Alignment-Whac…
Tuhin Chakrabarty@TuhinChakr

🚨New paper on AI & Copyright 👨‍⚖️Courts have credited LLM companies' claims that safety alignment prevents reproduction of copyrighted expression. But what if fine-tuning on a simple writing task ruins it all? Worse : Fine-tuning on a single author's books (e.g., Murakami) unlocks verbatim recall of copyrighted books from 30+ unrelated authors, sometimes as high as 90%. Joint work with @niloofar_mire (@LTIatCMU), Jane Ginsburg ( @ColumbiaLaw) and my amazing PhD student @irisiris_l (@sbucompsc ) (1/n)🧵

English
1
5
41
6.1K
Xinyue Liu retweetet
Niloofar
Niloofar@niloofar_mire·
Fine-tuning *commercial models* (GPT5, Gemini, ...) on one author's data unlocks regurgitation of other authors copyrighted material!! in our new preprint, alignment whack-a-mole🦫 we show emergent misalignment for copyright and memorization! Anyone who has talked to me in the past few weeks has heard my schpeel on how memorization in LLMs is transitive with respect to some latent variable, that models learn shared representations during pretraining, and finetuning on one side of the latent unlocks the other. The latent could be anything such as authorship, style, 'badness' (emergent misalignment), or it could be "copyrighted literary text.", any co-occuring content in pre-training. Amazing work led by @irisiris_l, @TuhinChakr and with Jane Ginsburg!
Niloofar tweet media
English
10
28
199
22.9K
Xinyue Liu retweetet
Thomas Wolf
Thomas Wolf@Thom_Wolf·
This is really cool. It got me thinking more deeply about personalized RL: what’s the real point of personalizing a model in a world where base models can become obsolete so quickly? The reality in AI is that new models ship every few weeks, each better than the last. And the pace is only accelerating, as we see on the Hugging Face Hub. We are not far away from better base models dropping daily. There’s a research gap in RL here that almost no one is working on. Most LLM personalization research assumes a fixed base model, but very few ask what happens to that personalization when you swap the base model. Think about going from Llama 3 to Llama 4. All the tuned preferences, reward signals, and LoRAs are suddenly tied to yesterday’s model. As a user or a team, you don’t want to reteach every new model your preferences. But you also don’t want to be stuck on an older one just because it knows you. We could call this "RL model transferability": how can an RL trace, a reward signal, or a preference representation trained on model N be distilled, stored, and automatically reapplied to model N+1 without too much user involvement? We solved that in SFT where a training dataset can be stored and reused to train a future model. We also tackled a version of that in RLHF phases somehow but it remain unclear more generally when using RL deployed in the real world. There are some related threads (RLTR for transferable reasoning traces, P-RLHF and PREMIUM for model-agnostic user representations, HCP for portable preference protocols) but the full loop seems under-studied to me. Some of these questions are about off-policy but other are about capabilities versus personalization: which of the old customizations/fixes does the new model already handle out of the box, and which ones are actually user/team-specific to ever be solved by default? That you would store in a skill for now but that RL allow to extend beyond the written guidance level. I have surely missed some work so please post any good work you’ve seen on this topic in the comments.
Ronak Malde@rronak_

This paper is almost too good that I didn't want to share it Ignore the OpenClaw clickbait, OPD + RL on real agentic tasks with significant results is very exciting, and moves us away from needing verifiable rewards Authors: @YinjieW2024 Xuyang Chen, Xialong Jin, @MengdiWang10 @LingYang_PU

English
33
64
740
117.5K
Xinyue Liu retweetet
Anthropic
Anthropic@AnthropicAI·
We invited Claude users to share how they use AI, what they dream it could make possible, and what they fear it might do. Nearly 81,000 people responded in one week—the largest qualitative study of its kind. Read more: anthropic.com/features/81k-i…
English
578
973
6.6K
2.7M
Xinyue Liu retweetet
张小珺 Xiaojun Zhang
张小珺 Xiaojun Zhang@zhang_benita·
@sainingxie 一起挑战7小时播客!他刚和Yann LeCun踏上“世界模型”的创业旅程(AMI Labs)。这是他第一次Podcast、第一次访谈。 2026年2月雪后的一天,我们在纽约布鲁克林,从下午2点,开启了一场始料未及的马拉松式访谈,直到凌晨时分散去。 这篇访谈的中文标题叫做《逃出硅谷》,但他又不厌其烦地枚举了影响他学术生涯的每一个人,并反反复复口头描摹这些人的人物特征(侯晓迪、何恺明、杨立昆、李飞飞…)正是这些,让这篇“逃出硅谷”的对话充斥着人性的温度。 By the way, 下面是访谈的YouTube版本,我们提供了中英字幕。 And yes, 我们是在用播客给这个世界建模😎 A 7-hour podcast with Saining Xie. He has just begun a new journey on world models with Yann LeCun at AMI Labs. This was his first podcast appearance and his first long-form interview. A day after the snowfall in February 2026, in Brooklyn, New York, we started recording at 2 p.m. What followed became an unexpected marathon conversation that lasted until the early hours of the morning. The Chinese title of the interview is “Escaping Silicon Valley.” Yet throughout the conversation, he patiently listed the people who shaped his academic life, repeatedly sketching their personalities in vivid detail: Hou Xiaodi, Kaiming He, Yann LeCun, Fei-Fei Li, and others. These portraits are what give this “escape from Silicon Valley” conversation its human warmth. By the way, the YouTube version of the interview is below, with Chinese and English subtitles. And yes, we are using podcasts to model the world 😎 A 7-hour marathon interview with Saining Xie: World Models, AMI Labs, Ya... youtu.be/rIwgZWzUKm8?si… 来自 @YouTube
YouTube video
YouTube
中文
54
180
1.2K
800.2K
Xinyue Liu retweetet
Niloofar
Niloofar@niloofar_mire·
Super cool phenomena, in my head it relates to semantic memorization and leakage, and even cross-modal leakage that i like to term “transitive” memorization. arxiv.org/abs/2408.06518 arxiv.org/abs/2507.17937
Neel Nanda@NeelNanda5

Out of context reasoning is one of the most fascinating developments in the science of how LLMs work. This primer by @OwainEvans_UK, one of the main discoverers of the phenomena, is a great introduction

English
7
11
91
13.8K
Xinyue Liu retweetet
Guri Singh
Guri Singh@heygurisingh·
🚨 Stanford just analyzed the privacy policies of the six biggest AI companies in America. Amazon. Anthropic. Google. Meta. Microsoft. OpenAI. All six use your conversations to train their models. By default. Without meaningfully asking. Here's what the paper actually found. The researchers at Stanford HAI examined 28 privacy documents across these six companies not just the main privacy policy, but every linked subpolicy, FAQ, and guidance page accessible from the chat interfaces. They evaluated all of them against the California Consumer Privacy Act, the most comprehensive privacy law in the United States. The results are worse than you think. Every single company collects your chat data and feeds it back into model training by default. Some retain your conversations indefinitely. There is no expiration. No auto-delete. Your data just sits there, forever, feeding future versions of the model. Some of these companies let human employees read your chat transcripts as part of the training process. Not anonymized summaries. Your actual conversations. But here's where it gets genuinely dangerous. For companies like Google, Meta, Microsoft, and Amazon companies that also run search engines, social media platforms, e-commerce sites, and cloud services your AI conversations don't stay inside the chatbot. They get merged with everything else those companies already know about you. Your search history. Your purchase data. Your social media activity. Your uploaded files. The researchers describe a realistic scenario that should make you pause: You ask an AI chatbot for heart-healthy dinner recipes. The model infers you may have a cardiovascular condition. That classification flows through the company's broader ecosystem. You start seeing ads for medications. The information reaches insurance databases. The effects compound over time. You shared a dinner question. The system built a health profile. It gets worse when you look at children's data. Four of the six companies appear to include children's chat data in their model training. Google announced it would train on teenager data with opt-in consent. Anthropic says it doesn't collect children's data but doesn't verify ages. Microsoft says it collects data from users under 18 but claims not to use it for training. Children cannot legally consent to this. Most parents don't know it's happening. The opt-out mechanisms are a maze. Some companies offer opt-outs. Some don't. The ones that do bury the option deep inside settings pages that most users will never find. The privacy policies themselves are written in dense legal language that researchers people whose job is reading these documents found difficult to interpret. And here's the structural problem nobody is addressing. There is no comprehensive federal privacy law in the United States governing how AI companies handle chat data. The patchwork of state laws leaves massive gaps. The researchers specifically call for three things: mandatory federal regulation, affirmative opt-in (not opt-out) for model training, and automatic filtering of personal information from chat inputs before they ever reach a training pipeline. None of those exist today. The uncomfortable truth is this: every time you type something into ChatGPT, Gemini, Claude, Meta AI, Copilot, or Alexa, you are contributing to a training dataset. Your medical questions. Your relationship problems. Your financial details. Your uploaded documents. You are not the customer. You are the curriculum. And the companies doing this have made it as hard as possible for you to stop.
Guri Singh tweet media
English
329
3.9K
8.6K
1.7M
Niloofar
Niloofar@niloofar_mire·
@dylan522p U think i got any sleep last night?
English
3
0
34
2.6K
Dylan Patel
Dylan Patel@dylan522p·
I was literally just about to go to sleep... now I'ma be up for the next few hours
Dylan Patel tweet media
English
19
39
759
42K
Xinyue Liu retweetet
Augmented Mind Podcast
Augmented Mind Podcast@augmind_fm·
"Maybe human is becoming one of the bottleneck for how much human can benefit from AI" For our second guest we welcome @tongshuangwu, professor at Carnegie Mellon University, whose research sits at the intersection of human-computer interaction and natural language processing. From making AI work for imperfect humans to making humans work better with AI — Sherry's work challenges us to rethink both sides of the equation. 0:00 - Teaser 1:13 - Prelude: Introducing Sherry Wu 2:30 - How the AI Field Has Changed in the Last Four Years 4:22 - Making AI Systems Work for Imperfect Humans 6:54 - Models vs. Scaffolding 10:36 - Understanding Human Imperfection in Teaching Contexts arxiv.org/abs/2509.21890 19:28 - AI Literacy Skills 22:04 - How AI Is Changing CS Education 25:38 - Suppose We Have AGI, What Does It Mean to Be Human? 29:14 - Training Models to be More Human-centered 31:46 - Checklists Are Better Than Reward Models arxiv.org/abs/2507.18624 36:56 - Challenge in Aligning Models 43:22 - Advice for Interdisciplinary Research 45:37 - Reflection on Her Own Research
English
0
6
21
21.5K
Xinyue Liu retweetet
Jiaxin Wen
Jiaxin Wen@jiaxinwen22·
recently I've been chatting with a few new PhD admits about offer decisions. I'm surprised by how often I get asked "will this university/advisor get me a top-tier industry/academia job, or introduce me to investors who'll drop $100M so I can build a startup" With the exponential trend of AI, such long-term forecasting feels increasingly fake. I might be dumb but my honest advice is to pick an environment that: 1) lets you adapt fast (yes including quitting your PhD) 2) makes you happy: vibe check the city, your potential advisor and your labmates. you're living there, not just working there. AI might obsolete the prestige game by 2028, but I don't think it will solve human happiness by 2036.
English
3
10
215
34.2K