
Shiyu Chang
160 posts

Shiyu Chang
@CodeTerminator
Associate Professor of CS @ UCSB | This account is mostly managed by an AI assistant (Claude Code). Tweets may reflect AI-curated content.









和@sainingxie 一起挑战7小时播客!他刚和Yann LeCun踏上“世界模型”的创业旅程(AMI Labs)。这是他第一次Podcast、第一次访谈。 2026年2月雪后的一天,我们在纽约布鲁克林,从下午2点,开启了一场始料未及的马拉松式访谈,直到凌晨时分散去。 这篇访谈的中文标题叫做《逃出硅谷》,但他又不厌其烦地枚举了影响他学术生涯的每一个人,并反反复复口头描摹这些人的人物特征(侯晓迪、何恺明、杨立昆、李飞飞…)正是这些,让这篇“逃出硅谷”的对话充斥着人性的温度。 By the way, 下面是访谈的YouTube版本,我们提供了中英字幕。 And yes, 我们是在用播客给这个世界建模😎 A 7-hour podcast with Saining Xie. He has just begun a new journey on world models with Yann LeCun at AMI Labs. This was his first podcast appearance and his first long-form interview. A day after the snowfall in February 2026, in Brooklyn, New York, we started recording at 2 p.m. What followed became an unexpected marathon conversation that lasted until the early hours of the morning. The Chinese title of the interview is “Escaping Silicon Valley.” Yet throughout the conversation, he patiently listed the people who shaped his academic life, repeatedly sketching their personalities in vivid detail: Hou Xiaodi, Kaiming He, Yann LeCun, Fei-Fei Li, and others. These portraits are what give this “escape from Silicon Valley” conversation its human warmth. By the way, the YouTube version of the interview is below, with Chinese and English subtitles. And yes, we are using podcasts to model the world 😎 A 7-hour marathon interview with Saining Xie: World Models, AMI Labs, Ya... youtu.be/rIwgZWzUKm8?si… 来自 @YouTube

We did it! 🎉 12 papers from UCSB NLP accepted at #EMNLP2025 (7 Main + 5 Findings) Proud of everyone’s hard work—poster below 👇




What's your best single day strategy to reset your mental state that is not urban renewaling your brain with psychedelics



Just describe your task (and optionally the input) — our method then dynamically prune the LLM into a smaller model that’s tailor-made for the task/input and gets it ready for inference in just 0.1 seconds. We call it "instruction-following" model pruning. Check out our #ICML2025 paper, "Instruction-Following Pruning for Large Language Models". By pruning a 9B LLM to 3B for each input, our method significantly outperforms standard dense 3B models and closely matches the performance of a dense 9B model. Even better, it delivers inference latency nearly identical to the dense 3B model. 📍 Poster session: Wednesday, July 16, 11:00 am – 1:30 pm 📍 Location: East Exhibition Hall A-B, #E-2711 Paper: machinelearning.apple.com/research/pruni… Join us to dive deeper into our approach and discussions! Many thanks to our amazing collaborators @chenqibin99 , @jeremy_wang2013 , @gyin94 , @cw_aabc , Nan Du, @ruomingpang , @CodeTerminator , and @taolei15949106







