Xiang Lisa Li

40 posts

Xiang Lisa Li

Xiang Lisa Li

@XiangLisaLi2

PhD student at Stanford

Katılım Mayıs 2019
240 Takip Edilen3.3K Takipçiler
Xiang Lisa Li retweetledi
Percy Liang
Percy Liang@percyliang·
What would truly open-source AI look like? Not just open weights, open code/data, but *open development*, where the entire research and development process is public *and* anyone can contribute. We built Marin, an open lab, to fulfill this vision:
Percy Liang tweet media
English
64
221
1.2K
190.6K
Xiang Lisa Li retweetledi
Percy Liang
Percy Liang@percyliang·
When @XiangLisaLi2 built diffusion LMs in 2022 (arxiv.org/abs/2205.14217), we were interested in more powerful controllable generation (inference-time conditioning on an arbitrary reward), but inference was slow. Interestingly, the main advantage now is speed. Impressive to see how far diffusion LMs have come!
Inception@_inception_ai

We are excited to introduce Mercury, the first commercial-grade diffusion large language model (dLLM)! dLLMs push the frontier of intelligence and speed with parallel, coarse-to-fine text generation.

English
1
17
122
18.5K
Xiang Lisa Li retweetledi
John Hewitt
John Hewitt@johnhewtt·
I’m hiring PhD students in computer science at Columbia! Our lab will tackle core challenges in understanding and controlling neural models that interact with language. for example, - methods for LLM control - discoveries of LLM properties - pretraining for understanding
English
18
154
873
106.6K
Xiang Lisa Li retweetledi
Percy Liang
Percy Liang@percyliang·
This year, I have 4 exceptional students on the academic job market, and they couldn’t be more diffferent, with research spanning AI policy, robotics, NLP, and HCI. Here’s a brief summary of their research, along with one representative work each:
English
7
43
679
122.5K
Xiang Lisa Li retweetledi
Percy Liang
Percy Liang@percyliang·
Lisa Li (@XiangLisaLi2) changes how people fine-tune (prefix tuning, the original PEFT), generate (diffusion LM, non-autoregressively), improve (GV consistency fine-tuning without supervision), and evaluate language models (using LMs). Prefix tuning: arxiv.org/abs/2101.00190
English
1
8
80
13.6K
Xiang Lisa Li retweetledi
Transluce
Transluce@TransluceAI·
Eliciting Language Model Behaviors with Investigator Agents We train AI agents to help us understand the space of language model behaviors, discovering new jailbreaks and automatically surfacing a diverse set of hallucinations. Full report: transluce.org/automated-elic…
Transluce tweet media
English
6
15
83
22K
Xiang Lisa Li retweetledi
Chunting Zhou
Chunting Zhou@violet_zct·
Introducing *Transfusion* - a unified approach for training models that can generate both text and images. arxiv.org/pdf/2408.11039 Transfusion combines language modeling (next token prediction) with diffusion to train a single transformer over mixed-modality sequences. This allows us to leverage the strengths of both approaches in one model. 1/5
Chunting Zhou tweet mediaChunting Zhou tweet media
English
24
213
1K
236.6K
Xiang Lisa Li
Xiang Lisa Li@XiangLisaLi2·
Other than knowledge-intensive QA, we use AutoBencher to create datasets for math and multilingual. The scalability of AutoBencher allows it to test fine-grained categories and tail knowledge, creating datasets that are more novel and more difficult than existing benchmarks.
Xiang Lisa Li tweet media
English
1
0
15
2.2K
Xiang Lisa Li
Xiang Lisa Li@XiangLisaLi2·
arxiv.org/abs/2407.08351 LM performance on existing benchmarks is highly correlated. How do we build novel benchmarks that reveal previously unknown trends? We propose AutoBencher: it casts benchmark creation as an optimization problem with a novelty term in the objective.
Xiang Lisa Li tweet media
English
9
67
304
58.1K
Xiang Lisa Li retweetledi
Kelvin Guu
Kelvin Guu@kelvin_guu·
New from @GoogleDeepMind: When can you trust your LLM? We show that LLMs consistently overestimate their own accuracy on some topics (eg nutrition) while underestimating it on others (eg math). Our Few-shot Recalibrator fixes LLM over/under-confidence: arxiv.org/abs/2403.18286 🧵
English
2
14
78
7.5K
Xiang Lisa Li retweetledi
Jesse Mu
Jesse Mu@jayelmnop·
Prompting is cool and all, but isn't it a waste of compute to encode a prompt over and over again? We learn to compress prompts up to 26x by using "gist tokens", saving memory+storage and speeding up LM inference: arxiv.org/abs/2304.08467 (w/ @XiangLisaLi2 and @noahdgoodman) 🧵
English
14
113
582
160.6K
Xiang Lisa Li retweetledi
Stanford NLP Group
Stanford NLP Group@stanfordnlp·
And here at the #cs224n #NLProc with Deep Learning poster session at Tressider (@Stanford) is almost all of the (large!) teaching team.
Stanford NLP Group tweet media
English
0
7
45
18.3K
Xiang Lisa Li retweetledi
Stanford NLP Group
Stanford NLP Group@stanfordnlp·
The #cs224n poster session is happening now! We are super excited about amazing, cutting-edge NLP posters from ~650 students!
Stanford NLP Group tweet mediaStanford NLP Group tweet mediaStanford NLP Group tweet mediaStanford NLP Group tweet media
English
2
10
56
25.7K
Xiang Lisa Li retweetledi
Omar Khattab
Omar Khattab@lateinteraction·
Introducing Demonstrate–Search–Predict (𝗗𝗦𝗣), a framework for composing search and LMs w/ up to 120% gains over GPT-3.5. No more prompt engineering.❌ Describe a high-level strategy as imperative code and let 𝗗𝗦𝗣 deal with prompts and queries.🧵 arxiv.org/abs/2212.14024
Omar Khattab tweet media
English
31
186
958
224.5K
Xiang Lisa Li retweetledi
Mina Lee
Mina Lee@MinaLee__·
Language models (LMs) are already deployed in many real-world applications and used to interact with users 👩‍🦰, but these models are primarily evaluated non-interactively. How can we evaluate LMs interactively and why is it important? (1/8)
Mina Lee tweet media
English
5
57
325
64.4K
Xiang Lisa Li retweetledi
Weijia Shi
Weijia Shi@WeijiaShi2·
🙋‍♀️How to present the same text in diff. tasks/domains as diff. embeddings W/O training? We introduce Instructor👨‍🏫, an instruction-finetuned embedder that can generate text embeddings tailored to any task given the task instruction➡️sota on 7⃣0⃣tasks👇! instructor-embedding.github.io
Weijia Shi tweet media
English
11
111
585
100.2K