

Dingli Yu
25 posts

@dingli_yu
Researcher @ OpenAI | PhD from Princeton
















Fine-tuning can improve chatbots (e.g., Llama 2-Chat, GPT-3.5) on downstream tasks — but may unintentionally break their safety alignment. Our new paper: Adding a safety prompt is enough to largely mitigate the issue, but be cautious about when to add it! arxiv.org/abs/2402.18540

We are excited to introduce the PLI Blog! pli.princeton.edu/blog First post by @prfsanjeevarora, "Are Language Models Mere Stochastic Parrots? The SkillMix Test Says NO." bit.ly/47PpKp4



Nontrivial ∞width neural nets are either kernel machines or feature learners. Latter's scaling makes optimal hyperparams invariant to width What if depth→∞as well? 🆕 Feature diversity is key; maxed out by abs (not relu); gives invariance to depth! But GPT flawed 🧵
