Tal Linzen
16.2K posts

Tal Linzen
@tallinzen
Professor @nyuling and @NYUDataScience, research scientist @GoogleAI, inventor of the word "bertology"


London absolutely slaps in the sun — after almost 30 years of being here, it still doesn’t fail to amaze me


There’s been plethora of work going on to understand models through lenses of science. The Neuroscience of Transformers: arxiv.org/pdf/2603.15339… Biology of LLMs: transformer-circuits.pub/2025/attributi… Physics of LLMs: physics.allen-zhu.com Neuroscientific approach to interpretability of LLMs: arxiv.org/pdf/2502.12131 Entropy, Thermodynamics and the Geometrization of LLMs: arxiv.org/pdf/2407.21092 LLMs and Cognitive Science: arxiv.org/pdf/2409.02387 A Statistical Physics of LLM Reasoning :arxiv.org/pdf/2506.04374

1/ New paper from @ylecun et al on alternative approach for AI to learn more biologically... paper basically says AI is super smart but still can't learn like a toddler can... the main critique







I never want to review an ML paper ever again. Most of the good ML researchers go work in industry instead of submitting public papers, so ML conference papers are adversely selected and are on average terrible. This mood brought to you by: having to review ICML.



I think we're going to need CS PhD students to do far more than provide accountability, by which I think Sayash means do code review for AI agents and make sure the agent isn't making silly mistakes. The main value of a strong PhD student for a PI is that they're immersed in a problem, a method, an application, a collaboration with another field; they are obsessed with finding the next question to ask, not just executing the experiments their advisor asks them to do. I simply wouldn't be able to work on the range of things I'm able to work on if I were going it on my own, even if all of my code was generated instantaneously by an agent.

Coding is only a small fraction of what a CS PhD student actually does; perhaps 10–30% of their time. The real goal of a PhD, and of being a professor, is not to outsource research work but to educate and train the next generation of scientists: people who deeply understand their field, can think critically about it, and ultimately become experts capable of pushing the frontier of knowledge forward. Coding is probably one of the least interesting part of being a CS PhD. We need human experts even more than before at the age of AI.


In the last few months, I've spoken to many CS professors who asked me if we even need CS PhD students anymore. Now that we have coding agents, can't professors work directly with agents? My view is that equipping PhD students with coding agents will allow them to do work that is orders of magnitude more impressive than they otherwise could. And they can be *accountable* for their outcomes in a way agents can't (yet). For example, who checks the agent's outputs are correct? Who is responsible for mistakes or errors?


In the last few months, I've spoken to many CS professors who asked me if we even need CS PhD students anymore. Now that we have coding agents, can't professors work directly with agents? My view is that equipping PhD students with coding agents will allow them to do work that is orders of magnitude more impressive than they otherwise could. And they can be *accountable* for their outcomes in a way agents can't (yet). For example, who checks the agent's outputs are correct? Who is responsible for mistakes or errors?

In the last few months, I've spoken to many CS professors who asked me if we even need CS PhD students anymore. Now that we have coding agents, can't professors work directly with agents? My view is that equipping PhD students with coding agents will allow them to do work that is orders of magnitude more impressive than they otherwise could. And they can be *accountable* for their outcomes in a way agents can't (yet). For example, who checks the agent's outputs are correct? Who is responsible for mistakes or errors?





