
August Wester
432 posts

August Wester
@augustwester
ML research & engineering at @veotechnologies. I like deep learning, and I occasionally blog about it on https://t.co/Nt7ezYxQ6Q.






This is precisely what statisticians used to say. "If you can predict well the next data point, they you understand the reality or the process that generates that point". They said it and said it until Causal Inference proved them wrong. Understanding takes more than prediction. #Bookofwhy










Does GPT understand the world? Here is what @ilyasut, co-founder of OpenAI, says during a discussion with Jensen Huang, CEO of Nvidia: (1) When we train a large neural network to accurately predict the next word in lots of different texts from the internet, the AI is learning a world model. (2) On the surface, it may look like learning correlations in text, but it turns out that to 'just learn' statistical correlations in text, to compress information really well, what the neural network learns is some representation of the process that produced the text. (3) This text is a projection of the world...what the neural network is learning is aspects of the world, of people, of the human conditions, their hopes, dreams, motivations, their interactions...the situations we are in. The neural network learns a compressed, abstract, usable representation." Do you think learning representations = understanding? Are large language models simply stochastic parrots, or are they much more?

















