Brandon Simonsen
9.9K posts

Brandon Simonsen
@B__Simo
numbers and stuff

A mindworm that has caused great damage among AI researchers is the implicit, universal assumption that every piece of data is a (random) "sample" from a (static) "distribution". This is a valid way of modeling some phenomena, but it isn't applicable to the vast majority of real-world signals.

It’s over (for now). A majority of senators caved to the fearmongering and bush league tactics of the administration and surveillance hawks in Congress, and they sold out Americans’ civil liberties. Section 702 has been reauthorized, not just without any meaningful reforms… 1/10


this is actually hilarious. Quora SEO'd themselves to the top of every search result, and is now serving chatGPT answers on their page, so that's propagating to the answers google gives the internet is dying




So far all evidence that LLMs can perform few-shot reasoning on novel problems seems to boil down to "LLMs store patterns they can reapply to new inputs", i.e. it works for problems that follow a structure the model has seen before, but doesn't work on new problems.





A shocking revelation: Companies don't have "datasets" waiting for you. It's your job to create and maintain them. This is time-consuming and hard, but primarily responsible for the success of your machine learning system. Here are some thoughts: 1 of 7





