okay (they/them) 🐀🇵🇸
11.5K posts

okay (they/them) 🐀🇵🇸
@OkayHughes
diet clever, sugar-cane sweet. they/them. 🏳️⚧️ climate+math. Ph.D student in climate. abolish the prison industrial complex. spicier acct: @kayohughes

Midwesterners are nice. Southerners are polite and may or may not be nice. Many of the most polite southerners I know are decidedly not nice.


I occasionally have my doubts about the Bay Area flavoured monoculture of Al hyper-bullishness, but occasionally I look at what the smarmy skeptics are offering and remind myself the alternative is even bleaker. All the confidence, none of the imagination.






A lot of people confused this AI wave for marketing and didn’t know that 4 enormous experimental breakthroughs happened over the last decades: 1) Deep neural nets generalize 2) Dot product attention works well for sequence modeling 3) Deep neural nets trained to model text sequences learn general concepts 4) Such nets are easily fine-tuned to use and apply general reasoning patterns and tools 1 - 2012 and before, AlexNet & many more 2 - 2016-2018 Attention is all you need, BeRT 3 - 2020 Language Models are Few Shot Learners, GPT3 4 - 2021-2024 Instruct-GPT, SFT, o1/R1

IJBOL the Pitt producers said the ICE storyline would be “balanced”



We've reached an agreement to acquire Astral. After we close, OpenAI plans for @astral_sh to join our Codex team, with a continued focus on building great tools and advancing the shared mission of making developers more productive. openai.com/index/openai-t…

I’ve been at a small conference this week, one where the AI people have been presenting early in the week and the domain science people will be presenting later in the week. At the end of the talks last night, the conversation turned very doomer with all the AI people talking about how well Claude Code or Codex can do hill-climbing AI research and how we (the AI people) are maybe all about to lose our jobs! The domain science people expressed their shock at this attitude because, though Claude Code can be let loose to complete lots of banal hill-climbing AI research projects, basically no experimental science is hill-climbing or even metric driven. Most scientific fields are about much more taste-driven exploration that is incredibly difficult to make metrics for or to parameterize, and this misunderstanding from the AI community is one of the most damaging things to the realization of great science with AI. Seems like we’re actually pretty far from having AI models do that… Over the summer, @evijit and I wrote about this (and some other things hindering AI for science) at a bit more length, and today that work is out in Patterns! So, if you care about these problems and the real challenges in bringing AI to science in the real work, I recommend giving it a read!






