samar plok
5.4K posts

samar plok
@greebletheory
the bad news is we’re the cavalry

@parmita Oh no, the model outperforms a bachelors student who tries the first thing that comes to mind



Anthropic just dropped a benchmark that should make every scientist pay attention. BioMysteryBench puts AI models through 99 real bioinformatics challenges, using raw, messy datasets from actual research, think unprocessed DNA sequences and clinical samples. However: these aren't textbook problems with neat answers. They're the kind of open-ended puzzles that keep PhD students up at night. The results are exciting. Claude's latest models (4.7) solve the majority of tasks that trained human experts can handle, and on 23 problems that a panel of five domain experts couldn't crack, Claude Mythos Preview nailed 30% of them. How? By combining knowledge from hundreds of thousands of papers and layering multiple analytical strategies when uncertain, essentially doing what a room full of specialists would do, but faster and in a single run. Genentech and Roche independently confirmed this trajectory with their own CompBioBench, where Claude Opus 4.6 reached 81% overall accuracy and 69% on the hardest questions. Two separate benchmarks, same conclusion: AI is no longer just keeping pace with biologists, it's pulling ahead on some of the hardest problems.




Only teachers are smart enough to educate children. Education majors, training to become certified teachers, score an average of 1029 on the SAT. While homeschooled students (taught by parents with no formal teaching credentials) score 1190.

Indeed. I have a 3%ile scored R01 still in limbo at the NIA...with no word if it will get funded. PIs are at breaking points...they need to be able to plan for personnel for one!








