Pierre

725 posts

Pierre

Pierre

@therealpeterobi

Statistical Physics x AI

United States Katılım Kasım 2012
834 Takip Edilen231 Takipçiler
Pierre
Pierre@therealpeterobi·
@curiouswavefn 2-3 weeks to set up and run MD is not typical.
English
0
0
0
117
Ash Jogalekar
Ash Jogalekar@curiouswavefn·
🧵: Just constructed a complex, multistep agentic pipeline using multiple LLMs for a real-world biotech / drug discovery problem. ~12 tools 7 steps (docking → MD → ML ADME models) CPU + GPU workloads Total runtime: ~2 hours Equivalent traditional time: 2–3 weeks.
English
8
3
51
6K
Chris Hayduk
Chris Hayduk@ChrisHayduk·
@therealpeterobi Not yet but putting it together soon! Need to compile a training and val set first and then will be good to go
English
1
0
2
41
Chris Hayduk
Chris Hayduk@ChrisHayduk·
Fantastic idea and I think this will actually be the better direction to take my NanoAF2 competition. Data is the scarce resource when building frontier bio ML models, so better per-sample efficiency will be a gamechanger.
Samip@industriaalist

1/ Introducing NanoGPT Slowrun 🐢: an open repo for state-of-the-art data-efficient learning algorithms. It's built for the crazy ideas that speedruns filter out -- expensive optimizers, heavy regularization, SGD replacements like evolutionary search.

English
1
0
18
2K
Frank Noe
Frank Noe@FrankNoeBerlin·
@therealpeterobi @MSFTResearch You can't, but once you can sample the rare events efficiently, that gives you a tool to train the model to be good at rare events too. TBH, I would not expect very rare events to be correctly predicted by BioEmu, which has been trained on semi-equilibrium data.
English
1
0
2
156
Pierre
Pierre@therealpeterobi·
Nothing’s about to happen. A lot of AI workflows fail outside in the wild even in the simplest domains. Look at code generation which should be ‘easy’ as it is more structured and look at all the harnessing and tooling being done around it. At this pace, doctors will be fine for a very long time
English
0
0
0
273
Pierre
Pierre@therealpeterobi·
@otis_reid It is a whole field of research. Worked with some circadian rhythm researchers studying these effects, fascinating stuff.
English
0
0
3
106
Elliot Arledge
Elliot Arledge@elliotarledge·
"The Physics of LLM Inference" is out for 5 dollars. This is NOT the other CUDA book I've been hyping up.
Elliot Arledge tweet media
English
9
12
169
20.3K
Pierre
Pierre@therealpeterobi·
Reading the Titans memory paper from Goggle and wondering, it’s that easy?
English
0
0
2
37
Pierre
Pierre@therealpeterobi·
One of the best feelings: doing something totally unrelated and suddenly seeing it differently. I ran a reshape in JAX and my brain went straight to memory access patterns (kernel-writing will do that to you).
English
0
0
3
40
Pierre
Pierre@therealpeterobi·
Randomly remembered Le Chatelier's principle. Chemistry was sometimes fun
English
0
0
0
37
Pierre
Pierre@therealpeterobi·
Learnt some Jax over the holidays and rewrote part of my numpy-heavy MD workflow. Started running a benchmark, went to a NYE party, came back and saw this for CPU and GPU fold speed increase. Still need to clean up the code and recheck timings but seems like another win for parallelism
Pierre tweet media
English
0
0
1
46
Pierre
Pierre@therealpeterobi·
Membrane-embedded proteins account for ~1/3 of approved drugs, yet many binding sites sit within the membrane itself. How do you design ligands that are lipophilic enough to enter the membrane, yet polar enough to bind specifically? Out now in the Journal of Medicinal Chemistry, our work studies TRPA1 antagonists and explores chameleonic efficiency as a quantitative guide for designing ligands targeting lipid-facing sites. #MedicinalChemistry #DrugDesign #MembraneProteins #IonChannels #TRPA1 pubs.acs.org/doi/10.1021/ac…
English
0
1
1
43
Pierre
Pierre@therealpeterobi·
Learning CUDA has changed the way I see GPUs and my everyday scientific work. I had a protein-ligand interaction script that was fine for a single MD simulation, but once I started running many replicates, the post-processing alone could take ~4 hours on CPU. The key realization was that most MD analyses are snapshot-independent. You don’t need the previous frame to decide whether a hydrogen bond exists in the current one. That makes it textbook data parallelism. Batch the snapshots and let GPU threads work in parallel. Same analysis, different mental model and a massive speedup.
English
0
0
0
35
Pierre
Pierre@therealpeterobi·
Day 28 - Focused on improving tiling and reducing memory access while working through Chapter 7 of PMPP. Chapters 1–6 really are the foundation; everything after (including 7) depends heavily on those core concepts. Also started prototyping an FP4 GEMM kernel for the second Nvidia challenge due Dec 19. Solved a few GPU puzzles from @LeetGPU, these were surprisingly fun! It’s one thing to understand a matrix transpose, but thinking through the global implications beyond element-wise behavior while tiling was a great exercise. Finally, began reading through the OpenFold 3 codebase. I’ll share more soon, but I’m exploring a very exciting direction here!
Pierre tweet media
English
0
0
1
53
Pierre
Pierre@therealpeterobi·
Finally got to the 3rd book of the series. Need to allocate some time for CUDA kernels.
Pierre tweet media
English
0
0
0
24
Pierre
Pierre@therealpeterobi·
11/29/26 update reposted, deleted in error. Day 21 – Travel and holidays meant mostly reading (bank conflicts, thread coarsening, FP4 dequant etc.) Wrote a GEMV kernel for the NVFP4 competition, not a prize winner, but it inspired the whole learning streak. Working toward more advanced kernels for the future rounds
Pierre tweet media
English
0
0
1
65
Pierre
Pierre@therealpeterobi·
Day 13. Busy in IRL, slower progress. Studied CUDA memory + tiling, wrote a tiled GEMM with dynamic shared memory. Hits ~10% of cuBLAS on large matrices and up to 2× on tiny ones (overhead wins here). New strategy: just make the matrices smaller.
Pierre tweet media
English
0
0
1
55
Pierre
Pierre@therealpeterobi·
Day 7 of CUDA/C++. @GPU_MODE convinced me this was a good idea for the competition. Now I’m hand drawing MatMul tiles like a confused architect and rewriting kernels from memory every 2 days. Somehow… it’s working?
Pierre tweet media
English
0
0
0
58
Pierre
Pierre@therealpeterobi·
Watched this video a couple of years ago, still the best video explaining tensors.
Pierre tweet media
English
0
0
0
29