Sam Blouir

26 posts

Sam Blouir

Sam Blouir

@SamBlouir_NLP

Intern @ Amazon AGI Foundations Thanks for coming to our AAAI 2025 Foundation Models for Biology workshop!

Fairfax, VA Katılım Mayıs 2022
180 Takip Edilen99 Takipçiler
Sam Blouir retweetledi
Delta Institute @ ICLR
Delta Institute @ ICLR@DeltaInstitutes·
We hosted our first NeurIPS reading group with @Devvrit_Khatri, including lunch sponsored by @AGI_Inc! Huge thanks to Devvrit for presenting his breakthrough paper on RL Scaling Laws from his time at Meta Superintelligence Labs!
Delta Institute @ ICLR tweet media
English
1
1
6
472
Sam Blouir retweetledi
Delta Institute @ ICLR
Delta Institute @ ICLR@DeltaInstitutes·
We hosted our first NeurIPS reading group with @Devvrit_Khatri, including lunch sponsored by @AGI_Inc! Huge thanks to Devvrit for presenting his breakthrough paper on RL Scaling Laws from his time at Meta Superintelligence Labs!
Delta Institute @ ICLR tweet media
English
1
3
10
1.2K
Sam Blouir retweetledi
Defne Circi
Defne Circi@DCirci·
Excited to share that we will be hosting a Duke-site for the 3rd Annual LLM Hackathon for Applications in Materials & Chemistry from September 11–12, 2025. Registration: RSVP for the Duke Hub: lnkd.in/gRPPPDSW Register for the Global Event: lnkd.in/eXD8cmhS
Defne Circi tweet media
English
2
4
29
2K
Sam Blouir
Sam Blouir@SamBlouir_NLP·
#ICML2025: Agents bring NLP advances to genomics - come see "Agents for Genomics” (ACM HCI '25 STAIG) Wed, July 16, 11:30-12 PT at Amazon’s booth. See the evolution from Birdie (EMNLP ’24) to BirdieDNA (ICLR ’25 MLGenX) and AMR (antimicrobial resistance, upcoming).
English
1
1
5
989
Sam Blouir retweetledi
Raj Dabre
Raj Dabre@prajdabre·
🚨🚨🚨New paper alert🚨🚨🚨 Cycle distill can help you go from a pretrained models requiring few shots and unsupervised data to high quality fine tuned models which can work in zero shot. The idea is simple: Stage 0: Prompt a LLM with a few handcrafted shots and generate synthetic fine tuning data. Stage 1 to K: Fine tune on the synthetic data and generate next iteration of synthetic data. We see gains everywhere even after 2 cycles on a small dataset. Much future work to be done. Paper: arxiv.org/abs/2506.19952
Raj Dabre tweet mediaRaj Dabre tweet media
English
5
6
106
17.1K
Sam Blouir retweetledi
Antonis Anastasopoulos
Antonis Anastasopoulos@anas_ant·
📢 I am looking for a postdoc for the next academic year! (Due to the funding source, US persons preferred) Interested in multimodal LLMs and their application to education domains (plus multilinguality, cross-lingual, and low-resource learning)? Contact me here/email if yes!
English
0
13
37
3.9K
Sam Blouir
Sam Blouir@SamBlouir_NLP·
#ICLR2025 & #ACMHCI are a wrap! 🎉 Couldn't ask for better vibes, conversations, and meeting so many other researchers! Thanks for checking out: 🧬 BirdieDNA 🗣️ SLP Sidekick, and 🤖 Agents for Genomics Big thanks to my co-authors @DCirci , Flavia Negrete, Celeste Watkins, Asher Moldwin & Amarda Shehu, absolute rockstars 🤝🙏
English
0
0
5
135
Sam Blouir
Sam Blouir@SamBlouir_NLP·
🚀Thrilled to present 3 papers this weekend! BirdieDNA (MLGenX-ICLR 2025) 4/27 4-5:25 PM SGT SLP Sidekick (Staig-ACM HCI) 4/27 10:15-11:15 JST Agents for Genomics (Staig-ACM HCI) 4/27 13:30-14:30 JST Deeply grateful for my co-authors @DCirci, Flavia Negrete, Celeste Watkins, Asher Moldwin & Amarda Shehu! 🙏 #ACM #ICLR2025
English
0
2
4
442
Roy Xie
Roy Xie@RoyXie_·
Excited to share that I will be joining @Apple Seattle as a Research Intern, working on LLM reasoning/efficiency. Let me know if you would like to connect/meet up!
English
6
2
88
10.4K
Sidharth Pulipaka
Sidharth Pulipaka@psidharth567·
@prajdabre1 Didn't Google figure out long context window almost 1 and a half year ago?? The fact that they are serving 2 million context window at such a low price indicates that DeepSeek isn't even close to what those big US labs have behind the doors. Open source is so behind 😢
English
1
0
0
91
Thomas Ahle
Thomas Ahle@thomasahle·
I find Meta’s original approach to hallucinations delightfully counter intuitive: 1. Extract factoid from training dataset 2. LLM generates (question, answer) pair based on factoid 3. Ask the the question (without fact in context) and judge it against true answer. 4. If answer is wrong, train model to say "I don't know". In a way this is obvious in hindsight, but it goes against ML engineers natural tendency when detecting a wrong answer: Teaching the model the right answer. At least appreciate why this wasn't an obvious thing to do for generation 1 LLMs.
Aran Komatsuzaki@arankomatsuzaki

Meta presents Improving Factuality with Explicit Working Memory Presents EWE, a novel approach that enhances factuality in long-form text generation by integrating a working memory that receives real-time feedback from external resources EWE outperforms strong baselines on four fact-seeking long-form generation datasets, increasing the factuality metric, VeriScore, by 2 to 10 points absolute without sacrificing the helpfulness of the responses. arxiv.org/abs/2412.18069

English
8
21
356
39.3K
Sam Blouir
Sam Blouir@SamBlouir_NLP·
General benchmark scores remain intact across 21 tasks on the EleutherAI LM Eval harness, and greatly improve on our new infilling task. 💡 With smarter training, we maintain SSMs’ efficiencies while dramatically enhancing their capabilities.
Sam Blouir tweet media
English
1
0
5
264