Daniel P Jeong

113 posts

Daniel P Jeong banner
Daniel P Jeong

Daniel P Jeong

@danielpjeong

PhD student @mldcmu @SCSatCMU + intern @AbridgeHQ | ML for Healthcare | Previously @MSFTResearch, @columbia, @nasajpl

Pittsburgh, PA Katılım Kasım 2017
455 Takip Edilen165 Takipçiler
Sabitlenmiş Tweet
Daniel P Jeong
Daniel P Jeong@danielpjeong·
🧵 Are "medical" LLMs/VLMs *adapted* from general-domain models, always better at answering medical questions than the original models? In our oral presentation at #EMNLP2024 today (2:30pm in Tuttle), we'll show that surprisingly, the answer is "no". arxiv.org/abs/2411.04118
English
2
34
104
24.1K
Daniel P Jeong retweetledi
Shahriar Noroozizadeh
Shahriar Noroozizadeh@ShNoroozi·
1/ Excited to present our #ICLR2026 paper today at the poster session! Come say hi: Pavilion 3 P3-#101 (10:30 AM – 1:00 PM 🇧🇷). “SURVHTE-BENCH: A Benchmark for Heterogeneous Treatment Effect Estimation in Survival Analysis” Here is why we built it, and why it matters 🧵👇
Shahriar Noroozizadeh tweet mediaShahriar Noroozizadeh tweet media
English
2
4
7
563
Dylan Sam
Dylan Sam@dylanjsam·
I defended my PhD thesis! Also, a very (~4 month) late life update, but I've joined @OpenAI to work on safety research and pretraining safer language models! 📈 Thank you to my advisor @zicokolter and my committee: Matt Fredrikson, @andrew_ilyas, and @furongh! 🙏
Dylan Sam tweet media
English
24
9
220
21.5K
Sachin Goyal @ ICLR’26 🇧🇷🏖️
Some long-pending life updates 🎉 Super excited to share that I defended my PhD thesis, “From Scaling Efficiently to Rethinking Pretraining of Foundation Models”, a couple of months ago. What a ride this journey has been.
Sachin Goyal @ ICLR’26 🇧🇷🏖️ tweet media
English
17
5
104
4.9K
Daniel P Jeong retweetledi
Shahriar Noroozizadeh
Shahriar Noroozizadeh@ShNoroozi·
Do deep sequence models store knowledge associatively, like a lookup table of atomic facts? We show the answer is, (surprisingly) no! Even without explicit pressure, they organize memory into global geometric maps. 🗺️ Excited to share our work from my internship last summer. Now on arXiv: arxiv.org/abs/2510.26745
Vaishnavh Nagarajan@_vaishnavh

1/ We found that deep sequence models memorize atomic facts "geometrically" -- not as an associative lookup table as often imagined. This opens up practical questions on reasoning/memory/discovery, and also poses a theoretical "memorization puzzle."

English
0
5
17
1.6K
Daniel P Jeong retweetledi
Emily Byun
Emily Byun@yewonbyun_·
I’m at NeurIPS this week (12/2-12/8) to present our work on when/how synthetic data (e.g., LLM simulations) can help scientists make inferences with less real data, improving the efficiency of costly experiments. Come by Poster #904 on Thursday 4:30PM (Exhibit Hall C,D,E)!🙂
Emily Byun@yewonbyun_

💡Can we trust synthetic data for statistical inference? We show that synthetic data (e.g. LLM simulations) can significantly improve the performance of inference tasks. The key intuition lies in the interactions between the moments of synthetic data and those of real data

English
2
4
31
12.7K
Daniel P Jeong retweetledi
Eunkyu Eunice Park
Eunkyu Eunice Park@uunicee_·
🧵Sharing our most-recent work! Critical or Compliant? The Double-Edged Sword of Reasoning in Chain-of-Thought Explanations
English
1
4
54
36.1K
Daniel P Jeong retweetledi
Sachin Goyal @ ICLR’26 🇧🇷🏖️
📢 Multi-token prediction has long struggled with defining the right “auxiliary target,” leading to tons of heuristics. We show a core limitation of these and propose a simple & sweet idea: future summary prediction. Introducing what I call 🚀TL;DR token pretraining🚀
Sachin Goyal @ ICLR’26 🇧🇷🏖️ tweet media
Divyat Mahajan@divyat09

[1/9] While pretraining data might be hitting a wall, novel methods for modeling it are just getting started! We introduce future summary prediction (FSP), where the model predicts future sequence embeddings to reduce teacher forcing & shortcut learning. 📌Predict a learned embedding of the future sequence, not the tokens themselves

English
4
36
242
28.8K
Daniel P Jeong retweetledi
Emily Byun
Emily Byun@yewonbyun_·
💡Can we trust synthetic data for statistical inference? We show that synthetic data (e.g. LLM simulations) can significantly improve the performance of inference tasks. The key intuition lies in the interactions between the moments of synthetic data and those of real data
Emily Byun tweet media
English
2
36
143
31K
Daniel P Jeong retweetledi
Sachin Goyal @ ICLR’26 🇧🇷🏖️
1/Excited to share the first in a series of my research updates on LLM pretraining🚀. Our new work shows *distilled pretraining*—increasingly used to train deployable models—has trade-offs: ✅ Boosts test-time scaling ⚠️ Weakens in-context learning ✨ Needs tailored data curation
Sachin Goyal @ ICLR’26 🇧🇷🏖️ tweet media
English
6
66
338
33.5K
Daniel P Jeong
Daniel P Jeong@danielpjeong·
Excited to talk about our work on evaluating *medically adapted* LLMs and VLMs, covering additional results since our oral presentation at EMNLP 2024. If you ever wondered whether Med-* LLMs/VLMs are really better than their general-domain counterparts, come check it out on 8/4!
MedAI Group@MedaiStanford

This Monday, Daniel Jeong from Carnegie Mellon University will be joining us to talk about his work on the limited impact of medical adaptation of LLMs and VLMs. Catch it at 1-2pm PT this Monday on Zoom! Subscribe to mailman.stanford.edu/mailman/listin… #ML #AI #medicine #healthcare

English
1
9
42
4.5K
Youngseog Chung
Youngseog Chung@YoungseogC·
Life update!! 📣🎉 I defended my PhD thesis! Big thanks to my wonderful advisor Jeff Schneider and thesis committee @zicokolter, Aarti Singh, and Jasper Snoek. Next up: I'm joining @MicrosoftAI as a Member of Technical Staff - stoked to be back in the Bay Area!! Idemooooo 🔥💪
Youngseog Chung tweet media
English
25
7
198
16.8K
Brandon Trabucco
Brandon Trabucco@brandontrabucco·
I'm excited to share that I've joined Thinking Machines to work on Multimodal Data, and Agents! It's a blessing to learn from so many thoughtful, passionate, and kind people. I can't wait to explore new kinds of data, build some models, and see where this journey takes us. Thanks to the many friends and mentors who helped me become a researcher worth his salt!
Mira Murati@miramurati

Thinking Machines Lab exists to empower humanity through advancing collaborative general intelligence. We're building multimodal AI that works with how you naturally interact with the world - through conversation, through sight, through the messy way we collaborate. We're excited that in the next couple months we’ll be able to share our first product, which will include a significant open source component and be useful for researchers and startups developing custom models. Soon, we’ll also share our best science to help the research community better understand frontier AI systems. To accelerate our progress, we’re happy to confirm that we’ve raised $2B led by a16z with participation from NVIDIA, Accel, ServiceNow, CISCO, AMD, Jane Street and more who share our mission. We’re always looking for extraordinary talent that learns by doing, turning research into useful things. We believe AI should serve as an extension of individual agency and, in the spirit of freedom, be distributed as widely and equitably as possible.  We hope this vision resonates with those who share our commitment to advancing the field. If so, join us. thinkingmachines.paperform.co

English
21
8
436
50.5K
Daniel P Jeong retweetledi
Rattana Pukdee
Rattana Pukdee@rpukdeee·
In our #AISTATS2025 paper, we ask: when it is possible to recover a consistent joint distribution from conditionals? We propose path consistency and autoregressive path consistency—necessary and easily verifiable conditions. See you at Poster session 3, Monday 5th May.
Rattana Pukdee tweet media
English
1
7
15
1.1K
Daniel P Jeong retweetledi
Pratyush Maini
Pratyush Maini@pratyushmaini·
Looking forward to giving a talk this Friday @OpenAI with @zhilifeng on some of our privacy & memorization research + how it applies to production LLMs! We've been gaining momentum on detecting, quantifying & erasing memorization; excited to explore its real-world impact!
Pratyush Maini tweet media
English
0
10
105
8.1K
Daniel P Jeong retweetledi
Sachin Goyal @ ICLR’26 🇧🇷🏖️
Excited to be at #ICLR2025 🇸🇬 to talk about my recent works👇that uncover key pitfalls & inefficiencies in pretraining & inference🚨. Final PhD lap —thinking a lot about how pretraining interventions can shape downstream behaviors (like reasoning & safety). DM to chat or vibe!
Sachin Goyal @ ICLR’26 🇧🇷🏖️ tweet media
English
1
4
45
2.8K
Daniel P Jeong retweetledi
Pratyush Maini
Pratyush Maini@pratyushmaini·
1/Being in academia is such a privilege: You get to collaborate with insanely talented & passionate students on their journey to upskill themselves. Very excited to share *OpenUnlearning*: a unified, easily extensible framework for unlearning led by @anmol_mekala @VineethDorna🧵
Pratyush Maini tweet media
English
4
26
143
27.5K