Dashun Wang

3.2K posts

Dashun Wang

Dashun Wang

@dashunwang

Kellogg Chair of Technology at Kellogg, Founding director, Center for Science of Science and Innovation, Northwestern University

Evanston, IL Katılım Eylül 2008
1.1K Takip Edilen6K Takipçiler
Dashun Wang retweetledi
Ari Holtzman
Ari Holtzman@universeinanegg·
Dashun Wang talking 'airplanes for the mind' 🛩️ are heavier than air, but can fly faster than sound—remind you of another technology that "shouldn't" be able to work, but does?
Ari Holtzman tweet media
Chenhao Tan@ChenhaoTan

Excited to announce the 2026 iteration of the Communication & Intelligence Symposium at UChicago! We have an amazing lineup of speakers @Diyi_Yang @johnhewtt @dashunwang @TomerUllman We have a simple call for abstract that is due on Apr 15 (links 👇). Please come and share your research! Co-organized with the awesome @universeinanegg and @divingwithorcas

English
0
3
12
2.7K
Dashun Wang
Dashun Wang@dashunwang·
Fascinating - Newton's gravity law turns out to be ... correct! largest scale validation. 3 centuries later, are we still doing Newton's homework for us? science.org/content/articl…
English
0
0
2
303
Dashun Wang retweetledi
Yian Yin
Yian Yin@yian_yin·
Fantastic analysis from journal submission data. It is also encouraging to see key findings from our recent @ScienceMagazine paper on #AI impact on science — increased productivity, more complex scientific writing, and emerging quality concerns — echoed in this setting.
Yian Yin tweet media
Pangram Labs@pangramlabs

How are large language models impacting the submission and review process at high-impact journals? Severely. Since the release of ChatGPT in 2022, AI-generated and AI-assisted papers, identified by Pangram, drove a 42% increase in submission volume at Organization Science (figure below). While the journal rejected the majority of these submissions, there is a human cost to reviewing papers, which volunteer reviewers are shouldering. AI-generated content is also showing up in reviews, which similarly suffer in quality because of it -- editors at Organization Science found that AI-generated reviews are lower quality, less specific, and less topically diverse than human-written ones. The problem is not isolated. Earlier this year, ICML desk-rejected 497 papers from authors who submitted AI-generated reviews, after those authors opted into a policy that disallowed the use of AI. Grant funders also saw a surge in applications: the Marie Skłodowska-Curie Actions, a set of major research fellowships for the EU, received 142% more proposals in 2025 compared to 2022. Many scientific and academic systems implicitly rely on friction as a barrier to entry. LLMs have removed that friction, allowing for a deluge of AI slop that is straining the capacity of these institutions.

English
5
50
240
47.2K
Dashun Wang
Dashun Wang@dashunwang·
New paper led by Minsu Park, showing that high impact interdisciplinary papers disproportionately come from deeply disciplinary grants. Check it out!
Minsu Park@likeateenspirit

A belated self-promotion of our new paper in @PNASNexus. We ask how the interdisciplinarity of a supporting grant and that of the focal paper jointly shape the paper's scientific impact (academic.oup.com/pnasnexus/arti…). Much assumed, rarely tested; so we tested it at scale! (1/n)

English
0
6
18
4.3K
Dashun Wang
Dashun Wang@dashunwang·
Bipartisan-cited science thus highlights both the limits of science in bridging polarization and its potential to provide a shared evidentiary foundation for addressing pressing challenges. Key Q going forward is how to improve the evidentiary basis for policymaking. Can gen AI help? or make it much worse? Stay tuned for more! 6/n
English
1
1
0
242
Dashun Wang
Dashun Wang@dashunwang·
🚨Bipartisan-cited science is rare, unevenly distributed, and disproportionately influential. 🚨 Check out our latest paper in PNAS, led by @zfurnas 🧵1/n
Dashun Wang tweet media
English
1
7
26
2.1K
Dashun Wang retweetledi
Sukh Sroay
Sukh Sroay@sukh_saroy·
Northwestern just ran a test nobody wanted to run. They took two LLMs with nearly identical MMLU-Pro scores and compared their answers question by question. The models disagreed on up to 66% of items. Same benchmark. Same accuracy. Completely different error profiles. Even among frontier models scoring above 60%, disagreement stayed between 16% and 38%. Two models that each "get 7 out of 10 right" systematically get a different 7 right. Then they did the part that should have ended the conversation. They took published studies in education and political science — real peer-reviewed research — and re-ran the analyses using 8 different LLMs as annotators. In an education study on literacy intervention, the estimated treatment effect varied by 84% depending on which model scored the student essays. 0.19 with one model. 0.35 with another. Same essays. Same rubric. In a political science study on Russian state media, switching the model reversed the sign of the finding. The original study concluded officials were more likely to be credited with good economic news. Two of the models concluded the opposite — that they were more likely to be blamed for bad news. Not "slightly different magnitude." The opposite conclusion. This is what the paper calls a benchmark illusion. MMLU-Pro and GPQA are the currency the entire industry uses to decide which model to deploy. The working assumption is that models with matching scores are interchangeable. They aren't. They're making different mistakes on different questions, and those mistakes happen to average out to the same number on a leaderboard. The villain here isn't any specific model. It's the benchmark itself. A single accuracy score compresses away the exact information that matters for science: which items you got wrong and why. Every paper that says "we used GPT-4 to annotate our dataset" was making an invisible methodological choice. Swap the model, potentially flip the result. The uncomfortable part isn't that LLMs disagree. It's that the tools we built to measure them were never designed to catch it. Paper in comments.
Sukh Sroay tweet media
English
4
19
47
4.7K
Dashun Wang retweetledi
Chenhao Tan
Chenhao Tan@ChenhaoTan·
Abstract is due in two days! Join us for an exciting day!
Chenhao Tan@ChenhaoTan

Excited to announce the 2026 iteration of the Communication & Intelligence Symposium at UChicago! We have an amazing lineup of speakers @Diyi_Yang @johnhewtt @dashunwang @TomerUllman We have a simple call for abstract that is due on Apr 15 (links 👇). Please come and share your research! Co-organized with the awesome @universeinanegg and @divingwithorcas

English
0
2
5
3.3K
Dashun Wang retweetledi
CASBS
CASBS@CASBSStanford·
CALL FOR APPLICATIONS | CASBS 2026 Summer Institute in AI Methods for Social Scientists (AIMS) July 26-31 | App deadline: May 8 AIMS brings together scholars from across the social sciences for a week of learning, practice & collaboration DETAILS & APP: bit.ly/4muwstd
CASBS tweet media
English
0
49
159
16.9K
Dashun Wang
Dashun Wang@dashunwang·
Excited to share that our SciSciGPT paper is featured on the cover of Nature Computational Science. Special credit to Hanna Renedo — this is also the first piece coming out of our newly launched CSSI art studio. More to come.
Dashun Wang tweet media
Nature Computational Science@NatComputSci

🚨Our March issue is now live, including an AI collaborator for science of science, a method for property-guided molecule generation, a Comment on the future of density functional theory, and much more! nature.com/natcomputsci/v… 📰Cover: nature.com/articles/s4358…

English
0
14
68
5.4K
Kevin A. Bryan
Kevin A. Bryan@Afinetheorem·
I am bringing a small group of innovation economists and AI policy folks to China Apr 12-19. Hangzhou, Beijing, Shenyang. Lots of meetings on AI and "real industry"/robotics - if we should talk to someone great in your network, please email and connect us!
English
2
6
47
5.5K