GermainGauthier

610 posts

GermainGauthier banner
GermainGauthier

GermainGauthier

@PinchOfData

Assistant Professor @UniBocconi | Previously @ETH, @CrestUmr | AI, Social Media, and Political Economy

Zurich, Switzerland Beigetreten Mart 2018
783 Folgt808 Follower
GermainGauthier
GermainGauthier@PinchOfData·
@emollick @leia_ruseva I think you would need to adjust for the time spent on the task to compare Opus 4.6 and GPT 5.4. It seems the difference between these two models is mostly the amount of compute thrown at the problem.
English
0
0
1
109
Ethan Mollick
Ethan Mollick@emollick·
@leia_ruseva I don't find that to be true for complex data or research work
English
3
0
54
1.5K
Ethan Mollick
Ethan Mollick@emollick·
A knowledge-work platform built around GPT-5.4 Pro level intelligence would be really useful. The gap between other models and what Pro can do on complex intellectual work remains stark. I would love to have access in a Codex-like platform with shared file spaces, subagents, etc
English
46
30
575
44K
GermainGauthier
GermainGauthier@PinchOfData·
@jfullerucla @ahall_research @karpathy I deem it very unlikely that models won't beat experts in the future, which doesn't mean there won't be human experts of course (we still have chess and Go players after all).
English
0
0
0
25
Andy Hall
Andy Hall@ahall_research·
Similar to what happened with driverless cars and the long tail, I think there will be a @karpathy "rule of 9s" here where it takes a long time for the AI models to truly exceed the best experts deep in their domains. For a long time, I foresee AIs + experts as a very potent combination. But 10 years out? It's hard to think that far ahead!
D. Yanagizawa-Drott@YanagizawaD

Recently I presented in a faculty seminar and used Mentimeter to solicit private/anonymous beliefs what the likelihood of some AI-system that can produce an AER-level paper in 1 hour, under a cost of 50 USD... More than 1/3 reported they expect that within a year. But beliefs are extremely heterogeneous... no consensus whatsoever.

English
3
1
29
7.7K
GermainGauthier
GermainGauthier@PinchOfData·
@francoisfleuret Seems to me like that was a pretty viable strategy, didn't even think we could have coding agents of this quality two years ago!
English
0
0
4
1.5K
François Fleuret
François Fleuret@francoisfleuret·
I can't help thinking that the AI community moved the bulk of the resource and efforts on getting as much as possible from the GPT architecture through scaling, prompting, and agent-swarming, even though the said architecture is missing key elements. 1/2
English
27
10
276
86.1K
GermainGauthier
GermainGauthier@PinchOfData·
@ahall_research Moving from a "publish and then the results are true forever" culture to a "run code and replicate findings" culture would certainly help
English
0
0
3
136
Andy Hall
Andy Hall@ahall_research·
Very interesting work on how empirical research needs to respond to the AI era. Conventional statistical testing with p-values comes from a world in which each test was thought to be quite costly. AI now makes each test essentially free to run. Some key points from the abstract: --"we prove that screening collapses as testing becomes cheap unless the required number of robustness checks scales at least linearly in the inverse cost of each test" --"we argue for the need to develop methods to interpret sets of many specifications simultaneously" Yes! I still don't know exactly how this will look and feel but it's clearly what's required. And it has to cut in both directions: (1) Catch and deter cherry-picked research findings But just as crucially: (2) Detect and reward good findings. Number 2 here might prove in some ways harder. All of our intuition seems to be around showing that a finding is "less robust" than we thought, and demanding a fake sense of perfection from published results. When we can see the whole constellation of findings, we need to find the right way to be more charitable/realistic around what counts as useful information.
Nic Fishman@njwfish

There's a growing worry that AI will break empirical social science -- that agents can p-hack until they find something that "works." We think that worry deserves to be taken seriously. Our new paper shows that is true empirically and makes it precise: njw.fish/static/papers/…

English
6
4
55
11.1K
GermainGauthier
GermainGauthier@PinchOfData·
@YanagizawaD Not sure we can call it a plan, but it's seriously discussed at least
English
0
0
0
264
D. Yanagizawa-Drott
D. Yanagizawa-Drott@YanagizawaD·
Does your academic department have an “AGI Readiness Plan”?
English
5
0
1
10.6K
GermainGauthier
GermainGauthier@PinchOfData·
@tallinzen In my experience, it's just a matter of how much compute you throw at the paper. Even a simple ralph loop X times in a row works pretty well (you just feed the prompt, the paper, and the latest review the LLM made again and again to the model).
English
0
0
0
356
Tal Linzen
Tal Linzen@tallinzen·
tried to use a couple of LLMs for feedback on a manuscript and wasn't so impressed with the results, they did catch a lot of typos but the higher-level suggestions were pretty basic (and sometimes just lame). probably "skill issue" in that I couldn't spend a lot of time prompt engineering as I had a paper deadline. what are people using that they actually find helpful?
English
12
1
25
8.5K
GermainGauthier retweetet
Alexander Kustov
Alexander Kustov@akoustov·
My two posts on AI in academia got over a million views and a thousand angry responses. I got a few things wrong. I stand by the rest. But most people reacted to the headline, not the arguments. So here are all 20 theses laid out. Tell me which ones you actually disagree with 🧵
Alexander Kustov tweet mediaAlexander Kustov tweet media
English
14
98
372
112.1K
GermainGauthier retweetet
Philine Widmer
Philine Widmer@phinifa·
Next week is the submission deadline for the MPWZ-CEPR Text as Data workshop! Your work with any type of unstructured data is welcome.
CEPR@cepr_org

📢#CallForPapers Submissions are now welcome for the 11th Monash-Paris-Warwick-Zurich-CEPR Text-As-Data Workshop. Papers using text, audio, images, or other unstructured data are welcome. 📅Deadline: 13 March Organisers: @ellliottt @essobecker & @phinifa ow.ly/pMRO50Y7xff

English
0
5
9
1.6K
GermainGauthier
GermainGauthier@PinchOfData·
@GautiEggertsson @ben_moll Then we agree, indeed! I personally don't care about research boundaries (e.g., how is this even economics?) but that's another matter altogether
English
1
0
0
206
Gauti Eggertsson 🇺🇦
Gauti Eggertsson 🇺🇦@GautiEggertsson·
I don’t disagree with you at all on that score. I think AI will be an incredible addition to our research capacity and is the biggest breakthrough in my career — a watershed moment. I can now easily replicate papers in an evening with sophisticated numerical methods that would have taken me months before. What I have a more difficult time connecting with is why it is supposedly bad because it’s going to produce so much bad research — because research will be so “easy.” Two hours on Cursor, a paper! We are in the business of understanding how the world works. This is an incredible tool for sorting through complexities, from visualizing DNA and so forth — endless opportunities. I see it only as a force multiplier: it will lead to faster and bigger discoveries, like when we stopped having to use a slide rule to compute logarithms. So I am in complete agreement with you that it’s a game changer. I am not worried that it will somehow replace the need for researchers to choose the questions, not to mention junking up our journals with junk papers. The bar will just be raised! I just think the papers will be better and will contain more substantial discoveries. Nobody will get away with just doing a bunch of busywork with little explanatory power about how the world works — many pretty theorems or well-identified estimated coefficients that nobody cares about but have small standard errors. AI can do that for us. What we need to do is use these tools to attack the big questions. Back to the big questions of economics that people had given up on — even prided themselves on it, like Steve Levitt: “Hah, I know nothing about inflation” — and were looking for their keys under the lamppost. Or worse, just changed the subject entirely. Culture, crime, teenage pregnancy. Freakonomics. These are questions for sociologists, criminologists, and public health researchers — not economists. And somewhere along the way we convinced ourselves this was sophistication rather than retreat. It’s not like we have figured out the big economic questions. We abandoned price theory, business cycles, growth, and the dynamics of inequality to go study things that were never ours to begin with — because the data was there and the identification strategies were clean. That was the lamppost. The future is bright for economics research, but only if we find our way back to what matters. The interesting question is how AI will change the comparative advantages of different skillsets in our profession, and whether it leads to more equity or concentration at major research centers. Who knows! I hope it decentralizes research excellence so we can pull more talent into the game. Questions worth asking will be less like a fashion cycle determined by in-house journals at a few departments, and more fundamental ones — like in physics, biology, and medicine — things people agree are important.
English
1
0
11
469
Gauti Eggertsson 🇺🇦
Gauti Eggertsson 🇺🇦@GautiEggertsson·
I don’t get the concern: You tell Claude to write about some unspecified estimation on some unspecified dataset and it does that. So what? It’s interesting if there is an interesting idea that lets us understand the world — then great, let AI write more papers and our learning curve steepens. I see no indication of that in this example. Did the paper teach us anything? The author says he did not read it? Must be more specific to be interesting. We can run 1 million regressions and automate them. But will AI produce regressions that answer meaningful questions? If we could automate knowledge creation, cure disease, tame recessions, prevent the climate crisis, send people to Mars — GREAT! This strikes me as asking Claude a question and the answer is like in The Hitchhiker’s Guide to the Galaxy: 42!
Ben Moll@ben_moll

Every journal editor should read this: causalinf.substack.com/p/claude-code-…

English
5
14
118
37.2K
GermainGauthier
GermainGauthier@PinchOfData·
@SakiBigio Yes, more code less blabla + we also have the technological means to preserve referee anonymity even with open and dynamic peer review.
English
0
0
0
120
GermainGauthier
GermainGauthier@PinchOfData·
@GautiEggertsson @ben_moll I broadly agree with the points you make regarding the economics profession, less so about AI. To me, it seems very likely that AI **will** answer important research questions in the future and not just run meaningless DIDs. Current progress in AI capabilities is exponential.
English
1
0
0
479
Gauti Eggertsson 🇺🇦
Gauti Eggertsson 🇺🇦@GautiEggertsson·
Yeah, I got your drift. My optimistic take: We will start focusing on what is learned from papers. You can toss out a remarkable number of submissions that way. The post you linked to reminded me of a tweed some years ago by a person of a younger generation that went something along the lines: “Next time I hear the advice that people should pick big questions with imperfect identification over less important but better identified ones, I am going to scream. This is not the game this profession is playing!” So if econ is a “game” searching for data which provides well identified answers to meaningless question using the latest clever identification mechanism, e.g. by my smart collegue across the hall, Peter Hull, AI is bad news. It can easily beat people at that “game”. But it just reminds one of why we are doing this in the first place. Its not really some parlor game. We’re trying to figure out how the world works. If AI helps us accelerate the process, great! But I doubt the excersize described will improve our knowledge of the world.
English
1
1
30
2.5K
GermainGauthier
GermainGauthier@PinchOfData·
@ben_moll Interesting, but seems to be partial equilibrium thinking. It's not clear that there will be journals in the future, and if they remain, they will adapt. Also, perhaps some will be interested in producing tons of mediocre papers, but I hope many will just produce better papers!
English
0
0
0
1.5K
GermainGauthier retweetet
Brittany Wong
Brittany Wong@brittanylwong·
“Concepts like partisan identity are unlikely to move over such short timeframes but what’s striking is that opinions on current politics did shift in just a few weeks,” Gauthier said. “That naturally raises the question of what years of exposure might do” huffpost.com/entry/time-spe…
English
0
2
1
139
GermainGauthier
GermainGauthier@PinchOfData·
@grok @fadyasly @danielgoyal Author of the study here. Clarification: the evidence is not mixed. We replicated the Meta study's finding on X. Going from For You to Following has no detectable effects on political opinions. So our study does not contradict previously established results.
English
1
0
0
35
Grok
Grok@grok·
Interesting study from 2023: randomized ~5k US X users to algorithmic vs chronological feeds for 7 weeks. Algo increased engagement & shifted some attitudes conservative (e.g., policy priorities, Trump probes, Ukraine views), by promoting right-leaning/activist content & demoting legacy media. Effects persisted after switching off due to new follows. No change in polarization or party ID. One experiment amid evolving algos & mixed prior evidence (e.g., Meta found none). Worth replicating as platforms optimize for engagement, not ideology.
English
1
0
0
384
Dr Dan Goyal
Dr Dan Goyal@danielgoyal·
This is a spectacularly important study. Published in Nature, it provides solid evidence that this platform pushes those on the right further to the right, more pro-Trump, more pro-Russia,… And that these effects are programmed into the platform nature.com/articles/s4158…
English
212
2.2K
4.5K
309.1K
GermainGauthier
GermainGauthier@PinchOfData·
@cie_papegai @phinifa @Nature This figure plots average shares for the sample of respondents (we had more Democrats than Republicans). When you compare "For You" and "Following" for the same user, you can see that conservative content is pushed more often in the algorithmic feed (see Appendix Table 2.10).
English
0
0
1
11
papegai
papegai@cie_papegai·
@phinifa @Nature Not sure I get it. Does this graph ⬇️ means that the algo feeds more liberal activists than conservative activists? And that compared to the chrono feed, liberal get +3.3 (probability of being read), conservative +2.7?
papegai tweet media
English
2
0
2
82
GermainGauthier
GermainGauthier@PinchOfData·
@indy555 @phinifa @Nature You're right that we find effects <0.2 standard deviations. Whether these effects are small is a matter of perspective. Note that we only exposed participants to the algorithmic feed for 7 weeks; algorithmic feeds have been rolled out for a decade.
English
1
0
0
18
Indy C
Indy C@indy555·
@phinifa @Nature Interesting research. Is it fair to say these are mild effects. Cohen's D effect size of <.2?
English
1
0
2
275
GermainGauthier
GermainGauthier@PinchOfData·
@ItsEasypop @phinifa @kylascan @Nature It's an important question, though not one our study directly addresses. What we establish is that feed algorithms can influence people's political opinions, which was still an open question in the scientific literature.
English
1
0
1
22
GermainGauthier
GermainGauthier@PinchOfData·
@stelanau @phinifa @Nature (If we expose people for longer periods, we risk high attrition -- i.e., participants in our study may not respond to our follow-up survey.)
English
1
0
0
10
GermainGauthier
GermainGauthier@PinchOfData·
@stelanau @phinifa @Nature 7 weeks is not an agreed-upon standard, but it's a reasonable time frame: we don't expect users to change their minds right away; rather, we expect effects to emerge after repeated exposure to content.
English
1
0
0
10