Joshua Kazdan

32 posts

Joshua Kazdan

Joshua Kazdan

@JoshuaK92829

Katılım Ekim 2024
32 Takip Edilen71 Takipçiler
Joshua Kazdan retweetledi
Jessica Chudnovsky
Jessica Chudnovsky@jchudnov·
Your deduplication pipeline was built for small models. At scale, it's broken. New preprint: "Scale Dependent Data Duplication" 1/10
Jessica Chudnovsky tweet media
English
6
28
116
26.1K
Joshua Kazdan
Joshua Kazdan@JoshuaK92829·
@AlexanderSpangh @jchudnov yes! If you take a look at Fig 2 that's exactly what it shows. The longer you train the model, the more the gradients induced by semantically identical documents align.
English
0
0
0
22
Alex Spangher @ Neurips2025
Alex Spangher @ Neurips2025@AlexanderSpangh·
Super cool work! You mention "stronger models encode data better", but in your plots I mainly see # model parameters and model size. Does this also imply that the same-size model trained longer, better etc. will also make less effective use of its dataset? Does data effectiveness decrease throughout training?
English
1
0
0
194
Joshua Kazdan
Joshua Kazdan@JoshuaK92829·
3. Writing the majority of your review using a language model. It did such a great job! Thanks also to the AC for ignoring us when we reported this review for violating the @NeurIPSConf guidelines against LM reviewing.
Joshua Kazdan tweet media
English
1
0
3
233
Joshua Kazdan
Joshua Kazdan@JoshuaK92829·
@casper_hansen_ @RylanSchaeffer There's no contradiction. We don't claim that min-p is better or worse than other logit processors-- we contend only that the evidence in Minh et. al. does not meet scientific standards to claim superiority.
English
0
0
3
154
Casper Hansen
Casper Hansen@casper_hansen_·
@RylanSchaeffer I use min_p and it improves coherence in my experience. The claims made here are in direct conflict with my and other people’s experience
English
3
0
17
2.2K
Joshua Kazdan retweetledi
Rylan Schaeffer
Rylan Schaeffer@RylanSchaeffer·
🚨New preprint 🚨 Turning Down the Heat: A Critical Analysis of Min-p Sampling in Language Models We examine min-p sampling (ICLR 2025 oral) & find significant problems in all 4 lines of evidence: human eval, NLP evals, LLM-as-judge evals, community adoption claims 1/8
Rylan Schaeffer tweet media
English
12
33
285
75.2K
Joshua Kazdan retweetledi
Rylan Schaeffer
Rylan Schaeffer@RylanSchaeffer·
A bit late to the party, but our paper on predictable inference-time / test-time scaling was accepted to #icml2025 🎉🎉🎉 TLDR: Best of N was shown to exhibit power (polynomial) law scaling (left), but maths suggest one should expect exponential scaling (center). We show how to ... 1/3
Rylan Schaeffer tweet media
English
9
16
116
17.8K
Joshua Kazdan retweetledi
Jason Weston
Jason Weston@jaseweston·
🚨 New Paper 🚨 An Overview of Large Language Models for Statisticians 📝: arxiv.org/abs/2502.17814 - Dual perspectives on Statistics ➕ LLMs: Stat for LLM & LLM for Stat - Stat for LLM: How statistical methods can improve LLM uncertainty quantification, interpretability, trustworthiness & more. - LLM for Stat: How LLMs can enhance statistical workflows: from data collection, synthesis, annotation to statistical modeling, with applications to medical research Presents key LLM advances: Architecture, Training, Reasoning, and Self-Alignment: (1) 🧠Evolution of LLM architectures with Transformers and Self-Attention (2) LLM training pipeline from pre-training, SFT, to RLHF and Preference Optimization. (3) 💭 System 2 Prompting and Chain-of-Thought for test-time scaling . (4) 🚀 LLM Self-Alignment for achieving super-human intelligence Statisticians play a key role in the development of large-scale AI models: (1) 💡 Statistical insights improve LLM uncertainty quantification & interpretability (2) 🤖 Watermarking for AI-generated content detection (3) ⚖️ Privacy & algorithmic fairness to ensure responsible AI adoption LLMs can also empower statistical science by: (1) 📈 Scaling up data collection, synthesis, and annotation. (2) 🖥️ Automating statistical coding & exploratory analysis (3) 🔬 Facilitating medical research By bridging statistics & AI, we can: ✅ Improve better LLMs with statistical methodologies. ✅ Leverage LLMs for statistical applications in high-stakes domains
Jason Weston tweet media
English
0
55
220
18.7K
Joshua Kazdan retweetledi
Krishnamurthy (Dj) Dvijotham
(1/n) Fine tuning APIs create significant security vulnerabilities, breaking alignment in frontier models for under $100! Introducing NOICE, a fine-tuning attack that requires just 1000 training examples to remove model safeguards. The strangest part: we use ONLY harmless data.
Krishnamurthy (Dj) Dvijotham tweet media
English
1
7
33
2.7K
Belinda
Belinda@belindmo·
New package + paper drop 📄 - Introducing KGGen – a simple library to transform unstructured text into knowledge graphs. Text is abundant, but good knowledge graphs are scarce. Feed it raw text, and KGGen generates a structured network of entities and relationships. (1/7)
English
8
28
126
11K