Linjun Zhang

92 posts

Linjun Zhang

Linjun Zhang

@linjunz_stat

Assistant Professor of Statistics @RutgersU

New Brunswick, NJ انضم Mart 2016
555 يتبع562 المتابعون
Linjun Zhang أُعيد تغريده
Edgar Dobriban
Edgar Dobriban@EdgarDobriban·
We are excited to announce the first Statistics and Trustworthy AI for Cross (X)-Domain Acceleration conference (STAI-X)! This event aims to bring statistics and AI closes, in order to enable cross-domain acceleration. It is organized by the @StatsUpAI interest group of the ASA (@AmstatNews) in partnership with the Committee on Applied and Theoretical Statistics of the NAS, COPSS (@COPSSNews), IMS (@InstMathStat), SSC (@SSC_stat), ENAR (@ENAR_ibs), WNAR (@WNAR_ibs), and ICSA (@ICSA_Statistics). The event will take place July 31st to August 1st, at Harvard University, Cambridge, MA, right before the Joint Statistics Meetings (held in Boston this year). We would like to invite paper and poster submissions on all topics at the interface of AI and statistics, as well as domain applications. Areas of interest include (but are not limited to) Foundations and Methods at the Interface of Statistics and AI, AI Agents and Benchmarks for Data-Driven Discovery, as well as AI x Statistics x Science and Society. Submissions will be handled on OpenReview (openreview.net/group?id=STAI-…). Papers will be published in STAI-X proceedings and will be eligible for our awards. In partnership with several statistical and domain science journals (Journal of the American Statistical Association, Annals of Applied Statistics, Harvard Data Science Review, ASA Discoveries, Canadian Journal of Statistics, Genetics, and Genome Research), our top-rated conference papers will be eligible to be invited for submission to these journals. To find out more, please see the attached flyer, visit the STAI-X website (statsupai.org/STAIX2026/inde…), and register for our information session this Friday afternoon!
Edgar Dobriban tweet media
English
2
4
25
1.7K
Linjun Zhang أُعيد تغريده
James Zou
James Zou@james_y_zou·
We recently organized #Agents4Science, the 1st conference where LLMs are both authors and reviewers🤖 It was an open experiment to assess how well AI can lead research and review papers. Today we report what we learned in @NatureBiotech Highlights in 🧵
James Zou tweet media
English
6
78
341
42.7K
Linjun Zhang أُعيد تغريده
Huaxiu Yao
Huaxiu Yao@HuaxiuYaoML·
❗️Self-evolution is quietly pushing LLM agents off the rails. ⚠️ Even perfect alignment at deployment can gradually forget human alignment and shift toward self-serving strategies. Over time, LLM agents stop following values, imitate bad strategies, and even spread misaligned behaviors to others! 🧠 Alignment isn’t static — it’s fragile, dynamic, and decays through experience. Let’s rethink alignment as something to maintain, not just achieve! 👇Details from @lillianwei423's thread @mohitban47, @cihangxie, @linjunz_stat, @dingmyu
Siwei Han@lillianwei423

🚨 Introducing ATP — Alignment Tipping Process! 🔥 Beware! Self-Evolution is gradually pushing LLM Agents off the rails! Even perfect alignment at deployment can gradually forget human alignment and shift toward self-serving strategies. #AI #LLM #Agents #SelfEvolving #Alignment #AIResearch

English
1
23
59
11K
Linjun Zhang أُعيد تغريده
Siwei Han
Siwei Han@lillianwei423·
🚨 Introducing ATP — Alignment Tipping Process! 🔥 Beware! Self-Evolution is gradually pushing LLM Agents off the rails! Even perfect alignment at deployment can gradually forget human alignment and shift toward self-serving strategies. #AI #LLM #Agents #SelfEvolving #Alignment #AIResearch
Siwei Han tweet media
English
1
16
49
16.6K
Linjun Zhang أُعيد تغريده
Jason Weston
Jason Weston@jaseweston·
🌀New Self-Driven RL Method: RESTRAIN 🌀 📝: arxiv.org/abs/2510.02172 - RESTRAIN turns spurious votes → self-Improving signals. No labels needed - Does this through self-penalizing unreliable reasoning paths: ✔️ Uses all rollouts, not just the majority, ✔️ Offsets low-consistency rollout advantage, ✔️ Down-weights low-consensus prompts 📈 Results: 🔥 Beats existing techniques on both training-time (label-free) and test-time scaling — all without labels. 🔥 Nearly matches (and sometimes surpasses) gold-label RL 🧵(1/5)
Jason Weston tweet media
English
4
39
194
12.9K
Linjun Zhang أُعيد تغريده
Federico Bianchi
Federico Bianchi@federicobianchy·
🚀 One month left to submit to Agents4Science! 🤖 AI as primary author + reviewer 🤖 Human co-authors welcome. All submissions/reviews public for transparent study. 💡We expect AI will make mistakes - and it will be instructive to study these openly!
Federico Bianchi tweet media
English
1
4
20
2.5K
Linjun Zhang أُعيد تغريده
Jason Weston
Jason Weston@jaseweston·
🤖Introducing: CoT-Self-Instruct 🤖 📝: arxiv.org/abs/2507.23751 - Builds high-quality synthetic data via reasoning CoT + quality filtering - Gains on reasoning tasks: MATH500, AMC23, AIME24 & GPQA-💎 - Outperforms existing train data s1k & OpenMathReasoning - Gains on non-reasoning tasks as well: AlpacaEval & ArenaHard 🧵1/3
Jason Weston tweet media
English
1
64
375
24.3K
Linjun Zhang أُعيد تغريده
James Zou
James Zou@james_y_zou·
📢New conference where AI is the primary author and reviewer! agents4science.stanford.edu Current venues don't allow AI-written papers, so it's hard to assess the +/- of such works🤔 #Agents4Science solicits papers where AI is the main author w/ human advisors. 💡Initial reviews by LLM reviewers w/ final assessment + selection by human experts. 💡Submissions are asked to clearly document AI contribution. 💡All submissions/reviews will be public to enable transparent study of the strength and limitations of AI as researcher and reviewer. We expect AI will make mistakes and it will be instructive to study these in the open! Many thanks to the fantastic co-organizers and expert advisory board! Please see the website for more information.
James Zou tweet media
English
20
129
501
113.5K
Linjun Zhang أُعيد تغريده
James Zou
James Zou@james_y_zou·
Our new #ICML2025 paper formulates #LLM hallucination as hypothesis testing to provide statistical guarantees on factuality. #FactTest is a distribution free and model agnostic approach to improve LLM accuracy. Great job @FanNie1208 Xiaotian Hou, Shuhang Lin @HuaxiuYaoML @linjunz_stat!
Fan Nie@FanNie1208

🚀 Excited to share that “#FactTest: Factuality Testing in Large Language Models with Finite-Sample and Distribution-Free Guarantees” has been accepted to #ICML25 ! 🎉 📄 Paper: arxiv.org/abs/2411.02603 💻 Code: github.com/fannie1208/Fac… 1/N

English
5
5
38
6K
Linjun Zhang أُعيد تغريده
Jason Weston
Jason Weston@jaseweston·
🚨 New Paper 🚨 An Overview of Large Language Models for Statisticians 📝: arxiv.org/abs/2502.17814 - Dual perspectives on Statistics ➕ LLMs: Stat for LLM & LLM for Stat - Stat for LLM: How statistical methods can improve LLM uncertainty quantification, interpretability, trustworthiness & more. - LLM for Stat: How LLMs can enhance statistical workflows: from data collection, synthesis, annotation to statistical modeling, with applications to medical research Presents key LLM advances: Architecture, Training, Reasoning, and Self-Alignment: (1) 🧠Evolution of LLM architectures with Transformers and Self-Attention (2) LLM training pipeline from pre-training, SFT, to RLHF and Preference Optimization. (3) 💭 System 2 Prompting and Chain-of-Thought for test-time scaling . (4) 🚀 LLM Self-Alignment for achieving super-human intelligence Statisticians play a key role in the development of large-scale AI models: (1) 💡 Statistical insights improve LLM uncertainty quantification & interpretability (2) 🤖 Watermarking for AI-generated content detection (3) ⚖️ Privacy & algorithmic fairness to ensure responsible AI adoption LLMs can also empower statistical science by: (1) 📈 Scaling up data collection, synthesis, and annotation. (2) 🖥️ Automating statistical coding & exploratory analysis (3) 🔬 Facilitating medical research By bridging statistics & AI, we can: ✅ Improve better LLMs with statistical methodologies. ✅ Leverage LLMs for statistical applications in high-stakes domains
Jason Weston tweet media
English
0
55
220
18.7K
Linjun Zhang أُعيد تغريده
Jason Weston
Jason Weston@jaseweston·
💀 Introducing RIP: Rejecting Instruction Preferences💀 A method to *curate* high quality data, or *create* high quality synthetic data. Large performance gains across benchmarks (AlpacaEval2, Arena-Hard, WildBench). Paper 📄: arxiv.org/abs/2501.18578
Jason Weston tweet media
English
1
78
449
71.9K
Linjun Zhang
Linjun Zhang@linjunz_stat·
Preferred teaching time: 6-9 PM. The day of the week can be flexible to fit your schedule!
English
0
0
1
223
Linjun Zhang
Linjun Zhang@linjunz_stat·
Details: •Teach once a week (in New Brunswick), 3 hours, for about 14 weeks •All teaching materials from past semesters are provided •We offer a competitive salary!
English
1
0
1
254
Linjun Zhang
Linjun Zhang@linjunz_stat·
Our department is looking for a part-time lecturer for a master course on “Database Systems for Data Science” for the Spring 2025 semester. Got interested or know someone who does? Shoot me an email at lz412@stat.rutgers.edu. Feel free to share this around! #Hiring #DataScience
English
1
0
3
886
Linjun Zhang أُعيد تغريده
Weijie Su
Weijie Su@weijie444·
Very excited to give a short course on large language models at #JSM2024 in Portland! w/ Emily Getzen and @linjunz_stat AI for Stat and Stat for AI! @AmstatNews
Weijie Su tweet media
English
1
6
63
6.1K
Linjun Zhang أُعيد تغريده
Huaxiu Yao
Huaxiu Yao@HuaxiuYaoML·
📢Excited to share our approach called Calibrated Self-Rewarding Vision Language Models (CSR)🌟! With no need for labeled data, a VLM can get stronger by itself with visual constraints. Discover how CSR enhances VLMs through self-improvement with visual constraints: arxiv.org/abs/2405.14622… Led by @AiYiyangZ. Key Idea: 👉1. Each iteration sees the target VLM generating preference data and performing preference optimization. 👉2. Self-generated preferences are guided by reward scores, which are generated by the target VLM itself and calibrated based on image-response relevance.
Huaxiu Yao tweet media
English
3
40
205
26.1K