Meng Jiang

200 posts

Meng Jiang

Meng Jiang

@Meng_CS

Frank M. Freimann Collegiate Professor at Notre Dame CSE | Data Mining | NLP | AI

Notre Dame, IN Bergabung Ağustos 2012
536 Mengikuti1.6K Pengikut
Meng Jiang
Meng Jiang@Meng_CS·
Decentralized RAG allows your database to benefit all LLM clients. On the other side, not all data sources are reliable. Managing source reliability on blockchain can avoid third-party manipulation. Introducing dRAG + Blockchain + Truth Discovery: arxiv.org/abs/2511.07577
English
0
1
7
1.3K
Meng Jiang me-retweet
Peng Qi
Peng Qi@qi2peng2·
𝗕𝗲𝗰𝗮𝘂𝘀𝗲 𝟵.𝟭𝟭>𝟵.𝟵 𝗮𝗻𝗱 𝗮 𝘁𝗿𝗶𝗮𝗻𝗴𝗹𝗲 𝗵𝗮𝘀 𝗳𝗼𝘂𝗿 𝘀𝗶𝗱𝗲𝘀, 𝘁𝗵𝗲𝗿𝗲𝗳𝗼𝗿𝗲 𝟭+𝟭=𝟮. LLMs and Language Agents can sometimes generate correct answers from blatantly incorrect reasoning, which is more often in complex tasks, and exacerbated by reinforcement learning (RL), the commonly believed silver bullet to complex reasoning in LLMs. This is due to a well-known phenomenon called reward hacking, where if the only training signal LLMs are getting from the training data exclusively regards the final result, then LLMs are incentivized to match the correct final output through whatever means possible on its training data, leading to inconsistent and ungeneralizable reasoning processes in RL's wake. With our intern Mengzhao Jia, we (@ignaciocases and myself, plus folks from @Meng_CS s lab at Notre Dame) explore a simple fix: can we use the LLM's own reasoning to provide some additional supervision signal for the reasoning process itself, so that besides the final result, the LLM is also encouraged to stay consistent in its reasoning during training? We design an algorithm to automatically create rubrics for LLM reasoning processes, and train the model to adhere to these rubrics alongside generating correct final answers during RL. The resulting model not only produces significantly more consistent reasoning, but also generalizes better on a wide range of complex reasoning tasks we benchmarked, even with just 10% of the training data. We hope this technique helps pave the way to more powerful and generalizable reasoning models for complex tasks. Read more in our preprint: arxiv.org/pdf/2510.14738
Peng Qi tweet media
English
4
9
67
7.4K
Meng Jiang me-retweet
Hy Dang
Hy Dang@HyDang99·
Thrilled to share that “Improving Large Language Models Function Calling and Interpretability via Guided-Structured Templates” paper has been accepted to EMNLP 2025 (Main Conference)!🎉 📄 Check it out on arXiv: arxiv.org/abs/2509.18076 project page: hygiadang.com/publication/em… 1/3
English
1
1
3
752
Meng Jiang me-retweet
Tarannum Zaki
Tarannum Zaki@tarannum_zaki·
.@DomSoos from @WebSciDL and @oducs is presenting "Can LLMs Beat Humans on Discerning Human-written and LLM-generated Science News?" They explored whether LLMs can outperform humans for LLM-generated vs. human written news. 🔗doi: 10.1145/3720553.3746674 #LLM #NLP @fanchyna
Tarannum Zaki tweet mediaTarannum Zaki tweet mediaTarannum Zaki tweet mediaTarannum Zaki tweet media
English
0
4
2
461
Noah Ziems
Noah Ziems@NoahZiems·
@Meng_CS We have a foundation models lab?!
English
1
0
2
232
Meng Jiang me-retweet
Yining Lu
Yining Lu@Yining__Lu·
✴️ Pleased to introduce our new paper yining610.github.io/dynamic-reward… - Rebalance multiobjectives during training through dynamic reward weighting - Build Pareto-dominant front over static baselines across online RL algorithms, datasets, and model families - Faster convergence rate 1/8
Yining Lu tweet media
English
1
6
23
5.4K
Omar Khattab
Omar Khattab@lateinteraction·
@NoahZiems @MIT_CSAIL @Meng_CS @DSPyOSS Welcome Noah!! So great to have you as a founding member here of this new lab :D And I’m so excited to continue to collaborate with and learn more closely from Meng!
English
1
0
11
704
Noah Ziems
Noah Ziems@NoahZiems·
Quick update! This year I am on visit at @MIT_CSAIL working under the wonderful @lateinteraction while I am continuing to be advised by the wonderful @Meng_CS Right now my focus is to continue making Arbor a fantastic RL framework for optimizing @DSPyOSS programs
English
14
7
126
15.1K
Meng Jiang me-retweet
Gang Liu
Gang Liu@gliu0329·
🔥 Only 15 days left! 🔥 The Open Polymer Challenge already has 9,800+ entrants and 38,000+ submissions. If you have not joined yet, let’s jump in these last few days to 🌍 accelerate polymer discovery with ML and go for 💰 $50,000 in prizes. 👉 LINK: kaggle.com/competitions/n…
Gang Liu tweet media
English
0
3
6
1.5K
Meng Jiang me-retweet
Omar Khattab
Omar Khattab@lateinteraction·
New paper: Reflective Prompt Evolution Can Outperform GRPO. It's becoming clear that learning via natural-language reflection (aka prompt optimization) will long be a central learning paradigm for building AI systems. Great work by @LakshyAAAgrawal and team on GEPA and SIMBA.
Omar Khattab tweet media
Lakshya A Agrawal@LakshyAAAgrawal

How does prompt optimization compare to RL algos like GRPO? GRPO needs 1000s of rollouts, but humans can learn from a few trials—by reflecting on what worked & what didn't. Meet GEPA: a reflective prompt optimizer that can outperform GRPO by up to 20% with 35x fewer rollouts!🧵

English
11
91
492
70.4K
Michael Bernstein
Michael Bernstein@msbernst·
Thank you to everyone for your energy and enthusiasm in joining this adventure with me so far!
GIF
English
85
14
762
44.6K
Meng Jiang me-retweet
Gang Liu
Gang Liu@gliu0329·
🎓💰🔬 Want to learn machine learning, win a cash prize (USD 50K in total!!), and help drive real progress in discovering new polymer materials? All available at our NeurIPS 2025 Open Polymer Challenge: open-polymer-challenge.github.io 🚀 Join now (Kaggle): kaggle.com/competitions/n… 📈
Gang Liu tweet media
English
2
9
37
11K
Meng Jiang
Meng Jiang@Meng_CS·
Open Polymer Challenge: Leveraging Machine Learning for Polymer Informatics was accepted to NeurIPS 2025 Competition Track and is now LAUNCHed on Kaggle! JOIN US AND WIN $50,000 Awards! YES, FOUR "0"s - it's $50,000! Soooo what are YOU waiting for???
Gang Liu@gliu0329

🎓💰🔬 Want to learn machine learning, win a cash prize (USD 50K in total!!), and help drive real progress in discovering new polymer materials? All available at our NeurIPS 2025 Open Polymer Challenge: open-polymer-challenge.github.io 🚀 Join now (Kaggle): kaggle.com/competitions/n… 📈

English
0
4
16
3.4K
Meng Jiang me-retweet
Gang Liu
Gang Liu@gliu0329·
Introducing🔥torch-molecule🔥: A single line of code for molecular property prediction, generation & representation learning: > 30 deep learning methods + models, sklearn-style. All available at: `pip install torch-molecule` Code: github.com/liugangcode/to…
Gang Liu tweet media
English
5
46
301
19.7K
Meng Jiang me-retweet
Gang Liu
Gang Liu@gliu0329·
Our team, with @Meng_CS, Yihan, and Eric, is actively building and enriching the library. We welcome feedback, suggestions, and contributions from the community. Please feel free to reach out with any feedback.
English
1
1
4
823
Meng Jiang me-retweet
Omar Khattab
Omar Khattab@lateinteraction·
So many things in the run-up to DSPy 3. Here's a first, EXPERIMENTAL one: 🚨We're releasing dspy.GRPO, an online RL optimizer for DSPy programs Your DSPy code as-is can be dspy.GRPO'ed. Yes, even compound multi-module programs. Led by @NoahZiems @LakshyAAAgrawal @dilarafsoylu.
English
25
85
639
217.5K
Meng Jiang me-retweet
Yining Lu
Yining Lu@Yining__Lu·
Quick reminder that our paper, Benchmarking Language Model Creativity: A Case Study on Code Generation, will be presented today! 📅 11AM-12:30PM, Fri, May 2 📍 Hall 3 📝 arxiv.org/abs/2407.09007 🎥 youtube.com/watch?v=v1cHyC…
YouTube video
YouTube
Yining Lu tweet media
English
0
5
17
964