Parameter Lab

140 posts

Parameter Lab banner
Parameter Lab

Parameter Lab

@parameterlab

Empowering individuals and organisations to safely use foundational AI models.

Tübingen, Germany شامل ہوئے Mart 2023
116 فالونگ280 فالوورز
Parameter Lab
Parameter Lab@parameterlab·
‼️ New paper from Parameter Lab! ⛓️‍💥 We identify privacy collapse, a silent failure mode of LLMs: LLMs fine-tuned on seemingly benign data can lose their ability to respect contextual privacy norms. Done by @anmgoel during his internship! Check-out 👇
Parameter Lab tweet media
Anmol Goel@anmgoel

🚨 Fine-tuning your model to be more helpful or empathetic might be making it less private, without you noticing. In our latest work, we show that benign fine-tuning can silently break contextual privacy in language models while safety & general capabilities appear intact. ⬇️

English
0
2
3
414
Parameter Lab ری ٹویٹ کیا
Anmol Goel
Anmol Goel@anmgoel·
🚨 Fine-tuning your model to be more helpful or empathetic might be making it less private, without you noticing. In our latest work, we show that benign fine-tuning can silently break contextual privacy in language models while safety & general capabilities appear intact. ⬇️
Anmol Goel tweet media
English
1
2
7
1.7K
Parameter Lab
Parameter Lab@parameterlab·
👏 Proud to share that the paper that Ahmed Heakl authored during his internship at Parameter Lab was accepted at #ICLR2026! See how 🩺Dr.LLM increases accuracy and decreases inference computations of frozen LLMs: lnkd.in/dqzByRkT
English
1
0
5
202
Parameter Lab ری ٹویٹ کیا
Martin Gubri
Martin Gubri@framart1·
🎉Delighted to announce that our 🫗Leaky Thoughts paper about contextual privacy with reasoning models is accepted to #EMNLP main! Huge congrats to the amazing team @tommasogreen @HaritzPuerto @coallaoh @oodgnas
Martin Gubri@framart1

Delighted by this great thread from @omarsar0 presenting our new Leaky Thoughts paper! We show that reasoning models pose serious privacy risks when used as personal agents. Reasoning traces are a new attack vector. Work led by @tommasogreen during his internship @parameterlab!

English
0
4
12
1.3K
Parameter Lab
Parameter Lab@parameterlab·
🧪 Our latest research: Does SEO boost the visibility of content in LLM-based conversational search? We present C-SEO Bench, a benchmark to evaluate conversational SEO strategies. Key takeaway: SEO methods that target LLM do not work. But surprisingly, traditional SEO is not dead: it still matters, as LLMs tend to favour content already ranked higher in their input.
Parameter Lab tweet media
Haritz Puerto@HaritzPuerto

🔎 Does Conversational SEO (C-SEO) actually work? Our new benchmark has an answer. Excited to announce C-SEO Bench: Does Conversational SEO Work? 🌐 RTAI: researchtrend.ai/papers/2506.11… 📄 Paper: arxiv.org/abs/2506.11097 💻 Code: github.com/parameterlab/c… 📊 Data: huggingface.co/datasets/param…

English
0
0
0
176
Parameter Lab
Parameter Lab@parameterlab·
🎉 Very excited to see our new Leaky Thoughts 🫗 paper featured among last week's top AI papers by both @dair_ai and @TheAITimeline! - x.com/dair_ai/status… - x.com/TheAITimeline/… ➡️ Learn more about the paper in this great thread by @omarsar0: x.com/omarsar0/statu… ➡️ ArXiv link: arxiv.org/abs/2506.15674
Parameter Lab tweet media
elvis@omarsar0

Leaky Thoughts Hey AI devs, be careful how you prompt reasoning models. This work shows that reasoning traces frequently contain sensitive user data. More of my notes below:

English
1
3
6
783
Parameter Lab ری ٹویٹ کیا
Haritz Puerto
Haritz Puerto@HaritzPuerto·
Do you want to prove that your copyrighted document/corpus was trained by an LLM? Come to poster 46 #NAACL2025
Haritz Puerto tweet media
English
0
6
41
2.7K
Parameter Lab ری ٹویٹ کیا
Haritz Puerto
Haritz Puerto@HaritzPuerto·
#NAACL2025 has started! I’ll be presenting my work at @parameterlab about detecting pretraining data on Friday 🗓️ May 2, 11:00 AM - May 2, 12:30 PM 🗺️ Poster Session 8 - APP: NLP Applications Location: Hall 3 Work with @framart1 @oodgnas @coallaoh
Haritz Puerto@HaritzPuerto

🧵 It is assumed that Membership Inference Attacks (MIA) do not work on LLMs, but our new paper shows it can work at the right scale! MIA is effective if the number of input tokens is large enough, such as in long documents and collections of them. 📃arxiv.org/abs/2411.00154

English
1
1
12
486
Parameter Lab ری ٹویٹ کیا
Haritz Puerto
Haritz Puerto@HaritzPuerto·
I will be in person at #NAACL2025 🌵🇺🇸 to present Scaling Up Membership Inference: When and How Attacks Succeed on LLMs. Come and say hi 👋 if you want to know how to proof if an LLM was trained on a data point!
Haritz Puerto@HaritzPuerto

🧵 It is assumed that Membership Inference Attacks (MIA) do not work on LLMs, but our new paper shows it can work at the right scale! MIA is effective if the number of input tokens is large enough, such as in long documents and collections of them. 📃arxiv.org/abs/2411.00154

English
1
5
21
1.7K
Parameter Lab ری ٹویٹ کیا
Min Choi
Min Choi@minchoi·
GPT-4o image gen is seriously impressive. People are unlocking new creative ways to use it. 10 wild examples
Min Choi tweet media
English
87
467
5.5K
1.3M
Parameter Lab
Parameter Lab@parameterlab·
👥 We're Hiring: Senior/Junior Data Engineer! 📍 Remote or Local | Full-Time or Part-Time At ResearchTrend.AI, we’re building a platform that connects researchers and AI engineers worldwide—helping them stay ahead with daily digests, insightful summaries, and interactive events. Our LLM-powered ecosystem also bridges the gap between cutting-edge research and industry leaders. If you're passionate about data, AI, and making an impact, we’d love to have you on board! What You’ll Do: ✔ Build Scalable Data Pipelines – Design and optimize workflows using tools like Airflow. ✔ Work Closely with AI Experts & Engineers – Collaborate to solve real-world data challenges. ✔ Optimize and Maintain Systems – Keep our data infrastructure fast, secure, and adaptable. What You Bring: ✅ Proficiency in Airflow & PostgreSQL – You know your way around complex workflows and databases. ✅ Strong Python Skills – Clean, efficient, and maintainable code is your thing. ✅ (Bonus) Experience with LLMs – A huge plus as we integrate AI-driven solutions. ✅ Problem-Solving Mindset – You enjoy tackling challenges with real impact. ✅ Team Spirit – Excellent collaboration and communication. Why Join Us? 🚀 Make a Difference – Your work directly enhances how research is shared and discovered. 🌍 Flexibility – Choose full-time or part-time, work remotely or locally. ⚡ Innovative Environment – AI, research, and data-driven solutions all in one place. 🤝 Great Team – Work with passionate, talented people shaping the future of research. Ready to Join? Send your resume + a short note on why you’re a great fit to recruit@parameterlab.de. Be part of a team that’s redefining research with AI! #Hiring #DataEngineer #AI #RemoteJobs
Parameter Lab tweet media
English
0
0
2
773
Parameter Lab
Parameter Lab@parameterlab·
🔎 Wonder how to prove an LLM was trained on a specific text? The camera ready of our Findings of #NAACL 2025 paper is available! 📌 TLDR: longs texts are needed to gather enough evidence to determine whether specific data points were included in training of LLMs: arxiv.org/abs/2411.00154
Haritz Puerto@HaritzPuerto

🧵 It is assumed that Membership Inference Attacks (MIA) do not work on LLMs, but our new paper shows it can work at the right scale! MIA is effective if the number of input tokens is large enough, such as in long documents and collections of them. 📃arxiv.org/abs/2411.00154

English
0
0
4
477
Parameter Lab ری ٹویٹ کیا
Seong Joon Oh
Seong Joon Oh@coallaoh·
We just wanted to say: Membership inference is unlikely to succeed on n-grams or even paragraphs. Language models require **multiple documents** to gather enough evidence to determine whether specific data points were included in training. Accepted to #NAACL2025 Findings.
Haritz Puerto@HaritzPuerto

I'm excited to announce that my internship paper at @parameterlab was accepted to Findings of #NAACL2025 🎉 Huge thanks to @framart1 @coallaoh and @oodgnas! Amazing team!!

English
0
3
10
1.1K
Parameter Lab ری ٹویٹ کیا
Parameter Lab ری ٹویٹ کیا
Haritz Puerto
Haritz Puerto@HaritzPuerto·
techcrunch.com/2025/01/09/mar… From time to time we hear news like this. However, proving that an LLM was trained on a specific document is very challenging 🥴 This motivated my latest work, where we show that current methods can be effective if we use enough data 🧐
English
1
2
5
554