‼️New paper from Parameter Lab!
⛓️💥 We identify privacy collapse, a silent failure mode of LLMs: LLMs fine-tuned on seemingly benign data can lose their ability to respect contextual privacy norms.
Done by @anmolgoel.bsky.social during his internship!
Check-out 👇
03.02.2026 19:40 — 👍 3 🔁 1 💬 0 📌 0
Our #EMNLP2025 paper Leaky Thoughts 🫗 shows that Large Reasoning Models (LRMs) can unintentionally leak sensitive information hidden in their internal thoughts.
📍 Come chat with Tommaso at our poster on Friday 7th, 10:30–12:00 in Hall C3
📄 aclanthology.org/2025.emnlp-m...
04.11.2025 21:45 — 👍 2 🔁 1 💬 0 📌 0
Leaky Thoughts: Large Reasoning Models Are Not Private Thinkers
We study privacy leakage in the reasoning traces of large reasoning models used as personal agents. Unlike final outputs, reasoning traces are often assume...
We challenge the view that reasoning traces are a safe internal part of a model’s process. Our work shows they can leak information, through both deliberate attacks and accidental leakage.
RTAI: researchtrend.ai/papers/2506....
ArXiv: arxiv.org/abs/2506.15674
Code: github.com/parameterlab...
2/2
21.08.2025 15:14 — 👍 1 🔁 0 💬 0 📌 0
Overall diagram about contextual privacy & LRMs
🫗 An LLM's "private" reasoning may leak your sensitive data!
🎉 Excited to share our paper "Leaky Thoughts: Large Reasoning Models Are Not Private Thinkers" was accepted at #EMNLP main!
1/2
21.08.2025 15:14 — 👍 5 🔁 1 💬 1 📌 2
Work done with: Haritz Puerto, Martin Gubri @mgubri.bsky.social , Tommaso Green, Sangdoo Yun and Seong Joon Oh @coallaoh.bsky.social
#SEO #AI #LLM #GenerativeAI #Marketing #DigitalMarketing #Perplexity #NLProc
23.06.2025 16:38 — 👍 1 🔁 0 💬 0 📌 0
Key takeaways:
❌ C-SEO doesn’t help improve visibility in AI answers.
🔎 Traditional SEO is your tool for online visibility.
🚀 Our benchmark sets the stage to develop C-SEO methods that might work in the future.
23.06.2025 16:38 — 👍 0 🔁 0 💬 1 📌 0
🔎 The results are clear: current C-SEO strategies don’t work. This challenges the recent hype and suggests that creators don’t need to game LLMs and create even more clickbaits. Just focus on producing genuinely good content and let traditional SEO do its work.
23.06.2025 16:38 — 👍 0 🔁 0 💬 1 📌 0
C-SEO Bench evaluates Conversational Search Engine Optimization (C-SEO) techniques on two key tasks:
🔍 Product Recommendation
❓ Question Answering
Spanning multiple domains, it tests both domain-specific performance and the generalization of C-SEO methods.
23.06.2025 16:38 — 👍 0 🔁 0 💬 1 📌 0
Illustration of a conversational search engine for product recommendation. After applying a C-SEO method on the third document, its ranking gets boosted by +2 positions.
💥 With the rise of conversational search, a new technique of "Conversational SEO" (C-SEO) emerged, claiming it can boost content inclusion in AI-generated answers. We put these claims to the test by building C-SEO Bench, the first comprehensive benchmark to rigorously evaluate these new strategies.
23.06.2025 16:38 — 👍 0 🔁 0 💬 1 📌 0
Paper thumbnail.
🔎Does Conversational SEO actually work? Our new benchmark has an answer!
Excited to announce our new paper: C-SEO Bench: Does Conversational SEO Work?
🌐 RTAI: researchtrend.ai/papers/2506....
📄 Paper: arxiv.org/abs/2506.11097
💻 Code: github.com/parameterlab...
📊 Data: huggingface.co/datasets/par...
23.06.2025 16:38 — 👍 2 🔁 1 💬 1 📌 1
Excited to share that our paper "Scaling Up Membership Inference: When and How Attacks Succeed on LLMs" will be presented next week at #NAACL2025!
🖼️ Catch us at Poster Session 8 - APP: NLP Applications
🗓️ May 2, 11:00 AM - 12:30 PM
🗺️ Hall 3
Hope to see you there!
26.04.2025 10:11 — 👍 2 🔁 1 💬 0 📌 0
Ready to Join? Send your resume + a short note on why you’re a great fit to recruit@parameterlab.de.
Be part of a team that’s redefining research with AI! #Hiring #DataEngineer #AI #RemoteJobs
14.02.2025 16:08 — 👍 0 🔁 0 💬 0 📌 0
Why Join Us?
🚀 Make a Difference – Your work directly enhances how research is shared and discovered.
🌍 Flexibility – Choose full-time or part-time, work remotely or locally.
⚡ Innovative Environment – AI, research, and data-driven solutions all in one place.
🤝 Great Team
14.02.2025 16:08 — 👍 0 🔁 0 💬 1 📌 0
What You Bring:
✅ Proficiency in Airflow & PostgreSQL – Complex workflows and databases.
✅ Strong Python Skills – Clean, efficient, and maintainable code is your thing.
✅ (Bonus) Experience with LLMs – A huge plus as we integrate AI-driven solutions.
✅ Problem-Solving Mindset
✅ Team Spirit
14.02.2025 16:08 — 👍 0 🔁 0 💬 1 📌 0
What You’ll Do:
✔ Build Scalable Data Pipelines – Design and optimize workflows using tools like Airflow.
✔ Work Closely with AI Experts & Engineers – Collaborate to solve real-world data challenges.
✔ Optimize and Maintain Systems – Keep our data infrastructure fast, secure, and adaptable.
14.02.2025 16:08 — 👍 0 🔁 0 💬 1 📌 0
Our LLM-powered ecosystem also bridges the gap between cutting-edge research and industry leaders. If you're passionate about data, AI, and making an impact, we’d love to have you on board!
14.02.2025 16:08 — 👍 0 🔁 0 💬 1 📌 0
ResearchTrend.AI
Explore the most trending research topics in AI
👥 We're Hiring: Senior/Junior Data Engineer!
📍 Remote or Local | Full-Time or Part-Time
At ResearchTrend.AI, we’re building a platform that connects researchers and AI engineers worldwide—helping them stay ahead with daily digests, insightful summaries, and interactive events.
14.02.2025 16:08 — 👍 2 🔁 0 💬 1 📌 1
🔎 Wonder how to prove an LLM was trained on a specific text? The camera ready of our Findings of #NAACL 2025 paper is available!
📌 TLDR: longs texts are needed to gather enough evidence to determine whether specific data points were included in training of LLMs: arxiv.org/abs/2411.00154
06.02.2025 13:02 — 👍 5 🔁 1 💬 0 📌 0
We are delighted to announce that our research paper on the scale of LLM membership inference has been accepted for publication in the Findings of #NAACL2025! 🎉
23.01.2025 14:16 — 👍 4 🔁 0 💬 0 📌 0
🎉We’re pleased to share the release of the models from our Apricot🍑 paper, accepted at ACL 2024!
At Parameter Lab, we believe openness and reproducibility are essential for advancing science, and we've put in our best effort to ensure it.
🤗 huggingface.co/collections/...
🧵 bsky.app/profile/dnns...
20.11.2024 23:55 — 👍 9 🔁 3 💬 0 📌 0
🙌 Team Credits: This research was conducted by Haritz Puerto @mgubri.bsky.social @oodgnas.bsky.social and @coallaoh.bsky.social with support from NAVER AI Lab. Stay tuned for more updates! 🚀
19.11.2024 09:15 — 👍 1 🔁 0 💬 1 📌 0
🤓 Want More? Check out the community page of MIA for LLMs in http://ReserachTrend.AI https://researchtrend.ai/communities/MIALM You can see related works, the evolution of the community, and top authors!
19.11.2024 09:15 — 👍 0 🔁 0 💬 1 📌 0
💬 What Do You Think? Could MIA reach a level where data owners use it as legal evidence? How might this affect LLM deployment? Let us know! #AI #LLM #NLProc
19.11.2024 09:15 — 👍 0 🔁 0 💬 1 📌 0
🌐 Implications for Data Privacy: Our findings have real-world relevance for data owners worried about unauthorized use of their content in model training. It can also be used to support accountability of LLM evaluation in end-tasks.
19.11.2024 09:15 — 👍 0 🔁 0 💬 1 📌 0
🔎 Better Results in Fine-Tuning: Fine-tuned models show even stronger MIA results. The table shows the performance at sentence level and for collections of 20 sentences, evaluated on Phi-2 fine-tuned for QA (https://huggingface.co/haritzpuerto/phi-2-dcot ).
19.11.2024 09:15 — 👍 0 🔁 0 💬 1 📌 0
🔬 Our Testing Setup: We ran experiments using Pythia models (2.8B and 6.9B parameters) with training samples from The Pile dataset, comparing them to validation and test sets. This setup avoids data leakage to ensure a reliable evaluation of MIA.
19.11.2024 09:15 — 👍 0 🔁 0 💬 1 📌 0
🚀 The Key? Number of tokens & Aggregation: MIA’s accuracy improves as we aggregate MIA scores across multiple paragraphs. Longer documents or larger document collections significantly boost MIA effectiveness.
19.11.2024 09:15 — 👍 0 🔁 0 💬 1 📌 0