Heading to the EMNLP BlackboxNLP Workshop this Sunday? Don’t miss @nfel.bsky.social and @lkopf.bsky.social poster on „Interpreting Language Models Through Concept Descriptions: A Survey“
aclanthology.org/2025.blackbo...
#EMNLP #BlackboxNLP #XAI #Interpretapility
08.11.2025 10:55 — 👍 9 🔁 3 💬 0 📌 0
Counterfactual explainers for dynamic graphs by Qu et al. by @scadsai.bsky.social
arxiv.org/abs/2403.16846
Explainable Biomedical Claim Verification by Liang et al. from @dfki.bsky.social
arxiv.org/abs/2502.21014
06.11.2025 14:59 — 👍 3 🔁 0 💬 0 📌 0
We were happy to see other explainability-themed posters:
A study of monosemanticity of SAE features in VLMs by Pach et al. from @munichcenterml.bsky.social
arxiv.org/abs/2504.02821
User-centered research for data attribution by Nguyen et al. from @tuebingen-ai.bsky.social
arxiv.org/abs/2409.16978
06.11.2025 14:59 — 👍 4 🔁 0 💬 1 📌 0
Manuel Welte presented ongoing work on intrinsic interpretability of transformer models through a novel approach for restructuring internal representations.
06.11.2025 14:59 — 👍 2 🔁 0 💬 1 📌 0
@lkopf.bsky.social and @eberleoliver.bsky.social presented the PRISM framework for multi-concept feature descriptions in LLMs.
arxiv.org/abs/2506.15538
06.11.2025 14:59 — 👍 2 🔁 0 💬 1 📌 0
We are grateful for the opportunity to present some of our work at the All Hands Meeting of the German AI Centers, hosted by @dfki.bsky.social in Saarbrücken.
Andreas Lutz @eberleoliver.bsky.social Manuel Welte @lorenzlinhardt.bsky.social @lkopf.bsky.social
#AI #XAI #Interpretability
06.11.2025 14:59 — 👍 6 🔁 3 💬 1 📌 0
Happy to share that our PRISM paper has been accepted at #NeurIPS2025 🎉
In this work, we introduce a multi-concept feature description framework that can identify and score polysemantic features.
📄 Paper: arxiv.org/abs/2506.15538
#NeurIPS #MechInterp #XAI
19.09.2025 12:01 — 👍 30 🔁 4 💬 1 📌 3
a black background with green text that says `` hello , world ''
ALT: a black background with green text that says `` hello , world ''
This is the eXplainable AI research channel of the machine learning group of Prof. Klaus-Robert Müller at Technische Universität Berlin @tuberlin.bsky.social & BIFOLD @bifold.berlin.
Let's connect!
#XAI #ExplainableAI #MechInterp #MachineLearning #Interpretability
03.11.2025 11:43 — 👍 22 🔁 6 💬 0 📌 0
Using computers to better understand languages, texts, and music
OG Web, Python, Corpus Linguistics, DataViz, Philology, Ancient Greek, Music Theory, Tolkien, Space, Health, Retro Computing
Perseus, Greek Learner Texts, @digitaltolkien.com
CompLing group (CLAUSE) at Bielefeld U (PI: Sina Zarrieß). We work on: NLG, Language & Vision, Pragmatics & Dialogue, HateSpeech, BabyLMs, DH, and more!
clause-bielefeld.github.io
Postdoc @ TakeLab, UniZG | previously: Technion; TU Darmstadt | PhD @ TakeLab, UniZG
Faithful explainability, controllability & safety of LLMs.
🔎 On the academic job market 🔎
https://mttk.github.io/
Computational linguist trying to understand how humans and computers learn and use language 👶🧠🗣️🖥️💬
PhD @clausebielefeld.bsky.social, Bielefeld University
https://bbunzeck.github.io
@Lamarr Institute for Machine Learning and Artificial Intelligence
Sharing insights, news, and findings from the web
https://lamarr-institute.org/
Research in XAI @ Max Plack Institute for Human Cognitive & Brain Sciences (MPI CBS) | in the cinema @ Berlinale Generation
The MCML is a joint research initiative of LMU München and TU München. It is institutionally funded by the Federal Ministry of Education and Research and the Free State of Bavaria.
We are a joint partnership of University of Tübingen and Max Planck Institute for Intelligent Systems. We aim at developing robust learning systems and societally responsible AI. https://tuebingen.ai/imprint
https://tuebingen.ai/privacy-policy#c1104
Deutsches Forschungszentrum für Künstliche Intelligenz
Impressum: https://www.dfki.de/web/impressum
The Center for Scalable Data Analytics and Artificial Intelligence Dresden/Leipzig is an institutionalized national center of excellence for Artificial Intelligence.
👩👩👧👧 scads.ai/about-us/people/
💥 scads.ai/blog/
🧪 scads.ai/research/
PhD student @ Fraunhofer HHI. Interpretability, incremental NLP, and NLU. https://pkhdipraja.github.io/
Post-Doctoral Researcher at @eml-munich.bsky.social, in
@www.helmholtz-munich.de and @tumuenchen.bsky.social.
Optimal Transport, Explainability, Robustness, Deep Representation Learning, Computer Vision.
https://qbouniot.github.io/
Independent Researcher, NLP/ASR Engineer looking for a PhD position in Computational Neuro/Psycho/linguistics. He/him.
josellarena.github.io
Learning and hacking! 🌈⚡️🔥💧🌱
Postdoc @aau.dk
Run for life, drift for dream!
https://machuangtao.github.io/
Postdoc @ CMU • Metascience of AI: safeguarding scientific integrity in the age of AI • Explainable multi-agent systems @ University of Edinburgh • gbalint.me • 🇭🇺🏴
Computer Science PhD Student at @humboldtuni.bsky.social and @mdc-berlin.bsky.social | Data Science | Machine learning | AI | Bioinformatics | Genomics | Single-Cell Biology
Professor for AI4Health @UniOldenburg uol.de/en/ai4health Former head of Applied ML Group ML Group @FraunhoferHHI Former theoretical physicist
Visual Inference Lab of @stefanroth.bsky.social at @tuda.bsky.social - Research in Computer Vision and Machine Learning.
See https://www.visinf.tu-darmstadt.de/visual_inference
Decoding the world through bioimage analysis. Training models by day, writing papers by night. Cephalopod enthusiast. UAB MD/PhD ‘30.
Views my own