Sebastian Bordt's Avatar

Sebastian Bordt

@sbordt.bsky.social

Language models and interpretable machine learning. Postdoc @ Uni Tรผbingen. https://sbordt.github.io/

469 Followers  |  247 Following  |  66 Posts  |  Joined: 10.10.2023  |  1.7067

Latest posts by sbordt.bsky.social on Bluesky

Preview
Theory of XAI Workshop, Dec 2, 2025 Explainable AI (XAI) is now deployed across a wide range of settings, including high-stakes domains in which misleading explanations can cause real harm. For example, explanations are required by law ...

๐Ÿšจ Workshop on the Theory of Explainable Machine Learning

Call for โ‰ค2 page extended abstract submissions by October 15 now open!

๐Ÿ“ Ellis UnConference in Copenhagen
๐Ÿ“… Dec. 2
๐Ÿ”— More info: sites.google.com/view/theory-...

@gunnark.bsky.social @ulrikeluxburg.bsky.social @emmanuelesposito.bsky.social

30.09.2025 14:00 โ€” ๐Ÿ‘ 7    ๐Ÿ” 4    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

I am hiring PhD students and/or Postdocs, to work on the theory of explainable machine learning. Please apply through Ellis or IMPRS, deadlines end october/mid november. In particular: Women, where are you? Our community needs you!!!

imprs.is.mpg.de/application
ellis.eu/news/ellis-p...

17.09.2025 06:17 โ€” ๐Ÿ‘ 23    ๐Ÿ” 15    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

We need new rules for publishing AI-generated research. The teams developing automated AI scientists have customarily submitted their papers to standard refereed venues (journals and conferences) and to arXiv. Often, acceptance has been treated as the dependent variable. 1/

14.09.2025 17:15 โ€” ๐Ÿ‘ 73    ๐Ÿ” 23    ๐Ÿ’ฌ 4    ๐Ÿ“Œ 4
Preview
Center for the Alignment of AI Alignment Centers We align the aligners

This new center strikes the right tone in approaching the AI alignment problem. alignmentalignment.ai

11.09.2025 20:47 โ€” ๐Ÿ‘ 58    ๐Ÿ” 14    ๐Ÿ’ฌ 4    ๐Ÿ“Œ 4
Pascual Restrepo - Research Pascual Restrepo Official Website. Economist, MIT.

I dont know if it's a good point to start, but you might want to take a look at the works by Daron Acemoglu and Pascual Restrepo pascual.scripts.mit.edu/research/

03.09.2025 21:41 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
How much can we forget about Data Contamination? - [Sebastian Bordt]
YouTube video by Friday Talks Tรผbingen How much can we forget about Data Contamination? - [Sebastian Bordt]

A new recording of our FridayTalks@Tรผbingen series is online!

How much can we forget about Data Contamination?
by
@sbordt.bsky.social

Watch here: youtu.be/T9Y5-rngOLg

29.08.2025 07:05 โ€” ๐Ÿ‘ 2    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
What Formal Languages Can Transformers Express? A Survey Abstract. As transformers have gained prominence in natural language processing, some researchers have investigated theoretically what problems they can and cannot solve, by treating problems as forma...

Thanks for your comments. I don't think that neural networks are just a form of memory (though they can store a large number of memories). For example, transformers with unbounded steps are Turing-complete direct.mit.edu/tacl/article...

23.07.2025 05:10 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I see the point of the original post, but I think it's also important to keep in mind this other aspect.

20.07.2025 21:13 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

www.inference.vc/we-may-be-su...

20.07.2025 20:54 โ€” ๐Ÿ‘ 7    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

The stochastic parrot is now an IMO gold medalist parrot

19.07.2025 20:50 โ€” ๐Ÿ‘ 57    ๐Ÿ” 7    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 2

Wednesday: Position: Rethinking Explainable Machine Learning as Applied Statistics icml.cc/virtual/2025...

14.07.2025 14:49 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I'm at #ICML in Vancouver this week, hit me up if you want to chat about pre-training experiments or explainable machine learning.

You can find me at these posters:

Tuesday: How Much Can We Forget about Data Contamination? icml.cc/virtual/2025...

14.07.2025 14:49 โ€” ๐Ÿ‘ 1    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Great to hear that you like it, and thank you for the feedback! I agree that stakeholders are important, although you are not going to find much about it in this paper. We might argue, though, that similar aspects with stakeholders arise in data science with large datasets, hence the analogy :)

11.07.2025 22:39 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Our #ICML position paper: #XAI is similar to applied statistics: it uses summary statistics in an attempt to answer real world questions. But authors need to state how concretely (!) their XAI statistics contributes to answer which concrete (!) question!
arxiv.org/abs/2402.02870

11.07.2025 07:35 โ€” ๐Ÿ‘ 6    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

There are many more interesting aspects to this, so take a look at our paper!

arxiv.org/abs/2402.02870

We would also be happy for questions and comments on why we got it completely wrong.๐Ÿ˜Š

If you are at ICML, I will present this paper on Wed 16 Jul 4:30 in the East Exhibition Hall A-B #E-501.๐Ÿ“

10.07.2025 18:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

We think the literature on explainable machine learning can learn a lot from looking at these papers!๐Ÿ“š

10.07.2025 18:02 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

As I learned from our helpful ICML reviewers, there is a lot of existing research at the intersection of machine learning and statistics that takes the matter of interpretation quite seriously.

10.07.2025 18:02 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

In this framework, another way to formulate the initial problems is: For many popular explanation algorithms, it is not clear whether they have an interpretation.

10.07.2025 18:01 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Having an interpretation means that the explanation formalizes an intuitive human concept, which is a fancy philosophical way of saying that it is clear what aspect of the function the explanation describes.๐Ÿง 

10.07.2025 18:01 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

In addition, the way to develop explanations that are useful "in the world" is to develop explanations that have an interpretation.

10.07.2025 18:01 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

This has several important implications. Most importantly, explainable machine learning has often been trying to reinvent the wheel when we already have a robust framework for discussing complex objects in the light of pressing real-world questions.

10.07.2025 18:00 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

It took us a while to recognize it, but once you see it, you can't unsee it: Explainable machine learning is applied statistics for learned functions.โœจ

10.07.2025 18:00 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Concretely, researchers in applied statistics study complex datasets by mapping their most important properties into low-dimensional structures. Now think:

Machine learning model ~ Large dataset
Explanation algorithm ~ Summary statistics, visualization

10.07.2025 18:00 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Here comes our key realization: This question has occurred in other disciplines before, specifically in applied statistics research.

10.07.2025 18:00 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

So, how can we seriously discuss whether an explanation algorithm can be used to answer relevant questions about our trained model or the world?๐ŸŒ

10.07.2025 17:59 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

I have actually encountered this point in my own research before, where we did a detailed mathematical analysis of SHAP, but all the math could not reveal the right way to use the explanations in practice (arxiv.org/abs/2209.040...

10.07.2025 17:59 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

And then they never make clear how the technical properties of the algorithm tie back to the grander goals of explainability research (meaning: how do we actually use the algorithm in practice?)๐ŸŽฏ

10.07.2025 17:59 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

The problem: Papers about explanation algorithms often start with grand promises, but then quickly dive into technical details.

10.07.2025 17:59 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

During the last couple of years, we have read a lot of papers on explainability and often felt that something was fundamentally missing๐Ÿค”

This led us to write a position paper (accepted at #ICML2025) that attempts to identify the problem and to propose a solution.

arxiv.org/abs/2402.02870
๐Ÿ‘‡๐Ÿงต

10.07.2025 17:58 โ€” ๐Ÿ‘ 12    ๐Ÿ” 5    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

joint work with Suraj Srinivas, Valentyn Boreiko, and @ulrikeluxburg.bsky.social

Link to paper: arxiv.org/pdf/2410.03249
Link to code: github.com/tml-tuebinge...

We would be happy for feedback or questions about this work!

08.07.2025 06:46 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

@sbordt is following 20 prominent accounts