Worried about reliability?
CausalPFN has a built-in calibration, and can make reliable estimations even for datasets that fall outside of its pretraining prior. 
Try it using: pip install causalpfn
Made with β€οΈ for better causal inference
[7/7]
#CausalInference #ICML2025
               
            
            
                11.06.2025 13:13 β π 1    π 0    π¬ 0    π 0                      
            
         
            
        
            
            
            
            
            
    
    
    
    
            When does it work?
Our theory shows that posterior distribution of causal effects is consistent if and only if the pretraining data only includes identifiable causal structures. 
π We show how to carefully design the prior, one of the key differences in our work relative to predictive PFNs. [6/7]
               
            
            
                11.06.2025 13:13 β π 0    π 0    π¬ 1    π 0                      
            
         
            
        
            
            
            
            
                                                 
                                                
    
    
    
    
            Real-world uplift modelling:
CausalPFN works out of the box on real-world data. On 5 real RCTs in marketing (Hillstrom, Criteo, Lenta, etc.), it outperforms baselines like X-/S-/DA-Learners on policy evaluation (Qini score). [5/7]
               
            
            
                11.06.2025 13:13 β π 0    π 0    π¬ 1    π 0                      
            
         
            
        
            
            
            
            
                                                 
                                                
    
    
    
    
            Benchmarks:
On IHDP, ACIC, Lalonde:
β Best avg. rank across many tasks
β Faster than all baselines
β No tuning needed compared to the baselines (that were tuned via cross-validation)
[4/7]
               
            
            
                11.06.2025 13:13 β π 0    π 0    π¬ 1    π 0                      
            
         
            
        
            
            
            
            
            
    
    
    
    
            Why does it matter?
Causal inference traditionally needs domain expertise + hyperparameter tuning across dozens of estimators. CausalPFN flips this paradigm: we pay the cost once (at pretraining), then itβs ready to use out-of-the-box! [3/7]
               
            
            
                11.06.2025 13:13 β π 0    π 0    π¬ 1    π 0                      
            
         
            
        
            
            
            
            
            
    
    
    
    
            What is it?
CausalPFN transforms effect estimation to a supervised learning problem. It's a transformer trained on millions of simulated datasets. It learns to map from data to treatment effect distributions directly. At test time, no finetuning and manual estimator selection are required. [2/7]
               
            
            
                11.06.2025 13:13 β π 1    π 0    π¬ 1    π 0                      
            
         
            
        
            
            
            
            
                                                 
                                                
    
    
    
    
            π¨ Introducing CausalPFN, a foundation model trained on simulated data for in-context causal effect estimation, based on prior-fitted networks (PFNs). Joint work with Hamid Kamkari, Layer6AI & @rahulgk.bsky.social π§΅[1/7]
π arxiv.org/abs/2506.07918
π github.com/vdblm/Causal...
π£οΈOral@ICML SIM workshop
               
            
            
                11.06.2025 13:13 β π 4    π 1    π¬ 1    π 2                      
            
         
            
        
            
        
            
            
            
            
                                                 
                                                
    
    
    
    
            To do so, we consider all prior distributions on the unobserved factors (e.g. the distribution over each arm's mean reward) that align with the expert data. We then choose the prior with the maximum entropy (least information) and apply posterior sampling to guide the exploration (4/5)
               
            
            
                12.12.2024 16:05 β π 0    π 0    π¬ 1    π 0                      
            
         
            
        
            
            
            
            
            
    
    
    
    
            Online exploration can eventually identify unobserved factors but requires trial and error. Instead, we use expert data to limit the exploration space. In a billion-armed bandit with expert data spanning only the first ten actions, the learner should only explore those ten arms (3/5)
               
            
            
                12.12.2024 16:05 β π 0    π 0    π¬ 1    π 0                      
            
         
            
        
            
            
            
            
                                                 
                                                
    
    
    
    
            Unobserved confounding factors affect the expert policy in ways that are not understood by the learner. An important example is experts acting with privileged information. Naive imitation leads to single aggregated policies for each observed state and fails to generalize (2/5)
               
            
            
                12.12.2024 16:05 β π 0    π 0    π¬ 1    π 0                      
            
         
            
        
            
            
            
            
            
    
    
    
    
            How can we use offline expert data with unobserved confounding to guide exploration in RL? Our approach is to learn prior distributions from expert data and follow posterior sampling
Come to our poster #NeurIPS2024 today to learn more!
ποΈ Thu 12 Dec 4:30 - 7 pm PST 
π West Ballroom A-D #6708
(1/5)
               
            
            
                12.12.2024 16:05 β π 0    π 0    π¬ 1    π 1                      
            
         
    
         
        
            
        
                            
                    
                    
                                            Professor a NYU; Chief AI Scientist at Meta.
Researcher in AI, Machine Learning, Robotics, etc.
ACM Turing Award Laureate. 
http://yann.lecun.com
                                     
                            
                    
                    
                                            Machine learning researcher, working on causal inference and healthcare applications
                                     
                            
                    
                    
                                            Welcome to #UofT! Canadaβs top university, and a catalyst for discovery, innovation, and progress. #UofTDefyGravity
π utoronto.ca
                                     
                            
                    
                    
                                            Professor and Head of Machine Learning Department at Carnegie Mellon. Board member OpenAI. Chief Technical Advisor Gray Swan AI. Chief Expert Bosch Research.
                                     
                            
                    
                    
                                    
                            
                    
                    
                                            Assistant Prof of CS at the University of Waterloo, Faculty and Canada CIFAR AI Chair at the Vector Institute. Joining NYU Courant in September 2026. Co-EiC of TMLR. My group is The Salon. Privacy, robustness, machine learning.
http://www.gautamkamath.com
                                     
                            
                    
                    
                                            Cofounder & CTO @ Abridge, Raj Reddy Associate Prof of ML @ CMU, occasional writer, relapsing π·, creator of d2l.ai & approximatelycorrect.com
                                     
                            
                    
                    
                                            Professor at UW; Researcher at Meta. LMs, NLP, ML. PNW life.
                                     
                            
                    
                    
                                            MIT researcher β’ π» I study teams, communication, & computational social science β’ π Stanford, Wharton
X: @xemilyhu
                                     
                            
                    
                    
                                            β·οΈ ML Theorist carving equations and mountain trails | π΄ββοΈ Biker, Climber, Adventurer | π§  Reinforcement Learning: Always seeking higher peaks, steeper walls and better policies.
https://ualberta.ca/~szepesva
                                     
                            
                    
                    
                                            AI, sociotechnical systems, social purpose. Research director at Google DeepMind. Cofounder and Chair at Deep Learning Indaba. FAccT2025 co-program chair. shakirm.com
                                     
                            
                    
                    
                                            Machine learning prof at U Toronto.  Working on evals and AGI governance.
                                     
                            
                    
                    
                                            Research Director, Founding Faculty, Canada CIFAR AI Chair @VectorInst.
Full Prof @UofT - Statistics and Computer Sci. (x-appt) danroy.org
I study assumption-free prediction and decision making under uncertainty, with inference emerging from optimality.
                                     
                            
                    
                    
                                            PhD student @UofT of Machine Learning, Healthcare. 
http://www.cs.toronto.edu/~nikita/
                                     
                            
                    
                    
                                            scaling lawver @ EIT from BHπΊπ§π·
http://cottascience.github.io
                                     
                            
                    
                    
                                            Asst. Prof. in Machine Learning at UofT. #LongCOVID patient.
https://www.cs.toronto.edu/~cmaddis/
                                     
                            
                    
                    
                                            Assistant Professor at the University of Toronto 
βοΈ  π₯ Deep learning and causal inference for computational medicine
                                     
                            
                    
                    
                                            official Bluesky account (check usernameπ)
Bugs, feature requests, feedback: support@bsky.app