MCM 2025 - Program | MCM 2025 Chicago
MCM 2025 -- Program.
Flying today towards Chicago ๐ for MCM 2025
fjhickernell.github.io/mcm2025/prog...
Will give a talk on our recent/ongoing works on self-normalized importance sampling, including learning a proposal with MCMC and ratio diagnostics.
www.branchini.fun/pubs
24.07.2025 09:06 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0
Really cool work : ) @alexxthiery.bsky.social
www.tandfonline.com/doi/full/10....
16.07.2025 08:57 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
agree; you should check out @yfelekis.bsky.social 's work on this line ๐
08.07.2025 11:03 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0
Just don't see that the PPD_q of the original post leads somewhere useful.
Anyway, thanks for engaging @alexlew.bsky.social : )
06.07.2025 11:03 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
I agree, except I think it can be ok to shift the criteria of "good q" to instead some well-defined measure of predictive performance (under no model misspecification, let's say). Ofc Bayesian LOO-CV is one. We could discuss to use other quantities, and how to estimate them, ofc.
06.07.2025 11:03 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Genuine question: what is the estimated value used for then ?
06.07.2025 10:46 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
(computed with the inconsistent method)
06.07.2025 10:38 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Well, re: [choose q1 or q2 based on whether P_q1 > P_q2]
I seem to understand that many VI papers say: here's a new VI method, it produces q1; old VI method gives q2. q1 is better than q2 because test-log PPD is higher !
06.07.2025 10:36 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Not entirely obvious to me, but I see the intuition !
05.07.2025 13:31 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
Am definitely at least *trying* to think carefully about the evaluation here ๐
๐
05.07.2025 13:30 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
Right ! Definitely not sure if necessary, but I like to think there would be value / would be interesting if we wanted to somehow speak formally about generalizing over unseen test points
05.07.2025 13:28 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
It still seems "dangerous" to use the numerical value of (an estimate of) โซ p(y|ฮธ) q(ฮธ) to decide which approximate q is better.
(Of course, you may argue we maybe shouldn't use even any MC estimates of the original โซ p(y|ฮธ) p(ฮธ|D) with q as proposal, but the above is even less justified)
05.07.2025 13:25 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
I don't see that it needs to get that philosophical ?
It is totally possible to formally estimate the pdf itself, since we have some 'test' samples of y, and consider MISE type errors, even if in this case pointwise evaluations of the pdfs have the intractable integral.
05.07.2025 13:20 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
thatโs why I like to just say posterior predictive integral(s) instead
05.07.2025 13:13 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
This is an aside btw, but nobody in practice actually in principle estimates the marginal ppd density, right ?
Just the pointwise density evaluations at some specific yโs (whose randomness is not taken into account)
05.07.2025 13:12 โ ๐ 1 ๐ 0 ๐ฌ 2 ๐ 0
(I mean it is obvious in the unnormalized weights case, less so when self-normalizing, which was what I refer to here)
05.07.2025 12:54 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
(also, with with approach using q as proposal, it is theoretically possible to choose q that is better than iid sampling from the true posterior ;) )
05.07.2025 12:39 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
(not sure where that link came from)
I could even be somewhat ok with an inconsistent estimator, but we canโt then make a decision comparing values with different proposals in any mathematically sound way ?
05.07.2025 11:54 โ ๐ 0 ๐ 0 ๐ฌ 2 ๐ 0
but then you can just tune the underlying P_q to make it high, by changing / optimizing q !
It doesnโt make any sense
05.07.2025 11:53 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Thatโs my point about 2.
If you donโt use it as proposal, you have an inconsistent estimator of P*. With the inconsistent approach, youโre changing the true underlying quantity !
(They argue itโs of interest by itself)
This is bad as I see, since people choose q1 or q2 based on whether P_q1 > P_q2
05.07.2025 11:51 โ ๐ 3 ๐ 0 ๐ฌ 2 ๐ 0
Understanding the difficulties of posterior predictive estimation
Predictive posterior densities (PPDs) are essential in approximate inference for quantifying predictive uncertainty and comparing inference methods. Typically, PPDs are estimated by simple Monte...
This new ICML paper (openreview.net/forum?id=Tzf...) reinforces and insists on the notion that one would want to replace the posterior predictive P* := โซ p(y|ฮธ) p(ฮธ|D) ,
with P^{q} := โซ p(y|ฮธ) q(ฮธ) , with q(ฮธ) โ p(ฮธ|D), then estimate _that_ with MC.
You know me. I don't get it.
What do I miss?
05.07.2025 11:03 โ ๐ 8 ๐ 0 ๐ฌ 2 ๐ 0
It will seem very silly, but every single time I have a poster, I have at least one person being very surprised that you can do better Monte Carlo than "perfect" i.i.d. sampling from p, for E_p[f(ฮธ)] ๐
17.06.2025 20:17 โ ๐ 7 ๐ 0 ๐ฌ 0 ๐ 0
Here's how the gradient flow for minimizing KL(pi, target) looks under the Fisher-Rao metric. I thought some probability mass would be disappearing on the left and appearing on the right (i.e. teleportation), like a geodesic under the same metric, but I was very wrong... What's the right intuition?
13.06.2025 16:29 โ ๐ 23 ๐ 6 ๐ฌ 4 ๐ 0
Wish I had found this when I got started : P (or, maybe not)
13.06.2025 15:16 โ ๐ 4 ๐ 0 ๐ฌ 1 ๐ 0
it is to mimick the name of their coffee place, which is โcafetoriaโ
04.06.2025 12:52 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0
As Edinburgh's appointed minister of Gelato by @nolovedeeplearning.bsky.social, I look forward to try ๐
04.06.2025 12:48 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
I have cleaned up the notebooks for my course on Optimal Transport for Machine Learners and added links to the slides and lecture notes. github.com/gpeyre/ot4ml
25.05.2025 09:12 โ ๐ 58 ๐ 9 ๐ฌ 1 ๐ 0
Towards Adaptive Self-Normalized Importance Samplers
The self-normalized importance sampling (SNIS) estimator is a Monte Carlo estimator widely used to approximate expectations in statistical signal processing and machine learning.
The efficiency of S...
๐จ New paper: โTowards Adaptive Self-Normalized ISโ, @ IEEE Statistical Signal Processing Workshop.
TLDR;
To estimate ยต = E_p[f(ฮธ)] with SNIS, instead of doing MCMC on p(ฮธ) or learning a parametric q(ฮธ), we try MCMC directly on p(ฮธ)| f(ฮธ)-ยต | (variance-minimizing proposal).
arxiv.org/abs/2505.00372
02.05.2025 13:29 โ ๐ 31 ๐ 11 ๐ฌ 1 ๐ 0
If one of the two distributions is an isotropic Gaussian, then flow matching is equivalent to a diffusion model. This is known as Tweedie's formula. In particular, the vector field is a gradient vector, as in optimal transport. speakerdeck.com/gpeyre/compu...
31.05.2025 10:16 โ ๐ 45 ๐ 4 ๐ฌ 1 ๐ 0
Unofficial bot by @vele.bsky.social w/ http://github.com/so-okada/bXiv https://arxiv.org/list/stat.CO/new
List https://bsky.app/profile/vele.bsky.social/lists/3lim7ccweqo2j
ModList https://bsky.app/profile/vele.bsky.social/lists/3lim3qnexsw2g
PhD student at @oxfordstatistics.bsky.social working on generative models
Research fellow @ Oxford Statistics Department
jrmcornish.github.io
Professor of Mathematical Sciences, working mainly on epidemiology although partial to a bit of non-commutative algebra, social science and basic biology.
https://personalpages.manchester.ac.uk/staff/thomas.house/about.html
PhD Student at the University of Tรผbingen
Interested in ML in Science, Probabilistic Inference & Simulation
Postdoc @ EPFL, working on the theory of deep learning. Previously Polytechnique and Sorbonne Universitรฉ.
Assoc. prof. in mathematics at Chalmers/University of Gothenburg. Mainly probability, analysis and computational methods, but I like to visit most areas of mathematics. Elected member Young Academy of Sweden. Mostly here for the science, some for the jokes
Our mission is to raise awareness of queer issues in AI, foster a community of queer researchers and celebrate the work of queer scientists. More about us: queerinai.com
machine learning asst prof at @cs.ubc.ca and amii
statistical testing, kernels, learning theory, graphs, active learningโฆ
she/her, ๐ณ๏ธโโง๏ธ, @queerinai.com
PhD student at Mila | Diffusion models and reinforcement learning ๐ง | hyperpotatoneo.github.io
https://malkin1729.github.io/ / Edinburgh, Scotland / theyโฅshe>heโฅ0
Mathematician/informatician thinking probabilistically, expecting the same of you.
โTis categories in the mind and guns in their hands which keep us enslaved.
Incoming postdoc @ProbAIHub, hosted @warwickstats @Princeton @orfe. PhD @GatsbyUCL. Works on universality and ML theory. He/him. ๐
kevinhanhuang.page
Assistant Professor at the Department of Computer Science, University of Liverpool.
https://lutzoe.github.io/
Foundations of AI. I like simple and minimal examples and creative ideas. I also like thinking about the next token ๐งฎ๐งธ
Google Research | PhD, CMU |
https://arxiv.org/abs/2504.15266 | https://arxiv.org/abs/2403.06963
vaishnavh.github.io
PhD student @probcompproj and @MITCoCoSci, working on scaling data science using probabilistic programming.
Assistant professor in Statistics at UWโMadison.
Interested in #Bayesian statistics, sports analytics, causal inference. Also cocktails and Dallas sports. #mffl
Bringing together Gen AI researchers to collaborate on models driving impacts for science, industry & society. https://linktr.ee/genaihub
PhD student in statistics at Uni. Pompeu Fabra => Postdoc Bocconi | high-dimensional inference | Bayesian statistics | and moreeee https://sites.google.com/view/paulrognonvael
Statistical Machine Learning PhD student at University of Oxford