Aditya Arun's Avatar

Aditya Arun

@adityaarun1.bsky.social

Research Scientist at Adobe | PhD Computer Vision and Machine Learning https://adityaarun1.github.io/

26 Followers  |  222 Following  |  1 Posts  |  Joined: 19.11.2024  |  2.0356

Latest posts by adityaarun1.bsky.social on Bluesky

Video thumbnail

As AI agents face increasingly long and complex tasks, decomposing them into subtasks becomes increasingly appealing.

But how do we discover such temporal structure?

Hierarchical RL provides a natural formalism-yet many questions remain open.

Here's our overview of the field🧡

27.06.2025 20:15 β€” πŸ‘ 34    πŸ” 10    πŸ’¬ 1    πŸ“Œ 3
Post image

What is the probability of an image? What do the highest and lowest probability images look like? Do natural images lie on a low-dimensional manifold?
In a new preprint with Zahra Kadkhodaie and @eerosim.bsky.social, we develop a novel energy-based model in order to answer these questions: 🧡

06.06.2025 22:11 β€” πŸ‘ 70    πŸ” 23    πŸ’¬ 1    πŸ“Œ 1
Post image Post image

pretty sweet:

chewisinho.github.io/opt_notes_fi...

23.05.2025 15:33 β€” πŸ‘ 55    πŸ” 12    πŸ’¬ 2    πŸ“Œ 0
Preview
Optimal Transport for Machine Learners Optimal Transport is a foundational mathematical theory that connects optimization, partial differential equations, and probability. It offers a powerful framework for comparing probability distributi...

I have cleaned a bit my lecture notes on Optimal Transport for Machine Learners arxiv.org/abs/2505.06589

13.05.2025 05:18 β€” πŸ‘ 120    πŸ” 29    πŸ’¬ 0    πŸ“Œ 0

Happy to be recognized among outstanding reviewers πŸ™‚

10.05.2025 15:16 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Enjoying reading this. Clarifies some nice connections between scoring rules, probabilistic divergences, convex analysis, and so on. Should read it even more closely, to be honest!

01.05.2025 22:22 β€” πŸ‘ 33    πŸ” 3    πŸ’¬ 6    πŸ“Œ 1
Post image

Nice tutorial arxiv.org/abs/2503.21673

29.03.2025 18:26 β€” πŸ‘ 12    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0
Post image Post image Post image

"Inductive Moment Matching" by Linqi Zhou et al. I like the use of multiple particles to apply a loss similar to consistency models, but on distributions. Training is stable and gives high-quality generated images in very few sampling steps

πŸ“„ arxiv.org/abs/2503.07565
🌍 lumalabs.ai/news/inducti...

13.03.2025 15:05 β€” πŸ‘ 7    πŸ” 2    πŸ’¬ 0    πŸ“Œ 1
Preview
A Bear Case: My Predictions Regarding AI Progress β€” LessWrong This isn't really a "timeline", as such – I don't know the timings – but this is my current, fairly optimistic take on where we're heading. …

Smashing the endorse button as fast as I can
www.lesswrong.com/posts/oKAFFv...

09.03.2025 21:09 β€” πŸ‘ 88    πŸ” 20    πŸ’¬ 10    πŸ“Œ 6
Post image

Yet another short paper on randomness predictors (of which conformal predictors are a subclass) by Vovk, focusing on an inductive variant. arxiv.org/abs/2503.02803

06.03.2025 06:01 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 1    πŸ“Œ 1
Post image

I already advertised for this document when I posted it on arXiv, and later when it was published.

This week, with the agreement of the publisher, I uploaded the published version on arXiv.

Less typos, more references and additional sections including PAC-Bayes Bernstein.

arxiv.org/abs/2110.11216

05.03.2025 01:16 β€” πŸ‘ 109    πŸ” 22    πŸ’¬ 1    πŸ“Œ 3
Post image

1/nπŸš€If you’re working on generative image modeling, check out our latest work! We introduce EQ-VAE, a simple yet powerful regularization approach that makes latent representations equivariant to spatial transformations, leading to smoother latents and better generative models.πŸ‘‡

18.02.2025 14:26 β€” πŸ‘ 18    πŸ” 8    πŸ’¬ 1    πŸ“Œ 1
Post image

Slides for a general introduction to the use of Optimal Transport methods in learning, with an emphasis on diffusion models, flow matching, training 2 layers neural networks and deep transformers. speakerdeck.com/gpeyre/optim...

15.01.2025 19:08 β€” πŸ‘ 126    πŸ” 27    πŸ’¬ 4    πŸ“Œ 1

I don’t get the DeepSeek freak-out. Chinese orgs have been making models of all sorts that were on par with those of US orgs for a while now.

28.01.2025 00:48 β€” πŸ‘ 46    πŸ” 5    πŸ’¬ 2    πŸ“Œ 0
Preview
Convergence of Statistical Estimators via Mutual Information Bounds Recent advances in statistical learning theory have revealed profound connections between mutual information (MI) bounds, PAC-Bayesian theory, and Bayesian nonparametrics. This work introduces a novel...

We uploaded yesterday on arXiv a paper on a variant of the "Mutual Information bound" taylored to analyze statistical estimators (MLE, Bayes and variational Bayes, etc).

I assume I should advertise for it after the holidays, but in case you are still online today:
arxiv.org/abs/2412.18539

25.12.2024 11:08 β€” πŸ‘ 34    πŸ” 4    πŸ’¬ 1    πŸ“Œ 1

100%

which reminded me of @ardemp.bsky.social rule #1 on how to science: Don’t be too busy

Being too busy (with noise) = less time to read papers, less time to think and to connect the dots, less time for creative work!

15.12.2024 14:28 β€” πŸ‘ 30    πŸ” 12    πŸ’¬ 0    πŸ“Œ 4
Post image

Reinforcement Learning: An Overview

This manuscript gives a big-picture, up-to-date overview of the field of (deep) reinforcement learning and sequential decision making, covering value-based RL, policy-gradient methods, model-based methods, and various other topics.

arxiv.org/abs/2412.05265

09.12.2024 08:37 β€” πŸ‘ 54    πŸ” 8    πŸ’¬ 0    πŸ“Œ 1

I've been around the block a few times. When deep learning first became hot, many older colleagues bemoaned it as just tinkering + chain rule, and not intellectually satisfying. Then came SSL, equivariance, VAEs, GANs, neural ODEs, transformers, diffusion, etc. The richness was staggering.

πŸ§΅πŸ‘‡

06.12.2024 17:06 β€” πŸ‘ 61    πŸ” 7    πŸ’¬ 2    πŸ“Œ 0
Post image Post image Post image Post image

Detection, Pose Estimation and Segmentation for Multiple Bodies: Closing the Virtuous Circle

Miroslav Purkrabek, Jiri Matas

tl;dr: detect bbox -> mask -> estimate human pose -> mask them and repeat. SAM-enabled method :)
arxiv.org/abs/2412.01562

03.12.2024 18:53 β€” πŸ‘ 14    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Samples y | x from Treeffuser vs. true densities, for multiple values of x under three different scenarios. Treeffuser captures arbitrarily complex conditional distributions that vary with x.

Samples y | x from Treeffuser vs. true densities, for multiple values of x under three different scenarios. Treeffuser captures arbitrarily complex conditional distributions that vary with x.

I am very excited to share our new Neurips 2024 paper + package, Treeffuser! 🌳 We combine gradient-boosted trees with diffusion models for fast, flexible probabilistic predictions and well-calibrated uncertainty.

paper: arxiv.org/abs/2406.07658
repo: github.com/blei-lab/tre...

🧡(1/8)

02.12.2024 21:48 β€” πŸ‘ 156    πŸ” 23    πŸ’¬ 4    πŸ“Œ 4
Post image

Proud to announce our NeurIPS spotlight, which was in the works for over a year now :) We dig into why decomposing aleatoric and epistemic uncertainty is hard, and what this means for the future of uncertainty quantification.

πŸ“– arxiv.org/abs/2402.19460 🧡1/10

03.12.2024 09:45 β€” πŸ‘ 74    πŸ” 12    πŸ’¬ 3    πŸ“Œ 2
Video thumbnail

Optimal transport computes an interpolation between two distributions using an optimal coupling. Flow matching, on the other hand, uses a simpler β€œindependent” coupling, which is the product of the marginals.

02.12.2024 12:46 β€” πŸ‘ 197    πŸ” 31    πŸ’¬ 9    πŸ“Œ 6
Post image

Our big_vision codebase is really good! And it's *the* reference for ViT, SigLIP, PaliGemma, JetFormer, ... including fine-tuning them.

However, it's criminally undocumented. I tried using it outside Google to fine-tune PaliGemma and SigLIP on GPUs, and wrote a tutorial: lb.eyer.be/a/bv_tuto.html

03.12.2024 00:18 β€” πŸ‘ 118    πŸ” 19    πŸ’¬ 3    πŸ“Œ 2
Video thumbnail

Anne Gagneux, Ségolène Martin, @quentinbertrand.bsky.social Remi Emonet and I wrote a tutorial blog post on flow matching: dl.heeere.com/conditional-... with lots of illustrations and intuition!

We got this idea after their cool work on improving Plug and Play with FM: arxiv.org/abs/2410.02423

27.11.2024 09:00 β€” πŸ‘ 356    πŸ” 102    πŸ’¬ 12    πŸ“Œ 11
Post image

Updated: 6 benchmarks testing spatial and agent reasoning of LLM/VLMs
arxiv.org/abs/2410.06468 does spatial cognition
arxiv.org/abs/2307.06281 MMBench
arxiv.org/abs/2411.13543 BALROG
arxiv.org/abs/2410.07765 GameTraversalBenchmark
3dsrbench.github.io 3DSRBenchmark
open-eqa.github.io Open-EQA

26.11.2024 08:25 β€” πŸ‘ 62    πŸ” 9    πŸ’¬ 2    πŸ“Œ 1
Post image

Yeah, looks like different image resolutions will give you entirely different depth maps.

26.11.2024 07:24 β€” πŸ‘ 70    πŸ” 4    πŸ’¬ 7    πŸ“Œ 1
Preview
Theoretical Foundations of Conformal Prediction This book is about conformal prediction and related inferential techniques that build on permutation tests and exchangeability. These techniques are useful in a diverse array of tasks, including hypot...

This new textbook by Anastasios Angelopoulos, Rina Barber and Stephen Bates looks really neat! arxiv.org/abs/2411.11824

19.11.2024 13:47 β€” πŸ‘ 58    πŸ” 6    πŸ’¬ 2    πŸ“Œ 2
Post image

now, if we think of p(output | prompt, a few examples) as a predictive distribution p(y|x, D) ... it looks very much like learning to me :)

see e.g. my slide deck on drive.google.com/file/d/1B-Ka...

21.11.2024 00:43 β€” πŸ‘ 25    πŸ” 4    πŸ’¬ 3    πŸ“Œ 1
Preview
GitHub - kuleshov-group/awesome-discrete-diffusion-models: A curated list for awesome discrete diffusion models resources. A curated list for awesome discrete diffusion models resources. - kuleshov-group/awesome-discrete-diffusion-models

Discrete diffusion has become a very hot topic again this year. Dozens of interesting ICLR submissions and some exciting attempts at scaling. Here's a bibliography on the topic from the Kuleshov group (my open office neighbors).

github.com/kuleshov-gro...

21.11.2024 18:39 β€” πŸ‘ 76    πŸ” 10    πŸ’¬ 1    πŸ“Œ 0

@adityaarun1 is following 20 prominent accounts