Krishnakumar Balasubramanian, Nathan Ross
Finite-Dimensional Gaussian Approximation for Deep Neural Networks: Universality in Random Weights
https://arxiv.org/abs/2507.12686
@krizna.bsky.social
https://sites.google.com/view/kriznakumar/ Associate professor at @ucdavis #machinelearning #deeplearning #probability #statistics #optimization #sampling
Krishnakumar Balasubramanian, Nathan Ross
Finite-Dimensional Gaussian Approximation for Deep Neural Networks: Universality in Random Weights
https://arxiv.org/abs/2507.12686
link to paper:
arxiv.org/abs/2505.15059
New theory for simulated tempering using restricted spectral gap with arbitrary local MCMC samplers under multi-modality.
When applied to simulated tempering Metropolis-Hasting algorithm for sampling from Gaussian mixture models, we obtain high-accuracy TV guarantees.
We implement these oracles using heat-kernel truncation & Varadhan's asymptotics, linking our method to entropy-regularized proximal point method on Wasserstein spaces, in the latter case.
Joint work with Yunrui Guan and @shiqianma.bsky.social
New work on Riemannian Proximal Sampler, to sample on Riemannian manifolds:
arxiv.org/abs/2502.07265
Comes with high-accuracy (i.e., log(1/eps), where eps is tolerance) guarantees with exact and inexact oracles for Manifold Brownian Increments and Riemannian Heat-kernels
Happy to have this paper on Improved rates for Stein Variational Gradient Descent accepted as an oral presentation at #ICLR2025
arxiv.org/abs/2409.08469
Only theory, No deep learning (although techniques useful for DL), No experiments in this time of scale and AGI :)
Our bounds show how key factorsβlike the number of matches and treatment balanceβimpact Gaussian approximation accuracy.
We also introduce multiplier bootstrap bounds for obtaining finite-sample valid, data-driven confidence intervals.
Matching-based ATE estimators align treated and control units to estimate causal effects without strong parametric assumptions.
Using Malliavin-Stein method we establish Gaussian Approximation bounds for these estimators.
Got this paper out in 2024, just in time before AGI takes over in 2025:
arxiv.org/abs/2412.17181
We develop Gaussian approximation bounds and non-asymptotically valid confidence intervals for matching-based Average Treatment Effect (ATE) estimators.
It seems that OpenAI's latest model, o3, can solve 25% of problems on a database called FrontierMath, created by EpochAI, where previous LLMs could only solve 2%. On Twitter I am quoted as saying, "Getting even one question right would be well beyond what we can do now, let alone saturating them."
20.12.2024 23:15 β π 86 π 8 π¬ 8 π 1Von Neumann: With 4 parameters, I can fit an elephant. With 5, I can make it wiggle its trunk.
OpenAI: Hold my gazillion parameter Sora model - Iβll make the elephant out of leaves and teach it to dance.
youtu.be/4QG_MGEBQow?...
thanks, resent the email now!
03.12.2024 01:11 β π 0 π 0 π¬ 0 π 0@iclr-conf.bsky.social Would greatly appreciate any guidance on what to do if reviewer, AC and PC did not respond. Thanks a lot!
cc:
@yisongyue.bsky.social
How well RF performs in these settings? Thatβs still an open question.
Bottom-line: Time to compare SGD-trained NNs with RF and not kernel methods!
Going beyond mean-field regime for SGD trained NNs certainly helps. Recent works connect learnability of SGD trained NNs with leap complexity and information exponent of function classes (like single and multi index models) with the goal of explaining feature learning.
27.11.2024 15:07 β π 0 π 0 π¬ 1 π 0It also creates an intriguing parallel with NNs: greedy-trained partitioning models and SGD-trained NNs (in the mean-field regime) both thrive under specific structural assumptions (eg. MSP) but struggle otherwise.
However, under MSP, greedy RFs are provably better that SGD-trained 2-NNs!
In our work:
arxiv.org/abs/2411.04394
we show that If the true regression function satisfies MSP, greedy training works well with π(log π) samples.
Otherwise, it struggles.
This settles the question of learnability for greedy recursive partitioning algorithms like CART.
MSP is used to argue that SGD trained 2-layer NNs are better than vanilla kernel methods.
But how do neural nets compare with random forest (RF) trained using greedy algorithms like CART?
How to characterize the learnability of local algorithms ?
The Merged Staircase Property (MSP) proposed by Abbe et al. (2022) is used to completely characterize the learnability of SGD-trained 2-layer neural networks (NN) in the regime where mean-field approximation holds for SGD.
add me please
π
Yes, but is the cover indicative of RL notations by any chance :P
24.11.2024 17:31 β π 1 π 0 π¬ 1 π 0