Today at NeurIPS, weโll be presenting our Noether's Razor paper! ๐โจ
๐
Today Fri, Dec 13
โฐ 11 a.m. โ 2 p.m. PST
๐ East Exhibit Hall A-C, #4710 (ALL the way in the back I believe!)
w/ @mvdw.bsky.social @pimdh.bsky.social
Come say hi! ๐
13.12.2024 16:38 โ ๐ 21 ๐ 3 ๐ฌ 2 ๐ 0
For those interested, I will be giving a short talk today about our recent work โPyramid Vector Quantization for LLMsโ at the Microsoft booth at NeurIPS, starting at 10:30.
12.12.2024 17:40 โ ๐ 4 ๐ 0 ๐ฌ 0 ๐ 0
Sure!
06.12.2024 16:51 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0
Noether's razor: Learning Conserved Quantities
Symmetries have proven useful in machine learning models, improving generalisation and overall performance. At the same time, recent advancements in learning dynamical systems rely on modelling the un...
Interested?
Come check out our poster at NeurIPS 2024 in Vancouver.
East Exhibit Hall A-C #4710, Fri 13 Dec 1-4 pm CST
Link: neurips.cc/virtual/2024...
Paper/code: arxiv.org/abs/2410.08087
Thankful to co-authors @mvdw.bsky.social and @pimdehaan for their joint supervision of this project
๐งต16/16
06.12.2024 13:42 โ ๐ 12 ๐ 1 ๐ฌ 1 ๐ 0
Noether's razor allows symmetries to be parameterised in terms of conserved quantities and enables automatic symmetry at train time. This results in more accurate Hamiltonians that obey symmetry and trajectory predictions that remain more accurate over longer time spans.๐งต15/16
06.12.2024 13:42 โ ๐ 5 ๐ 0 ๐ฌ 1 ๐ 0
Our work demonstrates that approximate Bayesian model selection can be useful in neural nets, even in sophisticated use cases. We aim to further improve the efficiency and usability of neural model selection, making it a more integral part of training deep neural nets. ๐งต14/16
06.12.2024 13:42 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
If one wishes, we could extend to non-quadratic conserved quantities to model non-affine actions. More free-form conserved quantities would require ODE-solving instead of the matrix exponential and have more parameters, which could complicate the Bayesian model selection.๐งต13/16
06.12.2024 13:42 โ ๐ 3 ๐ 0 ๐ฌ 1 ๐ 0
Further, this does not limit the shapes of symmetry groups we can learn. For instance, we find the Euclidean group, which is itself a curved manifold with non-trivial topology.๐งต12/16
06.12.2024 13:42 โ ๐ 3 ๐ 0 ๐ฌ 1 ๐ 0
We use quadratic conserved quantities, which can represent any symmetry with an affine (linear + translation) action on the state space. This encompasses nearly all cases studied in geometric DL. ๐งต12/16
06.12.2024 13:42 โ ๐ 3 ๐ 0 ๐ฌ 1 ๐ 0
On more complex N-body problems, our method correctly discovers the correct 7 linear generators which correspond to the correct linear symmetries of rotation around center of mass, rotation around the origin, translations, and momentum-dependent translations. ๐งต11/16
06.12.2024 13:42 โ ๐ 3 ๐ 0 ๐ฌ 2 ๐ 0
By learning the correct symmetries, the jointly learned Hamiltonians are more accurate, directly improving trajectory predictions at test time. We show this for n-harmonic oscillators, but also more complex N-body problems (see table below). ๐งต10/16
06.12.2024 13:42 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
We verify the correct symmetries and group dimensionality are learned by inspecting parallelism, singular vectors, and transformations associated with learned generators. For instance, we correctly learn the nยฒ dimensional unitary Lie group U(n) on N-harmonic oscillators.๐งต9/16
06.12.2024 13:42 โ ๐ 4 ๐ 0 ๐ฌ 1 ๐ 0
Our method discovers the correct symmetries from data. Learned Hamiltonians that obey the right symmetry generalise better as they remain more accurate in larger areas of the phase space, depicted here for a correctly learned SO(2) on a simple harmonic oscillator. ๐งต8/16
06.12.2024 13:42 โ ๐ 3 ๐ 0 ๐ฌ 1 ๐ 0
Noether's razor jointly learns conservation laws and the symmetrised Hamiltonian on train data, without requiring validation data. By relying on differentiable model selection, we do not introduce additional regularizers that require tuning.๐งต7/16
06.12.2024 13:42 โ ๐ 5 ๐ 0 ๐ฌ 1 ๐ 0
As far as we know, this is the first case in which Bayesian model selection with VI is successfully scaled to deep neural networks, an achievement in its own right, whereas most works so far have relied on Laplace approximations. ๐งต6/16
06.12.2024 13:42 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
We scale to deep neural networks using Variational Inference (VI). To obtain a sufficiently tight bound for model selection, we avoid mean-field and instead use efficient matrix normal posteriors and closed-form updates of prior and output variance (see tricks in App. D). ๐งต5/16
06.12.2024 13:42 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
We jointly learn the Hamiltonian and conserved quantities with Bayesian model selection by optimising marginal likelihood estimates. This yields a Noether's Razor effect favouring simpler symmetric solutions if this describe data well, avoiding collapse into no symmetry. ๐งต4/16
06.12.2024 13:42 โ ๐ 4 ๐ 0 ๐ฌ 1 ๐ 0
We let our Hamiltonian neural network be a flexible deep neural network that is symmetrized by integrating it over the flow induced by a set of learnable conserved quantities. This construction allows us to define a flexible function with learnable symmetry. ๐งต3/16
06.12.2024 13:42 โ ๐ 3 ๐ 0 ๐ฌ 1 ๐ 0
Noether's theorem states that symmetries in a dynamical system have an associated conserved quantity, i.e. an observable that remains invariant over a trajectory. We use this result to parameterise symmetries in Hamiltonian neural networks in terms of conserved quantities. ๐งต2/16
06.12.2024 13:42 โ ๐ 5 ๐ 0 ๐ฌ 1 ๐ 0
๐Noether's razorโญ๏ธ Our NeurIPS 2024 paper connects ML symmetries to conserved quantities through a seminal result in mathematical physics: Noether's theorem. We can learn neural network symmetries from data by learning associated conservation laws. Learn more๐. 1/16๐งต
06.12.2024 13:42 โ ๐ 101 ๐ 13 ๐ฌ 3 ๐ 2
I made a starter pack for Bayesian ML and stats (mostly to see how this starter pack business works).
Let me know whom I missed!
go.bsky.app/2Bqtn6T
18.11.2024 16:19 โ ๐ 96 ๐ 33 ๐ฌ 54 ๐ 2
I made a starter pack for researchers in probabilistic machine learning.
DM/reply if you want to be added!
go.bsky.app/DuCtJqC
17.11.2024 23:09 โ ๐ 105 ๐ 37 ๐ฌ 67 ๐ 2
Hi ๐ I'm a postdoc in the #Neuroimmunology and #Imaging group at the @dzne.science Bonn ๐งช๐ฌ Passionate about #ComputationalNeuroscience ๐ง ๐ป and #NeuralModeling ๐งฎ
๐ fabriziomusacchio.com
๐จโ๐ป github.com/FabrizioMusacchio
๐ sigmoid.social/@pixeltracker
Physicist, PhD network science & co-founder www.dribia.com. Strictly personal & sci thoughts in this account. Interested in #ML #Sport #Politics
President of the @ec.europa.eu
Mother of seven. Brussels-born. European by heart. ๐ช๐บ
Human/AI interaction. ML interpretability. Visualization as design, science, art. Professor at Harvard, and part-time at Google DeepMind.
Assistant professor (of mathematics) at the University of Toronto. Algebraic geometry, number theory, forever distracted and confused, etc. He/him.
Professor a NYU; Chief AI Scientist at Meta.
Researcher in AI, Machine Learning, Robotics, etc.
ACM Turing Award Laureate.
http://yann.lecun.com
Neuroscience, RL for motor learning, neural control of movement, NeuroAI.
Opinions stated here are my own, not those of my company.
PhD candidate @ University of Edinburgh
Bayesian Stats | Machine Learning | Uncertainty Quantification | ML4Science | Scientific Imaging
https://teresa-klatzer.github.io/
My opinions only here.
๐จโ๐ฌ RS DeepMind
Past:
๐จโ๐ฌ R Midjourney 1y ๐งโ๐ DPhil AIMS Uni of Oxford 4.5y
๐งโโ๏ธ RE DeepMind 1y ๐บ SWE Google 3y ๐ TUM
๐ค @nwspk
Chief Scientist at Inverted AI
Bahvioral modelling for autonomous driving
PhD physicist turned AI researcher
http://www.julian-rodemann.de | PhD student in statistics @LMU_Muenchen | currently @HarvardStats
AI Research Scientist @Rain.AI (he/him/his)
Machine Learning Professor
https://cims.nyu.edu/~andrewgw
Deep learner at FAIR. Into codegen, RL, equivariance, generative models. Spent time at Qualcomm, Scyfer (acquired), UvA, Deepmind, OpenAI.
Machine Learning at CuspAI - materials discovery for carbon capture
Enjoy learning new things. PhD student at Ghent University Global Campus in Korea.
Large Models, Multimodality, Continual Learning | ELLIS ML PhD with Oriol Vinyals & Zeynep Akata | Previously Google DeepMind, Meta AI, AWS, Vector, MILA
๐ karroth.com
Scientist, Inventor, author of the NTQR Python package for AI safety through formal verification of unsupervised evaluations. On a mission to eliminate Majority Voting from AI systems. E Pluribus Unum.