Flow-based extremal mathematical structure discovery. ~ Gergely BΓ©rczi, Baran Hashemi, Jonas KlΓΌver. arxiv.org/abs/2601.180... #AI4Math
04.02.2026 11:31 β π 3 π 1 π¬ 0 π 0Flow-based extremal mathematical structure discovery. ~ Gergely BΓ©rczi, Baran Hashemi, Jonas KlΓΌver. arxiv.org/abs/2601.180... #AI4Math
04.02.2026 11:31 β π 3 π 1 π¬ 0 π 0
1/ Are frontier LLMs the only path to AI math breakthroughs? I think not!
We introduce FlowBoost, A lightweight RL+Flow-Matching framework that discovers new extremal geometric structures, beating AlphaEvolve with 100β1000Γ less compute with zero-shot geometry-aware & reward-guided generation. π
6/
π Paper: arxiv.org/pdf/2601.18005
π» Code: github.com/berczig/Flow...
Built at
@au.dk
and
@maxplanck.de.
Happy to discuss methodology, results, or potential applications!
#AI4Mathematics #AI4Science #FlowMatching
5.2/ Key insight:
FlowBoost closes the loop: online reward-guided fine-tuning (with trust-region self-distillation) directly optimizes the flow policy β continual learning that converges in ~5 steps!
5.1/ Key insight:
Frame extremal math discovery as Simulation-Based Optimization (SBO) with a true closed feedback loop. Open-loop methods iterate blindly: generate β filter/select β retrain on elites. --> No direct signal pushes the policy toward rarer, higher-reward solutions.
4.2/ Results:
β’ Sphere packing 12d: Discover denser configurations than those produced by classical heuristics!
β’ Sphere packing 3d: Match or exceed the best previously reported packing fractions.
4.1/ Results:
β’ Circle packing: New records n=26 & 32, surpassing AlphaEvolve!
β’ Heilbronn Problem: Improve the minimum triangle area over the training dataset --> OOD sampling.
3.3/ Our key innovations:
β’ Trust-region Self-Distillation: Prevents collapse while enabling extrapolative OOD discovery.
Result: Convergence in ~10 updates vs 100β1000 in open-loop systems.
3.2/ Our key innovations:
β’ Closed-loop Reward-guided Fine-tuning: Online reward weighting + action exploration directly optimizes the flow toward rare high-score samples.
3.1/ Our key innovations:
β’ Geometry-Aware Sampling (GAS): Interleaves ODE integration with geometric constraint projections for feasible zero-shot samples on-the-fly.
2/ Extremal combinatorial geometry problems are rugged continuous landscapes, perfect for modern generative models with strong inductive biases. We ditch LLM code evolution for direct continuous generation + closed-loop optimization. We call the new paradigm, "de novo Mathematical Structure Design."
30.01.2026 20:27 β π 0 π 0 π¬ 1 π 0
1/ Are frontier LLMs the only path to AI math breakthroughs? I think not!
We introduce FlowBoost, A lightweight RL+Flow-Matching framework that discovers new extremal geometric structures, beating AlphaEvolve with 100β1000Γ less compute with zero-shot geometry-aware & reward-guided generation. π
Soon I will write a thread on our new work. Stay tuned!
#AI4Math
Tnx Kyle π€
19.09.2025 05:41 β π 1 π 0 π¬ 0 π 0
We got accepted at #NeurIPS2025. I am very happy that I could merge my knowledge of Mathematics with AI to create sth new and useful for the community. βΊοΈ
The paper: arxiv.org/abs/2505.17190
The code: github.com/Baran-phys/T...
Current AI research vibes:
- Letβs use LLM to do a baby science/math, after it doesnβt work, headline: LLM is bad at the baby math task β> guaranteed virality π
- Meanwhile, you develope a novel (non-LLM) method to solve this issue, report success on a deep math problem
β> naa, not enough dramaπ€¦π»
Another new result from the #NeurIPS rebuttal/discussion phase, our Tropical Transformer achieves much better length OOD performance across all algorithmic tasks, while being 3x-9x faster at inference and using 20% fewer parameters than the Universal Transformer (UT) models.
04.08.2025 20:47 β π 3 π 0 π¬ 0 π 0
During #NeurIPS rebuttal, we have evaluatedπ΄Tropical Transformer on the Long Range Arena (LRA), achieving highly competitive results, placing 2ndπ₯ overall in average accuracy.
Check out our paper: arxiv.org/abs/2505.17190
Our code: github.com/Baran-phys/T...
Cool. Will definitely do π
27.05.2025 05:18 β π 1 π 0 π¬ 0 π 0Interesting. I was not aware of aware if the challenges in the video subfield. But that makes sense given the context. We will definitely explore those benchmarks in the future. Thanks for the suggestions.
27.05.2025 05:10 β π 0 π 0 π¬ 1 π 0Tnx. We did not test yet on any other benshmarks. You mean algorithmic or language type benchmarks?
27.05.2025 04:57 β π 0 π 0 π¬ 1 π 0Interesting. I was not aware of this study. However, we did not just used tropical operations, we tried to simulate a concrete tropical circuit and do the message passing in the tropical space with the Generalized Hilbert metric as the kernel.
27.05.2025 04:54 β π 0 π 0 π¬ 0 π 0
7/ Our message βοΈ
Better reasoning might come not from bigger models, but from choosing the right algebra/geometry π΄.
@petar-v.bsky.social @jalonso.bsky.social
#TropicalGeometry #NeuralAlgorithmicReasoning #AI4Math
6/ We also show that each Tropical attention head can function as a tropical gate in a tropical circuit, simulating any max-plus circuit.
26.05.2025 13:08 β π 1 π 0 π¬ 1 π 05/ We benchmarked on 11 canonical combinatorial tasks. Tropical attention beat vanilla & adaptive softmax attention on all three OOD axes, Length, value and Adversarial attack generalization:
26.05.2025 13:08 β π 1 π 0 π¬ 1 π 0
4/ Tropical Attention runs each head natively in max-plus. Result:
Strong OOD length generalization with sharp attention maps even in several algorithmic tasks, including the notorious Quickselect algorithm (Another settlement for the challenge identified by @mgalkin.bsky.social )
Image by Cowdery and Challas, featured in June 2009 Mathematics Magazine
3/ In the Tropical (max + ) geometry, βadditionβ is max, βmultiplicationβ is +. Many algorithms already live here, carving exact polyhedral decision boundaries --> so why force them through exponential probabilities?
Let's ditch softmax, embrace the tropical semiring π€―πΉ.
2/ We introduce Tropical Attention -- the first Neural Algorithmic reasoner that operates in the Tropical semiring, achieving SOTA OOD performance on executing several combinatorial algorithms
arxiv.org/abs/2505.17190
π§΅ Tropical Attention --> Softmax is out, Tropical max-plus is in π¦Ύ
1/ π₯Ever experinced softmax attention fade as sequences grow?
That blur is why many attention mechanisms stumble on algorithmic and reasoning tasks. Well, we have a Algebraic Geometric Tropical solution π΄
I'm speaking about AI for enumerative geometry at the CMSA New Technologies in Mathematics seminar, on Wednesday.
07.04.2025 18:40 β π 0 π 0 π¬ 0 π 0