Baran Hashemi's Avatar

Baran Hashemi

@rythian47.bsky.social

AI for Mathematics

52 Followers  |  106 Following  |  35 Posts  |  Joined: 01.12.2024
Posts Following

Posts by Baran Hashemi (@rythian47.bsky.social)

Preview
Flow-based Extremal Mathematical Structure Discovery The discovery of extremal structures in mathematics requires navigating vast and nonconvex landscapes where analytical methods offer little guidance and brute-force search becomes intractable. We intr...

Flow-based extremal mathematical structure discovery. ~ Gergely BΓ©rczi, Baran Hashemi, Jonas KlΓΌver. arxiv.org/abs/2601.180... #AI4Math

04.02.2026 11:31 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

1/ Are frontier LLMs the only path to AI math breakthroughs? I think not!
We introduce FlowBoost, A lightweight RL+Flow-Matching framework that discovers new extremal geometric structures, beating AlphaEvolve with 100–1000Γ— less compute with zero-shot geometry-aware & reward-guided generation. πŸš€

30.01.2026 20:27 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

6/
πŸ“„ Paper: arxiv.org/pdf/2601.18005
πŸ’» Code: github.com/berczig/Flow...
Built at
@au.dk
and
@maxplanck.de.
Happy to discuss methodology, results, or potential applications!
#AI4Mathematics #AI4Science #FlowMatching

30.01.2026 20:27 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

5.2/ Key insight:
FlowBoost closes the loop: online reward-guided fine-tuning (with trust-region self-distillation) directly optimizes the flow policy β€” continual learning that converges in ~5 steps!

30.01.2026 20:27 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

5.1/ Key insight:
Frame extremal math discovery as Simulation-Based Optimization (SBO) with a true closed feedback loop. Open-loop methods iterate blindly: generate β†’ filter/select β†’ retrain on elites. --> No direct signal pushes the policy toward rarer, higher-reward solutions.

30.01.2026 20:27 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

4.2/ Results:
β€’ Sphere packing 12d: Discover denser configurations than those produced by classical heuristics!
β€’ Sphere packing 3d: Match or exceed the best previously reported packing fractions.

30.01.2026 20:27 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

4.1/ Results:
β€’ Circle packing: New records n=26 & 32, surpassing AlphaEvolve!
β€’ Heilbronn Problem: Improve the minimum triangle area over the training dataset --> OOD sampling.

30.01.2026 20:27 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

3.3/ Our key innovations:
β€’ Trust-region Self-Distillation: Prevents collapse while enabling extrapolative OOD discovery.
Result: Convergence in ~10 updates vs 100–1000 in open-loop systems.

30.01.2026 20:27 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

3.2/ Our key innovations:
β€’ Closed-loop Reward-guided Fine-tuning: Online reward weighting + action exploration directly optimizes the flow toward rare high-score samples.

30.01.2026 20:27 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

3.1/ Our key innovations:
β€’ Geometry-Aware Sampling (GAS): Interleaves ODE integration with geometric constraint projections for feasible zero-shot samples on-the-fly.

30.01.2026 20:27 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

2/ Extremal combinatorial geometry problems are rugged continuous landscapes, perfect for modern generative models with strong inductive biases. We ditch LLM code evolution for direct continuous generation + closed-loop optimization. We call the new paradigm, "de novo Mathematical Structure Design."

30.01.2026 20:27 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

1/ Are frontier LLMs the only path to AI math breakthroughs? I think not!
We introduce FlowBoost, A lightweight RL+Flow-Matching framework that discovers new extremal geometric structures, beating AlphaEvolve with 100–1000Γ— less compute with zero-shot geometry-aware & reward-guided generation. πŸš€

30.01.2026 20:27 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

Soon I will write a thread on our new work. Stay tuned!
#AI4Math

30.01.2026 18:41 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Tnx Kyle 🀜

19.09.2025 05:41 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

We got accepted at #NeurIPS2025. I am very happy that I could merge my knowledge of Mathematics with AI to create sth new and useful for the community. ☺️

The paper: arxiv.org/abs/2505.17190
The code: github.com/Baran-phys/T...

19.09.2025 04:21 β€” πŸ‘ 21    πŸ” 7    πŸ’¬ 1    πŸ“Œ 1

Current AI research vibes:
- Let’s use LLM to do a baby science/math, after it doesn’t work, headline: LLM is bad at the baby math task β€”> guaranteed virality πŸ˜’
- Meanwhile, you develope a novel (non-LLM) method to solve this issue, report success on a deep math problem
β€”> naa, not enough drama🀦🏻

08.09.2025 09:16 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Another new result from the #NeurIPS rebuttal/discussion phase, our Tropical Transformer achieves much better length OOD performance across all algorithmic tasks, while being 3x-9x faster at inference and using 20% fewer parameters than the Universal Transformer (UT) models.

04.08.2025 20:47 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Tropical Attention: Neural Algorithmic Reasoning for Combinatorial Algorithms Dynamic programming (DP) algorithms for combinatorial optimization problems work with taking maximization, minimization, and classical addition in their recursion algorithms. The associated value functions correspond to convex polyhedra in the max plus semiring. Existing Neural Algorithmic Reasoning models, however, rely on softmax-normalized dot-product attention where the smooth exponential weighting blurs these sharp polyhedral structures and collapses when evaluated on out-of-distribution (OOD) settings. We introduce Tropical attention, a novel attention function that operates natively in the max-plus semiring of tropical geometry. We prove that Tropical attention can approximate tropical circuits of DP-type combinatorial algorithms. We then propose that using Tropical transformers enhances empirical OOD performance in both length generalization and value generalization, on algorithmic reasoning tasks, surpassing softmax baselines while remaining stable under adversarial attacks. We also present adversarial-attack generalization as a third axis for Neural Algorithmic Reasoning benchmarking. Our results demonstrate that Tropical attention restores the sharp, scale-invariant reasoning absent from softmax.

During #NeurIPS rebuttal, we have evaluated🌴Tropical Transformer on the Long Range Arena (LRA), achieving highly competitive results, placing 2ndπŸ₯ˆ overall in average accuracy.
Check out our paper: arxiv.org/abs/2505.17190
Our code: github.com/Baran-phys/T...

01.08.2025 20:05 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 1

Cool. Will definitely do πŸ‘

27.05.2025 05:18 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Interesting. I was not aware of aware if the challenges in the video subfield. But that makes sense given the context. We will definitely explore those benchmarks in the future. Thanks for the suggestions.

27.05.2025 05:10 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Tnx. We did not test yet on any other benshmarks. You mean algorithmic or language type benchmarks?

27.05.2025 04:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Interesting. I was not aware of this study. However, we did not just used tropical operations, we tried to simulate a concrete tropical circuit and do the message passing in the tropical space with the Generalized Hilbert metric as the kernel.

27.05.2025 04:54 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

7/ Our message ✍️
Better reasoning might come not from bigger models, but from choosing the right algebra/geometry 🌴.
@petar-v.bsky.social @jalonso.bsky.social
#TropicalGeometry #NeuralAlgorithmicReasoning #AI4Math

26.05.2025 13:08 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0
Post image

6/ We also show that each Tropical attention head can function as a tropical gate in a tropical circuit, simulating any max-plus circuit.

26.05.2025 13:08 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

5/ We benchmarked on 11 canonical combinatorial tasks. Tropical attention beat vanilla & adaptive softmax attention on all three OOD axes, Length, value and Adversarial attack generalization:

26.05.2025 13:08 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

4/ Tropical Attention runs each head natively in max-plus. Result:
Strong OOD length generalization with sharp attention maps even in several algorithmic tasks, including the notorious Quickselect algorithm (Another settlement for the challenge identified by @mgalkin.bsky.social )

26.05.2025 13:08 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Image by Cowdery and Challas, featured in June 2009 Mathematics Magazine

Image by Cowdery and Challas, featured in June 2009 Mathematics Magazine

3/ In the Tropical (max + ) geometry, β€œaddition” is max, β€œmultiplication” is +. Many algorithms already live here, carving exact polyhedral decision boundaries --> so why force them through exponential probabilities?
Let's ditch softmax, embrace the tropical semiring 🀯🍹.

26.05.2025 13:08 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Tropical Attention: Neural Algorithmic Reasoning for Combinatorial Algorithms Dynamic programming (DP) algorithms for combinatorial optimization problems work with taking maximization, minimization, and classical addition in their recursion algorithms. The associated value func...

2/ We introduce Tropical Attention -- the first Neural Algorithmic reasoner that operates in the Tropical semiring, achieving SOTA OOD performance on executing several combinatorial algorithms
arxiv.org/abs/2505.17190

26.05.2025 13:08 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

🧡 Tropical Attention --> Softmax is out, Tropical max-plus is in 🦾
1/ πŸ”₯Ever experinced softmax attention fade as sequences grow?
That blur is why many attention mechanisms stumble on algorithmic and reasoning tasks. Well, we have a Algebraic Geometric Tropical solution 🌴

26.05.2025 13:08 β€” πŸ‘ 10    πŸ” 4    πŸ’¬ 1    πŸ“Œ 1
Post image

I'm speaking about AI for enumerative geometry at the CMSA New Technologies in Mathematics seminar, on Wednesday.

07.04.2025 18:40 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0