Bernardo Esteves's Avatar

Bernardo Esteves

@esteveste.bsky.social

PhD at Técnico, University of Lisbon | Deep Learning + RL bernardoesteves.com

69 Followers  |  52 Following  |  15 Posts  |  Joined: 23.11.2024  |  1.813

Latest posts by esteveste.bsky.social on Bluesky

Preview
NeuralSolver: Learning Algorithms For Consistent and Efficient... We contribute NeuralSolver, a novel recurrent solver that can efficiently and consistently extrapolate, i.e., learn algorithms from smaller problems (in terms of observation size) and execute those...

Check out our paper: openreview.net/forum?id=IxR...
Project website: bernardoesteves.com/NeuralSolver

26.11.2024 14:55 — 👍 3    🔁 0    💬 0    📌 0
Post image

This method can also be applied to sequential decision problems, where small mistakes lead to catastrophic results. NeuralSolver can solve the doorkey environment of size 128x128 with a performance very close to the oracle, just by performing behavior cloning on smaller sizes.

26.11.2024 14:55 — 👍 3    🔁 0    💬 1    📌 0
Post image

With this approach, NeuralSolver achieves almost perfect algorithmic extrapolation by executing the same learned algorithm on much harder problems. While highly surpassing previous works.

26.11.2024 14:55 — 👍 2    🔁 0    💬 1    📌 0
Post image

We can notice this by looking at how the latent values converge to the final result and the number of recurrent iterations. NeuralSolver learns an algorithm that starts by solving the dead ends of the maze, until finding the correct path that goes from the player to the goal.

26.11.2024 14:37 — 👍 1    🔁 0    💬 1    📌 0

By keeping the recurrent module output size equal to the input size, the model learns small local algorithms that are executed in parallel and work with any input size.

26.11.2024 14:37 — 👍 0    🔁 0    💬 1    📌 0
Post image

We do this by leveraging a recurrent model with an optional pooling layer for different-size tasks.
The recurrent model keeps the input size constant at each iteration.
The pooling layer is then used to collapse the information from the latent state to the desired output size.

26.11.2024 14:37 — 👍 1    🔁 0    💬 1    📌 0

We developed NeuralSolver, a method capable of learning algorithms that solve much more complex tasks than the ones used for training. That works on tasks that have the same or different output sizes!

26.11.2024 14:37 — 👍 1    🔁 0    💬 1    📌 0
Preview
NeuralSolver: Learning Algorithms For Consistent and Efficient... We contribute NeuralSolver, a novel recurrent solver that can efficiently and consistently extrapolate, i.e., learn algorithms from smaller problems (in terms of observation size) and execute those...

Check out our paper: openreview.net/forum?id=IxR...
Project website: bernardoesteves.com/NeuralSolver

26.11.2024 14:30 — 👍 0    🔁 0    💬 0    📌 0
Post image

This method can also be applied to sequential decision problems, where a small mistake in the sequence can lead to catastrophic results. NeuralSolver can solve the doorkey environment of size 128x128 with a performance very close to the oracle, just by performing behavior cloning on smaller sizes.

26.11.2024 14:22 — 👍 0    🔁 0    💬 1    📌 0
Post image

With this approach, NeuralSolver achieves almost perfect algorithmic extrapolation by executing the same learned algorithm on much harder problems. While highly surpassing previous works.

26.11.2024 14:22 — 👍 0    🔁 0    💬 1    📌 0
Post image

We can notice this by looking at how the recurrent latent values converge to the final result, along a certain number of recurrent iterations. NeuralSolver learns an algorithm that starts by solving the dead ends of the maze, until finding the correct path that goes from the player to the goal.

26.11.2024 14:22 — 👍 0    🔁 0    💬 1    📌 0

By keeping the recurrent module output size equal to the input size, the model learns small local algorithms that are executed in parallel and work with any input size.

26.11.2024 14:22 — 👍 0    🔁 0    💬 1    📌 0
Post image

We are able to do this by leveraging a recurrent model with an optional pooling layer for different-size tasks.
The recurrent model keeps the input size constant at each iteration.
The pooling layer is then used to collapse the information from the latent state to the desired output size.

26.11.2024 14:22 — 👍 0    🔁 0    💬 1    📌 0
Post image

We explore problems that increase in complexity mainly through size. Our goal is to be able to train an agent on a small/easy set of tasks and then without any more training to be able to solve more complex tasks.

26.11.2024 14:22 — 👍 7    🔁 1    💬 1    📌 0
Post image

🚀 Excited to present NeuralSolver at NeurIPS 2024!

Humans can easily learn algorithms to solve much more complex tasks, from simple tasks. Machine learning methods usually fail to do this.
Can we build models that have similar algorithmic extrapolation abilities?

Thread 🧵:

26.11.2024 14:22 — 👍 50    🔁 8    💬 1    📌 2

@esteveste is following 20 prominent accounts