Bao Pham's Avatar

Bao Pham

@baopham.bsky.social

PhD Student at RPI. Interested in Hopfield or Associative Memory models and Energy-based models.

26 Followers  |  17 Following  |  4 Posts  |  Joined: 05.12.2024
Posts Following

Posts by Bao Pham (@baopham.bsky.social)

Post image

I am excited to announce the call for papers for the New Frontiers in Associative Memories workshop at ICLR 2025. New architectures and algorithms, memory-augmented LLMs, energy-based models, Hopfield nets, AM and diffusion, and many other topics.

Website: nfam.vizhub.ai

@iclr-conf.bsky.social

14.01.2025 16:56 — 👍 16    🔁 6    💬 0    📌 0

Most of the work on Dense Associative Memory (DenseAM) thus far has focused on the regime when the amount of data (number of memories) is below the critical memory storage capacity. We are beginning to explore the opposite limit, when the data is large.

05.12.2024 18:19 — 👍 16    🔁 3    💬 1    📌 0
On the left figure, it showcases the behavior of Hopfield models. Given a query (the initial point of energy descent), a Hopfield model will retrieve the closest memory (local minimum) to that query such that it minimizes the energy function. A perfect Hopfield model is able to store patterns in distinct minima (or buckets). In contrast, the right figure illustrates a bad Associative Memory system, where stored patterns share a distinctive bucket. This enables the creation of spurious patterns, which appear like mixture of stored patterns. Spurious patterns will have lower energy than the memories due to this overlapping.

On the left figure, it showcases the behavior of Hopfield models. Given a query (the initial point of energy descent), a Hopfield model will retrieve the closest memory (local minimum) to that query such that it minimizes the energy function. A perfect Hopfield model is able to store patterns in distinct minima (or buckets). In contrast, the right figure illustrates a bad Associative Memory system, where stored patterns share a distinctive bucket. This enables the creation of spurious patterns, which appear like mixture of stored patterns. Spurious patterns will have lower energy than the memories due to this overlapping.

Diffusion models create beautiful novel images, but they can also memorize samples from the training set. How does this blending of features allow creating novel patterns? Our new work in Sci4DL workshop #neurips2024 shows that diffusion models behave like Dense Associative Memory networks.

05.12.2024 17:29 — 👍 39    🔁 5    💬 1    📌 1

The work is done in collaboration with Gabriel Raya, Matteo Negri, Mohammed J. Zaki, @lucamb.bsky.social , @krotov.bsky.social

Lastly, join us at Sci4DL workshop at #NeurIPS2024 to learn more!

We will be giving an oral presentation there!

05.12.2024 17:29 — 👍 3    🔁 0    💬 0    📌 0

This work enables a positive perspective of spurious patterns. Unlike their usual perception in Associative Memory, such patterns play a role in signaling generalization in deep generative models, like diffusion models.

Here is a link to the paper: openreview.net/pdf?id=zVMMa....

05.12.2024 17:29 — 👍 2    🔁 0    💬 1    📌 0
This figure depicts some hand-selected examples of memorized, spurious, and generalized patterns in which our diffusion models generate, across various training data sizes. We show a target image along with its top-5 nearest neighbors from training set (top row) and synthetic set (bottom row). 

Memorized patterns are typically generated at low data sizes. Meanwhile, when the data size is large enough, but not substantially large enough for diffusion models to approximate the data distribution well, spurious patterns are typically generated from our models.

This figure depicts some hand-selected examples of memorized, spurious, and generalized patterns in which our diffusion models generate, across various training data sizes. We show a target image along with its top-5 nearest neighbors from training set (top row) and synthetic set (bottom row). Memorized patterns are typically generated at low data sizes. Meanwhile, when the data size is large enough, but not substantially large enough for diffusion models to approximate the data distribution well, spurious patterns are typically generated from our models.

In the low training data regime (number of memories), diffusion models memorize. As the data size increases, spurious states emerge, signaling the blending of stored features into new combinations which enables generalization. This is how such models create novel outputs in the high data regime.

05.12.2024 17:29 — 👍 3    🔁 0    💬 1    📌 0
On the left figure, it showcases the behavior of Hopfield models. Given a query (the initial point of energy descent), a Hopfield model will retrieve the closest memory (local minimum) to that query such that it minimizes the energy function. A perfect Hopfield model is able to store patterns in distinct minima (or buckets). In contrast, the right figure illustrates a bad Associative Memory system, where stored patterns share a distinctive bucket. This enables the creation of spurious patterns, which appear like mixture of stored patterns. Spurious patterns will have lower energy than the memories due to this overlapping.

On the left figure, it showcases the behavior of Hopfield models. Given a query (the initial point of energy descent), a Hopfield model will retrieve the closest memory (local minimum) to that query such that it minimizes the energy function. A perfect Hopfield model is able to store patterns in distinct minima (or buckets). In contrast, the right figure illustrates a bad Associative Memory system, where stored patterns share a distinctive bucket. This enables the creation of spurious patterns, which appear like mixture of stored patterns. Spurious patterns will have lower energy than the memories due to this overlapping.

Diffusion models create beautiful novel images, but they can also memorize samples from the training set. How does this blending of features allow creating novel patterns? Our new work in Sci4DL workshop #neurips2024 shows that diffusion models behave like Dense Associative Memory networks.

05.12.2024 17:29 — 👍 39    🔁 5    💬 1    📌 1