Xiaoxuan

Xiaoxuan

@xiaoxuanlei.bsky.social

CompNeuro & AI ❤️ FutureToBeBlackBoxBreaker👻 PhD candidate @McGill et @Mila

96 Followers 65 Following 11 Posts Joined Nov 2024
1 year ago

📌 Poster Session:
⏰ When: TODAY, Thu, Dec 12, 4:30 p.m. – 7:30 p.m. PST
📍 Where: East Exhibit Hall A-C, #3705
📄 What: Geometry of Naturalistic Object Representations in Recurrent Neural Network Models of Working Memory

Hope to see you there!
@bashivan.bsky.social @takuito.bsky.social

3 1 0 1
1 year ago

Excited to be at #NeurIPS2024 in #Vancouver! Our poster session is TODAY—if you're interested in naturalistic representations in dynamic working memory models, please drop by and let’s chat!

1 0 1 0
1 year ago
Preview
Geometry of naturalistic object representations in recurrent neural network models of working memory Working memory is a central cognitive ability crucial for intelligent decision-making. Recent experimental and computational work studying working memory has primarily used categorical (i.e., one-hot)...

👉 Check it out: arxiv.org/abs/2411.02685
📅 We’ll be at NeurIPS! Join us for our poster presentation on Thu 12 Dec, 7:30 p.m. EST — 10:30 p.m. EST.

#AI #CognitiveScience #WorkingMemory #DeepLearning #RepresentationGeometry #MultiTask

1 0 2 0
1 year ago

Our findings bridge cognitive science & AI, revealing how high-dimensional object information is encoded, retained, and recalled in recurrent models of working memory.

1 0 1 0
1 year ago
Post image

🎯 With training, RNNs implemented chronological memory subspaces allowing them to track object information using rotational dynamics—supporting resource-based models of working memory.

0 0 1 0
1 year ago
Post image

📐 Surprisingly, object features are less orthogonalized in RNN representations compared to perceptual space.

0 0 1 0
1 year ago
Post image

🧠 We found that multi-task RNNs (unlike single-task ones) retain both task-relevant & irrelevant info but reusable representations only emerged in simple gateless architectures.

0 0 1 0
1 year ago
Post image

🖥️ To answer this question, we trained multi-task RNNs (vanilla, GRU, LSTM) on 9 N-back tasks using naturalistic 3D object stimuli to study encoding, retention, & retrieval dynamics.

1 0 1 0
1 year ago

It’s unclear how high-dimensional naturalistic sensory information is encoded, retained and recalled in these models to accommodate various task demands.

0 0 1 0
1 year ago

Previous models of working memory have mainly focused on using abstract stimuli (Mante et al., 2013, Yang et al., 2019, Driscoll et al., 2024, Fascianelli et al., 2024, Piwek & Stokes, 2023 etc)

0 0 1 0
1 year ago
Post image

🌟 New Research Alert! 🌟
Excited to share our latest work (accepted to NeurIPS2024) on understanding working memory in multi-task RNN models using naturalistic stimuli!: with @takuito.bsky.social and @bashivan.bsky.social
#tweeprint below:

50 17 3 2