Topology-Aware Knowledge Propagation in Decentralized Learning
Topology-Aware Knowledge Propagation in Decentralized Learning
β¨ TL;DR: Making decentralized learning aware of network topology boosts performance & resilienceβvital for federated learning, edge AI, and IoT. Check out the paper for technical details & results! (6/6) mansisak.com/topology_awa...
19.06.2025 23:19 β π 1 π 0 π¬ 0 π 0
π Experiments show significant improvements: faster knowledge propagation compared to traditional, topology-agnostic approaches in various real-world network settings. (5/6)
19.06.2025 23:19 β π 1 π 0 π¬ 1 π 0
π§ The system prioritizes information from well-connected or "influential" nodes, ensuring that high-quality knowledge quickly reaches more isolated or less-informed devices. No one gets left behind! (4/6)
19.06.2025 23:19 β π 1 π 0 π¬ 1 π 0
π This work introduces topology-aware knowledge propagation, which tailors how models and information are shared based on each deviceβs place in the network, leading to more effective learning overall. (3/6)
19.06.2025 23:19 β π 1 π 0 π¬ 1 π 0
π Decentralized learning often involves many devices (like edge or mobile) collaborating without a central server. However, the way these devices connectβthe network topologyβcan heavily impact learning quality. (2/6)
19.06.2025 23:19 β π 0 π 0 π¬ 1 π 0
π€ New advances in decentralized learning! "Topology-Aware Knowledge Propagation in Decentralized Learning" proposes a novel way to improve how information flows in distributed machine learning systems. Letβs break it down! π§΅ (1/6) (arxiv.org/abs/2505.11760)
19.06.2025 23:19 β π 2 π 0 π¬ 1 π 1
Mansi, Aswathy, and Nathaniel are presenting their poster.
Mansi presenting
Mansi, Aswathy, and Nathaniel are standing infront of a nice building
@mansisakarvadia.bsky.social, Aswathy (PhD students), and @nathaniel-hudson.bsky.social (postdoc) presented their work on identifying and ablating memorization in #LLMs at the 2024 MSLD workshop! π Their research is also accepted to ICLR 2025 β check it out: mansisak.com/memorization/
29.04.2025 03:31 β π 3 π 2 π¬ 0 π 0
7/ π Special Thanks:
A huge shoutout to my incredible co-authors from multiple institutions for their contributions to this work:
Aswathy Ajith, Arham Khan, @nathaniel-hudson.bsky.social , @calebgeniesse.bsky.social, Yaoqing Yang, @kylechard.bsky.social , @ianfoster42.bsky.social, Michael Mahoney
04.03.2025 18:15 β π 2 π 0 π¬ 0 π 0
6/ π Scalable Impact:
Our methods arenβt just for small models! We show that they scale effectively to larger LMs, providing robust memorization mitigation without compromising performance across different sizes of models. Exciting progress for real-world applications!
04.03.2025 18:15 β π 1 π 0 π¬ 1 π 0
5/π‘Best Approach:
Our proposed unlearning method, BalancedSubnet, outperforms others by effectively removing memorized info while maintaining high accuracy.
04.03.2025 18:15 β π 1 π 0 π¬ 1 π 0
4/π§ͺ Key Findings:
Unlearning-based methods are faster and more effective than regularization or fine-tuning in mitigating memorization.
04.03.2025 18:15 β π 1 π 0 π¬ 1 π 0
3/β‘Introducing TinyMem:
We created TinyMem, a suite of small, efficient models designed to help test and benchmark memorization mitigation techniques. TinyMem allows for quick experiments with lower computational costs.
04.03.2025 18:15 β π 1 π 0 π¬ 1 π 0
2/ π¨ Main Methods:
We test 17 methodsβregularization, fine-tuning, and unlearningβ5 of which we propose. These methods aim to remove memorized info from LMs while preserving performance.
04.03.2025 18:15 β π 1 π 0 π¬ 1 π 0
1/π§΅ICLR 2025 Spotlight Research on LM & Memorization!
Language models (LMs) often "memorize" data, leading to privacy risks. This paper explores ways to reduce that!
Paper: arxiv.org/pdf/2410.02159
Code: github.com/msakarvadia/...
Blog: mansisak.com/memorization/
04.03.2025 18:15 β π 4 π 2 π¬ 1 π 1
How can we stop LLMs from memorizing information? @mansisakarvadia.bsky.social's paper, which develops new methods for unlearning memorized information and introduces a suite of small LMs for testing these methods, was just accepted to ICLR '25!
arxiv.org/pdf/2410.02159
28.01.2025 02:14 β π 2 π 1 π¬ 1 π 0
A Year In Review: C.S. PhD Student Edition | Mansi Sakarvadia
A recap of my academic journey in 2024.
Reflecting on my 2024 PhD journey: passed my qualifying exam, spent the summer at Berkeley, mentored undergrad students, and tackled the fast pace of AI/ML research. Itβs been a year of milestones and growth! Read more here: mansisak.com/blog/2025/ye... #PhDJourney #AIResearch
05.01.2025 03:13 β π 9 π 2 π¬ 0 π 0
Computer Science PhD Student at the University of Chicago. Genome-scale language models. AI steered molecular dynamics. AI4Science.
First Workshop on Large Language Model Memorization.
Visit our website at https://sites.google.com/view/memorization-workshop/
π¨π½βπ» Ph.D. Computer Scientist
π« Assistant Professor at Illinois Tech
π https://nathaniel-hudson.github.io/
Assistant Professor of Linguistics, and Harrington Fellow at UT Austin. Works on computational understanding of language, concepts, and generalization.
πΈοΈποΈ: https://kanishka.website
Deep computing for deep science. UChicago & Argonne Nat Lab. Work with amazing people. Globus. AI, HPC, robotics for discovery. And sailing. And books. And music. And beauty.
Globus Labs is a research group led by @ianfoster42.bsky.social and @kylechard.bsky.social that aims to realize a world in which all research data are reliably, rapidly, and securely accessible, discoverable, and usable.