In any case, the work is featuring at an interesting-looking workshop this weekend, put on by @katherinelee.bsky.social, @vaidehipatil.bsky.social, and others. More info here: mugenworkshop.github.io
15.07.2025 13:27 β π 2 π 1 π¬ 0 π 0@vaidehipatil.bsky.social
Ph.D. Student at UNC NLP | Prev: Apple, Amazon, Adobe (Intern) vaidehi99.github.io | Undergrad @IITBombay
In any case, the work is featuring at an interesting-looking workshop this weekend, put on by @katherinelee.bsky.social, @vaidehipatil.bsky.social, and others. More info here: mugenworkshop.github.io
15.07.2025 13:27 β π 2 π 1 π¬ 0 π 0UT Austin campus
Extremely excited to announce that I will be joining 
@utaustin.bsky.social Computer Science in August 2025 as an Assistant Professor! π
Thanks to my amazing collaborators Yi-Lin Sung , @peterbhase.bsky.social , Jie Peng, Tianlong Chen , @mohitbansal.bsky.social for a wonderful collaboration!
07.05.2025 18:54 β π 0 π 0 π¬ 0 π 0π Check it out here!
π Paper: arxiv.org/abs/2505.01456
 π» Code and Dataset: github.com/Vaidehi99/Un...
huggingface.co/datasets/vai...
 π€ HuggingFace: huggingface.co/papers/2505....
Key Findings
π₯ Multimodal attacks are the most effective 
π‘οΈ Our strongest defense is deleting info from hidden states
π Larger models are more robust to extraction attacks post-editing compared to smaller ones
π― UnLOK-VQA enables targeted evaluations of unlearning defenses
βοΈ Benchmarking Multimodal Unlearning Defenses
Multimodal data opens up new attack vectors.
 We benchmark 6 unlearning defenses against 7 attack strategies, including:
β
White-box attacks
β
Black-box paraphrased multimodal prompts
This enables two key types of evaluation:
β
Generalization Evaluation
βοΈRephrased questions
βοΈRephrased images
β
Specificity Evaluation
βοΈNeighboring questions (same image, new question)
βοΈNeighboring images (same concept, different image)
π¦ What Is UnLOK-VQA?
UnLOK-VQA focuses on unlearning pretrained knowledge and builds on OK-VQA, a visual QA dataset. We extend it w/ an automated question-answer generation and image generation pipeline:
β
Forget samples from OK-VQA
β
New samples at varying levels of proximity (easy, medium, hard)
This is essential for:
 π Legal compliance (e.g., GDPR, CCPA, the right to be forgotten)
 π Multimodal Privacy (e.g., faces, locations, license plates)
 π· Trust in real-world image-grounded systems
π Why Does Multimodal Unlearning Matter?
Existing unlearning benchmarks focus only on text.
But multimodal LLMs are trained on web-scale dataβimages + captionsβmaking them highly vulnerable to leakage of sensitive or unwanted content.
Unlearning must hold across modalities, not just in language.
We study:
β How effectively can we erase multimodal knowledge?
β How should we measure forgetting in multimodal settings?
β
We benchmark 6 unlearning defenses against 7 whitebox and blackbox attack strategies
π¨ Introducing our @tmlrorg.bsky.social paper βUnlearning Sensitive Information in Multimodal LLMs: Benchmark and Attack-Defense Evaluationβ
We present UnLOK-VQA, a benchmark to evaluate unlearning in vision-and-language models, where both images and text may encode sensitive or private information.
In Singapore for #ICLR2025 this week to present papers + keynotes π, and looking forward to seeing everyone -- happy to chat about research, or faculty+postdoc+phd positions, or simply hanging out (feel free to ping)! π
 
Also meet our awesome students/postdocs/collaborators presenting their work.
Come chat about unlearning with us!!
02.04.2025 16:57 β π 5 π 1 π¬ 0 π 0Call for PC Members!
Weβre looking for program committee members!
π Submit your Expression of Interest here: forms.gle/ZPEHeymJ4t5N...
#ICML2025
π©βπ» Organizers:
Mantas Mazeika, Yang Liu, @katherinelee.bsky.social, @mohitbansal.bsky.social, Bo Li and myself (@vaidehipatil.bsky.social) π
π₯ Speakers & Panelists:
We're lucky to have an incredible lineup of speakers and panelists covering diverse topics in our workshop:
 Nicholas Carlini, Ling Liu, Shagufta Mehnaz, @peterbhase.bsky.social , Eleni Triantafillou, Sijia Liu, @afedercooper.bsky.social, Amy Cyphert
We invite contributions exploring key challenges and advancements at the intersection of machine unlearning and generative AI!
π Full details & updates: mugenworkshop.github.io
π
 Key Dates:
 π Submission Deadline: May 19
 β
 Acceptance Notifications: June 9
 π€ Workshop Date: July 18 or 19
π¨Exciting @icmlconf.bsky.social workshop alert π¨
Weβre thrilled to announce the #ICML2025 Workshop on Machine Unlearning for Generative AI (MUGen)! 
β‘Join us in Vancouver this July to dive into cutting-edge research on unlearning in generative AI with top speakers and panelists! β‘
 π₯³π₯³ Honored and grateful to be awarded the 2025 Apple Scholars in AI/ML PhD Fellowship! β¨
Huge shoutout to my advisor @mohitbansal.bsky.social, & many thanks to my lab mates @unccs.bsky.social , past collaborators + internship advisors for their support βΊοΈπ
machinelearning.apple.com/updates/appl...
π¨UPCORE is our new method for balancing unlearning/forgetting with maintaining model performance.
Best part is it works by selecting a coreset from the data rather than changing the model, so it is compatible with any unlearning method, with consistent gains for 3 methods + 2 tasks!
Huge thanks to my co-authors 
@esteng.bsky.social , and @mohitbansal.bsky.social for a great collaboration!
π Check it out here:
π Paper: arxiv.org/abs/2502.15082
π» Code: github.com/Vaidehi99/UP...
π€ @huggingface page: huggingface.co/papers/2502....
UPCORE consistently outperforms baselines across all methods:
βοΈ Less unintended degradation
βοΈ Deletion transferred to pruned points
UPCORE provides a practical, method-agnostic approach that improves the reliability of unlearning techniques.
Instead of evaluating at a single training checkpoint, we introduce AUC (Area Under the Curve) across deletion effectiveness and utility.
This provides a complete picture of the trade-off between forgetting and knowledge retention over the unlearning trajectory.
We apply UPCORE across three unlearning methods:
π Gradient Ascent
π« Refusal
π Negative Preference Optimization (NPO)
We measure:
βοΈ Deletion effectiveness β How well the target is removed
βοΈ Unintended degradation β Impact on other abilities
βοΈ Positive transfer β How well unlearning generalizes
Even after pruning, the pruned points in the forget set still become unlearned -- thanks to positive collateral transfer from the core forget set.
Thus, UPCORE reduces negative collateral effects while maintaining effective deletion.
UPCORE constructs a core forget set by identifying and removing outlier points using Isolation Forest.
β
 Minimizes unintended degradation
β
 Preserves model utility
β
 Compatible with multiple unlearning methods
Our key insight: Not all forget set points degrade the model equally.
Points contributing to high variance cause more collateral damage when unlearned.
By pruning these outliers, UPCORE reduces unintended forgetting while ensuring effective deletion.
LLMs train on vast datasets, often with sensitive or unwanted info. Regulations like GDPR, CCPA mandate removal.
Yet, standard unlearning can degrade unrelated knowledge, making it unreliable.
Effective unlearning is key for:
π Compliance (GDPR, CCPA)
π Privacy & security
βοΈ Ethical AI development
π¨ Introducing UPCORE, to balance deleting info from LLMs with keeping their other capabilities intact.
UPCORE selects a coreset of forget data, leading to a better trade-off across 2 datasets and 3 unlearning methods.
π§΅π