Our approach shows strong generalization and versatility in generating accurate prompts for objects, styles and images across multiple T2I models, including Stable Diffusion, DALL-E, and Midjourney. It also enables easy editing and multi-concept prompt generation.
28.04.2025 22:52 β π 0 π 0 π¬ 0 π 0
Prompt engineering for personalized image generation is labor-intensive, requires model-specific tuning, limiting generalization.
PRISM uses VLMs and iterative in-context learning to automatically generate effective, human-readable prompts using only black-box access to image generation models.
28.04.2025 22:52 β π 0 π 0 π¬ 1 π 0
New work on automated prompt engineering for personalized text-to-image generation:
PRISM: Automated Black-box Prompt Engineering for Personalized Text-to-Image Generation
Paper + Code: kellyyutonghe.github.io/prism/
28.04.2025 22:50 β π 0 π 0 π¬ 1 π 0
Llama
The open-source AI models you can fine-tune, distill and deploy anywhere. Choose from our collection of models: Llama 4 Maverick and Llama 4 Scout.
www.llama.com
Llama4 models are out! Open sourced! Check them out:
βNative multimodality, mixture-of-experts models, super long context windows, step changes in performance, and unparalleled efficiency. All in easy-to-deploy sizes custom fit for how you want to use itβ
05.04.2025 19:18 β π 6 π 1 π¬ 0 π 0
With small perturbations (less than 5% of total web page pixels), attackers can execute targeted adversarial goals with up to 67% success rates.
We also find that inference-time compute that is often used to improve model performance can introduce new vulnerabilities and harm robustness.
19.02.2025 22:18 β π 0 π 0 π¬ 0 π 0
Dissecting Adversarial Robustness of Multimodal LM Agents
Dissecting Adversarial Robustness of Multimodal LM Agents
New work #ICLR2025 on βDissecting Adversarial Robustness of Multimodal LM Agentsβ that shows that one can successfully break latest agents that use black-box frontier LLMs, including agents that perform reflection and tree search.
Paper + Code + Data: chenwu.io/attack-agent/
19.02.2025 22:16 β π 1 π 0 π¬ 1 π 0
GenAI Summit 2025
#GenAIUCSD25
Excited to be at the GenAI Summit at UCSD!
I'll be sharing our latest work on VisualWebArena, inference-time tree search, and Internet-scale training of LLM Agents.
genaisummit2025.ucsd.edu
19.02.2025 17:19 β π 0 π 0 π¬ 0 π 0
4/4 Llama 3.1 70B agents successfully complete 16.7% of tasks on 150k websites. Agents trained on human-annotated data from Mind2Web and WebLINX struggle to generalize to real-world websites. Adding synthetic data significantly improves generalization.
With B Trabucco, G Sigurdsson, R Piramuthu
12.02.2025 02:22 β π 0 π 0 π¬ 0 π 0
3/4 Language models perform competitively with human annotators, achieving:
- 97% accuracy in detecting and filtering harmful content
- 89% success rate in generating feasible tasks
- 82% accuracy in judging successful task completions
12.02.2025 02:21 β π 0 π 0 π¬ 1 π 0
2/4 The pipeline follows a three-step process:
- LLM generates tasks for 150k websites
- LLM agents complete these tasks and produce trajectories
- LLM reviews the trajectories and evaluates their success
12.02.2025 02:21 β π 0 π 0 π¬ 1 π 0
1/4 New Work on InSTA: A pipeline for Internet-scale training of web agents across 150k diverse websites without human annotations.
Paper + Code: data-for-agents.github.io
Environment: github.com/data-for-age...
12.02.2025 02:21 β π 2 π 0 π¬ 1 π 0
3/3 Joint work with Tiffani Min, Yue Wu, Jimin Sun, Max Kaufmann, Fahim Tajwar, Yonatan Bisk
10.02.2025 22:30 β π 0 π 0 π¬ 0 π 0
2/3 Offline-collected state transitions are evaluated using PRMs to determine optimal intervention timing, creating labeled trajectories for training the helper model.
This minimizes costly intervention calls during training while leveraging PRMs to enhance robustness to off-policy data.
10.02.2025 22:29 β π 0 π 0 π¬ 1 π 0
1/3 New work on Self-Regulation and Requesting Interventions: Enabling agents with a limited intervention budget to decide when to seek help:
Paper: soyeonm.github.io/self_reg/
We develop an offline framework that trains a helper policy to request interventions by combining LLM-based PRMs with RL
10.02.2025 22:28 β π 4 π 0 π¬ 1 π 0
π² Ruslan Salakhutdinov (@rsalakhu.bsky.social) from CMU (@scsatcmu.bsky.social) opened the workshop with a talk on Tree Search for Language Model Agents.
Timestamp 36:20 in neurips.cc/virtual/2024...
π arxiv.org/abs/2407.01476
#NeurIPS2024 #AdaptiveFoundationModels
19.12.2024 04:59 β π 1 π 1 π¬ 1 π 0
π Had fun at #NeurIPS2024 Workshop on #AdaptiveFoundationModels!
π Speakers: @rsalakhu.bsky.social @sedielem.bsky.social Kate Saenko, Matthias Bethge / @vishaalurao.bsky.social Minjoon Seo, Bing Liu, Tianqi Chen
πPosters: adaptive-foundation-models.org/papers
π¬ neurips.cc/virtual/2024...
π§΅Recap!
19.12.2024 04:59 β π 10 π 2 π¬ 1 π 0
With my amazing students and collaborators at @neuripsconf.bsky.social in Vancouver!
15.12.2024 17:05 β π 0 π 0 π¬ 0 π 0
Carnegie Mellon University at NeurIPS 2024
Carnegie Mellon University is proud to present 194 papers at the 38th conference on Neural Information Processing Systems (NeurIPS 2024), held from December 10-15 at the Vancouver Convention Center. H...
Carnegie Mellon University at NeurIPS 2024 β Machine Learning Blog | ML@CMU | Carnegie Mellon University
Carnegie Mellon University is proud to present 194 papers at the 38th conference on Neural Information Processing Systems (NeurIPS 2024)
blog.ml.cmu.edu/2024/12/02/c...
03.12.2024 15:34 β π 2 π 0 π¬ 0 π 0
2/2 Our findings show that even when unlearning a single fact, current methods either fail to properly unlearn with high recall or end up unlearning many other irrelevant facts.
Paper: arxiv.org/abs/2410.15153
Code+Dataset: github.com/wrh14/deep_u...
joint work R Wu, C Yadav, K Chaudhuri.
03.12.2024 14:43 β π 0 π 0 π¬ 0 π 0
Evaluating Deep Unlearning in Large Language Models
Machine unlearning is a key requirement of many data protection regulations such as GDPR. Prior work on unlearning has mostly considered superficial unlearning tasks where a single or a few related pi...
1/2 New work on Evaluating Deep Unlearning in Large Language Models.
Paper: arxiv.org/abs/2410.15153
Unlearning specific facts in LLMs is challenging because the facts in LLMs can be deduced from each other. This work proposes a framework for deep unlearning of facts that are interrelated.
03.12.2024 14:42 β π 0 π 0 π¬ 1 π 0
What is happening?! Who is this? π
24.11.2024 00:22 β π 0 π 0 π¬ 0 π 0
Hello BlueSky
24.11.2024 00:21 β π 2 π 0 π¬ 1 π 0
Research Scientist @ Apple
Work: Modeling intelligence, generalization, representation learning.
Play: Music, games, outdoors
Machine Learning Professor
https://cims.nyu.edu/~andrewgw
AI professor. Director, Foundations of Cooperative AI Lab at Carnegie Mellon. Head of Technical AI Engagement, Institute for Ethics in AI (Oxford). Author, "Moral AI - And How We Get There."
https://www.cs.cmu.edu/~conitzer/
Director of Artificial Intelligence at https://helsing.ai. Fellow @ellis.eu. Ex Principal Applied Scientist at AWS. Living in absurdism.
http://www0.cs.ucl.ac.uk/staff/c.archambeau/
Assistant Professor of Machine Learning, Carnegie Mellon University (CMU)
Building a Natural Science of Intelligence π§ π€β¨
Prev: ICoN Postdoctoral Fellow @MIT, PhD @Stanford NeuroAILab
Personal Website: https://cs.cmu.edu/~anayebi
Interpretable Deep Networks. http://baulab.info/ @davidbau
official Bluesky account (check usernameπ)
Bugs, feature requests, feedback: support@bsky.app