Aditya Vashistha's Avatar

Aditya Vashistha

@imadityav.bsky.social

Assistant Professor at Cornell. Research in HCI4D, Social Computing, Responsible AI, and Accessibility. https://www.adityavashistha.com/

1,476 Followers  |  56 Following  |  33 Posts  |  Joined: 15.11.2024  |  2.1264

Latest posts by imadityav.bsky.social on Bluesky

Aditya and Joy wearing regalia and posing in front of a bright red background!

Aditya and Joy wearing regalia and posing in front of a bright red background!

Big day today with Joy Ming graduating!! A new doctor in town! Can’t wait to see all the incredible things Joy will take on next. So proud!

23.05.2025 21:28 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Thank you to all our participants, co-organizers, student volunteers, funders, and partners who made this possible. And to Joy Ming for the beautiful visual summaries.

23.05.2025 16:00 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Our conversations spanned:
πŸ”· Meaningful use cases of AI in high-stakes global settings
πŸ”· Interdisciplinary methods across computing and humanities
πŸ”· Partnerships between academia, industry, and civil society
πŸ”· The value of local knowledge, lived experiences, and participatory design

23.05.2025 16:00 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Over three days, we explored what it means to design and govern pluralistic and humanistic AI technologies β€” ones that serve diverse communities, respect cultural contexts, and center social well-being. The summit was part of the Global AI Initiative at Cornell.

23.05.2025 16:00 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Yesterday we wrapped up the Thought Summit on LLMs and Society at Cornell β€” an energizing and deeply reflective gathering of researchers, practitioners, and policymakers from across institutions and geographies.

23.05.2025 16:00 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Thank you Dhanaraj for attending the Thought Summit and sharing your thoughts on how we can design AI for All!

23.05.2025 15:56 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

This was the week of reflection, new ideas, and a renewed sense of urgency to design AI systems that serve marginalized communities globally. Can't wait for what's next.

02.05.2025 01:05 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
ASHABot: An LLM-Powered Chatbot to Support the Informational Needs of Community Health Workers | Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems You will be notified whenever a record that you have chosen has been cited.

Pragnya Ramjee presented work (with Mohit Jain at MSR India) on deploying LLM tools for community health workers in India. In collaboration with Khushi Baby, we show how thoughtful AI design can (and cannot) bridge critical informational gaps in low-resource settings.
dl.acm.org/doi/10.1145/...

02.05.2025 01:05 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0
Preview
"Who is running it?" Towards Equitable AI Deployment in Home Care Work | Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems You will be notified whenever a record that you have chosen has been cited.

Ian RenΓ© Solano-Kamaiko presented our study on how algorithmic tools are already shaping home care workβ€”often invisibly. These systems threaten workers’ autonomy and safety, underscoring the need for stronger protections and democratic AI governance.
dl.acm.org/doi/10.1145/...

02.05.2025 01:05 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Exploring Data-Driven Advocacy in Home Health Care Work | Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems

Joy Ming presented our award-winning paper on designing advocacy tools for home care workers. In this work, we unpack tensions between individual and collective goals and highlight how to use data responsibly in frontline labor organizing.
dl.acm.org/doi/10.1145/...

02.05.2025 01:05 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
AI Suggestions Homogenize Writing Toward Western Styles and Diminish Cultural Nuances | Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems You will be notified whenever a record that you have chosen has been cited.

Dhruv presented our cross-cultural study on AI writing tools and their Western-centric biases. We found that AI suggestions disproportionately benefit American users and subtly nudge Indian users toward Western writing normsβ€”raising concerns about cultural homogenization.
dl.acm.org/doi/10.1145/...

02.05.2025 01:05 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
"Ignorance is not Bliss": Designing Personalized Moderation to Address Ableist Hate on Social Media | Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems

Sharon Heung presented our work on personalizing moderation tools to help disabled users manage ableist content online. We showed how users want control over filtering and framingβ€”while also expressing deep skepticism toward AI-based moderation.
dl.acm.org/doi/10.1145/...

02.05.2025 01:05 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Dhruv, Sharon, Aditya, Jiamin, Ian, and Joy in front of buildings and trees surrounded by a lush green landscape.

Dhruv, Sharon, Aditya, Jiamin, Ian, and Joy in front of buildings and trees surrounded by a lush green landscape.

Just wrapped up an incredible week at #CHI2025 in Yokohama with Joy Ming, @sharonheung.bsky.social, Dhruv Agarwal, and Ian RenΓ© Solano-Kamaiko! We presented several papers that push the boundaries of what Globally Equitable AI could look like in high-stakes contexts.

02.05.2025 01:05 β€” πŸ‘ 11    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Why writing with ChatGPT makes you sound like an American A new study from Cornell University suggests some large generative AI models have a Western bias that strips away cultural nuance.

www.fastcompany.com/91324551/cha...

Kudos to Dhruv Agarwal for leading this work and such fun collaboration with @informor.bsky.social!

02.05.2025 00:41 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
The Great Language Flattening Chatbots learned from human writing. Now it’s their turn to influence us.

As these tools become more common, it’s critical to ask: Whose voice is being amplifiedβ€”and whose is being erased? www.theatlantic.com/technology/a...

02.05.2025 00:41 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Excited to see our research in The Atlantic and Fast Company!

Our work, presented at #CHI2025 this week, shows how AI writing suggestions often nudge people toward Western styles, unintentionally flattening cultural expression and nuance.

02.05.2025 00:41 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Excited to be at #CHI2025 with Joy Ming, Sharon Heung, Dhruv Agarwal, and Ian Rene Solano-Kamaiko!

Our lab will be presenting several papers Globally Equitable AI β€” centering equity, culture, and inclusivity in high-stakes contexts. 🌎

If you’ll be there, would love to connect! πŸ–οΈ

26.04.2025 09:32 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Huge congratulations to Mahika Phutane for leading this work, and Ananya Seelam for her contributions!

We’re thrilled to share this at ACM FAccT 2025.

Read the full paper: lnkd.in/eCsAupvK

12.04.2025 20:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Our findings make a clear case: AI moderation systems must center disabled people’s expertise, especially when defining harm and safety.

This isn’t just a technical problemβ€”it’s about power, voice, and representation.

12.04.2025 20:57 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Disabled participants frequently described these AI explanations as β€œcondescending” or β€œdehumanizing.”

The models reflect a clinical, outsider gazeβ€”rather than lived experience or structural understanding.

12.04.2025 20:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

AI systems often underestimate ableismβ€”even in clear-cut cases of discrimination or microaggressions.

And when they do explain their decisions? The explanations are vague, euphemistic, or moralizing.

12.04.2025 20:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Methodology of our paper, starting with creating a dataset containing ableist and non-ableist post, followed by collecting and analyzing ratings and explanations from AI models and disabled and non-disabled participants.

Methodology of our paper, starting with creating a dataset containing ableist and non-ableist post, followed by collecting and analyzing ratings and explanations from AI models and disabled and non-disabled participants.

We studied how AI systems detect and explain ableist contentβ€”and how that compares to judgments from 130 disabled participants.

We also analyzed explanations from 7 major LLMs and toxicity classifiers. The gaps are stark.

12.04.2025 20:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
The image of our Arxiv Photo with paper title and author list: Mahika Phutane, Ananya Seelam, and Aditya Vashistha

The image of our Arxiv Photo with paper title and author list: Mahika Phutane, Ananya Seelam, and Aditya Vashistha

Our paper, β€œCold, Calculated, and Condescending”: How AI Identifies and Explains Ableism Compared to Disabled People, has been accepted at ACM FAccT 2025!

A quick thread on what we found:

12.04.2025 20:57 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Excited to be at @umich.edu this week to speak at Democracy’s Information Dilemma event, and at the Social Media and Society conference! Hard to believe it’s been since 2016 that I was last here. Can’t wait for engaging conversations, new ideas, and reconnecting with colleagues old and new!

02.04.2025 21:47 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image Post image Post image Post image

NEW YEAR, NEW PLATFORM? πŸ‘€

Bowers is joining the Bluesky community! Follow to stay updated on technology innovation, collaborative research, and faculty expertise.

06.01.2025 16:45 β€” πŸ‘ 12    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

The FATE group at @msftresearch.bsky.social NYC is accepting applications for 2025 interns. πŸ₯³πŸŽ‰

For full consideration, apply by 12/18.

jobs.careers.microsoft.com/global/en/jo...

Interested in AI evaluation? Apply for the STAC internship too!

jobs.careers.microsoft.com/global/en/jo...

25.11.2024 13:31 β€” πŸ‘ 73    πŸ” 35    πŸ’¬ 4    πŸ“Œ 1

Thank you Shannon!

23.11.2024 16:32 β€” πŸ‘ 7    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Hi Shannon, I’d love to be added!

23.11.2024 16:10 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

An exciting opportunity for PhD students to spend a summer at our campus in NYC and be part of our new Security, Trust, and Safety (SETS) Initiative! Deadline: January 21, 2025.

22.11.2024 15:32 β€” πŸ‘ 9    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0

I'd love to be added too! Thank you!

22.11.2024 15:24 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@imadityav is following 20 prominent accounts