Academic paper titled un-straightening generative ai: how queer artists surface and challenge the normativity of generative ai models
The piece is written by Jordan Taylor, Joel Mire, Franchesca Spektor, Alicia DeVrio, Maarten Sap, Haiyi Zhu, and Sarah Fox.
As an image titled 24 attempts at intimacy showing 24 ai generated images with the word intimacy, none of which seems to include same gender couples
๐ณ๏ธโ๐๐จ๐ป๐ข Happy to share our workshop study on queer artistsโ experiences critically engaging with GenAI
Looking forward to presenting this work at #FAccT2025 and you can read a pre-print here:
arxiv.org/abs/2503.09805
14.05.2025 18:38 โ ๐ 25 ๐ 4 ๐ฌ 0 ๐ 0
When it comes to text prediction, where does one LM outperform another? If you've ever worked on LM evals, you know this question is a lot more complex than it seems. In our new #acl2025 paper, we developed a method to find fine-grained differences between LMs:
๐งต1/9
09.06.2025 13:47 โ ๐ 64 ๐ 19 ๐ฌ 1 ๐ 2
An overview of the work โResearch Borderlands: Analysing Writing Across Research Culturesโ by Shaily Bhatt, Tal August, and Maria Antoniak. The overview describes that We survey and interview interdisciplinary researchers (ยง3) to develop a framework of writing norms that vary across research cultures (ยง4) and operationalise them using computational metrics (ยง5). We then use this evaluation suite for two large-scale quantitative analyses: (a) surfacing variations in writing across 11 communities (ยง6); (b) evaluating the cultural competence of LLMs when adapting writing from one community to another (ยง7).
๐๏ธ Curious how writing differs across (research) cultures?
๐ฉ Tired of โculturalโ evals that don't consult people?
We engaged with interdisciplinary researchers to identify & measure โจcultural normsโจin scientific writing, and show thatโLLMs flatten themโ
๐ arxiv.org/abs/2506.00784
[1/11]
09.06.2025 23:29 โ ๐ 72 ๐ 29 ๐ฌ 1 ๐ 5
This looks incredible! Thanks for sharing the syllabus!
04.06.2025 17:23 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
Iโm thrilled to share RewardBench 2 ๐โ We created a new multi-domain reward model evaluation that is substantially harder than RewardBench, we trained and released 70 reward models, and we gained insights about reward modeling benchmarks and downstream performance!
02.06.2025 23:41 โ ๐ 22 ๐ 6 ๐ฌ 2 ๐ 1
Wisconsin-Madison's tree-filled campus, next to a big shiny lake
A computer render of the interior of the new computer science, information science, and statistics building. A staircase crosses an open atrium with visibility across multiple floors
I'm joining Wisconsin CS as an assistant professor in fall 2026!! There, I'll continue working on language models, computational social science, & responsible AI. ๐ฒ๐ง๐ฃ๐ปโโ๏ธ Apply to be my PhD student!
Before then, I'll postdoc for a year in the NLP group at another UW ๐๏ธ in the Pacific Northwest
05.05.2025 19:54 โ ๐ 145 ๐ 14 ๐ฌ 16 ๐ 3
When interacting with ChatGPT, have you wondered if they would ever "lie" to you? We found that under pressure, LLMs often choose deception. Our new #NAACL2025 paper, "AI-LIEDAR ," reveals models were truthful less than 50% of the time when faced with utility-truthfulness conflicts! ๐คฏ 1/
28.04.2025 20:36 โ ๐ 25 ๐ 9 ๐ฌ 1 ๐ 3
A bar plot comparing the storytelling rates for different topics in the example dataset of congressional speeches. There are often large differences between storytelling and non-storytelling for individual topics. For example, the topic whose top words read "NUM, years, service, great, state" has much more storytelling that non-storytelling.
The top five congressional speeches for the topic "NUM, years, service, great state." All of the documents honor the lives of important people.
I updated our ๐ญStorySeeker demo. Aimed at beginners, it briefly walks through loading our model from Hugging Face, loading your own text dataset, predicting whether each text contains a story, and topic modeling and exploring the results. Runs in your browser, no installation needed!
โณ
15.04.2025 12:05 โ ๐ 20 ๐ 6 ๐ฌ 1 ๐ 0
New work on multimodal framing! ๐ซ
Some fun results: comparisons of the same frame when expressed in images vs texts. When the "crime" frame is expressed in the article text, there are more political words in the text, but when the frame is expressed in the article image, more police words.
07.04.2025 09:48 โ ๐ 44 ๐ 10 ๐ฌ 0 ๐ 2
This was joint work with my co-author Zubin Aysola; collaborators @dchechel.bsky.social, Nick Deas, and @chryssazrv.bsky.social; and advisor @maartensap.bsky.social at @ltiatcmu.bsky.social @scsatcmu.bsky.social @columbiauniversity.bsky.social, and the @istecnico.bsky.social! (10/10)
06.03.2025 19:49 โ ๐ 4 ๐ 0 ๐ฌ 0 ๐ 0
Our work builds on sociolinguistic and NLP research on AAL and recent translation methods. Check out the paper for details! We hope others extend this work, e.g., to investigate or mitigate reward model biases against more dialects. (9/10)
06.03.2025 19:49 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
These results point to representational and quality-of-service harms for AAL speakers. โ ๏ธThey also highlight complex ethical questions about the desired behavior of LLMs concerning AAL. (8/10)
06.03.2025 19:49 โ ๐ 0 ๐ 1 ๐ฌ 1 ๐ 0
Bar chart showing the results from t-tests comparing rewards assigned to dialect mirroring (completion dialect matches prompt dialect) vs non-mirroring conditions (completion dialect differs from prompt dialect) inputs. The results show statistically significant preferences for responding in WME--regardless of whether the prompt was WME or AAL--for all models.
Finally, we show that the reward models strongly incentivize steering conversations toward WME, even when prompted with AAL. ๐ฃ๏ธ๐ (7/10)
06.03.2025 19:49 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Bar chart showing Pearson correlation coefficients between reward model score and AAL-ness score from a pre-existing dialect detection tool. The chart shows a statistically significant negative correlation between these variables for most models.
Also, for most models, rewards are negatively correlated with the predicted AAL-ness of a text (based on a pre-existing dialect detection tool). (6/10)
06.03.2025 19:49 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Bar chart showing the cohen's d effect sizes from t-tests comparing raw reward scores assigned to WME vs. AAL texts. All results show a significant dispreference for AAL texts.
Next, we show that most reward models predict lower rewards for AAL texts โฌ๏ธ (5/10)
06.03.2025 19:49 โ ๐ 0 ๐ 1 ๐ฌ 1 ๐ 0
Line chart showing that reward models are less accurate at assigning higher rewards to human-preferred completions when processing paired WME vs. AAL texts.
First, we see a significant drop in performance (-4% accuracy on average) in assigning higher rewards to human-preferred completions when processing AAL texts vs. WME texts. ๐ (4/10)
06.03.2025 19:49 โ ๐ 0 ๐ 1 ๐ฌ 1 ๐ 0
Diagram depicting several ways we combine prompts and completions in White Mainstream English (WME) and African American Language (AAL) to evaluate dialect biases in reward models. Also, the image contains text summaries of our main findings: accuracy drop for AAL, moderate dispreference for AAL-aligned texts, and WME responses for AAL prompts.
We introduce morphosyntactic & phonological features of AAL into WME texts from the RewardBench dataset using validated automatic translation methods. Then, we test 17 reward models for implicit anti-AAL dialect biases. ๐ (3/10)
06.03.2025 19:49 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0
We develop a framework for evaluating dialect biases in reward models and conduct a case study on biases against African American Language (AAL) relative to White Mainstream English (WME). ๐ (2/10)
06.03.2025 19:49 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Screenshot of Arxiv paper title, "Rejected Dialects: Biases Against African American Language in Reward Models," and author list: Joel Mire, Zubin Trivadi Aysola, Daniel Chechelnitsky, Nicholas Deas, Chrysoula Zerva, and Maarten Sap.
Reward models for LMs are meant to align outputs with human preferencesโbut do they accidentally encode dialect biases? ๐ค
Excited to share our paper on biases against African American Language in reward models, accepted to #NAACL2025 Findings! ๐
Paper: arxiv.org/abs/2502.12858 (1/10)
06.03.2025 19:49 โ ๐ 37 ๐ 11 ๐ฌ 1 ๐ 2
Assistant Professor of Sociology at Northwestern
Culture, political sociology, NLP, social networks, computational social science
oscarstuhler.org
go bears!!!
jessicad.ai
kernelmag.io
Professor at the University of Copenhagen. Explainable AI, Natural Language Processing, ML. Head of copenlu.bsky.social lab.
#NLProc #NLP #XAI
http://isabelleaugenstein.github.io/
Postdoc @ UofT. PhD from UBC.
Multilingual NLP x Social Science x Phonetics. Previously Lab126, NVIDIA, Ai2, UW. Check out my work at https://farhansamir.notion.site/samir.
Applied Scientist @Outreach / ex {UW, IIIT-H}
Poet @HugoHouse
I care about how AI can help readers and writers.
PhD student in @nerdsitu.bsky.social @itu.dk. ๐ฎ
Latest work on the impact of generative AI on social media, an experimental study: https://ai-research.andersgiovanni.com/
Computational social scientist researching human-AI interaction and machine learning, particularly the rise of digital minds. Visiting scholar at Stanford, co-founder of Sentience Institute, and PhD candidate at University of Chicago. jacyanthis.com
Director, Max Planck Center for Humans & Machines http://chm.mpib-berlin.mpg.de | Former prof. @MIT | Creator of http://moralmachine.net | Art: http://instagram.com/iyad.rahwan Web: rahwan.me
Associate Professor @ UBC
computational sociology
machine learning is feminist
You only have to look at the Medusa straight on to see her. And sheโs not deadly. Sheโs beautiful and sheโs laughing.
www.lauraknelson.com
Professor of Digital Literature and Methods - FAU Erlangen-Nรผrnberg
Psycholinguist at UMass Amherst Linguistics.
people.umass.edu/bwdillon
Blog: https://argmin.substack.com/
Webpage: https://people.eecs.berkeley.edu/~brecht/
Director, Language Development Dept, MPI for Psycholinguistics. She/her. "All models are wrong but some are useful" (George Box 1978). https://www.mpi.nl/people/rowland-caroline
Incoming PhD @ MIT EECS. ugrad philosophy & CS @ UWa
Assistant Professor of Communication @ Stanford. Histories of AI/ML, NLP, speech recognition, and related data practices; general tomfoolery.
University of Washington CSE PhD Student
Building my data praxis and researching how data scientists build theirs
PhD student @ Cornell info sci | Sociotechnical fairness & algorithm auditing | Previously MSR FATE, Penn | https://emmaharv.github.io/
I was bad at studying Victorian books, and now Iโm bad at ML research. Help!
Heading to:
Foundation Models @ Apple
Prev:
LLMs/World Models @ Riot Games
RecSys @ Apple
Heading to Seattle ๐ณ๏ธโ๐ he/him
Knowledge-monger at Carnegie Mellon University: clandestine print, weird data, book history, computational humanities, political thought. wrote a book called literature & the law of nations, co-founder of six degrees of francis bacon. humanities for all.