Hope Schroeder's Avatar

Hope Schroeder

@hopeschroeder.bsky.social

Studying NLP, CSS, and Human-AI interaction. PhD student @MIT. Previously at Microsoft FATE + CSS, Oxford Internet Institute, Stanford Symbolic Systems hopeschroeder.com

316 Followers  |  279 Following  |  54 Posts  |  Joined: 20.11.2024  |  2.458

Latest posts by hopeschroeder.bsky.social on Bluesky

Search Jobs | Microsoft Careers

We may have the chance to hire an outstanding researcher 3+ years post PhD to join Tarleton Gillespie, Mary Gray and me in Cambridge MA bringing critical sociotechnical perspectives to bear on new technologies.

jobs.careers.microsoft.com/global/en/jo...

28.07.2025 17:26 — 👍 87    🔁 47    💬 0    📌 2

Thanks for sharing- not just our paper but also learned a lot from this list! :)

24.07.2025 09:06 — 👍 1    🔁 0    💬 0    📌 0

Awesome work and great presentation! Congrats!! ⚡️

23.07.2025 13:42 — 👍 0    🔁 0    💬 0    📌 0

Talking about this work tomorrow (Wed, July 23rd) at #IC2S2 in Norrköping during the 11 am session on LLMs, Annotation, and Synthetic Data! Come hear about this and more!

22.07.2025 21:10 — 👍 14    🔁 2    💬 0    📌 0

👋 ☺️

22.07.2025 21:03 — 👍 0    🔁 0    💬 0    📌 0
Post image

Implications vary by task and domain. Researchers should clearly define their annotation constructs before reviewing LLM annotations. We are subject to anchoring bias that can affect our evaluations, or even our research findings!
Read more: arxiv.org/abs/2507.15821

22.07.2025 08:35 — 👍 5    🔁 0    💬 0    📌 0

Using LLM-influenced labels, even when a crowd of humans reviews them and is aggregated into a set of crowd labels, can lead to 1) different findings when used in data analysis and 2) different results when used as a basis of evaluating LLM performance on the task.

22.07.2025 08:34 — 👍 2    🔁 0    💬 1    📌 0
Post image

What happens if we use LLM-influenced labels as ground truth when evaluating LLM performance on these tasks? We can seriously overestimate LLM performance on these tasks. F1 scores for some tasks were +.5 higher when evaluated using LLM-influenced labels as ground truth!

22.07.2025 08:34 — 👍 3    🔁 0    💬 1    📌 0
Post image

However… annotators STRONGLY took LLM suggestions: just 40% of human crowd labels overlap with LLM baselines, but overlap jumps to over 80% when LLM suggestions are given (varied crowd thresholds and conditions shown in graph). Beware: Humans are subject to anchoring bias!

22.07.2025 08:33 — 👍 6    🔁 1    💬 1    📌 0

Some findings: ⚠️ reviewing LLM suggestions did not make annotators go faster, and often slowed them down! OTOH, having LLM assistance made annotators ❗more self-confident❗in their task and content understanding at no identified cost to their tested task understanding.

22.07.2025 08:33 — 👍 5    🔁 2    💬 1    📌 0

We conducted experiments where over 410 unique annotators generated with over 7,000 annotations across three LLM assistance conditions of varying strengths against a control, using two different models, and two different complex, subjective annotation tasks.

22.07.2025 08:33 — 👍 3    🔁 1    💬 1    📌 0

LLMs can be fast and promising annotators, so letting human annotators "review" first-pass LLM annotations in interpretive tasks is tempting. How does this impact productivity, annotators, evaluating LLM performance on subjective tasks and downstream data analysis?

22.07.2025 08:32 — 👍 2    🔁 0    💬 1    📌 0
Preview
Just Put a Human in the Loop? Investigating LLM-Assisted Annotation for Subjective Tasks LLM use in annotation is becoming widespread, and given LLMs' overall promising performance and speed, simply "reviewing" LLM annotations in interpretive tasks can be tempting. In subjective annotatio...

🗣️ Excited to share our new #ACL2025 Findings paper: “Just Put a Human in the Loop? Investigating LLM-Assisted Annotation for Subjective Tasks” with Jad Kabbara and Deb Roy. Arxiv: arxiv.org/abs/2507.15821
Read about our findings ⤵️

22.07.2025 08:32 — 👍 40    🔁 8    💬 4    📌 1

Thanks for attending and for your comments!!

25.06.2025 08:28 — 👍 1    🔁 0    💬 0    📌 0
Preview
Disclosure without Engagement: An Empirical Review of Positionality Statements at FAccT | Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency You will be notified whenever a record that you have chosen has been cited.

Come hear much more at our paper talk tomorrow (Wednesday, 6/25) in Meta Research and Critiques, 9:12 am at New Stage A! Read the paper here: dl.acm.org/doi/10.1145/...

24.06.2025 14:50 — 👍 4    🔁 1    💬 0    📌 0

*What should FAccT do?* We discuss a need for the conference to clarify its policies next year, engage scholars from different disciplines when considering policy on this delicate subject, and engage authors in reflexive practice upstream of paper-writing, potentially through CRAFT.

24.06.2025 14:50 — 👍 1    🔁 0    💬 1    📌 0
Post image

*Are disclosed features connected to described impacts?* Disclosed features are much less commonly described in terms of impacts the feature had on the research, which may leave a gap for readers to jump to conclusions about how the disclosed feature impacted the work.

24.06.2025 14:50 — 👍 3    🔁 1    💬 1    📌 0

*What do authors disclose in positionality statements?* We conducted fine-grained annotation of the statements. We find academic background and training are disclosed most often, but identity features like race and gender are also common.

24.06.2025 14:49 — 👍 2    🔁 0    💬 1    📌 0
Post image

We reviewed papers from the entire history of FAccT for the presence of positionality statements. We find 2024 marked a significant proportional increase in papers that included positionality statements, likely as a result of PC recommendations:

24.06.2025 14:48 — 👍 0    🔁 0    💬 1    📌 0
Responsible AI Research Needs Impact Statements Too Alexandra Olteanu, Michael Ekstrand, Carlos Castillo, and Jina Suh

In 2024, the FAccT PCs recommended that authors write positionality statements (blog post: medium.com/@alexandra.o...)

24.06.2025 14:47 — 👍 1    🔁 0    💬 1    📌 0

With ongoing reflection on the impact of computing on society, and the role of researchers in shaping impacts, positionality statements have become more common in computing venues, but little is known about their contents or the impact of conference policy on their presence.

24.06.2025 14:46 — 👍 1    🔁 0    💬 1    📌 0

1) Thrilled to be at #FAccT for the first time this week, representing a meta-research paper on positionality statements at FAccT from 2018-2024, in collaboration with @s010n.bsky.social and Akshansh Pareek, "Disclosure without Engagement: An Empirical Review of Positionality Statements at FAccT"

24.06.2025 14:46 — 👍 10    🔁 1    💬 1    📌 0

Democracy needs you! Super excited to be co-organizing this NLP for Democracy workshop at #COLM2025 with a wonderful group. Abstracts due June 19th!

06.06.2025 15:59 — 👍 8    🔁 2    💬 0    📌 0
FATE Research Assistant (“Pre-doc”) - Microsoft Research The Fairness, Accountability, Transparency, and Ethics (FATE) Research group at Microsoft Research New York City (MSR NYC) is looking for a pre-doctoral research assistant (pre-doc) to start August 20...

Exciting news: The Fairness, Accountability, Transparency and Ethics (FATE) group at Microsoft Research NYC is hiring a predoctoral fellow!!! 🎉

www.microsoft.com/en-us/resear...

20.05.2025 13:47 — 👍 32    🔁 13    💬 2    📌 2

a whole cluster of postdocs and phd positions in Tartu in Digital Humanities / Computational Social Science / AI under the umbrella of big European projects.

consider sharing please!

07.05.2025 11:17 — 👍 34    🔁 42    💬 0    📌 0

📣🧪
$25k grants for those who:

1. are working on research on STEM and education (including AI and CS, graduate education and MSIs, and scholarship that aims to reduce inequality), and

2. have had a recently terminated or cancelled grant from NSF.

Early-career scholars prioritized

03.05.2025 17:56 — 👍 112    🔁 80    💬 2    📌 2

Last poster session starting now in Pacifico North! Come say hi 👋

29.04.2025 06:30 — 👍 0    🔁 0    💬 0    📌 0
Preview
Forage: Understanding RAG-based Sensemaking for Community Conversations | Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems

Paper: dl.acm.org/doi/10.1145/...

Try the prototype: forage.ccc-mit.org

29.04.2025 06:28 — 👍 0    🔁 0    💬 1    📌 0
Post image

Great morning presenting LBW at #CHI2025: Forage: Understanding RAG-based Sensemaking for Community Conversations!

29.04.2025 06:27 — 👍 10    🔁 0    💬 1    📌 0

Glad we were in the same session! Thanks for the fun conversation!

28.04.2025 23:37 — 👍 1    🔁 0    💬 0    📌 0

@hopeschroeder is following 20 prominent accounts