Nari Johnson's Avatar

Nari Johnson

@narijohnson.bsky.social

researching AI [evaluation, governance, accountability]

192 Followers  |  210 Following  |  24 Posts  |  Joined: 18.11.2024
Posts Following

Posts by Nari Johnson (@narijohnson.bsky.social)

Post image Post image Post image

๐ŸšจThe 2025 AI Agent Index is out! ๐Ÿšจ
Amidst recent buzz over ๐Ÿฆ€ and NIST's new agent initiative, we find:
- Selective reporting โ€“ esp. on safety
- Almost all agents backend just 3 model families
- Many agents donโ€™t ID themselves as bots online
- Big US/China gaps
- And moreโ€ฆ

20.02.2026 14:04 โ€” ๐Ÿ‘ 3    ๐Ÿ” 3    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
CNTR Tech & Policy Summer School

We opened up applications for the Brown AI Policy Summer School! Please share with any computing or computational social science students who want to engage substantively with policymaking in the United States: cntr.brown.edu/summer-school.

Deadline March 27th!!! Funding available!

18.02.2026 12:46 โ€” ๐Ÿ‘ 9    ๐Ÿ” 7    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

The deadline for the 2026 FAccT DC is next Tuesday, February 24! If you are a student working on topics relevant to the FAccT's scope, this is an opportunity to interact with a diverse set of peers and mentors! #facct2026 #facct26 #facct

Details here: facctconference.org/2026/callfor...

18.02.2026 14:52 โ€” ๐Ÿ‘ 7    ๐Ÿ” 10    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Botender helps communities iteratively align their AI agents with their collective intents through case-based provocations.

Botender helps communities iteratively align their AI agents with their collective intents through case-based provocations.

๐Ÿ”ฎ How can we empower online communities to design AI agents tailored to their unique needs and norms?

In our #CHI2026 paper, we introduce #Botender, a system that enables collaborative design of AI agents through ๐Ÿ”ฅcase-based provocation๐Ÿ”ฅ

17.02.2026 15:01 โ€” ๐Ÿ‘ 15    ๐Ÿ” 5    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

PhD admissions visits/open houses are starting to happen, and I got a comment on an old Reddit post where I was offering advice, and realized that it's actually really good advice. So here it is! (And this applies whether you've already been admitted to the program or not.) ๐Ÿงต

05.02.2026 17:26 โ€” ๐Ÿ‘ 30    ๐Ÿ” 8    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 3
Maintainer told to remove malware skills, he says "There's about 1 Million things people want me to do, I don't have a magical team that verifies user generated content"

Maintainer told to remove malware skills, he says "There's about 1 Million things people want me to do, I don't have a magical team that verifies user generated content"

The attack tricks the LM by having it run a base64 string which is obviously malicious (curl bash script at this random IP and run it)

The attack tricks the LM by having it run a base64 string which is obviously malicious (curl bash script at this random IP and run it)

Yep and it gets worse! Owner doesn't even care to remove hundreds of skills which directly instruct the model to install malware

opensourcemalware.com/blog/clawdbo...

06.02.2026 01:05 โ€” ๐Ÿ‘ 6    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Our call for craft and tutorial sessions for #FAccT2026 is now live!

โ–ถ๏ธ Craft CfP: facctconference.org/2026/cfpcraf...
โ–ถ๏ธ Tutorials CfP: facctconference.org/2026/cft.html

Both kinds of proposals are due March 25!

05.02.2026 17:39 โ€” ๐Ÿ‘ 2    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

๐ŸŽญ How do LLMs (mis)represent culture?
๐Ÿงฎ How often?
๐Ÿง  Misrepresentations = missing knowledge? spoiler: NO!

At #CHI2026 we are bringing โœจTALESโœจ a participatory evaluation of cultural (mis)reps & knowledge in multilingual LLM-stories for India

๐Ÿ“œ arxiv.org/abs/2511.21322

1/10

02.02.2026 21:38 โ€” ๐Ÿ‘ 45    ๐Ÿ” 21    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 2

Microsoft Research NYC is hiringย a researcher in the space of AI and society!

29.01.2026 23:27 โ€” ๐Ÿ‘ 62    ๐Ÿ” 40    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 2
Preview
Making Sense of AI Policy Using Computational Tools | TechPolicy.Press A new report examines how to use computational tools to evaluate policy, with AI policy as a case study.

A new report by the Center for Tech Responsibility at Brown University and the ACLU uses computational tools to analyze legislative trends on AI across 1,804 state and federal bills, while offering recommendations for how to integrate the technology into policy analysis.

08.01.2026 20:56 โ€” ๐Ÿ‘ 13    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

We are studying the sentiments of visual artists towards generative AI in the workplace and their impacts on creative careers. If you're an artist, please consider filling out this recruitment form for access to our survey!
cmu.ca1.qualtrics.com/jfe/form/SV_...

19.12.2025 01:58 โ€” ๐Ÿ‘ 6    ๐Ÿ” 3    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
Post image

Most LLM evals use API calls or offline inference, testing models in a memory-less silo. Our new Patterns paper shows this misses how LLMs actually behave in real user interfaces, where personalization and interaction history shape responses: arxiv.org/abs/2509.19364

12.12.2025 20:42 โ€” ๐Ÿ‘ 38    ๐Ÿ” 11    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
Post image

US CAISI is hiring -- the internal govt name for the role is "IT Specialist" but it is effectively a research scientist role!

Salary is $120,579 to - $195,200 per year, and you get to work on AI evaluation within government agencies!

Job posting (**closes EOD 12/28/2025**): lnkd.in/exJgkqr5

11.12.2025 22:01 โ€” ๐Ÿ‘ 24    ๐Ÿ” 10    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

Also, our team is hiring an AI Research Scientist!

www.usajobs.gov/job/851528400

08.12.2025 14:47 โ€” ๐Ÿ‘ 10    ๐Ÿ” 7    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Accelerating AI Innovation Through Measurement Science Building gold-standard AI systems requires gold-standard AI measurement science โ€“ the scientific study of methods used to assess AI systemsโ€™ properties and impacts. NIST works to improve measurements ...

Our team at NIST's Center for AI Standards and Innovation (CAISI) just released a blog post with open questions for AI measurement science:

www.nist.gov/blogs/caisi-...

04.12.2025 20:17 โ€” ๐Ÿ‘ 5    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 2
Post image

Did you know that one base model is responsible for 94% of model-tagged NSFW AI videos on CivitAI?

This new paper studies how a small number of models power the non-consensual AI video deepfake ecosystem and why their developers could have predicted and mitigated this.

04.12.2025 17:32 โ€” ๐Ÿ‘ 6    ๐Ÿ” 3    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

I appreciate this sympathetic position

people's feelings of emotional dependency on these "human-like" bots is real. ridiculing them doesn't help anyone

28.11.2025 23:48 โ€” ๐Ÿ‘ 55    ๐Ÿ” 8    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
How public involvement can improve the science of AI | PNAS As AI systems from decision-making algorithms to generative AI are deployed more widely, computer scientists and social scientists alike are being ...

Can public involvement in AI evaluation improve the science? Or does it compromise quality, speed, cost?

In @pnas.org, Megan Price & I summarize challenges of AI evaluation, review strengths/weaknesses, & suggest how participatory methods can improve the science of AI
www.pnas.org/doi/10.1073/...

17.11.2025 12:47 โ€” ๐Ÿ‘ 19    ๐Ÿ” 13    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

it exists โ€” several AI vendors and US local governments have negotiated short term pilot contracts with a pay-only-if-it-works model. happy to chat and connect you if youโ€™re interested!

16.11.2025 18:13 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Performance of a sweep of models on Oolong-synth and Oolong-real. Performance decreases with increasing context length, sometimes steeply.

Performance of a sweep of models on Oolong-synth and Oolong-real. Performance decreases with increasing context length, sometimes steeply.

Can LLMs accurately aggregate information over long, information-dense texts? Not yetโ€ฆ

We introduce Oolong, a dataset of simple-to-verify information aggregation questions over long inputs. No model achieves >50% accuracy at 128K on Oolong!

07.11.2025 17:07 โ€” ๐Ÿ‘ 50    ๐Ÿ” 20    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 3
Post image

๐Ÿ“ฃ Our method for conducting community-based algorithmic impact assessments is now available! Weโ€™ve just launched a new section on our website where you can find an extensive toolkit, documentation of our pilots, and a series of reflections on lessons learned. datasociety.net/research/alg...

29.10.2025 19:10 โ€” ๐Ÿ‘ 21    ๐Ÿ” 8    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Supporting Industry Computing Researchers in Assessing, Articulating, and Addressing the Potential Negative Societal Impact of Their Work | Proceedings of the ACM on Human-Computer Interaction Recent years have witnessed increasing calls for computing researchers to grapple with the societal impacts of their work. Tools such as impact assessments have gained prominence as a method to uncover potential impacts, and a number of publication ...

๐’๐จ๐œ๐ข๐ž๐ญ๐š๐ฅ ๐ˆ๐ฆ๐ฉ๐š๐œ๐ญ ๐€๐ฌ๐ฌ๐ž๐ฌ๐ฌ๐ฆ๐ž๐ง๐ญ ๐Ÿ๐จ๐ซ ๐ˆ๐ง๐๐ฎ๐ฌ๐ญ๐ซ๐ฒ ๐‚๐จ๐ฆ๐ฉ๐ฎ๐ญ๐ข๐ง๐  ๐‘๐ž๐ฌ๐ž๐š๐ซ๐œ๐ก๐ž๐ซ๐ฌ
๐Ÿ… Best Paper Honorable Mention (Top 3% Submissions)
๐Ÿ”— dl.acm.org/doi/10.1145/...
๐Ÿ“† Wed, 22 Oct | 9:00 AM, CET: Toward More Ethical and Transparent Systems and Environments

19.10.2025 13:49 โ€” ๐Ÿ‘ 6    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1
Post image

๐Ÿ’กCan we trust synthetic data for statistical inference?

We show that synthetic data (e.g., LLM simulations) can significantly improve the performance of inference tasks. The key intuition lies in the interactions between the moment residuals of synthetic data and those of real data

10.10.2025 16:12 โ€” ๐Ÿ‘ 36    ๐Ÿ” 9    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 5
Preview
Machine Learning / AI Internships - Jobs - Careers at Apple Apply for a Machine Learning / AI Internships job at Apple. Read about the role and find out if itโ€™s right for you.

Our Responsible AI team at Apple is looking for spring/summer 2026 PhD research interns! Please apply at jobs.apple.com/en-us/detail... and email rai-internship@group.apple.com. Do not send extra info (e.g., CV), just drop us a line so we can find your application in the central pool!

10.10.2025 02:28 โ€” ๐Ÿ‘ 29    ๐Ÿ” 11    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Cella M. Sum โ€“

โœจIโ€™m on the academic job market โœจ

Iโ€™m a PhD candidate at @hcii.cmu.edu studying tech, labor, and resistance ๐Ÿ‘ฉ๐Ÿปโ€๐Ÿ’ป๐Ÿ’ช๐Ÿฝ๐Ÿ’ฅ

I research how workers and communities contest harmful sociotechnical systems and shape alternative futures through everyday resistance and collective action

More info: cella.io

09.10.2025 14:39 โ€” ๐Ÿ‘ 72    ๐Ÿ” 36    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 4
Carnegie Mellon University School of Computer Science Graduate Application Support Program. Apply by October 13, 2025.

Carnegie Mellon University School of Computer Science Graduate Application Support Program. Apply by October 13, 2025.

๐ŸŒŸ If youโ€™re applying to CMU SCS PhD programs, and come from a background that would bring additional dimensions to the CMU community, our PhD students are here to help!

Apply to the Graduate Applicant Support Program by Oct 13 to receive feedback on your application materials:

24.09.2025 16:00 โ€” ๐Ÿ‘ 7    ๐Ÿ” 4    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
Preview
Stephen Casper Visit the post for more.

๐Ÿ“Œ๐Ÿ“Œ๐Ÿ“Œ
I'm excited to be on the faculty job market this fall. I just updated my website with my CV.
stephencasper.com

04.09.2025 03:39 โ€” ๐Ÿ‘ 18    ๐Ÿ” 4    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1
Post image

๐Ÿ“ข2026 Fellowship applications are OPEN!๐Ÿ“ข
If you are someone looking to inform technology policy through rigorous original reporting or policy analyses, we want to hear from you!
Apply here: airtable.com/appIrc1F9M5d...

04.09.2025 11:47 โ€” ๐Ÿ‘ 18    ๐Ÿ” 10    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

We also have a position paper under review that's in the exact same situation. Thanks for your post - it's been super illuminating to help us make sense of what's happening.

31.08.2025 21:29 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Screenshot of the CSCW 2025 paper "The Future of Tech Labor: How Workers are Organizing and Transforming the Computing Industry" 

CELLA M. SUM, Carnegie Mellon University, USA
ANNA KONVICKA, Princeton University, USA
MONA WANG, Princeton University, USA
SARAH E. FOX, Carnegie Mellon University, USA

Abstract: The tech industryโ€™s shifting landscape and the growing precarity of its labor force have spurred unionization efforts among tech workers. These workers turn to collective action to improve their working conditions and to protest unethical practices within their workplaces. To better understand this movement, we interviewed 44 U.S.-based tech worker-organizers to examine their motivations, strategies, challenges, and future visions for labor organizing. These workers included engineers, product managers, customer support specialists, QA analysts, logistics workers, gig workers, and union staff organizers. Our findings reveal that, contrary to popular narratives of prestige and privilege within the tech industry, tech workers face fragmented and unstable work environments which contribute to their disempowerment and hinder their organizing efforts. Despite these difficulties, organizers are laying the groundwork for a more resilient tech worker movement through community building and expanding political consciousness. By situating these dynamics within broader structural and ideological forces, we identify ways for the CSCW community to build solidarity with
tech workers who are materially transforming our field through their organizing efforts.

Screenshot of the CSCW 2025 paper "The Future of Tech Labor: How Workers are Organizing and Transforming the Computing Industry" CELLA M. SUM, Carnegie Mellon University, USA ANNA KONVICKA, Princeton University, USA MONA WANG, Princeton University, USA SARAH E. FOX, Carnegie Mellon University, USA Abstract: The tech industryโ€™s shifting landscape and the growing precarity of its labor force have spurred unionization efforts among tech workers. These workers turn to collective action to improve their working conditions and to protest unethical practices within their workplaces. To better understand this movement, we interviewed 44 U.S.-based tech worker-organizers to examine their motivations, strategies, challenges, and future visions for labor organizing. These workers included engineers, product managers, customer support specialists, QA analysts, logistics workers, gig workers, and union staff organizers. Our findings reveal that, contrary to popular narratives of prestige and privilege within the tech industry, tech workers face fragmented and unstable work environments which contribute to their disempowerment and hinder their organizing efforts. Despite these difficulties, organizers are laying the groundwork for a more resilient tech worker movement through community building and expanding political consciousness. By situating these dynamics within broader structural and ideological forces, we identify ways for the CSCW community to build solidarity with tech workers who are materially transforming our field through their organizing efforts.

What can #CSCW learn from tech workers who have been involved in collective action and unionization about how to make transformative change within our field?

My new #CSCW2025 paper with Mona Wang, Anna Konvicka, and Sarah Fox seeks to answer this question.

Pre-print: arxiv.org/pdf/2508.12579

28.08.2025 14:14 โ€” ๐Ÿ‘ 43    ๐Ÿ” 17    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 4