's Avatar

@prakharg.bsky.social

30 Followers  |  163 Following  |  63 Posts  |  Joined: 27.11.2024
Posts Following

Posts by (@prakharg.bsky.social)

๐Ÿšจ Deadline Extended to Feb 5 (AoE)!
CFP still OPEN for the #AFAA2026 Workshop at @iclr-conf.bsky.social โ€” on fairness across alignment & agentic AI systems.
Full & tiny papers welcome โ€ข Interdisciplinary work encouraged!
๐Ÿ”— afciworkshop.org

#ICLR2026 #AFAA2026

02.02.2026 17:48 โ€” ๐Ÿ‘ 1    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1
Preview
AFAA 2026 The Algorithmic Fairness Across Alignment Procedures and Agentic Systems (AFAA) workshop aims to spark discussions on rethinking fairness in AI alignment procedures and agentic system development.

๐Ÿšจ CFP OPEN! Weโ€™re launching the #AFAA2026 Workshop at @iclr-conf.bsky.social on ๐—ณ๐—ฎ๐—ถ๐—ฟ๐—ป๐—ฒ๐˜€๐˜€ ๐—ฎ๐—ฐ๐—ฟ๐—ผ๐˜€๐˜€ ๐—ฎ๐—น๐—ถ๐—ด๐—ป๐—บ๐—ฒ๐—ป๐˜ ๐—ฎ๐—ป๐—ฑ ๐—ฎ๐—ด๐—ฒ๐—ป๐˜๐—ถ๐—ฐ ๐—”๐—œ ๐˜€๐˜†๐˜€๐˜๐—ฒ๐—บ๐˜€.
Submit your latest ideas (full or tiny papers!)
Interdisciplinary work especially welcome :D
๐Ÿ—“ Deadline: Jan 31 (AoE) | ๐Ÿ”— www.afciworkshop.org

#AFAA2026 #ICLR2026

06.01.2026 02:39 โ€” ๐Ÿ‘ 6    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 4

Four case studies with the gap between the reality of model use and their sandbox evaluations in audits... Definitely need to take a deeper dive, great presentation by Emily Black!

25.06.2025 08:52 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Evaluations in the way the model would be deployed vs evaluations in only controlled unrealistic settings!

25.06.2025 08:52 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Allowing companies to do isolated audits can lead to D-Hacking!! More robust testing is needed...

25.06.2025 08:52 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Legal frameworks tend to have control over allocative decisions (Yes/No outcomes), which fit well with traditional ML systems... But not with GenAI systems

25.06.2025 08:52 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Zollo et al: Towards Effective Discrimination Testing for Generative AI
#FAccT2025

25.06.2025 08:43 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Nuance of stereotype errors is so important to understand their true harms... Insightful presentation by @angelinawang.bsky.social

25.06.2025 08:43 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Women tend to report stereotype-reinforcing errors as more harmful while men tend to report stereotype-violating errors as more harmful...

25.06.2025 08:43 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Some items are more associated with men vs women (not surprising), but not all of them are equally harmful!!

25.06.2025 08:43 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Cognitive beliefs, attitudes and behaviours... Three ways to measure harms ('pragmatic harms')

25.06.2025 08:43 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Are all errors equally harmful? No! Stereotype-reinforcing errors vs stereotype-violating errors

25.06.2025 08:43 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Our understanding of stereotypes sometimes isn't indicative of reality.... they can appear in both directions, or might exist simply without harm

25.06.2025 08:43 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Wang et al: Measuring Machine Learning Harms from Stereotypes Requires Understanding Who Is Harmed by Which Errors in What Ways
#FAccT2025

25.06.2025 08:34 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Clear narrative and a great presentation by Cecilia Panigutti

25.06.2025 08:33 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Risk-measuring studies - Bringing it back to risk measurement, but this time with a clearly defined objective instead of risk-uncovering as before... Not just whether a risk exists, but 'how severe' is it?

25.06.2025 08:33 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Interface-design studies - Focus on UI design elements which impact user interaction

25.06.2025 08:33 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Reverse-engineering studies - Narrower scope and in-depth studies of how algorithms work... Methodological precision in the key!

25.06.2025 08:33 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Risk-uncovering studies - Typical starts from anecdotal evidence and help surface new risks

25.06.2025 08:33 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

A review organized not by data collection technique, but by DSA risk management framework categories

25.06.2025 08:33 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Narrative review of algorithmic auditing studies, practical recommendation for best practices, and mapping to DSA obligations...

25.06.2025 08:33 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Panigutti et al: How to investigate algorithmic-driven risks in online platforms and search engines? A narrative review through the lens of the EU Digital Services Act
#FAccT2025

25.06.2025 08:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Such a broad topic... Excellent presentation by @feliciajing.bsky.social

25.06.2025 08:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Historical methods working alongside many other ways of auditing these models can help us take advantage of the broader scope of historical evaluations....

25.06.2025 08:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

AI Audits have moved from bottom-up external evaluations to new age 'auditing companies'. While this has increased speed and scale, they have significantly narrowed the scope of auditing.

25.06.2025 08:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Why the history of AI assessments? A study through the lens of historical methods can help us understand neglected areas of auditing.

25.06.2025 08:22 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Sandoval and Jing: Historical Methods for AI Evaluations, Assessments, and Audits
#FAccT2025

25.06.2025 08:10 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Important recommendations on standardization of report creation and storage to allow better meta-analysis in the future... Eye opening presentation by @mkgerchick.bsky.social

25.06.2025 08:10 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Applicants impacted by these tools, whose demographic data is missing, are completely removed from these audits!

25.06.2025 08:10 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Serious issues with the data usage... most weird for me: 'simulated test data'!

25.06.2025 08:10 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0