🚨New paper:
Current reports on AI audits/evals often omit crucial details, and there are huge disparities between the thoroughness of different reports. Even technically rigorous evals can offer little useful insight if reported selectively or obscurely.
Audit cards can help.
21.04.2025 17:10 — 👍 2 🔁 2 💬 1 📌 0
Some researchers are rethinking how to measure AI intelligence
Current popular benchmarks are often inadequate or too easy to game, experts say.
A recent Stanford paper reveals that many popular AI benchmarks are fundamentally flawed: They can be outdated, easily gamed, or inaccurate. Stanford HAI Graduate Fellow
@ankareuel.bsky.social talks about how researchers are rethinking AI benchmarks: www.emergingtechbrew.com/stories/2025...
25.03.2025 21:26 — 👍 9 🔁 3 💬 1 📌 0
Hey Kabir! A lot of it is applicable for different types of evals, especially when it comes to reporting considerations. Would you mind sharing more infos here or via DM on the hackathon? Sounds like this would be a cool opportunity to extend the BetterBench work!
28.01.2025 00:45 — 👍 3 🔁 0 💬 1 📌 0
Submitting a benchmark to
ICML? Check out our NeurIPS Spotlight paper BetterBench! We outline best practices for benchmark design, implementation & reporting to help shift community norms. Be part of the change! 🙌
+ Add your benchmark to our database for visibility: betterbench.stanford.edu
27.01.2025 22:02 — 👍 12 🔁 3 💬 1 📌 0
This is such a hard one :D And I think it extends beyond being patient with the students but also being patient with yourself knowing that you won't get everything perfect the first time around (or ever 🥲)
05.01.2025 17:45 — 👍 6 🔁 0 💬 1 📌 0
🔄 Sharing is caring! Help us reach as wide of an audience as possible by spreading the word. Your support is key in crafting an insightful, community-driven chapter and help key researchers in the field get their work promoted! Thank you! 🙏#StanfordHAI #AIIndex x/
05.01.2025 17:42 — 👍 5 🔁 2 💬 0 📌 0
The AI Index is an initiative by @stanfordhai.bsky.social. The annual report showcases AI research to enable decision-makers to advance AI responsibly. Previous versions have been cited 300+ times; it's been featured in top media outlets like the @nytimes.com & the @financialtimes.com. 4/
05.01.2025 17:42 — 👍 5 🔁 0 💬 1 📌 0
Our chapter will cover fairness & non-discrimination, transparency, explainability, data governance & privacy, security, societal impact, and more. Plus, a special subchapter on responsible AI agents! 🤖 3/
05.01.2025 17:42 — 👍 0 🔁 0 💬 1 📌 0
📢 Excited to share: I'm again leading the efforts for the Responsible AI chapter for Stanford's 2025 AI Index, curated by @stanfordhai.bsky.social. As last year, we're asking you to submit your favorite papers on the topic for consideration (including your own!) 🧵 1/
05.01.2025 17:42 — 👍 13 🔁 8 💬 1 📌 0
This is all awesome advice, thank you so much for sharing! This is an in-person course but we’ll make all lectures publicly available.
04.01.2025 04:30 — 👍 1 🔁 0 💬 0 📌 0
I‘m teaching my first own course starting next week (Intro to AI Governance at Stanford). Super proud but also nervous 🥹 Any advice from more seasoned instructors? 😬 #AcademicTwitter #AcademicChatter #TeachingTips #AcademicAdvice
04.01.2025 03:14 — 👍 11 🔁 0 💬 2 📌 0
The regular reminder of my starter packs full of amazing folks / accounts to follow. I am trying to keep them up to date but let me know if I missed you.
24.12.2024 08:28 — 👍 5 🔁 1 💬 0 📌 0
Thank you, Stefanie! ❤️
19.12.2024 18:46 — 👍 0 🔁 0 💬 0 📌 0
In our latest brief, Stanford scholars present a novel assessment framework for evaluating the quality of AI benchmarks and share best practices for minimum quality assurance. @ankareuel.bsky.social @chansmi.bsky.social @mlamparth.bsky.social hai.stanford.edu/what-makes-g...
11.12.2024 18:08 — 👍 11 🔁 4 💬 0 📌 0
Looking forward to your talk! :)
09.12.2024 20:36 — 👍 0 🔁 0 💬 0 📌 0
Thanks a ton, Federico! :)
07.12.2024 19:37 — 👍 1 🔁 0 💬 0 📌 0
Thanks so much, Lorena!
07.12.2024 19:37 — 👍 1 🔁 0 💬 0 📌 0
Thanks so much, Daniel!
07.12.2024 19:37 — 👍 1 🔁 0 💬 0 📌 0
Thanks a lot, Stephan 😊
07.12.2024 06:42 — 👍 1 🔁 0 💬 0 📌 0
Thank you Karen 🦋
07.12.2024 02:02 — 👍 1 🔁 0 💬 0 📌 0
Thanks so much! And yes, very much looking forward to the weekend 😁🫶
06.12.2024 23:50 — 👍 1 🔁 0 💬 0 📌 0
Thanks a lot!
06.12.2024 23:40 — 👍 1 🔁 0 💬 1 📌 0
In the same boat as @mlamparth.bsky.social, would appreciate if you could add me, too, please! Thanks so much 😊
06.12.2024 22:47 — 👍 2 🔁 0 💬 2 📌 0
In the same boat as @mlamparth.bsky.social, would appreciate if you could add me, too, please! Thanks so much 😊
06.12.2024 22:47 — 👍 2 🔁 0 💬 1 📌 0
In the same boat as @mlamparth.bsky.social, would appreciate if you could add me, too! Thanks so much 😊
06.12.2024 22:46 — 👍 2 🔁 0 💬 1 📌 0
In the same boat as @mlamparth.bsky.social, would appreciate if you could add me, too! Thanks so much 😊
06.12.2024 22:46 — 👍 2 🔁 0 💬 0 📌 0
Would appreciate if you could add me to the Responsible AI and the Security starter packs, similar to @mlamparth.bsky.social, I’m moving here from X 😊
06.12.2024 22:45 — 👍 2 🔁 0 💬 1 📌 0
In the same boat as @mlamparth.bsky.social, would appreciate if you could add me, too! Thanks so much 😊
06.12.2024 22:42 — 👍 2 🔁 0 💬 1 📌 0
🏥 Doctor
🎓 Harvard ‘25
💼 Founder ApplicAid
💯 Forbes 30 under 30
📍 Boston
———
Lift as you climb 🏔
———
The official account of the Stanford Institute for Human-Centered AI, advancing AI research, education, policy, and practice to improve the human condition.
On the quest to understand the fundamental mathematics of intelligence and of the universe with curiosity. http://burnyverse.com Upskilling @StanfordOnline
EM doc. Politics nerd 🇷🇴🇬🇧🇫🇷
European at core
Into military stuff nowadays, out of necessity
We build secure, scalable, and private enterprise-grade AI technology to solve real-world business problems. Join us: http://cohere.com/careers
Asst Prof at Cornell Info Sci and Cornell Tech. Responsible AI
https://angelina-wang.github.io/
JD/PhD in Computer Science @ Stanford
https://neelguha.github.io/
PhD student @MIT & @medialab; Building a secure & robust computational future; 🇦🇺🦘; Fulbrighter; he/him
Ingeniera salsera cascabelera 🚀. Responsable de la Comisión de Inteligencia Artificial de la Universidad de Deusto. Coautora del juego de mesa Nobel Run y del libro "La IA explicada a todos los públicos".
#WomeninSTEM, #AIEthics
http://www.loretahur.net
Catedrática de psicología. Investigo creencias erróneas y sesgos cognitivos en humanos y en inteligencias artificiales. También la influencia de la IA en las decisiones humanas.
Autora de "Nuestra Mente nos Engaña".
https://helenamatute.wordpress.com/
AI Geopolicy | Researcher on the Democratic Governance of Cyberspace | Jurist | Let’s break monopolies | Tax the oligarchs | The Force is with me | @SecuoyaGroup
En Bikolabs Estudio la complejidad de personas y tecnología y busco la luna de Anarres
Blog: https://www.karlosgliberal.com/
Vicedean @ UPF Barcelona School of Management. Leveraging Artificial Intelligence for Social Good.
Investigadora en el Laboratorio de Psicología Experimental de Deusto y en Jakala. Interesada en el impacto de la IA en las decisiones humanas
Abogada aunque buena persona. Solo agredo por motivos profesionales. Humana de Rosca. Me aberroncho en señal de protesta.
Me dedico al #Derecho Digital y Entretenimiento. Mi pasión es la 🚲 y los #datos #Data #DPO #IA #Legaltech
Me puedes encontrar en www.iurismatica.com www.eventosjuridicos.es y https://www.datos.personales.es/
Somos una red colaborativa para reivindicar los #DerechosHumanos en el ámbito de la #InteligenciaArtificial.
[bridged from https://mastodon.social/@algorights on the fediverse by https://fed.brid.gy/ ]
Lecturer in AI, Government & Policy at the OII (University of Oxford) | Author of "The Materiality of AI" (2027, Bristol University Press) | Associate Editor at Big Data & Society | Investigating algorithmic accountability and environmental impacts
researcher answer.ai
interested in education, immunology, & AI
fast.ai co-founder, math PhD, data scientist
Writing: https://rachel.fast.ai/