Seth Lazar's Avatar

Seth Lazar

@sethlazar.org.bsky.social

Philosopher working on normative dimensions of computing and sociotechnical AI safety. Lab: https://mintresearch.org Self: https://sethlazar.org Newsletter: https://philosophyofcomputing.substack.com

5,141 Followers  |  1,109 Following  |  129 Posts  |  Joined: 06.06.2023  |  1.8926

Latest posts by sethlazar.org on Bluesky

Preview
Artificial Intelligence and Democratic Freedoms

@caseynewton.bsky.social in re an old discussion about AI denialists. , hope you’ve caught knightcolumbia.org/events/artif...

11.04.2025 15:59 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

🚨 UPCOMING EVENT: Artificial Intelligence and Democratic Freedoms, April 10-11 at @columbiauniversity.bsky.social & online. In collaboration with Senior AI Advisor @sethlazar.org & co-sponsored by the Knight Institute and @columbiaseas.bsky.social. RSVP: knightcolumbia.org/events/artif...

28.02.2025 16:38 β€” πŸ‘ 26    πŸ” 12    πŸ’¬ 1    πŸ“Œ 1
Preview
Normative Philosophy of Computing Newsletter Welcome to February!

New Philosophy of Computing newsletter: share with your philosophy friends. Lots of CFPs, events, opportunities, new papers.

philosophyofcomputing.substack.com/p/normative-...

25.02.2025 05:06 β€” πŸ‘ 8    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Preview
These Strange New Minds Stunning advances in digital technology have given us a new wave of disarmingly human-like AI systems. The march of this new technology is set to upturn our economies, challenge our democracies, and r...

I am a bit bashful about sharing this profile www.thetimes.com/uk/technolog... of me in @thetimes.com, but will do so because it kindly refers to my new book which is coming out in early March. www.penguin.co.uk/books/460891.... The tech titans pictured seem to be decoration (and not my co-authors)

22.02.2025 14:41 β€” πŸ‘ 59    πŸ” 12    πŸ’¬ 4    πŸ“Œ 0
Graph of web tasks along difficulty and severity (cost of errors)

Graph of web tasks along difficulty and severity (cost of errors)

I spent a few hours with OpenAI's Operator automating expense reports. Most corporate jobs require filing expenses, so Operator could save *millions* of person-hours every year if it gets this right.

Some insights on what worked, what broke, and why this matters for the future of agents 🧡

03.02.2025 18:04 β€” πŸ‘ 34    πŸ” 10    πŸ’¬ 6    πŸ“Œ 3
LM Agents: Prospects and Impacts (FAccT tutorial)
YouTube video by Seth Lazar LM Agents: Prospects and Impacts (FAccT tutorial)

Since Agents are now on everyone's minds, do check out this tutorial on the ethics of Language Model Agents, from June last year.

Looks at what 'agent' means, how LM agents work, what kinds of impacts we should expect, and what norms (and regulations) should govern them.

24.01.2025 07:29 β€” πŸ‘ 14    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Preview
Knight Institute Symposium on AI and Democratic Freedoms to Feature Leading Scholars and Technologists

We're excited to announce that our upcoming symposium on #AI and democracy w/ @sethlazar.org (4/10-4/11, at @columbiauniversity.bsky.social & online) will feature papers by a highly accomplished group of authors from a wide range of disciplines. Check them out: knightcolumbia.org/blog/knight-...

23.01.2025 15:03 β€” πŸ‘ 8    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Preview
Normative Philosophy of Computing - January Happy New Year!

January update from the normative philosophy of computing newsletter: new CFPs, papers, workshops, and resources for philosophers working on normative questions raised by AI and computing.

16.01.2025 06:48 β€” πŸ‘ 17    πŸ” 5    πŸ’¬ 1    πŸ“Œ 0
Preview
Artificial Intelligence and Democratic Freedoms

EVENT: Artificial Intelligence and Democratic Freedoms, 4/10-11, at @columbiauniversity.bsky.social & online. We're hosting a symposium w/ @sethlazar.org exploring the risks advanced #AI systems pose to democratic freedoms and interventions to mitigate them. RSVP: knightcolumbia.org/events/artif...

09.01.2025 21:08 β€” πŸ‘ 19    πŸ” 5    πŸ’¬ 0    πŸ“Œ 1

πŸ“’ Excited to share: I'm again leading the efforts for the Responsible AI chapter for Stanford's 2025 AI Index, curated by @stanfordhai.bsky.social. As last year, we're asking you to submit your favorite papers on the topic for consideration (including your own!) 🧡 1/

05.01.2025 17:42 β€” πŸ‘ 13    πŸ” 8    πŸ’¬ 1    πŸ“Œ 0
Preview
Trying out QvQβ€”Qwen’s new visual reasoning model I thought we were done for major model releases in 2024, but apparently not: Alibaba’s Qwen team just dropped the Apache2 2 licensed QvQ-72B-Preview, β€œan experimental research model focusing on …

Turns out we weren't done for major LLM releases in 2024 after all... Alibaba's Qwen just released QvQ, a "visual reasoning model" - the same chain-of-thought trick as OpenAI's o1 applied to running a prompt against an image

Trying it out is a lot of fun: simonwillison.net/2024/Dec/24/...

24.12.2024 20:52 β€” πŸ‘ 174    πŸ” 27    πŸ’¬ 5    πŸ“Œ 1
deepseek-ai/DeepSeek-V3-Base No model card or announcement yet, but this new model release from Chinese AI lab DeepSeek (an arm of Chinese hedge fund [High-Flyer](https://en.wikipedia.org/wiki/High-Flyer_(company))) looks very si...

Here are my collected notes on DeepSeek v3 so far: simonwillison.net/2024/Dec/25/...

25.12.2024 19:03 β€” πŸ‘ 37    πŸ” 8    πŸ’¬ 3    πŸ“Œ 1
Preview
deepseek-ai/DeepSeek-V3-Base Β· Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

deepseek-ai/DeepSeek-V3-Base

huggingface.co/deepseek-ai/...

25.12.2024 15:52 β€” πŸ‘ 35    πŸ” 4    πŸ’¬ 1    πŸ“Œ 2

Defo @natolambert.bsky.social but you’ll already know that :)

27.12.2024 02:20 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Did OpenAI Just Solve Abstract Reasoning? OpenAI’s o3 model aces the "Abstraction and Reasoning Corpus" β€” but what does it mean?

Some of my thoughts on OpenAI's o3 and the ARC-AGI benchmark

aiguide.substack.com/p/did-openai...

23.12.2024 14:38 β€” πŸ‘ 342    πŸ” 99    πŸ’¬ 17    πŸ“Œ 27
Post image Post image

OpenAI skips o2, previews o3 scores, and they're truly crazy. Huge progress on the few benchmarks we think are truly hard today. Including ARC AGI.
Rip to people who say any of "progress is done," "scale is done," or "llms cant reason"
2024 was awesome. I love my job.

20.12.2024 18:08 β€” πŸ‘ 113    πŸ” 14    πŸ’¬ 11    πŸ“Œ 5
Preview
o3: The grand finale of AI in 2024 A step change as influential as the release of GPT-4. Reasoning language models are the current big thing.

OpenAI's o3: The grand finale of AI in 2024
A step change as influential as the release of GPT-4. Reasoning language models are the current and next big thing.

I explain:
* The ARC prize
* o3 model size / cost
* Dispelling training myths
* Extreme benchmark progress

20.12.2024 23:34 β€” πŸ‘ 82    πŸ” 12    πŸ’¬ 8    πŸ“Œ 1

There's quite a lot! But it's all over on the other place. Crickets here...

26.12.2024 08:46 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I'm not seeing (here) much discussion of o3. If you are, point me to who's on here that I'm missing? If you're not: just registering that o3's performance on SWE-bench verified is *bananas*, and likely to have massive impacts in 2025.

26.12.2024 06:04 β€” πŸ‘ 20    πŸ” 1    πŸ’¬ 5    πŸ“Œ 0
Video thumbnail
23.12.2024 03:04 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail
23.12.2024 03:04 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail
23.12.2024 03:04 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image Post image Post image

Busy shopping day in Causeway Bay (long exposures handheld with Spectre App)

23.12.2024 03:04 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Feeling good (after o3) about some of the bets made in these papers… Human level software agents now seem nailed on for the near-term.

21.12.2024 05:30 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Two papers on anticipating and evaluating AI agent impacts now ready for (private) comments: if you're interested in how language agents might reshape democracy, or in how *platform agents* might intensify the worst features of the platform economy (but could also fix it), lmk.

20.12.2024 08:28 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 3    πŸ“Œ 1
Second Draft of the General-Purpose AI Code of Practice published, written by independent experts Independent experts present the second draft of the General-Purpose AI Code of Practice, based on the feedback received on the first draft, published on 14 November 2024.

As one of the vice chairs of the EU GPAI Code of Practice process, I co-wrote the second draft which just went online – feedback is open until mid-January, please let me know your thoughts, especially on the internal governance section!

digital-strategy.ec.europa.eu/en/library/s...

19.12.2024 16:59 β€” πŸ‘ 14    πŸ” 5    πŸ’¬ 0    πŸ“Œ 1

πŸ€”

20.12.2024 01:01 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Title card: Alignment Faking in Large Language Models by Greenblatt et al.

Title card: Alignment Faking in Large Language Models by Greenblatt et al.

New work from my team at Anthropic in collaboration with Redwood Research. I think this is plausibly the most important AGI safety result of the year. Cross-posting the thread below:

18.12.2024 17:46 β€” πŸ‘ 126    πŸ” 29    πŸ’¬ 7    πŸ“Œ 11

We're working hard behind the scenes to finalize the dates and venue for #FAccT2025! While final confirmation is still pending, our tentative conference dates are June 23-26. Expect more updates soon!

18.12.2024 15:54 β€” πŸ‘ 15    πŸ” 5    πŸ’¬ 0    πŸ“Œ 0

@sethlazar.org is following 19 prominent accounts