New podcast episode with Peter Salib and Simon Goldstein on their article ‘AI Rights for Human Safety’.
pnc.st/s/forecast/...
@forethought-org.bsky.social
Research nonprofit exploring how to navigate explosive AI progress. forethought.org
New podcast episode with Peter Salib and Simon Goldstein on their article ‘AI Rights for Human Safety’.
pnc.st/s/forecast/...
New podcast episode with @tobyord.bsky.social — on inference scaling, time horizons for AI agents, lessons from scientific moratoria, and more.
pnc.st/s/forecast/...
New report: “Will AI R&D Automation Cause a Software Intelligence Explosion?”
As AI R&D is automated, AI progress may dramatically accelerate. Skeptics counter that hardware stock can only grow so fast. But what if software advances alone can sustain acceleration?
x.com/daniel_2718...
Today we’re putting out our first paper, which gives an overview of these challenges. Read it here: www.forethought.org/research/pr...
11.03.2025 15:36 — 👍 1 🔁 0 💬 0 📌 1At Forethought, we’re doing research to help us understand the opportunities and challenges that AI-driven technological change will bring, and to help us figure out what we can do, now, to prepare.
11.03.2025 15:36 — 👍 2 🔁 0 💬 1 📌 0And this might happen blisteringly fast – our analysis suggests we’re likely to see a century’s worth of technological progress in less than a decade. Our current institutions were not designed for such rapid change. We need to prepare in advance.
11.03.2025 15:35 — 👍 1 🔁 0 💬 1 📌 0AI might help us to create many new technologies, and with them new opportunities – from economic abundance to enhanced collective decision-making – and new challenges – from extreme concentration of power to new weapons of mass destruction.
11.03.2025 15:35 — 👍 1 🔁 0 💬 1 📌 0Two years ago, AI systems were close to random guessing at PhD-level science questions. Now they beat human experts. As they continue to become smarter and more agentic, they may begin to significantly accelerate technological development. What happens next?
11.03.2025 15:35 — 👍 4 🔁 0 💬 1 📌 0