@ojdaniels.bsky.social, @jessicaji.bsky.social, @jacob-feldgoise.bsky.social, and @alicrawford.bsky.social weigh in on:
• Open questions raised by the Plan
• Security-related recommendations
• Export controls
• Workforce priorities
@ojdaniels.bsky.social
Writing on AI, security, & democracy. Associate Director of Analysis & Andrew W. Marshall Fellow at CSET. Probably behind deadline. Working on a book on AI and military affairs for Polity Press. https://cset.georgetown.edu/staff/owen-daniels/
@ojdaniels.bsky.social, @jessicaji.bsky.social, @jacob-feldgoise.bsky.social, and @alicrawford.bsky.social weigh in on:
• Open questions raised by the Plan
• Security-related recommendations
• Export controls
• Workforce priorities
“If Washington’s new AI strategy does not adequately account for open models, American AI companies, despite their world-leading models, will risk ceding international AI influence to China,” write @ojdaniels.bsky.social and @hannadohmen.bsky.social.
25.07.2025 18:44 — 👍 9 🔁 5 💬 1 📌 1
The new AI Action Plan touts the strategic importance of open models.
That's not news to China, where developers like DeepSeek have made the PRC the leader in open-weights AI.
In @foreignaffairs.com, CSET's @ojdaniels.bsky.social & @hannadohmen.bsky.social explain how the U.S. can respond.
I’m proud to share @csetgeorgetown.bsky.social’s Annual Report!
From Congressional testimony and groundbreaking research to essential data tools & translations, our team continues to shape critical emerging tech policy conversations.
Grateful to everyone who makes this work possible.
As AI continues to reshape our world, CSET remains committed to providing data-driven analysis on the security implications of emerging technologies.
Read our full response: cset.georgetown.edu/publication/...
The future of AI leadership requires thoughtful policy. @CSETGeorgetown just submitted our response to @NSF's RFI on the Development of an Artificial Intelligence Action Plan. Here's what we recommend: 🧵[1/]
17.03.2025 13:28 — 👍 5 🔁 5 💬 1 📌 0Mia Hoffmann and Owen J. Daniels from Georgetown’s Center for Security and Emerging Technology say Europe's apparent shift on AI policy could change the global landscape for AI governance. www.techpolicy.press/out-of-balan...
10.03.2025 22:58 — 👍 21 🔁 3 💬 2 📌 0
What does the EU's shifting strategy mean for AI?
CSET's @miahoffmann.bsky.social & @ojdaniels.bsky.social have a new piece out for @techpolicypress.bsky.social.
Read it now 👇
🇪🇺🤖🐘
10.03.2025 14:05 — 👍 2 🔁 0 💬 0 📌 0
U.S. AI firms, take note! An op-ed by @sambresnick.bsky.social and @colemcfaul.bsky.social in @barrons.com argues that China's "good enough" AI strategy (think DeepSeek) could disrupt the market.
It's about affordable tech. Huawei playbook 2.0? Open-source Chinese models are closing the gap, fast.
Are definitional clarifications dramatically boring? Yes. Are they policy relevant? Also yes. Some thoughts on conflating AI governance, safety, & regulation, informed by @csetgeorgetown.bsky.social work.
TLDR: regulation !=safety
safety != "hand-wringing" www.linkedin.com/pulse/some-t...
📅Webinar today📅
At 3pm ET today, February 24, join AllaSeiffert of AWS, CSET’s
@moschinski.bsky.social, and Elizabeth Laird of @cendemtech.bsky.social for a discussion on how the U.S. government uses and acquires AI talent and tools. cset.georgetown.edu/event/how-th...
It’s been a busy month for AI policy. CSET's @alexfriedland.bsky.social recapped everything you need to know:
1️⃣ Trump’s AI strategy
2️⃣ The Paris AI Summit
3️⃣ OpenAI’s releases and the roadmap to GPT-5
4️⃣ Sifting through the DeepSeek implications
And much more! cset.georgetown.edu/newsletter/f...
DeepResearch prompt: "'The End of History and the Last Man' but make it AI"
20.02.2025 19:34 — 👍 1 🔁 0 💬 0 📌 0I've left a lot out: the profit opportunities and risks of increasingly agentic systems got a lot of air; emerging research on scheming, sabotage, and survival instincts of LLMs and frontier models was prominent; and practical ethics policy ideas abounded. Looking forward to sharing more ideas soon
13.02.2025 16:41 — 👍 2 🔁 0 💬 0 📌 0Across the risk spectrum, the question arose time and again: where do we actually need AI solutions? Is it actually helpful to have AI to try to help us find common ground around political disagreements, for example? Do we want more tech in our democratic processes? www.science.org/doi/10.1126/...
13.02.2025 16:41 — 👍 2 🔁 0 💬 1 📌 0France’s announcement at the summit that it was tapping its nuclear power industry for data centers grabbed headlines, but nuclear power is not necessarily a panacea for all of AI’s energy issues. It remains a globally significant space to watch. thebulletin.org/2024/12/ai-g...
13.02.2025 16:41 — 👍 2 🔁 0 💬 1 📌 0Environmental and energy concerns will only continue to grow with scaling, and rightfully earned much discussion. Even with model innovation’s like DeepSeek R1, which is cheaper and more efficient to train, consumption for inference will remain high.
13.02.2025 16:41 — 👍 0 🔁 0 💬 1 📌 0The AISIs have different structures and stakeholders and are attuned to particular research ecosystems, meaning they're not 1-1 matches from one nation to the next, but they can still facilitate exchange. They'll obviously face some geopolitical headwinds amid tech competition.
13.02.2025 16:41 — 👍 0 🔁 0 💬 1 📌 0Despite disappointment at executive messaging, the AI Safety Institutes leading safety work at the national level could be ideal vehicles for developing and disseminating testing, evaluation, and safety best practices. Saw some impressive presentations at side events www.aisi.gov.uk/work/safety-...
13.02.2025 16:41 — 👍 0 🔁 0 💬 1 📌 0JD Vance's comments on Europe's "excessive regulation" were well covered, but EC Pres von der Leyen and Macron also championed getting out of the private sector's way. My colleague @miahoffmann.bsky.social wrote a thread about why this attitude could be troubling for Europe bsky.app/profile/miah...
13.02.2025 16:41 — 👍 1 🔁 0 💬 1 📌 0
A few thoughts on the outcomes of the AI Action Summit in Paris. The summit laid out some grand goals for AI governance (and covered them at length in the civil society portion), but the government-led portion of the summit was largely about AI enthusiasm.
www.linkedin.com/pulse/ai-act...
There have been a ton of AI policy developments coming out of the EU these past weeks, but one deeply concerning one is the withdrawal of the AI Liability Directive (AILD) by the European Commission. Here’s why:
13.02.2025 15:35 — 👍 4 🔁 3 💬 1 📌 1Great, thought-provoking discussion and panels at AI Safety Connect in Paris today. How various national iterations of AI safety institutes cooperate (or don’t) in the year ahead will be a major area to watch.
09.02.2025 22:49 — 👍 1 🔁 0 💬 0 📌 0@miahoffmann.bsky.social , @ojdaniels.bsky.social, and I wrote a piece on key AI governance areas to watch in 2025 with the upcoming AI Action Summit in mind. Check it out here! thebulletin.org/2025/02/will...
07.02.2025 03:00 — 👍 5 🔁 3 💬 0 📌 0Read their piece for @thebulletin.org: thebulletin.org/2025/02/will...
06.02.2025 14:39 — 👍 2 🔁 1 💬 0 📌 0
Next week, Paris will host world leaders, experts, and CEOs of some of the biggest AI developers at the #AIActionSummit.
Whether they can agree on a unified approach to AI governance remains to be seen, write
@miahoffmann.bsky.social, @minanrn.bsky.social, and
@ojdaniels.bsky.social.
They also fail to advance progress on more ambitious goals, like safety measures, energy efficiency, or labor standards in data supply-chains. Legislators need to think more on how governments and firms can ensure voluntary steps entail real, significant investments toward harnessing AI responsibly.
06.02.2025 12:13 — 👍 0 🔁 0 💬 0 📌 0Last but not least, voluntary commitments: voluntary measures that simply reframe business practices as governance risk coming off as participation trophies.
06.02.2025 12:12 — 👍 0 🔁 0 💬 1 📌 0The EU AI Act has inspired some of this state legislation and is a global role model. As provisions go into enforcement imminently, it may serve as a role model for enforcing, not just developing, legislation
06.02.2025 12:06 — 👍 0 🔁 0 💬 1 📌 0