Zach Stein-Perlman

AI strategy & governance. ailabwatch.org.

Posts

Introducing AI Lab Watch
by Zach Stein-Perlman @ 2024-04-30 | +114 | 0 comments
Staged release
by Zach Stein-Perlman @ 2024-04-20 | +16 | 0 comments
DeepMind: Evaluating Frontier Models for Dangerous Capabilities
by Zach Stein-Perlman @ 2024-03-21 | +28 | 0 comments
OpenAI: Preparedness framework
by Zach Stein-Perlman @ 2023-12-18 | +24 | 0 comments
AI governance & China: Reading list
by Zach Stein-Perlman @ 2023-12-18 | +14 | 0 comments
Zach Stein-Perlman's Quick takes
by Zach Stein-Perlman @ 2023-11-06 | +7 | 0 comments
Anthropic, Google, Microsoft & OpenAI announce Executive Director of the...
by Zach Stein-Perlman @ 2023-10-25 | +32 | 0 comments
ARC Evals: Responsible Scaling Policies
by Zach Stein-Perlman @ 2023-09-28 | +16 | 0 comments
Anthropic's Responsible Scaling Policy & Long-Term Benefit Trust
by Zach Stein-Perlman @ 2023-09-19 | +25 | 0 comments
US public opinion on AI, September 2023
by Zach Stein-Perlman @ 2023-09-18 | +29 | 0 comments
How to think about slowing AI
by Zach Stein-Perlman @ 2023-09-17 | +74 | 0 comments
Cruxes on US lead for some domestic AI regulation
by Zach Stein-Perlman @ 2023-09-10 | +20 | 0 comments
AI labs' requests for input
by Zach Stein-Perlman @ 2023-08-19 | +7 | 0 comments
Frontier Model Forum
by Zach Stein-Perlman @ 2023-07-26 | +40 | 0 comments
My favorite AI governance research this year so far
by Zach Stein-Perlman @ 2023-07-23 | +81 | 0 comments
Incident reporting for AI safety
by Zach Stein-Perlman, SeLo, stepanlos, MvK @ 2023-07-19 | +18 | 0 comments
Frontier AI Regulation
by Zach Stein-Perlman @ 2023-07-10 | +56 | 0 comments
AI labs' statements on governance
by Zach Stein-Perlman @ 2023-07-04 | +28 | 0 comments
DeepMind: Model evaluation for extreme risks
by Zach Stein-Perlman @ 2023-05-25 | +48 | 0 comments
AI strategy career pipeline
by Zach Stein-Perlman @ 2023-05-22 | +72 | 0 comments
GovAI: Towards best practices in AGI safety and governance: A survey of expert...
by Zach Stein-Perlman @ 2023-05-15 | +68 | 0 comments
Ideas for AI labs: Reading list
by Zach Stein-Perlman @ 2023-04-24 | +28 | 0 comments
AI policy ideas: Reading list
by Zach Stein-Perlman @ 2023-04-17 | +60 | 0 comments
FLI report: Policymaking in the Pause
by Zach Stein-Perlman @ 2023-04-15 | +29 | 0 comments
FLI open letter: Pause giant AI experiments
by Zach Stein-Perlman @ 2023-03-29 | +220 | 0 comments
Operationalizing timelines
by Zach Stein-Perlman @ 2023-03-10 | +30 | 0 comments
Framing AI strategy
by Zach Stein-Perlman @ 2023-02-07 | +16 | 0 comments
AI safety milestones?
by Zach Stein-Perlman @ 2023-01-23 | +6 | 0 comments
Will MacAskill: The Beginning of History
by Zach Stein-Perlman @ 2022-08-13 | +36 | 0 comments
2022 AI expert survey results
by Zach Stein-Perlman @ 2022-08-04 | +88 | 0 comments
Ajeya's TAI timeline shortened from 2050 to 2040
by Zach Stein-Perlman @ 2022-08-03 | +59 | 0 comments
More funding is really good
by Zach Stein-Perlman @ 2022-06-25 | +46 | 0 comments
Choosing causes re Flynn for Oregon
by Zach Stein-Perlman @ 2022-05-18 | +50 | 0 comments
Great Power Conflict
by Zach Stein-Perlman @ 2021-09-15 | +11 | 0 comments
The Governance Problem and the "Pretty Good" X-Risk
by Zach Stein-Perlman @ 2021-08-28 | +23 | 0 comments
Peter Singer – Good Charity, Bad Charity
by Zach Stein-Perlman @ 2013-08-10 | +4 | 0 comments
Nick Bostrom – Existential Risk Prevention as Global Priority
by Zach Stein-Perlman @ 2013-02-01 | +15 | 0 comments
Scott Alexander – A Modest Proposal
by Zach Stein-Perlman @ 2008-11-26 | +19 | 0 comments
Peter Singer – Famine, Affluence, and Morality
by Zach Stein-Perlman @ 1972-03-01 | +42 | 0 comments