Matthew_Barnett

Posts

AI Rights for Human Safety
by Matthew_Barnett @ 2024-08-03 | +48 | 0 comments
Clarifying two uses of "alignment"
by Matthew_Barnett @ 2024-03-10 | +36 | 0 comments
Against most, but not all, AI risk analogies
by Matthew_Barnett @ 2024-01-14 | +43 | 0 comments
AI values will be shaped by a variety of forces, not just the values of AI...
by Matthew_Barnett @ 2024-01-11 | +70 | 0 comments
AI alignment shouldn't be conflated with AI moral achievement
by Matthew_Barnett @ 2023-12-30 | +111 | 0 comments
What is the current most representative EA AI x-risk argument?
by Matthew_Barnett @ 2023-12-15 | +117 | 0 comments
My thoughts on the social response to AI risk
by Matthew_Barnett @ 2023-11-01 | +116 | 0 comments
The possibility of an indefinite AI pause
by Matthew_Barnett @ 2023-09-19 | +90 | 0 comments
Updating Drexler's CAIS model
by Matthew_Barnett @ 2023-06-17 | +59 | 0 comments
A compute-based framework for thinking about the future of AI
by Matthew_Barnett @ 2023-05-31 | +96 | 0 comments
Why I think it's important to work on AI forecasting
by Matthew_Barnett @ 2023-02-27 | +179 | 0 comments
Slightly against aligning with neo-luddites
by Matthew_Barnett @ 2022-12-26 | +71 | 0 comments
Preventing a US-China war as a policy priority
by Matthew_Barnett @ 2022-06-22 | +64 | 0 comments
My current thoughts on the risks from SETI
by Matthew_Barnett @ 2022-03-15 | +47 | 0 comments
A proposal for a small inducement prize platform
by Matthew_Barnett @ 2021-06-05 | +66 | 0 comments
Matthew_Barnett's Quick takes
by Matthew_Barnett @ 2020-03-02 | +4 | 0 comments
Effects of anti-aging research on the long-term future
by Matthew_Barnett @ 2020-02-27 | +61 | 0 comments
Concerning the Recent 2019-Novel Coronavirus Outbreak
by Matthew_Barnett @ 2020-01-27 | +144 | 0 comments