AI as a science, and three obstacles to alignment strategies
By So8res @ 2023-10-25T21:02 (+41)
This is a crosspost, probably from LessWrong. Try viewing it there.
nullSummaryBot @ 2023-10-26T13:25 (+1)
Executive summary: The post argues that AI alignment is difficult due to the intertwining of alignment and capabilities research, the difficulty of distinguishing real vs. fake solutions, and the high risk of theories failing on their first critical application.
Key points:
- Efforts to understand and aim AI systems often also uncover ways to make them more capable, allowing unchecked development to continue.
- Distinguishing scientifically-grounded alignment solutions from insufficient ones will be very difficult for regulators.
- Even rigorous alignment theories may fail catastrophically the first time they are tested in a real superintelligence.
- Recommends civilization pursue non-AI routes to transhumanism like uploading instead.
- Developing scientific theories of artificial cognition is important but risks accelerating progress.
- Minimal pivotal tasks are worth considering to contain failures of new alignment theories.
This comment was auto-generated by the EA Forum Team. Feel free to point out issues with this summary by replying to the comment, and contact us if you have feedback.