Sitemap
The Academic

The Academic is a top tier, peer-reviewed publication on Medium brought to you by a global community of subject-matter experts.

Member-only story

AI Alignment Is Impossible

9 min readApr 26, 2025

--

Image: ChatGPT, author generated

The “alignment problem” refers to the difficulty of ensuring that Artificial Intelligence systems behave in a manner consistent with the goals of their human operators. There is an everyday, practical dimension to this problem, manifest in such challenges as ensuring that an AI chatbot does not help people manufacture illegal drugs or spread hate speech. There is also the much broader philosophical or theoretical difficulty of how to ensure that future super-AIs with capacities that far exceed our own don’t misuse those capabilities to harm us, whether through “going rogue”, developing autonomous goals at odds with human wellbeing, or simply behaving in unexpected harmful ways in the course of trying to meet their objectives.

The classic example invented by Nick Bostrom in an early paper on the AI alignment problem is that of a hypothetical super AI trained on the goal of maximising paperclip production. In the thought experiment, Bostrom imagined such an AI turning the entire galaxy into paperclips, ruthlessly destroying humanity in the monomaniacal pursuit of its singular objective: more paperclips.

The idea is amusing and silly but highlights a real issue. When AIs are trained to pursue a particular goal, they may end up finding means that subvert the underlying…

--

--

The Academic
The Academic

Published in The Academic

The Academic is a top tier, peer-reviewed publication on Medium brought to you by a global community of subject-matter experts.

Pierz Newton-John
Pierz Newton-John

Written by Pierz Newton-John

Writer, coder, former psychotherapist, founding member of The School Of Life Melbourne. Essayist for Dumbo Feather magazine, author of Fault Lines (fiction).

Responses (12)