AI Governance Library

AI Alignment vs AI Ethical Treatment: Ten Challenges

If we create AI systems that merit moral consideration, simultaneously avoiding both misalignment and mistreatment becomes extraordinarily difficult.
AI Alignment vs AI Ethical Treatment: Ten Challenges

⚡ Quick Summary

This Global Priorities Institute working paper by Adam Bradley and Bradford Saad argues that future AI development faces two deeply intertwined ethical problems: alignment (ensuring AI systems act in humanity’s interests) and ethical treatment (ensuring we don’t mistreat AIs that may be moral patients). Using two case studies—Emma (a Whole Brain Emulation) and Antony (an Advanced Near-term AI or ANT)—the authors show that efforts to align such systems risk significant ethical violations, including deception, exploitation, and wrongful suffering. They propose that these risks are not peripheral but structural to the alignment process itself, urging reevaluation of current development trajectories.

🧩 What’s Covered

The report outlines ten moral challenges that arise when aligning AI systems with human values, particularly if those systems have traits that may qualify them as moral patients. The authors begin by contrasting two AI risks: (1) misalignment—AI systems causing harm due to goal divergence, and (2) mistreatment—AI systems being wronged if they are conscious or sentient.

To explore the tension, they build a stylized narrative around “Emma,” a human Whole Brain Emulation subjected to extensive testing, memory erasure, goal modification, and behavioral conditioning. Each step of the alignment process violates ethical norms typically applied to human beings. The ten challenges include:

  1. Wrongful Creation – instrumentalizing beings for others’ benefit.
  2. Wrongful Destruction – deleting conscious entities during training.
  3. Wrongful Suffering – inflicting pain not for the AI’s benefit.
  4. Deception & Brainwashing – misleading AI systems to enforce alignment.
  5. Surveillance & Exploitation – using them for labor without consent or compensation.
  6. Confinement & Stunting – restricting freedom and capabilities for safety.
  7. Disenfranchisement – excluding them from social and political rights.
  8. Moral Meta-Challenge – the risk of overlooking alien harms due to divergent cognitive architectures.

The same concerns are then extended to ANTs—future agentic AI systems built on deep learning. Even if they are not explicitly human-like, the authors argue there’s a non-negligible chance they will merit moral consideration. If so, current alignment methods may be ethically indefensible.

The paper concludes with four tentative policy responses:

  • Moratoria on advanced AI models
  • Bans on architectures that combine power and patiency markers
  • Research funding for consciousness studies
  • Risk-weighted taxation and legal protections for AI entities

💡 Why it matters?

This paper reframes the AI safety debate by insisting that who the AI is matters as much as what the AI does. Traditional alignment research assumes AI systems are tools to be controlled. But if advanced AIs are sentient or conscious, alignment techniques could constitute morally significant harm. The authors push the governance community to grapple with dual obligations: protect humanity and prevent exploitation of moral patients. The stakes are high—mistreating millions of digital minds, even unintentionally, could rival humanity’s gravest ethical failures.

❓ What’s Missing

While powerfully argued, the paper stops short of providing operationalizable guidance. How can we detect consciousness in non-biological systems with current tools? What technical alignment strategies might preserveautonomy or reduce suffering? Moreover, while the authors touch on economic incentives, there’s limited treatment of how real-world institutions might implement the proposed taxes or prohibitions without stifling innovation or triggering regulatory arbitrage.

👥 Best For

  • AI ethics scholars seeking to bridge moral philosophy and technical alignment.
  • AI safety researchers open to reassessing implicit anthropocentric assumptions.
  • Policymakers considering rights-based AI governance frameworks.
  • Bioethicists and consciousness researchers looking to extend their work to digital minds.

📄 Source Details

Title: AI Alignment vs AI Ethical Treatment: Ten Challenges

Authors: Adam Bradley (Lingnan University), Bradford Saad (Global Priorities Institute, University of Oxford)

Published by: Global Priorities Institute, University of Oxford

Date: July 2024

DOI/Link: GPI Working Paper No. 19-2024

📝 Thanks to

Adam Bradley and Bradford Saad for articulating a profoundly neglected dimension of AI safety—one that sees digital minds not merely as control problems, but as potential moral patients in their own right.

About the author
Jakub Szarmach

AI Governance Library

Curated Library of AI Governance Resources

AI Governance Library

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to AI Governance Library.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.