Skip to main navigation Skip to search Skip to main content

Democratizing value alignment: from authoritarian to democratic AI ethics

  • Linus Ta-Lun HUANG
  • , Gleb PAPYSHEV*
  • , James K. WONG
  • *Corresponding author for this work

Research output: Journal PublicationsJournal Article (refereed)peer-review

Abstract

Value alignment is essential for ensuring that AI systems act in ways that are consistent with human values. Existing approaches, such as reinforcement learning with human feedback and constitutional AI, however, exhibit power asymmetries and lack transparency. These “authoritarian” approaches fail to adequately accommodate a broad array of human opinions, raising concerns about whose values are being prioritized. In response, we introduce the Dynamic Value Alignment approach, theoretically grounded in the principles of parallel constraint satisfaction, which models moral reasoning as a dynamic process that balances multiple value principles. Our approach also enhances users’ moral and epistemic agency by granting users greater control over the values that influence AI behavior. As a more user-centric, transparent, and participatory framework for AI ethics, our approach not only addresses the democratic deficits inherent in current practices but also ensures that AI systems are flexibly aligned with a diverse array of human values.
Original languageEnglish
Pages (from-to)11-18
Number of pages8
JournalAI & Ethics
Volume5
Early online date2 Dec 2024
DOIs
Publication statusPublished - Feb 2025
Externally publishedYes

Funding

No funding was received for conducting this study. Open access funding provided by Hong Kong University of Science and Technology

Keywords

  • Authoritarian AI ethics
  • Value alignment
  • Large language models (LLMs)
  • Democratization

Fingerprint

Dive into the research topics of 'Democratizing value alignment: from authoritarian to democratic AI ethics'. Together they form a unique fingerprint.

Cite this