Abstract
The alignment of artificial intelligence (AI) systems with societal values and the public interest is a critical challenge in the field of AI ethics and governance. Traditional approaches, such as Reinforcement Learning with Human Feedback (RLHF) and Constitutional AI, often rely on pre-defined high-level ethical principles. This article critiques these conventional alignment frameworks through the philosophical perspectives of pragmatism and public interest theory, arguing against their rigidity and disconnect with practical impacts. It proposes an alternative alignment strategy that reverses the traditional logic, focusing on empirical evidence and the real-world effects of AI systems. By emphasizing practical outcomes and continuous adaptation, this pragmatic approach aims to ensure that AI technologies are developed according to the principles that are derived from the observable impacts produced by technology applications.
| Original language | English |
|---|---|
| Article number | e30 |
| Number of pages | 15 |
| Journal | Data and Policy |
| Volume | 7 |
| Early online date | 10 Mar 2025 |
| DOIs | |
| Publication status | Published - 2025 |
| Externally published | Yes |
Bibliographical note
Publisher Copyright:© The Author(s), 2025. Published by Cambridge University Press.
Funding
No funding was received to conduct this study.
Keywords
- AI alignment
- constitutional AI
- pragmatism
- public interest
- reinforcement learning with human feedback