Alignment

Ensuring AI goals match human values

EthicsRisk
Back to Glossary
Updated 2 May 2025

Definition

Alignment in AI refers to the process of ensuring that an artificial intelligence system's goals and behaviors closely match human values and intentions. Proper alignment reduces the risk of unintended, harmful outcomes by guiding AI behavior toward socially beneficial objectives.