Alignment

Ensuring AI goals match human values

EthicsRisk
Updated 2 May 2025·Reviewed

Definition

Alignment in AI refers to the process of ensuring that an artificial intelligence system's goals and behaviors closely match human values and intentions. Proper alignment reduces the risk of unintended, harmful outcomes by guiding AI behavior toward socially beneficial objectives.

All TermsBack to GlossaryNext TermAlignment Drift