« Back to Glossary Index

AI alignment is the process of ensuring that AI systems operate in a way that aligns with what humans consider desirable behavior. It involves specifying the goals and values of the AI system, and making sure that the system follows them robustly. AI alignment is important for building safe and beneficial AI systems that can cooperate with humans and avoid harmful outcomes.

See also “Our approach to alignment research – OpenAI”.


« Back to Glossary Index