AI Alignment
25 articles about AI Alignment
AI safety research requires collaboration with social scientists
OpenAI
Article
AI & Machine Learning
A paper argues AI safety needs social scientists to help align advanced AI with real human values and behavior, and OpenAI plans to hire them to collaborate full time.
Key research problems in AI safety for modern machine learning systems
OpenAI
Report
AI & Machine Learning
A paper by Google Brain with Berkeley and Stanford co-authors outlines concrete research problems to ensure modern AI systems behave as intended.
AI safety technique using agent debates judged by humans
OpenAI
Article
AI Agents & Autonomous Workflows
An AI safety method that trains AI agents to debate each other while a human judge decides the winner.
OpenAI's technical goals for building safe and accessible AI
OpenAI
Article
AI & Machine Learning
OpenAI aims to build safe AI and share its benefits as widely and fairly as possible.
Algorithm infers human preferences to improve AI safety
OpenAI
Insight
AI & Machine Learning
An algorithm learns what humans want by comparing which of two behaviors people prefer, reducing reliance on hand-written AI goals for safer systems.
Showing page 2 of 2