AI Alignment

25 articles about AI Alignment

AI safety research requires collaboration with social scientists

OpenAI
Article
AI & Machine Learning

A paper argues AI safety needs social scientists to help align advanced AI with real human values and behavior, and OpenAI plans to hire them to collaborate full time.

Key research problems in AI safety for modern machine learning systems

OpenAI
Report
AI & Machine Learning

A paper by Google Brain with Berkeley and Stanford co-authors outlines concrete research problems to ensure modern AI systems behave as intended.

AI safety technique using agent debates judged by humans

OpenAI
Article
AI Agents & Autonomous Workflows

An AI safety method that trains AI agents to debate each other while a human judge decides the winner.

OpenAI's technical goals for building safe and accessible AI

OpenAI
Article
AI & Machine Learning

OpenAI aims to build safe AI and share its benefits as widely and fairly as possible.

Algorithm infers human preferences to improve AI safety

OpenAI
Insight
AI & Machine Learning

An algorithm learns what humans want by comparing which of two behaviors people prefer, reducing reliance on hand-written AI goals for safer systems.

Showing page 2 of 2