Random Tag
Contributors
You are viewing revision 1.0.0, last edited by Ben Pace

Outer Alignment in the context of machine learning is the property where the specified loss function is aligned with the intended goal of its designers. This is an intuitive notion, in part because human intentions are themselves not well-understood. This is what is typically discussed as the 'value alignment' problem. It is contrasted with inner alignment, which discusses if an optimizer is the production of a outer aligned system, then whether that optimizer is itself aligned.