This website requires javascript to properly function. Consider activating javascript to get access to all site functionality.
AI ALIGNMENT FORUM
Tags
AF
Login
Power Seeking (AI)
•
Applied to
Natural Abstraction: Convergent Preferences Over Information Structures
by
paulom
2mo
ago
•
Applied to
You can't fetch the coffee if you're dead: an AI dilemma
by
hennyge
3mo
ago
•
Applied to
The Game of Dominance
by
Karl von Wendt
3mo
ago
•
Applied to
Incentives from a causal perspective
by
Tom Everitt
5mo
ago
•
Applied to
George Hotz on AI safety: ~"centralized power is bad"
by
Chipmonk
5mo
ago
•
Applied to
Instrumental Convergence? [Draft]
by
Dan H
6mo
ago
•
Applied to
Categorical-measure-theoretic approach to optimal policies tending to seek power
by
Victoria Krakovna
6mo
ago
•
Applied to
My Overview of the AI Alignment Landscape: Threat Models
by
Michelle Viotti
7mo
ago
•
Applied to
Ideas for studies on AGI risk
by
dr_s
7mo
ago
•
Applied to
Instrumental convergence in single-agent systems
by
Jacob Pfau
8mo
ago
•
Applied to
Risks from GPT-4 Byproduct of Recursively Optimizing AIs
by
Ben Hayum
8mo
ago
•
Applied to
[Linkpost] Shorter version of report on existential risk from power-seeking AI
by
Ruben Bloom
8mo
ago
•
Applied to
The Waluigi Effect (mega-post)
by
Cleo Nardo
9mo
ago
•
Applied to
Power-seeking can be probable and predictive for trained agents
by
Victoria Krakovna
9mo
ago
•
Applied to
Power-Seeking = Minimising free energy
by
Jonas Hallgren
9mo
ago
•
Applied to
Parametrically retargetable decision-makers tend to seek power
by
Alex Turner
9mo
ago