Epistemic status: This is a step towards formalizing some intuitions about AI. It is closely related to Vanessa Kosoy’s “Descriptive Agent Theory” - but I want to concretize the question, explain the reason that it is true in some form, and try to think through and provide some intuition about...
Bottom line up front: If you are currently working on, or are interested working in any area of mathematical AI alignment, we are collecting names and basic contact information to find who to talk to about opportunities in these areas. If that describes you, please fill out the form! (Please...
Epistemic Status: Trying to clarify a confusion people outside of the AI safety community seem to have about what safety means for AI systems. In engineering and design, there is a process that includes, among other stages, specification, creation, verification and validation, and deployment. Verification and validation are where most...
ALTER, with the University of Haifa and the Technion, is excited to announce the details about its upcoming conference introducing AI safety in Israel. We’re going to be hosting Stuart Russell, as well as several other AI safety researchers, to speak about the field. We're also hosting a number of...
It has been suggested that in a rapid enough takeoff scenario, governance would not be useful, because the transition to superintelligence would be too rapid for human actors - whether governments, corporations, or individuals - to respond to. This seems to imply that we only care about takeoff speed. And...
This paper is a revised and expanded version of my blog post Plausible cases for HRAD work, and locating the crux in the "realism about rationality" debate, now with David Manheim as co-author. Abstract: > Several different approaches exist for ensuring the safety of future Transformative Artificial Intelligence (TAI) or...