Title
The Agi Containment Problem
Abstract
There is considerable uncertainty about what properties, capabilities and motivations future AGIs will have. In some plausible scenarios, AGIs may pose security risks arising from accidents and defects. In order to mitigate these risks, prudent early AGI research teams will perform significant testing on their creations before use. Unfortunately, if an AGI has human-level or greater intelligence, testing itself may not be safe; some natural AGI goal systems create emergent incentives for AGIs to tamper with their test environments, make copies of themselves on the internet, or convince developers and operators to do dangerous things. In this paper, we survey the AGI containment problem - the question of how to build a container in which tests can be conducted safely and reliably, even on AGIs with unknown motivations and capabilities that could be dangerous. We identify requirements for AGI containers, available mechanisms, and weaknesses that need to be addressed.
Year
DOI
Venue
2016
10.1007/978-3-319-41649-6_6
ARTIFICIAL GENERAL INTELLIGENCE (AGI 2016)
DocType
Volume
ISSN
Conference
9782
0302-9743
Citations 
PageRank 
References 
4
0.45
5
Authors
3
Name
Order
Citations
PageRank
James Babcock140.45
János Kramár2634.26
Roman V. Yampolskiy338249.79