Earth is ruled by master-machines but the Three Laws of Robotics have been designed to ensure humans maintain the upper hand: 1) A robot may not injure a human being or allow a human being to come to harm 2) A robot must obey orders given to it by human beings except where such orders would conflict with the First Law. 3) A robot must protect its own existence as long as such protection does not conflict with the First or Second Law. But what happens when a rogue robot's idea of what is good for society contravenes the Three Laws?