{"id":1249,"date":"2011-06-03T19:00:18","date_gmt":"2011-06-04T03:00:18","guid":{"rendered":"http:\/\/agi-conf.org\/2011\/"},"modified":"2011-06-06T06:29:58","modified_gmt":"2011-06-06T14:29:58","slug":"bill-hibbard-abstract","status":"publish","type":"page","link":"https:\/\/agi-conf.org\/2011\/bill-hibbard-abstract\/","title":{"rendered":"Abstract: Bill Hibbard"},"content":{"rendered":"<p><strong>The End of Rough Equality of Intelligence<\/strong><\/p>\n<p><em>Bill Hibbard<\/em><\/p>\n<p>Abstract: Natural humans all have roughly the same intelligence: we speak the same languages (English speakers can learn French, etc) and share the same\u00a0daily skills. But with AI and human enhancement will come a divergence of intelligence levels of humans, or of the intelligence levels of machines serving\u00a0different humans. If a person&#8217;s intelligence depends on how large a brain they can afford and their wealth depends on their intelligence, this positive feedback\u00a0loop will create a divergence of intelligence levels. The most intelligent will speak languages that the less intelligent can never learn, which will freeze most out of any role in social decisions. It is an historical observation of human nature that power corrupts and absolute power corrupts absolutely.\u00a0 The greatest AI threat will be the teaming of the power of superior intelligence with corruptable human\u00a0nature. There will be political will for AI designs that prevent AI from taking control over all humans and prevent AI from wireheading, but politics are likely to work in favor of teaming AI with wealthy and powerful humans.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>The End of Rough Equality of Intelligence Bill Hibbard Abstract: Natural humans all have roughly the same intelligence: we speak the same languages (English speakers can learn French, etc) and share the same\u00a0daily skills. But with AI and human enhancement will come a divergence of intelligence levels of humans, or of the intelligence levels of [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"open","ping_status":"open","template":"","meta":{"footnotes":""},"class_list":["post-1249","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/agi-conf.org\/2011\/wp-json\/wp\/v2\/pages\/1249","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/agi-conf.org\/2011\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/agi-conf.org\/2011\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/agi-conf.org\/2011\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/agi-conf.org\/2011\/wp-json\/wp\/v2\/comments?post=1249"}],"version-history":[{"count":3,"href":"https:\/\/agi-conf.org\/2011\/wp-json\/wp\/v2\/pages\/1249\/revisions"}],"predecessor-version":[{"id":1291,"href":"https:\/\/agi-conf.org\/2011\/wp-json\/wp\/v2\/pages\/1249\/revisions\/1291"}],"wp:attachment":[{"href":"https:\/\/agi-conf.org\/2011\/wp-json\/wp\/v2\/media?parent=1249"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}