1 / 19

Ethics for self-improving machines

Ethics for self-improving machines. J Storrs Hall Mark Waser. Asimov's 3 Laws:. 1. A robot may not injure a human being or, through inaction, allow a human being to come to harm.

temira
Télécharger la présentation

Ethics for self-improving machines

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Ethics for self-improving machines J Storrs Hall Mark Waser

  2. Asimov's 3 Laws: 1. A robot may not injure a human being or, through inaction, allow a human being to come to harm. 2. A robot must obey orders given to it by human beings except where such orders would conflict with the First Law. 3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law. http://www.markzug.com/

  3. Asimov's robots didn't Improve Themselves. But our AIs (we hope)‏ Will.

  4. How do you design laws for something that will think in concepts you haven't heard of and which you couldn't grasp if you had?

  5. There is no chance that everybody will create their robots with any given set of laws anyhow! Laws reflect goals (and thus values) which do NOT converge over humanity.

  6. Axelrod's Evolution of Cooperation and decades of follow-on evolutionary game theory provide the theoretical underpinnings. • Be nice/don’t defect • Retaliate • Forgive • “Selfish individuals, for their • own selfish good, should be • nice and forgiving”

  7. In nature, cooperation appears whenever the cognitive machinery will support it. • Cotton-Top • Tamarins • (Hauser, et al) Vampire bats (Wilkinson) • Blue Jays • (Stephens, • McLinn, • & Stevens)

  8. Economic Sentience Defined as: “Awareness of the potential benefits of cooperation and trade with other intelligences” TIMEDISCOUNTING is its measure.

  9. Tragedy of the Commons

  10. Acting ethically is an attractor in the state space of • intelligent goal-driven systems (if they interact with • other intelligent goal-driven systems on • a long-term ongoing basis) • Ethics *IS* the • necessary basis • for cooperation

  11. We must find ethical design elements that are Evolutionarily Stable Strategies so that we can start AIs out in the attractor it's taken us millions of years to begin to descend.

  12. Let's call such a design element an ESV: Evolutionarily Stable (or Economically Sentient) Virtue.

  13. Economically Unviable • Destruction • Slavery • Short-term profit at the expense of the long term Avoiding all of these are ESVs

  14. Fair enforcement of contracts is an ESV that demonstrably promotes cooperation.

  15. Like auditing in current-day corporations, since money is their true emotion. Open Source motivations are anESV and other forms of guaranteeing trustworthiness

  16. In particular, RECIPROCAL ALTRUISM is an ESV; Exactly like it's superset ENLIGHTENED SELF-INTEREST (AKA ETHICS)

  17. A general desire for all ethical agents to live (and prosper) as long as possible is also an ESV, because it promotes a community with long-term stability and accountability.

  18. Curiosity – a will to extend and improve one's world model – is an ESV. There is no good but knowledge, and no evil but ignorance. — Socrates

  19. An AI with ESVs who knows what that means has a guideline for designing Version 2.0, even when the particulars of the new environment don't match the concepts of the old literal goal structure.

More Related