James Miller on Unusual Incentives Facing AGI Companies

||Conversations

rsz_11james-d-millerJames D. Milleris an associate professor of economics atSmith College. He is the author ofSingularity Rising,Game Theory at Work,and aprinciples of microeconomics textbookalong with several academic articles.

He has a PhD in economics from the University of Chicago and a J.D. from Stanford Law School where he was onLaw Review. He is a member of cryonics providerAlcorand a research advisor to MIRI. He is currently co-writing a book on better decision making with theCenter for Applied Rationalityand will be probably be an editor on the next edition of theSingularity Hypothesesbook. He is a committed bio-hacker currently practicing or consuming apaleo diet,neurofeedback,cold thermogenesis,intermittent fasting,brain fitnessvideo games,smart drugs,bulletproof coffee, andrationalitytraining.

Luke Muehlhauser:Your book chapter inSingularity Hypothesisdescribes some unusual economic incentives facing a future business that is working to createAGI. To explain your point, you make the simplifying assumption that “a firm’s attempt to build an AGI will result in one of three possible outcomes”:

  • Unsuccessful: The firm fails to create AGI, losing value for its owners and investors.
  • Riches: The firm creates AGI, bringing enormous wealth to its owners and investors.
  • Foom: The firm creates AGI but this event quickly destroys the value of money, e.g. via anintelligence explosionthat eliminates scarcity, or creates a weird world without money, or exterminates humanity.

How does this setup allow us to see the unusual incentives facing a future business that is working to create AGI?


James Miller:A huge asteroid might hit the earth, and if it does it will destroy mankind. You should be willing to bet everything you have that the asteroid will miss our planet because either you win your bet or Armageddon renders the wager irrelevant. Similarly, if I’m going to start a company that will either make investors extremely rich or create aFoomthat destroys the value of money, you should be willing to invest a lot in my company’s success because either the investment will pay off, or you would have done no better making any other kind of investment.

Pretend I want to create a controllable AGI, and if successful I will earn greatRichesfor my investors. At first I intend to follow a research and development path in which if I fail to achieveRiches, my company will beUnsuccessfuland have no significant impact on the world. Unfortunately, I can’t convince potential investors that the probability of my achievingRichesis high enough to make my company worth investing in. The investors assign too large a likelihood that other potential investments would outperform my firm’s stock. But then I develop an evil alternative research and development plan under which I have the exact same probability of achievingRichesas before but now if I fail to create a controllable AGI, an unfriendlyFoomwill destroy humanity. Now I can truthfully tell potential investors that it’s highly unlikely any other company’s stock will outperform mine.

Read more »

MIRI’s July Newsletter: Fundraiser and New Papers

||Newsletters



Greetings from the Executive Director

Dear friends,

Another busy month! Since our last newsletter, we’ve published 3 new papers and 2 new “analysis” blog posts, we’ve significantly improved our website (especially the亚博体育官网 page), we’verelocatedto downtown Berkeley, and we’ve launchedour summer 2013 matching fundraiser!

MIRI also recently presented at theEffective Altruism Summit, a gathering of 60+ effective altruists in Oakland, CA. As philosopher Peter Singer explained in hisTED talk, effective altruism “combines both the heart and the head.” The heart motivates us to be empathic and altruistic toward others, while the head can “make sure that what [we] do is effective and well-directed,” so that altruists can do not justsomegood butas much good as possible.

As I explain inFriendly AI Research as Effective Altruism, MIRI was founded in 2000 on the premise that creating Friendly AI might be a particularly efficient way to do as much good as possible. Effective altruists focus on a variety of other causes, too, such as poverty reduction. As I say inFour Focus Areas of Effective Altruism, I think it’s important for effective altruists to cooperate and collaborate, despite their differences of opinion about which focus areas are optimal. The world needs more effective altruists, of all kinds.

MIRI engages in direct efforts — e.g. Friendly AI research — to improve the odds that machine superintelligence has a positive rather than a negative impact. But indirect efforts — such as spreading rationality and effective altruism — are also likely to play a role, for they will influence the context in which powerful AIs are built. That’s part of why we createdCFAR.

If you think this work is important, I hope you’llyabo体育官网下载ios to support our work. MIRI isentirelysupported by private funders likeyou. And if you donate before August 15th, your contribution will be matched by one of the generous backers ofour current fundraising drive.

Thank you,

Luke Muehlhauser

Executive Director

Read more »

2013 Summer Matching Challenge!

||News

Thanks to the generosity of several major donors,每一个机器智能研究捐赠h Institute made from now until August 15th, 2013 will be matched dollar-for-dollar, up to a total of $200,000!

yabo体育官网下载ios

$0

$50,000

$100,000

$150,000

$200,000

We have reached our goal of $200,000!

Now is your chance todouble your impactwhile helping us raise up to $400,000 (with matching) to fund亚博体育官网 .


Early this year we made a transition from movement-building to research, and we’vehit the ground runningwith six major new research papers, six new strategic analyses on our blog, and much more. Give now to support our ongoing work onthe future’s most important problem.

Accomplishments in 2013 so far

Future Plans You Can Help Support

  • We will host many more research workshops, includingone in September, and one in December (withJohn Baezattending, among others).
  • Eliezer will continue to publish about open problems in Friendly AI. (Here is#1and#2.)
  • We will continue to publish strategic analyses, mostly via our blog.
  • We will publish nicely-edited ebooks (Kindle, iBooks, and PDF) for more of our materials, to make them more accessible:The Sequences, 2006-2009andThe Hanson-Yudkowsky AI Foom Debate.
  • We will continue to set up the infrastructure (e.g.new offices, researcher endowments) required to host a productive Friendly AI research team, and (over several years) recruit enough top-level math talent to launch it.

(Other projects are still being surveyed for likely cost and strategic impact.)

我们感谢你的支持对我们的影响力吧k! Donate now, and seize a better than usual chance to move our work forward. If you have questions about donating, please contact Louie Helm at (510) 717-1477 or louie@www.hdjkn.com.

$200,000 of total matching funds has been provided by Jaan Tallinn, Loren Merritt, Rick Schwall, and Alexei Andreev.

MIRI Has Moved!

||News

For the past several months, MIRI and its child organizationCFARhave been working from a much-too-small office on the outskirts of Berkeley. At the end of June, MIRI and CFAR took over the 3rd floor of2030 Addison St.在伯克利的市中心,有足够的空间or both organizations.

Our new office is 0.5 blocks from the Downtown Berkeley BART exit at Shattuck & Addison, and 2 blocks from the UC Berkeley campus. Here’s a photo of the campus from our roof:

view of campus from roof (500px)

The proximity to UC Berkeley will make it easier for MIRI to network with Berkeley’s professors and students. Conveniently, UC Berkeley is ranked5th in the worldin mathematics, and1st in the worldin mathematical logic.

Sharing an office with CFAR carries many benefits for both organizations:

  1. CFAR and MIRI can “flex” into each other’s space for short periods as needed, for example when MIRI is holding a week-long亚博体育苹果app官方下载 .
  2. We can share resources (printers, etc.).
  3. Both organizations can benefit from interaction between our two communities.

Getting the new office was a team effort, but the personmostresponsible for this success was MIRI Deputy Director Louie Helm.

Note that MIRI isn’t yet able to accommodate “drop in” visitors, as we keep irregular hours throughout the week. So if you’d like to visit, pleasecontact usfirst.

We retain 2721 Shattuck Ave. #1023 as an alternate mailing address.

MIRI’s September 2013 Workshop

||News

Paul at April workshop

From September 7-13, MIRI will host its4th Workshop on Logic, Probability, and Reflection. The focus of this workshop will be the foundations ofdecision theory.

Participants confirmed so far include:

如果你有一个strong mathematics background and might like to attend this workshop, it’s not too late to亚博体育苹果app官方下载 ! And even ifthisworkshop doesn’t fit your schedule, pleasedo apply, so that we can notify you of other workshops (long before they are announced publicly).

Responses to Catastrophic AGI Risk: A Survey

||Papers

MIRI is self-publishing another technical report that was too lengthy (60 pages) for publication in a journal:Responses to Catastrophic AGI Risk: A Survey.

The report, co-authored by past MIRI researcherKaj Sotalaand University of Louisville’sRoman Yampolskiy, is a summary of the extant literature (250+ references) on AGI risk, and can serve either as a guide for researchers or as an introduction for the uninitiated.

Here is the abstract:

Many researchers have argued that humanity will create artificial general intelligence (AGI) within the next twenty to one hundred years. It has been suggested that AGI may pose a catastrophic risk to humanity. After summarizing the arguments for why AGI may pose such a risk, we survey the field’s proposed responses to AGI risk. We consider societal proposals, proposals for external constraints on AGI behaviors, and proposals for creating AGIs that are safe due to their internal design.

The preferred discussion page for the paper ishere.

Update:This report has now been published inPhysica Scripta, availablehere.

What is Intelligence?

||yabo app

When asked their opinions about “human-level artificial intelligence” —aka“artificial general intelligence” (AGI)1— many experts understandably reply that these terms haven’t yet been precisely defined, and it’s hard to talk about something that hasn’t been defined.2In this post, I want to briefly outline an imprecise but useful “working definition” forintelligencewe tend to use at MIRI. In a future post I will write about some useful working definitions forartificial general intelligence.

Imprecise definitions can be useful

Precise definitions are important, but I concur with Bertrand Russell that

[You cannot] start with anything precise. You have to achieve such precision… as you go along.

PhysicistMilan Ćirkovićagrees, andgivesan example:

The formalization of knowledge — which includes giving precise definitions — usually comes at the end of the original research in a given field, not at the very beginning. A particularly illuminating example is the concept ofnumber, which was properly defined in the modern sense only after the development of axiomatic set theory in the… twentieth century.3

For a more AI-relevant example, consider the concept of a “self-driving car,” which has been given a variety of vague definitionssince the 1930s. Would a carguided by a buried cablequalify? What about amodified 1955 Studebakerthat could use sound waves to detect obstacles and automatically engage the brakes if necessary, but could only steer “on its own” if each turn was preprogrammed? Does that count as a “self-driving car”?

What about the “VaMoRs” of the 1980s that could avoid obstacles and steer around turns using computer vision, but weren’t advanced enough to be ready for public roads? How about the 1995Navlabcar that drove across the USA and was fully autonomous for 98.2% of the trip, or the robotic cars which finished the 132-mile off-road course of the2005 DARPA Grand Challenge, supplied only with the GPS coordinates of the route? What about the winning cars of the2007 DARPA Grand Challenge, which finished an urban race while obeying all traffic laws and avoiding collisions with other cars? DoesGoogle’s driverless carqualify, given that it has logged more than 500,000 autonomous miles without a single accident under computer control, but still struggles with difficult merges and snow-covered roads?4

Our lack of a precise definition for “self-driving car” doesn’t seem to have hindered progress on self-driving cars very much.5And I’m glad we didn’t wait to seriously discuss self-driving cars until we had a precise definition for the term.

Similarly, I don’t think we should wait for a precise definition of AGI before discussing the topic seriously. On the other hand, the term is useless if it carriesnoinformation. So let’s work our way toward a stipulative, operational definition for AGI. We’ll start by developing an operational definition forintelligence.

Read more »


  1. I use the HLAI and AGI interchangeably, but lately I’ve been using AGI almost exclusively, because I’ve learned that many people in the AI community react negatively to any mention of “human-level” AI but have no objection to the concept of narrow vs. general intelligence. See also Ben Goertzel’s commentshere.
  2. Asked when he thought HLAI would be created, Pat Hayes (a past president ofAAAI)replied: “I do not consider this question to be answerable, as I do not accept this (common) notion of ‘human-level intelligence’ as meaningful.” Asked the same question, AI scientistWilliam Utherreplied: “You ask a lot about ‘human level AGI’. I do not think this term is well defined,” while AI scientistAlan Bundyreplied: “I don’t think the concept of ‘human-level machine intelligence’ is well formed.”
  3. Sawyer (1943)gives another example: “Mathematicians first used the sign √-1, without in the least knowing what it could mean, because it shortened work and led to correct results. People naturally tried to find out why this happened and what √-1 really meant. After two hundreds years they succeeded.”Dennett (2013)makes a related comment: “定义你的条款,先生!No, I won’t. That would be premature… My [approach] is an instance ofnibblingon a tough problem instead of trying to eat (and digest) the whole thing from the outset… InElbow Room, I compared my method to the sculptor’s method of roughing out the form in a block of marble, approaching the final surfaces cautiously, modestly, working by successive approximation.”
  4. With self-driving cars, researchers did use many precise external performance measures (e.g. accident rates, speed, portion of the time they could run unassisted, frequency of getting stuck) to evaluate progress, as well as internal performance metrics (speed of search, bounded loss guarantees, etc.). Researchers could see that these bits of progress were in the right direction, even if their relative contribution long-term was unclear. And so it is with AI in general. AI researchers use many precise external and internal performance measures to evaluate progress, but it is difficult to know the relative contribution of these bits of progress toward the final goal of AGI.
  5. Heck, we’ve had pornography for millennia andstillhaven’t been able to define it precisely. Encyclopedia entries for “pornography”oftensimplyquote Justice Potter Stewart: “I shall not today attempt further to define the kinds of material I understand to be [pornography]… but I know it when I see it.”

MIRI’s July 2013 Workshop

||News

亚博体育苹果app官方下载

From July 8-14, MIRI will host its3rd Workshop on Logic, Probability, and Reflection. The focus of this workshop will be theLöbian obstacle to self-modifying systems.

Participants confirmed so far include:

如果你有一个strong mathematics background and might like to attend this workshop, it’s not too late to亚博体育苹果app官方下载 ! And even ifthisworkshop doesn’t fit your schedule, pleasedo apply, so that we can notify you of other workshops (long before they are announced publicly).

Information on past workshops: