Elon Musk Gives $10 Million In Grants To Study Safe AI - InformationWeek
IoT
IoT
IT Life
News
7/6/2015
08:05 AM
Connect Directly
Twitter
RSS
E-Mail
50%
50%
RELATED EVENTS
Moving UEBA Beyond the Ground Floor
Sep 20, 2017
This webinar will provide the details you need about UEBA so you can make the decisions on how bes ...Read More>>

Elon Musk Gives $10 Million In Grants To Study Safe AI

Elon Musk puts his money where his mouth is by helping fund 37 projects that could hopefully make AI safer and more useful to humans.

Disney's Tomorrowland Past And Present: A Celebration
Disney's Tomorrowland Past And Present: A Celebration
(Click image for larger view and slideshow.)

Through a $10 million grant from Elon Musk, the Future of Life Institute is awarding 37 grants to fund research that they believe will keep AI "robust and beneficial."

Even if you aren't in the alarmist camp of Musk, Bill Gates, and Stephen Hawking and believe that AI is a danger to humanity, the grants represent the sort of basic, foundational research that we need to improve AI.

The Future of Life Institute was cofounded by MIT cosmologist Max Tegmark and Skype cofounder Jaan Tallinn. It includes such big-name advisors as Musk, Hawking, Alan Alda, and Morgan Freeman. It was founded with the mission to save humanity from the existential threats they perceive from AI.

To prove their point, the institute's website opens with the ominous phrase: "Technology has given life the opportunity to flourish like never before … or to self-destruct."

If it all sounds a little Hollywood, maybe that's on purpose. The press release for the new grants mentions the new Terminator movie.

(Image: Paramount)

(Image: Paramount)

Still, this isn't some Hollywood movie where a benevolent organization is out to stop what they perceive to be an evil idea. The goal seems to be to do AI right and to do it with good science. This is the institute's stated mission: "FLI catalyzes and supports research and initiatives for safeguarding life and developing optimistic visions of the future, including positive ways for humanity to steer its own course considering new technologies and challenges."

[So will AI really kill us all? No, AI Won't Kill Us All.]

So what are the 37 projects they funded? You can check out the full list on the institute's site.

One of the most interesting is one that could be colloquially described as, "What would John Doe do?"

Paul Christiano from UC Berkeley is researching ways to teach autonomous AI to respond to situations that it doesn't understand in ways a human would, without intervention. One of the biggest fears of those who think AI is a danger is that of what an AI might do if it encounters a situation it doesn't understand. Christiano is hoping to create efficient mechanisms to provide human oversight. There are two similar projects that revolve around the idea of allowing AIs to observe humans to help them understand what humans want from them.

Manuela Veloso of Carnegie Mellon was given a grant to study how to make AIs explain their actions so we can better understand why they are doing something and take corrective action. If an autonomous car, for example, took a right turn when you expected a left, you could ask it why in order to make sure that the decision made sense.

Michael Webb of Stanford University is being a bit more practical. He's studying the economic and social impact of how AI could eventually replace us all. How do you build an economy where most of us don't have to work to keep it running? How do you distribute wealth and other resources? Most importantly, how do you make the transition to an economy like that?

There are other studies, including one on what happens if an AI breaks the law, another that examines the ethical implications that exist for AI by judging all potential outcomes of a situation with no regard to ethics, and many on how to teach ethics to AI.

While some of these may seem a little silly at first, they are a necessary step in the programming of intelligence.

As Tom Dietterich, president of the Association of the Advancement of Artificial Intelligence, says in the press release:

"In its early days, AI research focused on the 'known knowns' by working on problems such as chess and blocks world planning, where everything about the world was known exactly. Starting in the 1980s, AI research began studying the 'known unknowns' by using probability distributions to represent and quantify the likelihood of alternative possible worlds. The FLI grant will launch work on the 'unknown unknowns': How can an AI system behave carefully and conservatively in a world populated by unknown unknowns -- aspects that the designers of the AI system have not anticipated at all?"

This and other research, if successful, should make AI safer and more effective.

David has been writing on business and technology for over 10 years and was most recently Managing Editor at Enterpriseefficiency.com. Before that he was an Assistant Editor at MIT Sloan Management Review, where he covered a wide range of business topics including IT, ... View Full Bio

Comment  | 
Print  | 
More Insights
Comments
Newest First  |  Oldest First  |  Threaded View
soozyg
50%
50%
soozyg,
User Rank: Ninja
7/7/2015 | 11:56:41 AM
AI economy
How do you build an economy where most of us don't have to work to keep it running? 

I always find questions and goals like this so amusing. Americans have been writing and talking about this for decades. We're still trying to get a good part of the population employed, not unemployed. We're so behind, I don't think this is work discussing anymore.
Brian.Dean
50%
50%
Brian.Dean,
User Rank: Ninja
7/6/2015 | 6:57:29 PM
Re: AI
It is a valid assumption -- there are things that are known, known unknowns and unknown unknowns. If FLI finds that AI is a real and present threat, how is it going to be stopped? Because, if I recall correctly, private investment in AI has crossed aggregate government investment in AI and for good reason as AI has started to deliver a lot of business and economic value. 
PedroGonzales
50%
50%
PedroGonzales,
User Rank: Ninja
7/6/2015 | 5:44:31 PM
Re: AI
I just hope they do not start bringing fictional scenarios for these grants.  If you are running a simulation there could be some many scenarios that one can go with AI.  I would like to know whether such scenarios would even had the chance to go from the lab to the real world.  I would like to see more practical studies on AI, such as decision support system on hospitals, improving AI on autonomous cars.
Ariella
50%
50%
Ariella,
User Rank: Author
7/6/2015 | 11:34:32 AM
AI
That's quite some progress: from known knowns to known unknowns to unknown unknowns!
How Enterprises Are Attacking the IT Security Enterprise
How Enterprises Are Attacking the IT Security Enterprise
To learn more about what organizations are doing to tackle attacks and threats we surveyed a group of 300 IT and infosec professionals to find out what their biggest IT security challenges are and what they're doing to defend against today's threats. Download the report to see what they're saying.
Register for InformationWeek Newsletters
White Papers
Current Issue
IT Strategies to Conquer the Cloud
Chances are your organization is adopting cloud computing in one way or another -- or in multiple ways. Understanding the skills you need and how cloud affects IT operations and networking will help you adapt.
Video
Slideshows
Twitter Feed
Sponsored Live Streaming Video
Everything You've Been Told About Mobility Is Wrong
Attend this video symposium with Sean Wisdom, Global Director of Mobility Solutions, and learn about how you can harness powerful new products to mobilize your business potential.
Flash Poll