Elon Musk Gives $10 Million In Grants To Study Safe AI - InformationWeek

InformationWeek is part of the Informa Tech Division of Informa PLC

This site is operated by a business or businesses owned by Informa PLC and all copyright resides with them.Informa PLC's registered office is 5 Howick Place, London SW1P 1WG. Registered in England and Wales. Number 8860726.

IT Life
08:05 AM
Connect Directly

Elon Musk Gives $10 Million In Grants To Study Safe AI

Elon Musk puts his money where his mouth is by helping fund 37 projects that could hopefully make AI safer and more useful to humans.

Disney's Tomorrowland Past And Present: A Celebration
Disney's Tomorrowland Past And Present: A Celebration
(Click image for larger view and slideshow.)

Through a $10 million grant from Elon Musk, the Future of Life Institute is awarding 37 grants to fund research that they believe will keep AI "robust and beneficial."

Even if you aren't in the alarmist camp of Musk, Bill Gates, and Stephen Hawking and believe that AI is a danger to humanity, the grants represent the sort of basic, foundational research that we need to improve AI.

The Future of Life Institute was cofounded by MIT cosmologist Max Tegmark and Skype cofounder Jaan Tallinn. It includes such big-name advisors as Musk, Hawking, Alan Alda, and Morgan Freeman. It was founded with the mission to save humanity from the existential threats they perceive from AI.

To prove their point, the institute's website opens with the ominous phrase: "Technology has given life the opportunity to flourish like never before … or to self-destruct."

If it all sounds a little Hollywood, maybe that's on purpose. The press release for the new grants mentions the new Terminator movie.

(Image: Paramount)

(Image: Paramount)

Still, this isn't some Hollywood movie where a benevolent organization is out to stop what they perceive to be an evil idea. The goal seems to be to do AI right and to do it with good science. This is the institute's stated mission: "FLI catalyzes and supports research and initiatives for safeguarding life and developing optimistic visions of the future, including positive ways for humanity to steer its own course considering new technologies and challenges."

[So will AI really kill us all? No, AI Won't Kill Us All.]

So what are the 37 projects they funded? You can check out the full list on the institute's site.

One of the most interesting is one that could be colloquially described as, "What would John Doe do?"

Paul Christiano from UC Berkeley is researching ways to teach autonomous AI to respond to situations that it doesn't understand in ways a human would, without intervention. One of the biggest fears of those who think AI is a danger is that of what an AI might do if it encounters a situation it doesn't understand. Christiano is hoping to create efficient mechanisms to provide human oversight. There are two similar projects that revolve around the idea of allowing AIs to observe humans to help them understand what humans want from them.

Manuela Veloso of Carnegie Mellon was given a grant to study how to make AIs explain their actions so we can better understand why they are doing something and take corrective action. If an autonomous car, for example, took a right turn when you expected a left, you could ask it why in order to make sure that the decision made sense.

Michael Webb of Stanford University is being a bit more practical. He's studying the economic and social impact of how AI could eventually replace us all. How do you build an economy where most of us don't have to work to keep it running? How do you distribute wealth and other resources? Most importantly, how do you make the transition to an economy like that?

There are other studies, including one on what happens if an AI breaks the law, another that examines the ethical implications that exist for AI by judging all potential outcomes of a situation with no regard to ethics, and many on how to teach ethics to AI.

While some of these may seem a little silly at first, they are a necessary step in the programming of intelligence.

As Tom Dietterich, president of the Association of the Advancement of Artificial Intelligence, says in the press release:

"In its early days, AI research focused on the 'known knowns' by working on problems such as chess and blocks world planning, where everything about the world was known exactly. Starting in the 1980s, AI research began studying the 'known unknowns' by using probability distributions to represent and quantify the likelihood of alternative possible worlds. The FLI grant will launch work on the 'unknown unknowns': How can an AI system behave carefully and conservatively in a world populated by unknown unknowns -- aspects that the designers of the AI system have not anticipated at all?"

This and other research, if successful, should make AI safer and more effective.

David has been writing on business and technology for over 10 years and was most recently Managing Editor at Enterpriseefficiency.com. Before that he was an Assistant Editor at MIT Sloan Management Review, where he covered a wide range of business topics including IT, ... View Full Bio

We welcome your comments on this topic on our social media channels, or [contact us directly] with questions about the site.
Comment  | 
Print  | 
More Insights
Newest First  |  Oldest First  |  Threaded View
User Rank: Ninja
7/7/2015 | 11:56:41 AM
AI economy
How do you build an economy where most of us don't have to work to keep it running? 

I always find questions and goals like this so amusing. Americans have been writing and talking about this for decades. We're still trying to get a good part of the population employed, not unemployed. We're so behind, I don't think this is work discussing anymore.
User Rank: Ninja
7/6/2015 | 6:57:29 PM
Re: AI
It is a valid assumption -- there are things that are known, known unknowns and unknown unknowns. If FLI finds that AI is a real and present threat, how is it going to be stopped? Because, if I recall correctly, private investment in AI has crossed aggregate government investment in AI and for good reason as AI has started to deliver a lot of business and economic value. 
User Rank: Ninja
7/6/2015 | 5:44:31 PM
Re: AI
I just hope they do not start bringing fictional scenarios for these grants.  If you are running a simulation there could be some many scenarios that one can go with AI.  I would like to know whether such scenarios would even had the chance to go from the lab to the real world.  I would like to see more practical studies on AI, such as decision support system on hospitals, improving AI on autonomous cars.
User Rank: Author
7/6/2015 | 11:34:32 AM
That's quite some progress: from known knowns to known unknowns to unknown unknowns!
How COVID is Changing Technology Futures
Jessica Davis, Senior Editor, Enterprise Apps,  7/23/2020
10 Ways AI Is Transforming Enterprise Software
Cynthia Harvey, Freelance Journalist, InformationWeek,  7/13/2020
IT Career Paths You May Not Have Considered
Lisa Morgan, Freelance Writer,  6/30/2020
White Papers
Register for InformationWeek Newsletters
2020 State of DevOps Report
2020 State of DevOps Report
Download this report today to learn more about the key tools and technologies being utilized, and how organizations deal with the cultural and process changes that DevOps brings. The report also examines the barriers organizations face, as well as the rewards from DevOps including faster application delivery, higher quality products, and quicker recovery from errors in production.
Current Issue
Special Report: Why Performance Testing is Crucial Today
This special report will help enterprises determine what they should expect from performance testing solutions and how to put them to work most efficiently. Get it today!
Flash Poll