When Silicon Valley entrepreneur Elon Musk is not trying to build rocket technology to colonize Mars or revolutionize energy storage on Earth, he worries about how artificial intelligence could someday slip its shackles and become a danger to humanity. Now some of Musk’s ample wealth is helping fund a newly-announced group of research projects aimed at keeping AI in check.
The Boston-based Future of Life Institute has awarded $7 million in funds from Elon Musk and the Open Philanthropy Project to 37 research teams around the world. The list of grant recipients includes teams developing AI that can explain its decisions to humans, studying how to keep the economic impact of AI beneficial, and figuring out how to keep AI-based military weapons under human control. Such research could help society continue to benefit from the development of smarter AI while keeping potential dangers to a minimum.
“Building advanced AI is like launching a rocket,” said Jaan Tallinn, one of the founders of the Future of Life Institute, in a statement. “The first challenge is to maximize acceleration, but once it starts picking up speed, you also need to focus on steering.”
About 300 research projects applied for grants from the Future of Life Institute after the latter issued an open letter in January that called for research on “keeping AI robust and beneficial.” The letter was signed by AI researchers from Facebook, IBM, and Microsoft and the founders of Google’s DeepMind Technologies, among other areas of academia, nonprofits and industry. That reportedly moved Musk to donate $10 million to the Future of Life Institute.
But not everyone necessarily worries about time-traveling robots destroying us all. G. Pascal Zachary, professor of practice at Arizona State University, recently wrote for IEEE Spectrum about how humanity can begin shaping AI before it shapes us.
Humanity’s “dark fantasies” about killer robots tend to distract from more pertinent questions such as how AI impacts employment, how robot writers and artists could change human consumption of creative content, and who is held accountable for accidents involving autonomous processes, Zachary said.
A glance at the list of newly-funded research on AI suggests that researchers could begin tackling some of those questions. The Future of Life Institute took a stance similar to Zachary’s as it stressed the difference between Hollywood’s “Terminator” fantasies and reality.
“The danger with the Terminator scenario isn’t that it will happen, but that it distracts from the real issues posed by future AI”, said Max Tegmark, president of the Future of Life Institute. “We’re staying focused, and the 37 teams supported by today’s grants should help solve such real issues.”
Jeremy Hsu has been working as a science and technology journalist in New York City since 2008. He has written on subjects as diverse as supercomputing and wearable electronics for IEEE Spectrum. When he’s not trying to wrap his head around the latest quantum computing news for Spectrum, he also contributes to a variety of publications such as Scientific American, Discover, Popular Science, and others. He is a graduate of New York University’s Science, Health & Environmental Reporting Program.