Much of the ongoing quantum computing battle among tech giants such as Google and IBM has focused on developing the hardware necessary to solve impossible classical computing problems. A Berkeley-based startup looks to beat those larger rivals with a one-two combo: a fab lab designed for speedy creation of better quantum circuits and a quantum computing cloud service that provides early hands-on experience with writing and testing software.
This is a guest post. The views expressed here are solely those of the author and do not represent positions of IEEE Spectrum or the IEEE.
We owe the success of numerous state-of-the-art artificial intelligence applications to artificial neural networks. First designed decades ago, they rocketed the AI field to success quite recently, when researchers were able to run them on much more powerful hardware and feed them with huge amounts of data. Since then, the field of deep learning has been flourishing.
The effect seemed miraculous and promising. While it was hard to interpret what exactly was happening inside the networks, they started reaching human performance on a number of tasks: such as image recognition, natural language processing, and data classification in general. The promise was that we would elegantly cross the border between data processing and intelligence by pure brute force of deep artificial neural networks: Just give it all the data in the world!
However, this is easier said than done. There are limits to state-of-the-art AI that separate it from human-like intelligence:
● We humans can learn a new skill without forgetting what we have already learned.
● We can build upon what we know already. For example, if we learn language skills in one context we can reuse them to communicate any of our experiences, dreams, or completely new ideas.
● We can improve ourselves and gradually become better learners. For instance, after you learn one foreign language, learning another is usually easier, because you already possess a number of heuristics and tricks for language-learning. You can keep discovering and improving these heuristics and use them to solve new tasks. This is how we’re able to work through completely new problems.
Some of these things may sound trivial, but today’s AI algorithms are very limited in how much previous knowledge they are able to keep through each new training phase, how much they can reuse, and whether they are able to devise any universal learning strategies at all.
In practice, this means that you need to build and fine tune a new algorithm for each new specific task—which is a form of very sophisticated data processing, rather than real intelligence.
To build a true general intelligence has been a lifelong dream of Marek Rosa, from his days as a teenage programmer until now, when he’s a successful entrepreneur. Rosa, therefore, invested the wealth he made in the video game business into his own general AI R&D company in Prague: GoodAI.
Rosa recently took steps to scale up the research on general AI by founding the AI Roadmap Institute and launching the General AI Challenge. The AI Roadmap Institute is an independent entity that promotes big-picture thinking by studying and comparing R&D roadmaps towards general intelligence. It also focuses on AI safety and considers roadmaps that represent possible futures that we either want to create or want to prevent from happening.
The General AI Challenge is a citizen-science project with a US $5 million prize fund provided by Rosa. His motivation is to incentivize talent to tackle crucial research problems in human-level AI development and to speed up the search for safe and beneficial general artificial intelligence.
The $5 million will be given out as prizes in various rounds of the multi-year competition. Each round will tackle an important milestone on the way to general AI. In some rounds, participants will be tasked with designing algorithms and programming AI agents. In other rounds, they will work on theoretical problems such as AI safety or societal impacts of AI. The Challenge will address general AI as a complex phenomenon.
The Challenge kicked off on 15 February with a six-month “warm-up” round dedicated to building gradually learning AI agents. Rosa and the GoodAI team believe that the ability to learn gradually lies at the core of our intelligence. It’s what enables us to efficiently learn new skills on top of existing knowledge without forgetting what we already know and to reapply our knowledge in various situations across multiple domains. Essentially, we learn how to learn better, enabling us to readily react to new problems.
At GoodAI’s R&D lab, AI agents will learn via a carefully designed curriculum in a gradual manner. We call it “school for AI,” since the progression is similar to human schooling, from nursery till graduation. We believe this approach will provide more control over what kind of behaviors and skills the AI acquires, which is of great importance for AI safety. Essentially, the goal is to bias the AI towards behaviors and abilities that we humans find useful and that are aligned with our understanding of the world and morality.
Nailing gradual learning is not an easy task, and so the Challenge breaks the problem into phases. The first round strips the problem down to a set of simplistic tasks in a textual environment. The tasks were specifically designed to test gradual learning potential, so they can serve as guidance for the developers.
The Challenge competitors are designing AI agents that can engage in a dialog within a textual environment. The environment will be teaching the agents to react to text patterns in a certain way. As an AI progresses through the set of roughly 40 tasks, they become harder. The final tasks are impossible to solve in a reasonable amount of time unless the agent has figured out the environment’s logic, and can reuse some of the skills it acquired on previous tasks.
More than 390 individuals and teams from around the world have already signed up to solve gradual learning in the first round of the General AI Challenge. (And enrollment is still open!) All participants must submit their solutions for evaluation by August 15 of this year. Then the submitted AI agents will be tested on a set of tasks which are similar, but not identical, to those provided as part of the first-round training tasks. That’s where the AI agents’ ability to solve previously unseen problems will really be tested.
We don’t yet know whether a successful solution to the Challenge’s first phase will be able to scale up to much more complex tasks and environments, where rich visual input and extra dimensions will be added. But the GoodAI team hopes that this first step will ignite new R&D efforts, spread new ideas in the community, and advance the search for more human-like algorithms.
Olga Afanasjeva is the director of the General AI Challenge and COO of GoodAI.
For the first time in 21 years, the United States no longer claimed even the bronze medal. With this week’s release of the latest Top 500 supercomputer ranking, the top three fastest supercomputers in the world are now run by China (with both first and second place finishers) and Switzerland. And while the supercomputer horserace is spectacle enough unto itself, a new report on the supercomputer industry highlights broader trends behind both the latest and the last few years of Top500 rankings.
The report, commissioned last year by the Japanese national science agency Riken, outlines a worldwide race toward exascale computers in which the U.S. sees R&D spending and supercomputer talent pools shrink, Europe jumps into the breach with increased funding, and China pushes hard to become the new global leader, despite a still small user and industry base ready to use the world’s most powerful supercomputers.
Steve Conway, report co-author and senior vice president of research at Hyperion, says the industry trend in high-performance computing is toward laying groundwork for pervasive AI and big data applications like autonomous cars and machine learning. And unlike more specialized supercomputer applications from years past, the workloads of tomorrow’s supercomputers will likely be mainstream and even consumer-facing applications.
“Ten years ago the rationale for spending on supercomputers was primarily two things: national security and scientific leadership, and I think there are a lot of people who still think that supercomputers are limited to problems like will a proton go left or right,” he says. “But in fact, there’s been strong recognition [of the connections] between supercomputing leadership and industrial leadership.”
“With the rise of big data, high-performance computing has moved to the forefront of research in things like autonomous vehicle design, precision medicine, deep learning, and AI,” Conway says. “And you don’t have to ask supercomputing companies if this is true. Ask Google and Baidu. There’s a reason why Facebook has already bought 26 supercomputers.”
As the 72-page Hyperion report notes, “IDC believes that countries that fail to fund development of these future leadership-class supercomputers run a high risk of falling behind other highly developed countries in scientific innovation, with later harmful consequences for their national economies.” (Since authoring the report in 2016 as part of the industry research group IDC, its authors this year formed the spin-off research firm Hyperion.)
Conway says that solutions to problems plaguing HPC systems today will be found in consumer electronics and industry applications of the future. So while operating massively parallel computers with multiple millions of cores may today only be a problem facing the world’s fastest and second-fastest supercomputers—China’s Sunway TaihuLight and Tianhe-2, running on 10.6 and 3.1 million cores, respectively—that fact won’t hold true forever. However, because China is the only country tackling this problem now means they are more likely to develop the technology first, technology that the world will want when cloud computing with multiple millions of cores approaches the mainstream.
The same logic applies to optimizing the ultra-fast data rates that today’s top HPC systems use and minimizing the megawatt electricity budgets they consume. And as the world’s supercomputers approach the exascale, that is, the 1 exaflop or 1000 petaflop mark, new challenges will no doubt arise too.
So, for instance, the report says that rapid shut-down and power-up of cores not in use will be one trick supercomputer designers use to trim back some of their systems’ massive power budgets. And, too, high-storage density—in the 100 petabyte range—will become paramount to house the big datasets the supercomputers consume.
“You could build an exascale system today,” Conway says. “But it would take well over 100 megawatts, which nobody’s going to supply, because that’s over a 100 million dollar electricity bill. So it has to get the electricity usage under control. Everybody’s trying to get it in the 20 to 30 megawatts range. And it has to be dense. Much denser than any computing today. It’s got to fit inside some kind of building. You don’t want the building to be 10 miles long. And also the denser the machine, the faster the machine is going to be too.”
Conway predicts that these and other challenges will be surmounted, and the first exaflop supercomputers will appear on the Top500 list around 2021, while exaflop supercomputing could become commonplace by 2023.
Yet for all the attention paid to supercomputer rankings, he also cautions about reading too much significance into any individual machine’s advertised speed, the basis for its rank on the Top500 list.
“The Top500 list is valuable, because the numbers there are kind of like what’s printed on the carton that the computer comes in,” he says. “It’s a not-to-exceed kind of number. But there are computers, I promise you, that don’t even make that list that can run huge and key problems, say, in the automotive industry, faster than anything on that list.”
In June, we can look forward to two things: the Belmont Stakes and the first of the twice-yearly TOP500 rankings of supercomputers. This month, a well-known gray and black colt named Tapwrit came in first at Belmont, and a well-known gray and black supercomputer named Sunway TaihuLight came in first on June’s TOP500 list, released today in conjunction with the opening session of the ISC High Performance conference in Frankfurt. Neither was a great surprise.
More of a surprise, and perhaps more of a disappointment for some, is that the highest-ranking U.S. contender, the Department of Energy’s Titan supercomputer (17.6 petaflops) housed at Oak Ridge National Laboratory, was edged out of the third position by an upgraded Swiss supercomputer called Piz Daint (19.6 petaflops), installed at the Swiss National Supercomputing Center, part of the Swiss Federal Institute of Technology (ETH) in Zurich.
Not since 1996 has a U.S. supercomputer not made it into one of the first three slots on the TOP500 list. But before we go too far in lamenting the sunset of U.S. supercomputing prowess, we should pause for a moment to consider that the computer that bumped it from the number-three position was built by Cray and is stuffed with Intel processors and NVIDIA GPUs, all the creations of U.S. companies.
Even the second-ranking Tianhe-2 is based on Intel processors and co-processors. It’s only the TaihuLight that is truly a Chinese machine, being based on the SW26010, a 260-core processor designed by the National High Performance Integrated Circuit Design Center in Shanghai. And U.S. supercomputers hold five of the 10 highest ranking positions on the new TOPS500 list.
Still, national rivalries seem to have locked the United States into a supercomputer arms race with China, with both nations vying to be the first to reach the exascale threshold—that is, to have a computer that can perform a 1018 floating-point operations per second. China hopes to do so by amassing largely conventional hardware and is slated to have a prototype system ready around the end of this year. The United States, on the other hand, is looking to tackle the problems that come with scaling to that level using novel approaches, which require more research before even a prototype machine can be built. Just last week, the U.S. Department of Energy announced that it was awarding Advanced Micro Devices, Cray, Hewlett Packard, IBM, Intel, and NVIDIA US $258 million to support research toward building an exascale supercomputer. Who will get there first, is, of course, up for grabs. But one thing’s for sure: It’ll be a horse race worth watching.
The top 10 supercomputers from the June 2017 Top500.org list.
“Processors are overdesigned for most applications,” says University of Illinois electrical and computer engineering professor Rakesh Kumar. It’s a well-known and necessary truth: In order to have programmability and flexibility, there’s simply going to be more stuff on a processor than any one application will use. That’s especially true of the type of ultralow power microcontrollers that drive the newest embedded computing platforms such as wearables and Internet of Things sensors. These are often running one fairly simple application and nothing else (not even an operating system), meaning that a large fraction of the circuits on a chip never, ever see a single bit of data.
Kumar, University of Minnesota assistant professor John Sartori (formerly a student of Kumar’s), and their students decided to do something about all that waste. Their solution is a method that starts by looking at the design of a general-purpose microcontroller. They identify which individual logic gates are never engaged for the application it’s going to run, and strip away all the excess gates. The result is what Kumar calls a “bespoke processor.” It’s a physically smaller, less complex version of the original microcontroller that’s designed to perform only the application needed.
The dream of a space-based nigh-unhackable quantum Internet may now be closer to reality thanks to new experiments with Chinese and European satellites, two new studies find.
Quantum physics makes a strange phenomenon known as entanglement possible. Essentially, two or more particles such as photons that get linked or “entangled”can in theory influence each other simultaneously no matter how far apart they are.
A previous distance record for quantum entanglement, 97 kilometers, was set in 2012 across Qinghai Lake in China by quantum physicist Jian-Wei Pan at the University of Science and Technology of China at Hefei and his colleagues. However, entanglement gets easily disrupted by interference from the environment, and this fragility has stymied efforts at greater distance records on Earth.
Now Pan and his colleagues have set a new record for entanglement by using a satellite to connect sites on Earth separated by up to 1,203 kilometers. They detailed their findings this week in the journal Science.
Machine learning, the field of AI that allows Alexa and Siri to parse what you say and self-driving cars to safely drive down a city street, could benefit from quantum computer-derived speedups, say researchers. And if a technology incubator program in Toronto, Canada has its way, there may even be quantum machine learning startup companies launching in a few years too.
Incoming messages for straight men on dating sites are… rare. Yet many of the dashing men who tried out Ashley Madison, a site aimed at the already-married, got messages soon after signing up. To see the messages, the men had to pay. The more perceptive among them soon noticed that their pen pals wrote similar come-ons, logged in and out at the same time every day, and oddest of all, had not visited the men's profiles. Ashley Madison was using more than 70,000 bots to lure in users, Gizmodo found in a 2015 investigation.
The message-sending profiles were one iteration of a growing army of bots that populate our online social networks, affecting everything from our wallets to our politics. Now they are attracting academic study and government research dollars.
“For the first time, humans are beginning to share their social ecosystem with a new species,” says computer science graduate student Gregory Maus, of Indiana University. And because not everybody is as attentive as the Ashley Madison user who blew the whistle on the fembots, human users of social networks are susceptible to everything from outright scams to subtler political influence by bots promoting fake news. In response, two years ago the Defense Advanced Research Projects Agency (DARPA) challenged researchers to identify "influence bots" and is now funding further research on social networks.
Maus will present one of a growing number of socialbot taxonomies at the ACM Web Science conference in Troy, New York, this June. The taxonomy seeks to expand on earlier taxonomies focused on identifying the different types of botnets and categorizing malicious socialbots that, for example, flood with spam a Twitter hashtag used to organize political protests. Another recent paper began mapping the existence of benign bots. Maus says he hopes his new taxonomy will be a more “broad, flexible framework useful for researchers” seeking both to understand and interact with bots.
“The interesting aspect of the current work is that it considers five different dimensions,” says computational social scientist Taha Yasseri of Oxford University in the United Kingdom, who earlier this year published a case study of an unexpected years-long conflict between Wikipedia maintenance-bots.
Maus' paper sketches out categories based on the degree to which a bot tries to pretend to be human, who its owner is, how the bot interacts with other bots, whether it hides its connection to its owner, and its mission. Some of these have their own sub-categories. Yasseri adds that it would be useful to examine how the different types interact with each other, rather than just studying each type in isolation. The interaction of human and machine networks is the focus of Yasseri’s European Union-funded project, HUMANE.
In fact, that has been one of the features of the human approach to studying bot taxonomies: variety and interactivity. Researchers come from a wide range of backgrounds. Maus, whose undergraduate degree is in philosophy, worked in marketing before joining the Networks & agents Network group at IU. His colleagues there have a mixture of backgrounds in psychology, mathematics, physics, and computer science.
Maus says students or others interested in working on social network taxonomy can get an immediate start by studying the APIs of a social network and reaching out to other researchers working on these problems. His supervisor, Filippo Menczer, accepts potential students through any of three different Ph.D. tracks. The area of bot taxonomy is young enough—and complex enough—that the variety of human profiles almost matches that of the bots.
The best artificial intelligence still has trouble visually recognizing people performing many of Homer Simpson’s favorite behaviors such as drinking beer, eating chips, eating doughnuts, yawning, and the occasional face-plant. Those findings from DeepMind, the pioneering London-based AI lab, also suggest the motive behind why DeepMind has created a huge new dataset of YouTube clips to help train AI on identifying human actions in videos that go well beyond “Mmm, doughnuts” or “Doh!”
In a Nature paper published today, researchers at the Karlsruhe Institute of Technology in Germany and the Swiss Federal Institute of Technology in Lausanne report using a pair of silicon-nitride microresonator frequency combs to transmit 50 terabits of data through 75 kilometers of single-mode fiber using 179 separate carrier wavelengths. They also showed that microresonator frequency combs could serve as local oscillators in receivers, which could improve transmission quality, and lead to petabit-per-second (1000 terabit) data rates inside data centers.
IEEE Spectrum’s general technology blog, featuring news, analysis, and opinions about engineering, consumer electronics, and technology and society, from the editorial staff and freelance contributors.