Edge-AI Startup Gets $60-million, Preps for Mass Production

Hailo sped from an idea to a 26-trillion-operations-per-second chip in three years

2 min read
Hailo's chip is designed to do deep learning in cars, robots, and other “edge” machines.
Photo: Hailo

Israeli startup Hailo says it has raised US $60 million in the second round of funding it will use to mass produce its Hailo-8 chip. The Hailo-8 is designed to do deep learning in cars, robots, and other “edge” machines. Such edge chips are meant to reduce the cost, size, and power consumption needs of using AI to process high-resolution information from sensors such as HD cameras.

The three-year-old company’s technology stemmed from a cold, hard look at what AI needed and what contemporary chips could deliver. “When you do a bottom up analysis of what should be the cost—[in terms of] power to run these applications—we saw an order of magnitude gap between what is possible and what is actually being done,” says Hailo CEO Orr Danon. “The problem lies in the architecture.”

In a typical GPU or CPU, there is a relatively narrow access pathway to memory but a very deep memory structure, Danon explains. But for AI “what you need is very wide access but shallow address space.” Another problem is that the broad programmability of CPUs and GPUs means that elements having to do with the control of data dominate those that have to do with the flow of data. These architectures are “very flexible and very robust, but that’s irrelevant to running neural networks,” he says.

Hailo calls its solution a structure-defined data flow architecture. Deep learning is called “deep” because its neural networks consist of a deep stack of neural layers between the input and the output. Hailo’s software tools take a trained neural net and examine how much memory, computing capability, and data flow control each layer needs. The software then maps this to the processor’s resources so that each layer is contiguous with its neighbors.

“You’re not only controlling how the data moves, but also the structure,” says Danon. “Being able to do so keeps the data movement on the device very local.”

The result is a chip that can perform 26 trillion operations per second and reach 672 frames per second on a ResNet-50 image classification benchmark—with an efficiency of 3.1 trillion operations per second per watt. (You can see the Hailo-8 compared to other AI accelerators on the MLPerf benchmarks here.)

The chips is targeted at edge systems—including partially autonomous vehicles, smart cameras, smartphones, drones and AR/VR gadgets—that need sophisticated deep learning.

The new cash infusion was led by existing investors, but Hailo also won new investment from the venture capital arms of ABB, NEC, and London VC firm Latitude Ventures.

The Conversation (0)

Two Startups Are Bringing Fiber to the Processor

Avicena’s blue microLEDs are the dark horse in a race with Ayar Labs’ laser-based system

5 min read
Diffuse blue light shines from a patterned surface through a ring. A blue cable leads away from it.

Avicena’s microLED chiplets could one day link all the CPUs in a computer cluster together.


If a CPU in Seoul sends a byte of data to a processor in Prague, the information covers most of the distance as light, zipping along with no resistance. But put both those processors on the same motherboard, and they’ll need to communicate over energy-sapping copper, which slow the communication speeds possible within computers. Two Silicon Valley startups, Avicena and Ayar Labs, are doing something about that longstanding limit. If they succeed in their attempts to finally bring optical fiber all the way to the processor, it might not just accelerate computing—it might also remake it.

Both companies are developing fiber-connected chiplets, small chips meant to share a high-bandwidth connection with CPUs and other data-hungry silicon in a shared package. They are each ramping up production in 2023, though it may be a couple of years before we see a computer on the market with either product.

Keep Reading ↓Show less