Do carmakers know what their self-driving car architecture will be in 2020? – Part 2

Article By : Junko Yoshida

Find out about different autonomous car platforms and how they combine deep learning, sensor fusion and surround vision in the second part of the series.

Determinism and C/C++

But isn’t this “supercomputing-on-chip” pitch for autonomous cars similar to what Nvidia is promoting with its Drive PX?

Nvidia calls Drive PX “the world's most advanced autonomous car platform,” combining deep learning, sensor fusion, and surround vision.

The big difference, as Baissus contends, lies in two things. First, Kalray’s solution is “certifiable.” By certifiable, he said, “I mean we can prove determinism, and we can guarantee timing,” he said. “In high-performance computing, one-second delay is OK. But in a critical embedded market — such as aerospace and automotive — 10 millisecond delay could be fatal.”

Second, for Nvidia’s chip, programmers need to know CUDA, he said. “Our chip can run standard C/C++ code using standard tools and Linux.” Automakers already have a lot of legacy code, algorithms written in C. Even when carmakers move onto the new autonomous car platform, legacy code will be important, Baissus explained.

Nvidia is not alone in anticipating the need for a lot more processing power. Mobileye upped the ante, recently, by “pre-announcing” EyeQ5. The company is promising to deliver engineering samples in 2018.

EyeQ5, designed in advanced 10nm or below-FinFET technology node, will feature eight multithreaded CPU cores coupled with eighteen cores of Mobileye's next-generation vision processors. The company said that the EyeQ5 will produce more than 12 Tera operations per second, while keeping power consumption below 5W.
Nobody, including Baissus, is taking Mobileye lightly. Unlike Nvidia’s Drive PX, which many industry observers regard as a “test platform” for autonomous cars, Mobileye is going after the commercial market with increased processing power at a much lower power consumption level.

By leveraging its proven vision processing algorithms, EyeQ5 is now taking on data fusion — combining 20 external sensors (camera, radar or lidar) — into one.

But can EyeQ5 master the ECU inside the autonomous car? A Mobileye spokesman explained to EE Times that EyeQ5 will do not just “data fusion” but also “decision making.” But where that decision translates into action will take place elsewhere — on a “low-level ECU” chosen by automakers, he added.

Kalray is positioning the role of its manycore processor somewhat differently from Mobileye or Nvidia.

Need for a Super ECU?

Baissus told EE Times, “There have been a lot of advances made in sensors and machine-learning algorithms” necessary for autonomous cars. “But nothing has been really done in the processor domain.” This is where Kalray sees its opening.

In his opinion, the next-generation processors in autonomous cars need to perform functions well beyond data fusion. “They have to act more as open platforms,” he said. Kalray hopes to provide an autonomous vehicle open processing hub, which he calls a “Super ECU.”

The super ECU is capable of integrating dozens of multi-domain functions on the same die. It will bring superior results in such critical segments as “sensing, learning, security, network, safety and cost,” he explained.

Without naming names, Baissus told EE Times that leading car OEMs and Tier Ones are using current the Kalray platform to build their first prototype fleets. Baissus acknowledged that the autonomous car’s system architecture is “still not mature.”

But through collaborations with key players, Baissus hopes to learn more about carmakers’ needs, in defining Kalray’s new generation of solution for autonomous cars.

Asked about his business model, Baissus acknowledged that licensing its MPPA architecture to other automotive chip suppliers is also an option.
System architecture

Asked to compare how system architecture might evolve, Baissus shared the following diagrams.

[System architecture of Car today (Level 1, Level 2)]
__Figure 1:__ *System architecture of Car today (Level 1, Level 2). (Source: Kalray)*

Today’s cars have a collection of localized ECUs. They combine sensor, processing and control functions.

[System architecture of Car tomorrow (Level 3, 4, 5)]
__Figure 2:__ *System architecture of Car tomorrow (Level 3, 4, 5). (Source: Kalray)*

Kalray hopes to offer a Master ECU that aggregates ECUs for density and cost and sensor data for smarter control. Note that he isn’t saying the Master ECU will do everything. If machine learning, for example, needs to accelerate its algorithm, the “master ECU will connect to accelerators when needed or a dedicated ECU,” Baissus explained.

In the end, Kalray believes its manycore architecture can shine in autonomous cars in numerous ways. It can run “dozens of different control and data processing algorithms in parallel and in real time.” Further, it can offer highly efficient machine learning. But most important, it offers very low latency.

Subscribe to Newsletter

Test Qr code text s ss