ON Semiconductor has announced that it is leveraging its sophisticated image sensor modeling technology to provide real-time data to the NVIDIA DRIVE Constellation™ simulation platform. The open, cloud-based platform performs bit-accurate simulation for large-scale, hardware-in-the-loop testing and validation of autonomous vehicles.
ON Semiconductor’s Image Sensor model receives both scene information and control signals from DRIVE Constellation to calculate and output a real-time image based on the inputs. It then transmits the simulated image back to DRIVE Constellation for processing. The complex sensor model will utilise all critical parameters in the path from converting photons to digital output (e.g. Quantum Efficiency, Noise, Gain, Analog-to-Digital Conversion, Black Level Correction and more) to provide an accurate output of a real-world image sensor.
“ON Semiconductor is excited to be one of NVIDIA’s ecosystem partners working on its DRIVE Constellation platform,” said Ross Jatou, Vice President and General Manager, Automotive Sensing Division, ON Semiconductor. “Our innovative, scalable image sensor solutions address the stringent and evolving needs of autonomous driving, so to be an integral part of such an important and enabling program that will help accelerate progress to safe, robust driverless vehicles, is really important to us.”
“The NVIDIA DRIVE platform delivers a full hardware and software ecosystem for developers and is supported by specialists in an array of different technologies and disciplines,” said Zvi Greenstein, general manager at NVIDIA. “ON Semiconductor’s advanced imaging solutions have demonstrated class-leading performance and their innovative Image Sensor model can benefit those engaged in the groundbreaking work to take autonomous driving from concepts under test to mainstream reality.”
For more information, visit http://www.onsemi.com.