announced the world’s first commercially available Level 2+ automated driving
system, NVIDIA DRIVE AutoPilot, which integrates multiple breakthrough AI
technologies that will enable supervised self-driving vehicles to go into
production by next year.
CES 2019, leading automotive suppliers Continental and ZF announced Level 2+
self-driving solutions based on NVIDIA DRIVE, with production starting in 2020.
As a Level 2+ self-driving solution, NVIDIA DRIVE AutoPilot uniquely provides
both world-class autonomous driving perception and a cockpit rich in AI
capabilities. Vehicle manufacturers can use it to bring to market sophisticated
automated driving features — as well as intelligent cockpit assistance and
visualization capabilities — that far surpass today’s ADAS offerings in
performance, functionality and road safety.
full-featured, Level 2+ system requires significantly more computational
horsepower and sophisticated software than what is on the road today,” said Rob
Csongor, vice president of Autonomous Machines at NVIDIA, adding that, “NVIDIA
DRIVE AutoPilot provides these, making it possible for carmakers to quickly
deploy advanced autonomous solutions by 2020 and to scale this solution to
higher levels of autonomy faster.”
AutoPilot integrates for the first time high-performance NVIDIA Xavier
system-on-a-chip (SoC) processors and the latest NVIDIA DRIVE Software to
process many deep neural networks (DNNs) for perception as well as complete
surround camera sensor data from outside the vehicle and inside the cabin. This
combination enables full self-driving autopilot capabilities, including highway
merge, lane change, lane splits and personal mapping. Inside the cabin,
features include driver monitoring, AI copilot capabilities and advanced
in-cabin visualization of the vehicle’s computer vision system.
AutoPilot is part of the open, flexible NVIDIA DRIVE platform, which is being
used by hundreds of companies worldwide to build autonomous vehicle solutions
that increase road safety while reducing driver fatigue and stress on long
drives or in stop-and-go traffic. The new Level 2+ system complements the
NVIDIA DRIVE AGX Pegasus system that provides Level 5 capabilities for robotaxis.
AutoPilot addresses the limitations of existing Level 2 ADAS systems, which a
recent Insurance Institute for Highway Safety study showed offer inconsistent
vehicle detections and poor ability to stay within lanes on curvy or hilly
roads, resulting in a high occurrence of system disengagements where the driver
abruptly had to take control.
keeping and adaptive cruise control systems on the market today are simply not
living up to the expectations of consumers,” said Dominique Bonte, vice president
of Automotive Research at ABI Research. “The high-performance AI solutions from
NVIDIA will deliver more effective active safety and more reliable automated
driving systems in the near future, ” Bonte said.
Xavier SoC: Processing at 30 Teraops a
to NVIDIA DRIVE AutoPilot is the Xavier SoC, which delivers 30 trillion
operations per second of processing capability. Architected for safety, Xavier
has been designed for redundancy and diversity, with six types of processors
and 9 billion transistors that enable it to process vast amounts of data in
NVIDIA Introduces DRIVE AutoPilot
is the world’s first automotive-grade processor for autonomous driving and is
in production today. Global safety experts have assessed its architecture and
development process as suitable for building a safe product.
DRIVE AutoPilot software stack integrates DRIVE AV software for handling
challenges outside the vehicle, as well as DRIVE IX software for tasks inside
AV uses surround sensors for full, 360-degree perception and features highly
accurate localization and path-planning capabilities. These enable supervised
self-driving on the highway, from on-ramp to off-ramp. Going beyond basic
adaptive cruise control, lane keeping and automatic emergency braking, its
surround perception capabilities handle situations where lanes split or merge,
and safely perform lane changes.
AV also includes a diverse and redundant set of advanced DNN technologies that
enable the vehicle to perceive a wide range of objects and driving situations,
including DriveNet, SignNet, LaneNet, OpenRoadNet and WaitNet. This
sophisticated AI software understands where other vehicles are, reads lane
markings, detects pedestrians and cyclists, distinguishes different types of
lights and their colors, recognizes traffic signs and understands complex
addition to providing precise localisation to the world’s HD maps for vehicle
positioning on the road, DRIVE AutoPilot offers a new personal mapping feature
called “My Route,” which remembers where you have driven and can create a
self-driving route even if no HD map is available.
the vehicle, DRIVE IX intelligent experience software enables occupant
monitoring to detect distracted or drowsy drivers and provide alerts, or take
corrective action if needed. It is also used to create intelligent user
experiences, including the new ability for augmented reality. Displaying a
visualization of the surrounding environment sensed by the vehicle, as well as
planned route, instills trust in the system.
next-generation user experiences in the vehicle, the AI capabilities of DRIVE
IX can also be used to accelerate natural language processing, gaze tracking or
ExxonMobil Lubricants Pvt Ltd announced the launch of its next-generation passenger vehicle lubricants - ‘Mobil Super,’ in new, improved packaging with refreshed labels. The improved packaging offers ...
Michelin, the world’s leading sustainable mobility company, has become the first tyre brand in the passenger vehicle segment in India to be accredited with the newly introduced star labelling programm...
With an aim to reduce manual intervention and enhance vehicle inspections, MyRaasta, a car and bike service aggregator, has introduced Artificial Intelligence (AI)-based vehicle inspection technology ...