January 7, 2019

Nissan Merges Real and Virtual Worlds to Help Drivers See the "Invisible"


As the next step in providing a future with “zero emissions and zero fatalities” through its Nissan Intelligent Mobility philosophy, Nissan is using augmented reality to create the ultimate connected-car experience.

Dubbed Invisible-to-Visible, or I2V, it will support drivers by merging information from sensors outside and inside the vehicle with data from the cloud. This enables the system not only to track the vehicle’s immediate surroundings but also to anticipate what’s ahead—even showing what’s behind a building or around the corner. To make driving more enjoyable, guidance is given in an interactive, human-like way, such as through avatars that appear inside the car.

By tapping into the virtual world, I2V opens up endless possibilities for service and communication – making driving more convenient, comfortable and exciting.


I2V is powered by Nissan’s Omni-Sensing technology, which acts as a hub gathering real-time data from the traffic environment and from the vehicle’s surroundings and interior. Nissan’s SAM (Seamless Autonomous Mobility) technology analyzes the road environment through relevant real-time information, and the ProPILOT semiautonomous driver support system provides information about the car’s surroundings.

The technology maps a 360-degree virtual space around the car to provide information about things like road and intersection status, visibility, signage, or nearby pedestrians. It can also monitor the people inside the vehicle by using interior sensors to better anticipate when they may need assistance with finding something or a coffee break to stay alert.

I2V can also connect drivers and passengers to people in the Metaverse virtual world. This makes it possible for family, friends, or others to appear inside the car as three-dimensional, augmented-reality avatars to provide company or assistance.


During autonomous driving, I2V can make the time spent in a car more comfortable and enjoyable. For example, when driving in the rain, the scenery of a sunny day can be projected inside the vehicle.

When visiting a new place, the system can search within the Metaverse for a knowledgeable local guide who can communicate with people in the vehicle in real time.

Information provided by the guide can be collected with Omni-Sensing and stored in the cloud so that others visiting the same area can access the useful guidance. It can also be used by the onboard artificial intelligence system to provide a more efficient drive through local areas.

During manual driving, I2V provides information from Omni-Sensing as an overlay in the driver’s full field of view. The information helps drivers assess and prepare for things like corners with poor visibility, irregular road surface conditions, or oncoming traffic.


The driver can also book a professional driver from the Metaverse to get personal instruction in real time. The professional driver appears as a projected avatar or as a virtual chase car in the driver’s field of vision to demonstrate the best way to drive.

Not only will I2V display information about congestion and estimated travel time; it can communicate unique details to better inform the driver about the road ahead, removing unknowns to ease driving stress. I2V will give alternative suggestions, even down to the best-moving lane in heavy traffic, by using real-time local data mapping via Omni-Sensing.

Upon arrival at a destination, I2V can access SAM to scan for parking spaces and park the vehicle for the driver in situations requiring difficult parking maneuvers.

No comments:

Post a Comment

Feel free to comment or share your views. Comments that are derogatory and/or spam will not be tolerated. We reserve the right to moderate and/or remove comments.