Tech enthusiasts tend to think of Nvidia for its PC components like the RTX 3000 line and gaming services like GeForce Now, but especially during the pandemic, these are far from the company’s only moneymakers. There’s also Nvidia’s enterprise level hardware, specifically its data centers. Nvidia’s data centers have been at the center of its story this year as the pandemic has forced businesses to move more of their operations into the cloud, but according to a new blog post the company posted earlier today, its data centers might also soon be bringing their calculations to the streets.
Earlier this week, Amazon’s Zoox, which develops driverless taxi, posted an in-depth breakdown of its latest robotaxi prototype. While it’s not publicly available quite yet, a testing fleet has been built and is currently on the streets on both closed courses and certain public roads. The Zoox robotaxi sets itself apart from competitors by focusing on a symmetrical design with no proper forward or backwards, which lets it give customers identical experiences regardless of where they sit and eliminates the need to reverse. It’s also got built-in touchscreens for communicating with the taxi and four-wheel steering to be able to get into tight spaces without parallel parking, but most importantly, it sports an Nvidia-powered machine-learning algorithm that incorporates cameras, radar, and lidar to constantly improve.
According to Nvidia, Zoox has been working directly with Nvidia to bring a level 5 autonomous vehicle- which is the highest tier of self-driving car and requires no human intervention- to streets since at least 2017. It’s also an alumnus of Nvidia Inception, which is the company’s incubator program for startups that focus on AI and data science.
So how does Zoox’s algorithm works? Essentially, while individual robotaxis will use their many sensors to drive safely in the moment, achieving a 360 degree field of view around all four corners of the vehicle, that data also gets sent back to an Nvidia data center equipped that helps beef up the company’s machine learning algorithm. That algorithm can then interpret that data to improve the driving AI, and beam updates out to its fleet of vehicles over-the-air every few weeks.
A human hand is also involved.