Massive Autonomous Vehicle Sensor Data – What Does It Mean?

2

It was reported by Ford that connected car sensors generate 25 gigabytes of data per hour, and then by the WSJ that a typical autonomous vehicle generates 4 terabytes of data in 90 minutes, and then by Intel of 45 terabits per hour.  All are massive numbers, but why so different?  It points to a bandwidth problem.  The raw data is beyond any auto OEM’s ability to manage, even in 5G, and so the amounts reported could be raw data or some prioritized data, that facilitate core vehicle diagnostics and operating behavior improvements.  

One mitigating factor is local bandwidth.  In-vehicle computing offers the potential for sharing the computational workload.  Sensor systems like Mobileye’s onboard camera are integrating sophisticated machine-learning vision algorithms.  Likewise, Lidar innovator Innoviz is building a software stack which has a similar strategy.  One also can speculate a future where non-critical, near-redundant data capturing is efficiently crowdsourced.

Nevertheless, today there are in excess of 100 sensors onboard, and the auto sensor market is expected to grow over 100% per year, and reach nearly 200m units in 2021.  Advances in the more powerful sensors (camera, lidar) will produce richer data, and require more bandwidth.  Furthermore, physical world updating in real-time, and sensor-fusion for collating and acting will add more bandwidth demands.  Taken together, we can be sure bandwidth issues will continue as a challenge for OEMs.

So, what does it mean for the OEM?

 In an autonomous vehicle future, OEMs will differentiate by onboard data processing.  OEMs become vehicle operators and software companies, and their success metrics will lean on navigation performance and incidence rates.  Theoretically, as software companies, with massive real-time data, they can choose to optimize data processing between vehicle operating behavior and the environment beyond the vehicle.  Practically, OEMs will be compelled to prioritize vehicle operating behavior.

So, what does it mean for real-time third party applications?

 Even strong third party applications may be deprioritized.  Weather is a good example.  Weather modeling today is coarse and heavily interpolated.  With onboard environmental and weather sensors, the ability to mine very detailed and very real-time data would enable ultra-precise weather models, of interest to utilities, infrastructure companies, smart cities.  Interestingly, Continental has been testing vehicle swarm crowdsourcing to capture hyper-local weather sensor data.

So, what does it mean for AV technology companies?

While AV technology companies will continue to innovate and differentiate based on performance and system architecture, they will also increasingly be tested on bandwidth consumption.  Here are some companies to watch, and their bandwidth strategies:



Todd Simon
Todd Simon is a geospatial specialist with London investment firm DN Capital, and founder of Geospatial Alpha, to stimulate innovation in geospatial intelligence. Todd has also served as an advisor to the European Space Agency. He is a former senior technology investment banker with DLJ, Credit Suisse, and Oppenheimer, and has held executive and board positions with several leading geospatial technology companies.