This blog post was originally published in the late March 2016 edition of BDTI's InsideDSP newsletter. It is reprinted here with the permission of BDTI.
It's no secret that sensors are proliferating. Our smartphones, for example, contain accelerometers, magnetometers, ambient light sensors, microphones – over a dozen distinct types of sensors. A modern automobile contains roughly 200 sensors.
As sensors proliferate, the amount of data generated by these sensors grows too, of course. But different types of sensors produce vastly different amounts of data. As Chris Rowen, CTO of Cadence's IP group, recently pointed out in an excellent presentation, image sensors occupy a unique position in the sensor world. While the number of image sensors deployed across all products is a small fraction of the total number of sensors deployed, the amount of data generated by deployed image sensors dwarfs the amount of data generated by all other types of sensors combined.
The math behind this is simple: image sensors generate a lot of data. Even an old-fashioned VGA resolution image sensor running at 30 frames per second generates over 25 megabytes of color video data per second. That's 1,000 to 1,000,000 times more data per second than most other common sensor types.
You may point out that data is not the same thing as information. Indeed, not every pixel streaming from an image sensor contains useful information. But, we're discovering that there's valuable information contained in more pixels than we might have suspected.
Most sensors are single-purpose: one type of sensor for temperature, another for magnetic field, another for ambient light, etc. Image sensors are unique in that – when coupled with the right algorithms and sufficient processing power – they can become "software-defined sensors," capable of measuring many different types of things.
For example, using video of a person's face and shoulders, it's possible to identify the person, estimate their emotional state, determine heart rate and respiration rate, detect intoxication and drowsiness, and determine where the person's gaze is directed.
Similarly, in cars and trucks, a single image sensor (or a small cluster of them) can detect other vehicles, brake lights, pedestrians, cyclists, lane markings, speed limit signs, and more. Demonstrating the versatility of vision sensors, the Mercedes-Benz "Magic Body Control" system measures the detailed topography the road surface. According to Mercedes: "The system is thus able to recognize an uneven road surface before you come to drive over it, thus enabling the suspension to adjust itself in order to counteract, as far as possible, the undulations in the road."
To paraphrase Chris Rowen's presentation: In the future, approximately 100% of the data generated by all types of deployed sensors will be image and video data. This creates enormous opportunities, and big challenges as well.
The opportunities stem from the fact that image and video data can provide so many different types of useful information. The challenges derive from the inherent complexity of reliably extracting valuable information from pixels. Fortunately, as an industry, we're making rapid progress on these challenges through a combination of more powerful and energy-efficient processors, more reliable computer vision and deep learning algorithms, and better development tools. These improvements are enabling us to build devices, systems and applications that are safer, more autonomous, more responsive and more capable.
The very real potential for practical computer vision technology to enable an enormous variety of smarter, easier-to-use products is what motivated me to dedicate a large portion of my time to organizing the Embedded Vision Summit conference, taking place May 2-4 in Santa Clara, California. The Summit is the only event in the world focused entirely on building better products using computer vision.
At the Summit, I'm eagerly looking forward to hearing what leading thinkers like Chris Rowen of Cadence, Gary Bradski of the OpenCV Foundation, Jeff Dean of the Google Brain project and Larry Matthies of NASA JPL have to say about the challenges and opportunities of deploying visual intelligence.
If you're interested in learning about computer vision business opportunities, capabilities, technology and techniques, I invite you to join me at the Embedded Vision Summit on May 2-4 in Santa Clara, California. This event, an educational forum for product creators interested in incorporating visual intelligence into electronic systems and software, is organized by the Embedded Vision Alliance. For details about this unique conference, and to register, please visit www.embedded-vision.com/summit.
Co-Founder and President, BDTI
Founder, Embedded Vision Alliance