Cameras and Sensors for Embedded Vision
WHILE ANALOG CAMERAS ARE STILL USED IN MANY VISION SYSTEMS, THIS SECTION FOCUSES ON DIGITAL IMAGE SENSORS
While analog cameras are still used in many vision systems, this section focuses on digital image sensors—usually either a CCD or CMOS sensor array that operates with visible light. However, this definition shouldn’t constrain the technology analysis, since many vision systems can also sense other types of energy (IR, sonar, etc.).
The camera housing has become the entire chassis for a vision system, leading to the emergence of “smart cameras” with all of the electronics integrated. By most definitions, a smart camera supports computer vision, since the camera is capable of extracting application-specific information. However, as both wired and wireless networks get faster and cheaper, there still may be reasons to transmit pixel data to a central location for storage or extra processing.
A classic example is cloud computing using the camera on a smartphone. The smartphone could be considered a “smart camera” as well, but sending data to a cloud-based computer may reduce the processing performance required on the mobile device, lowering cost, power, weight, etc. For a dedicated smart camera, some vendors have created chips that integrate all of the required features.
Cameras
Until recent times, many people would imagine a camera for computer vision as the outdoor security camera shown in this picture. There are countless vendors supplying these products, and many more supplying indoor cameras for industrial applications. Don’t forget about simple USB cameras for PCs. And don’t overlook the billion or so cameras embedded in the mobile phones of the world. These cameras’ speed and quality have risen dramatically—supporting 10+ mega-pixel sensors with sophisticated image processing hardware.
Consider, too, another important factor for cameras—the rapid adoption of 3D imaging using stereo optics, time-of-flight and structured light technologies. Trendsetting cell phones now even offer this technology, as do latest-generation game consoles. Look again at the picture of the outdoor camera and consider how much change is about to happen to computer vision markets as new camera technologies becomes pervasive.
Sensors
Charge-coupled device (CCD) sensors have some advantages over CMOS image sensors, mainly because the electronic shutter of CCDs traditionally offers better image quality with higher dynamic range and resolution. However, CMOS sensors now account for more 90% of the market, heavily influenced by camera phones and driven by the technology’s lower cost, better integration and speed.

Sony Semiconductor Demonstration of Its Open-source Edge AI Stack with the IMX500 Intelligent Sensor
JF Joly, Product Manager for the AITRIOS platform at Sony Semiconductor, demonstrates the company’s latest edge AI and vision technologies and products at the 2025 Embedded Vision Summit. Specifically, Joly demonstrates Sony’s fully open-source software stack that enables the creation of AI-powered cameras using the IMX500 intelligent vision sensor. In this demo, Joly illustrates how

First Sensor Module for High-resolution Image Sensor IMX811 (247 Mpixel) Available Soon
Embedded vision engineers and tech leads can start preparing for high-resolution applications. Munich – July 17th 2025 – The first on the market for engineers and tech leads in the embedded vision sector: FRAMOS will soon release the first image sensor module and development kit available on the market for Sony’s new high-resolution IMX811 image

Sony Semiconductor Demonstration of On-sensor YOLO Inference with the Sony IMX500 and Raspberry Pi
Amir Servi, Edge AI Product Manager at Sony Semiconductors, demonstrates the company’s latest edge AI and vision technologies and products at the 2025 Embedded Vision Summit. Specifically, Servi demonstrates the IMX500 — the first vision sensor with integrated edge AI processing capabilities. Using the Raspberry Pi AI Camera and Ultralytics YOLOv11n models, Servi showcases real-time

Namuga Vision Connectivity Demonstration of Compact Solid-state LiDAR for Automotive and Robotics Applications
Min Lee, Business Development Team Leader at Namuga Vision Connectivity, demonstrates the company’s latest edge AI and vision technologies and products at the 2025 Embedded Vision Summit. Specifically, Lee demonstrates a compact solid-state LiDAR solution tailored for automotive and robotics industries. This solid-state LiDAR features high precision, fast response time, and no moving parts—ideal for

How Driver Monitoring Cameras Improve Driving Safety and Their Key Features
This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. Driver monitoring cameras have become widely accepted as a force in improving road safety. They go a long way to address the risks associated with driver inattention and fatigue by helping continuously observe driver behavior.

Namuga Vision Connectivity Demonstration of an AI-powered Total Camera System for an Automotive Bus Solution
Min Lee, Business Development Team Leader at Namuga Vision Connectivity, demonstrates the company’s latest edge AI and vision technologies and products at the 2025 Embedded Vision Summit. Specifically, Lee demonstrates his company’s AI-powered total camera system. The system is designed for integration into public transportation, especially buses, enhancing safety and automation. It includes front-view, side-view,

Namuga Vision Connectivity Demonstration of a Real-time Eye-tracking Camera Solution with a Glasses-free 3D Display
Min Lee, Business Development Team Leader at Namuga Vision Connectivity, demonstrates the company’s latest edge AI and vision technologies and products at the 2025 Embedded Vision Summit. Specifically, Lee demonstrates a real-time eye-tracking camera solution that accurately detects the viewer’s eye position and angle. This data enables a glasses-free 3D display experience using an advanced

Teledyne FLIR Demonstration of an Advanced Thermal Imaging Camera Enabling Automotive Safety Improvements
Ethan Franz, Senior Software Engineer at Teledyne FLIR, demonstrates the company’s latest edge AI and vision technologies and products in Lattice Semiconductor’s booth at the 2025 Embedded Vision Summit. Specifically, Franz demonstrates a state-of-the-art thermal imaging camera for automotive safety applications, designed using Lattice FPGAs. This next-generation camera, also incorporating Teledyne FLIR’s advanced sensing technology,

RealSense Completes Spinout from Intel, Raises $50 Million to Accelerate AI-powered Vision for Robotics and Biometrics
The newly independent company is set to lead in computer vision and machine perception for physical AI and beyond SAN FRANCISCO — July 11, 2025 — RealSense, a pioneer in AI-powered computer vision, today announced its successful spinout from Intel Corporation and the close of a $50 million Series A funding round. With investment led by

Lattice Semiconductor Demonstration of NVIDIA Holoscan-compatible Cameras from Valued Partners
Jacob Mercado, Applications Engineer at Lattice Semiconductor, demonstrates the company’s latest edge AI and vision technologies and products at the 2025 Embedded Vision Summit. Specifically, Mercado demonstrates Holoscan camera solutions from ecosystem partner vendors Leopard Imaging and e-con Systems, designed with Lattice FPGAs for use with NVIDIA Orin AGX or IGX platforms. Holoscan is enabled