Resources
In-depth information about the edge AI and vision applications, technologies, products, markets and trends.
The content in this section of the website comes from Edge AI and Vision Alliance members and other industry luminaries.
All Resources

Texas Instruments, D3 Embedded, Lattice and NVIDIA Show a Practical Radar-Camera Fusion Stack for Robotics
TI’s new application brief and companion demo outline how mmWave radar, camera input, FPGA-based sensor bridging and NVIDIA Holoscan can be combined into a low-latency perception pipeline for humanoids and other autonomous machines. Texas

Upcoming Webinar on Akida Radar Reference Platform
On April 20, 2026, at 8:00 pm PDT (11:00 am EDT) BrainChip will deliver a webinar “Akida Radar Reference Platform: See the Evolution of Radar Intelligence with AI-Powered Object Classification” From the event page: Join

From Connected to Aware: How PSOC™ Edge enables the next wave of smart devices
This blog post was originally published at Infineon’s website. It is reprinted here with the permission of Infineon. Across home, retail, and industry, devices that once followed simple rules are now expected to understand people
Building Robotics Applications with Ryzen AI and ROS 2
This blog post was originally published at AMD’s website. It is reprinted here with the permission of AMD. This blog showcases how to deploy power-efficient Ryzen AI perception models with ROS 2 – the Robot Operating


BrainChip Unveils Radar Reference Platform to Bridge the ‘Identification Gap’ in Edge AI
LAGUNA HILLS, Calif. — April 6, 2026 — BrainChip Holdings Ltd (ASX: BRN, OTCQX: BRCHF, BCHPY), the world’s first commercial producer of ultra-low-power, neuromorphic AI technology, today announced the launch of its Radar Reference Platform. This fully validated hardware and AI

e-con Systems Launches STURDeCAM57: A 5MP Global Shutter RGB-IR Camera for In-cabin Monitoring Systems
California & Chennai (March 31, 2026): e-con Systems®, a global leader in embedded vision solutions, launches STURDeCAM57, a 5MP global shutter RGB-IR GMSL2 camera designed to deliver reliable, context-rich vision from day to night for
The Future of Security Is Already Running. Here Is What It Looks Like.
This blog post was originally published at Axelera AI’s website. It is reprinted here with the permission of Axelera AI. A camera sees everything and understands nothing. For decades, that has been the fundamental limitation of

Bringing AI Closer to the Edge and On-Device with Gemma 4
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. The Gemmaverse expands with the launch of the latest Gemma 4 multimodal and multilingual models, designed to scale across the
Technologies

Texas Instruments, D3 Embedded, Lattice and NVIDIA Show a Practical Radar-Camera Fusion Stack for Robotics
TI’s new application brief and companion demo outline how mmWave radar, camera input, FPGA-based sensor bridging and NVIDIA Holoscan can be combined into a low-latency perception pipeline for humanoids and other autonomous machines. Texas Instruments, D3 Embedded, Lattice Semiconductor and NVIDIA are outlining a concrete radar-camera fusion stack for robotics rather than just talking

Upcoming Webinar on Akida Radar Reference Platform
On April 20, 2026, at 8:00 pm PDT (11:00 am EDT) BrainChip will deliver a webinar “Akida Radar Reference Platform: See the Evolution of Radar Intelligence with AI-Powered Object Classification” From the event page: Join us on 20 April at 8:00 AM PT for an exclusive deep dive into BrainChip’s Radar Reference Platform — bringing

From Connected to Aware: How PSOC™ Edge enables the next wave of smart devices
This blog post was originally published at Infineon’s website. It is reprinted here with the permission of Infineon. Across home, retail, and industry, devices that once followed simple rules are now expected to understand people and context. A thermostat shouldn’t just follow a schedule; it should know if anyone is in the room and choose the preferred
Applications

Texas Instruments, D3 Embedded, Lattice and NVIDIA Show a Practical Radar-Camera Fusion Stack for Robotics
TI’s new application brief and companion demo outline how mmWave radar, camera input, FPGA-based sensor bridging and NVIDIA Holoscan can be combined into a low-latency perception pipeline for humanoids and other autonomous machines. Texas Instruments, D3 Embedded, Lattice Semiconductor and NVIDIA are outlining a concrete radar-camera fusion stack for robotics rather than just talking

From Connected to Aware: How PSOC™ Edge enables the next wave of smart devices
This blog post was originally published at Infineon’s website. It is reprinted here with the permission of Infineon. Across home, retail, and industry, devices that once followed simple rules are now expected to understand people and context. A thermostat shouldn’t just follow a schedule; it should know if anyone is in the room and choose the preferred
Building Robotics Applications with Ryzen AI and ROS 2
This blog post was originally published at AMD’s website. It is reprinted here with the permission of AMD. This blog showcases how to deploy power-efficient Ryzen AI perception models with ROS 2 – the Robot Operating System. We utilize the Ryzen AI Max+ 395 (Strix-Halo) platform, which is equipped with an efficient Ryzen AI NPU and
Functions

AI On: 3 Ways to Bring Agentic AI to Computer Vision Applications
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Learn how to integrate vision language models into video analytics applications, from AI-powered search to fully automated video analysis. Today’s computer vision systems excel at identifying what happens in physical spaces and processes, but lack the abilities to explain the

SAM3: A New Era for Open‑Vocabulary Segmentation and Edge AI
Quality training data – especially segmented visual data – is a cornerstone of building robust vision models. Meta’s recently announced Segment Anything Model 3 (SAM3) arrives as a potential game-changer in this domain. SAM3 is a unified model that can detect, segment, and even track objects in images and videos using both text and visual

TLens vs VCM Autofocus Technology
This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. In this blog, we’ll walk you through how TLens technology differs from traditional VCM autofocus, how TLens combined with e-con Systems’ Tinte ISP enhances camera performance, key advantages of TLens over mechanical autofocus systems, and applications
