Resources
In-depth information about the edge AI and vision applications, technologies, products, markets and trends.
The content in this section of the website comes from Edge AI and Vision Alliance members and other industry luminaries.
All Resources

Upcoming Webinar on NVIDIA IGX Thor
On April 15, 2026, at 9:00 pm PDT (12:00 pm EDT) NVIDIA will deliver a webinar “Unlock Real-Time Physical AI for the Industrial Edge” From the event page: Join us to learn how IGX Thor’s

Smart Home Connectivity: Trends, Challenges and the Role of Next-Gen IoT Technology
This blog post was originally published at Synaptics’ website. It is reprinted here with the permission of Synaptics. As the smart home market matures, consumers demand devices that are not just connected, but intelligent and interoperable.

Upcoming Webinar on Agentic Memory Systems
On April 16, 2026, at 1:00 pm EDT (10:00 am PDT) Boston.AI will deliver a webinar “Remembering to Forget: Agentic Memory Systems and Context Constraints” From the event page: As AI agents evolve from stateless

From hardware to intelligence: The Qualcomm AI camera platform for scalable security solutions
At ISC West 2026 Qualcomm Technologies showcases its vision for the future of smart camera and security Key Takeaways: End-to-end AI development tools and the Qualcomm Insight Platform to enable customers to develop AI features

Lightweight Keyword Spotting Solution from Microchip
Microchip presents a customizable, target-agnostic solution to program wake words and voice commands. The ML model, generated and tested using a custom application, has low latency and a minimal memory footprint, making it ideal for

Accelerating Multi-Die Innovation: How Synopsys and Samsung are Shaping Chip Design
This blog post was originally published at Synopsys’s website. It is reprinted here with the permission of Synopsys. With the semiconductor industry shifting from monolithic chip designs to multi-die architectures — which leverage chiplets for improved flexibility, scalability,

2026: The Year Intelligence Gets Physical
This article was originally published at Analog Devices’ website. It is reprinted here with the permission of Analog Devices. Artificial intelligence is entering a new phase where models interpret contextual data whilst interacting with the physical

Why Night HDR Is More Challenging Than Daytime HDR
This blog post was originally published at Visidon’s website. It is reprinted here with the permission of Visidon. High Dynamic Range (HDR) imaging has become a standard feature in modern cameras, from smartphones to automotive and

AI-Assisted Coding: The Next Step in Abstraction
I’ve been using AI-assisted coding for my work a lot recently, and I’ll admit, I wasn’t sure how I felt about it. Was I cheating? How do I know it’s right? Do I admit to
Technologies

Cadence and NVIDIA Expand Partnership to Reinvent Engineering for the Age of AI and Accelerated Computing
15 Apr 2026 Expanded collaboration combines agentic AI, physics-based simulation, and digital twins to accelerate engineering and unlock new levels of productivity across semiconductors, physical AI systems and AI factories SAN JOSE, Calif.— At CadenceLIVE Silicon Valley 2026, Cadence (Nasdaq: CDNS) announced an expanded partnership with NVIDIA to deliver accelerated solutions across agentic AI, physics-based simulation

Intel Launches Intel Core Series 3 Processors: Changing the Game for Everyday Computing
Intel® Core™ Series 3 brings advanced features and Intel’s latest architectures to value buyers, commercial and essential edge devices What’s New: Intel® today unveiled its new Intel Core™ Series 3 mobile processors, bringing advanced performance, exceptional battery life, and AI-ready to value buyers, commercial and essential edge devices. Purpose-engineered for value, Intel Core Series 3

Upcoming Webinar on Sony’s IMX925/935 Sensor Series and High Performance SLVS-EC Interface
On May 12, 2026, at 10:00 am CEST, RESTAR FRAMOS will deliver a webinar “Reaching High-Speed and High-Resolution Architecture with IMX925/935 and SLVS-EC” From the event page: From sensor architecture to real-world integration — join the engineers behind the technology High-speed and high-resolution machine vision systems are pushing the limits of data throughput, latency, and
Applications

MPEG-5 LCEVC: A practical shift for industrial AI video pipelines
This blog post was originally published at V-Nova’s website. It is reprinted here with the permission of V-Nova. In Industrial and Defense environments, I hear the same story. More cameras. Higher resolutions. Stricter latency targets. Infrastructure that cannot be replaced easily. And increasing pressure around storage, bandwidth, compute, and privacy. This is why MPEG-5 LCEVC is becoming even more relevant. It improves compression

“Non-Contact Vital Sign Monitoring Using Low-Cost WiFi Devices,” a Presentation from the University of California, Santa Cruz
Katia Obraczka, Professor of Computer Science and Engineering, University of California, Santa Cruz, Pranay Kocheta, Lexington High School, and Nayan Bhatia, University of California, Santa Cruz present the “Non-Contact Vital Sign Monitoring Using Low-Cost WiFi Devices” tutorial at the December 2025 Edge AI and Vision Innovation Forum.

Bosch and Qualcomm Expand Collaboration to Strategic ADAS Solutions
Highlights: High performance solutions: Bosch and Qualcomm aim to make ADAS solutions for enhanced safety and comfort available to everyone. Continued Business Momentum: Collaboration has secured significant new business wins for both next-generation ADAS and cockpit solutions. Proven Global Success: Bosch delivers over 10 million cockpit computers powered by Snapdragon® Cockpit Platforms. Global Market Penetration:
Functions

AI On: 3 Ways to Bring Agentic AI to Computer Vision Applications
This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. Learn how to integrate vision language models into video analytics applications, from AI-powered search to fully automated video analysis. Today’s computer vision systems excel at identifying what happens in physical spaces and processes, but lack the abilities to explain the

SAM3: A New Era for Open‑Vocabulary Segmentation and Edge AI
Quality training data – especially segmented visual data – is a cornerstone of building robust vision models. Meta’s recently announced Segment Anything Model 3 (SAM3) arrives as a potential game-changer in this domain. SAM3 is a unified model that can detect, segment, and even track objects in images and videos using both text and visual

TLens vs VCM Autofocus Technology
This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. In this blog, we’ll walk you through how TLens technology differs from traditional VCM autofocus, how TLens combined with e-con Systems’ Tinte ISP enhances camera performance, key advantages of TLens over mechanical autofocus systems, and applications
