TECHNOLOGIES

What is a Dust Denoising Filter in TOF Camera, and How Does it Remove Noise Artifacts in Vision Systems?

This article was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. Time-of-Flight (ToF) cameras with IR sensors are susceptible to performance variations caused by environmental dust. This dust can create ‘dust noise’ in the output depth map, directly impacting camera accuracy and, consequently, the reliability of critical […]

What is a Dust Denoising Filter in TOF Camera, and How Does it Remove Noise Artifacts in Vision Systems? Read More +

Low-Light Image Enhancement: YUV vs RAW – What’s the Difference?

This blog post was originally published at Visidon’s website. It is reprinted here with the permission of Visidon. In the world of embedded vision—whether for mobile phones, surveillance systems, or smart edge devices—image quality in low-light conditions can make or break user experience. That’s where advanced AI-based denoising algorithms come into play. At our company, we

Low-Light Image Enhancement: YUV vs RAW – What’s the Difference? Read More +

NVIDIA-Accelerated Mistral 3 Open Models Deliver Efficiency, Accuracy at Any Scale

This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. The new Mistral 3 open model family delivers industry-leading accuracy, efficiency, and customization capabilities for developers and enterprises. Optimized from NVIDIA GB200 NVL72 to edge platforms, Mistral 3 includes: One large state-of-the-art sparse multimodal and multilingual mixture of

NVIDIA-Accelerated Mistral 3 Open Models Deliver Efficiency, Accuracy at Any Scale Read More +

Ambarella’s CV3-AD655 Surround View with IMG BXM GPU: A Case Study

The CV3-AD family block diagram. This blog post was originally published at Imagination Technologies’ website. It is reprinted here with the permission of Imagination Technologies. Ambarella’s CV3-AD655 autonomous driving AI domain controller pairs energy-efficient compute with Imagination’s IMG BXM GPU to enable real-time surround-view visualisation for L2++/L3 vehicles. This case study outlines the industry shift

Ambarella’s CV3-AD655 Surround View with IMG BXM GPU: A Case Study Read More +

e-con Systems to Launch Darsi Pro, an NVIDIA Jetson-Powered AI Compute Box for Advanced Vision Applications

This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. This blog offers expert insights into Darsi Pro, how it delivers a unified vision solution, and what it brings to alleviate modern workloads. Darsi Pro comes with GMSL camera options, rugged design, OTA support, and

e-con Systems to Launch Darsi Pro, an NVIDIA Jetson-Powered AI Compute Box for Advanced Vision Applications Read More +

Overcoming the Skies: Navigating the Challenges of Drone Autonomy

This blog post was originally published at Inuitive’s website. It is reprinted here with the permission of Inuitive. From early military prototypes to today’s complex commercial operations, drones have evolved from experimental aircraft into essential tools across industries. Since the FAA issued its first commercial permit in 2006, applications have rapidly expanded—from disaster relief and

Overcoming the Skies: Navigating the Challenges of Drone Autonomy Read More +

NVIDIA Advances Open Model Development for Digital and Physical AI

This blog post was originally published at NVIDIA’s website. It is reprinted here with the permission of NVIDIA. NVIDIA releases new AI tools for speech, safety and autonomous driving — including NVIDIA DRIVE Alpamayo-R1, the world’s first open industry-scale reasoning vision language action model for mobility — and a new independent benchmark recognizes the openness and

NVIDIA Advances Open Model Development for Digital and Physical AI Read More +

OpenVINO 2025.4 Release Broadens Model Support

OpenVINO 2025.4 is very much an edge-first release: it tightens the loop between perception, language, and action across AI PCs, embedded devices, and near-edge servers. On the model side, Intel is clearly optimizing for “local RAG + agents.” CPUs and GPUs now get first-class support for Qwen3-Embedding-0.6B and Qwen3-Reranker-0.6B, plus Mistral-Small-24B-Instruct-2501, giving developers a compact

OpenVINO 2025.4 Release Broadens Model Support Read More +

AMD Spartan UltraScale+ FPGA Kit Adds Proven Infineon HyperRAM Support for Edge AI Designs

Somewhat eclipsed by last week’s announcement that the AMD Spartan™ UltraScale+™ FPGA SCU35 Evaluation Kit is now available, AMD and Infineon have disclosed successful validation of Infineon’s 64-Mb HYPERRAM™ memory and HYPERRAM controller IP on the platform. This collaboration expands the kit’s value for engineers building edge AI and computer vision systems. The SCU35 kit,

AMD Spartan UltraScale+ FPGA Kit Adds Proven Infineon HyperRAM Support for Edge AI Designs Read More +

Breaking the Human Accuracy Barrier in Computer Vision Labeling

This article was originally published at 3LC’s website. It is reprinted here with the permission of 3LC. Introduction There’s been a lot of excitement lately around how foundation models (such as CLIP, SAM, Grounding DINO, etc.) can come close to human-level performance when labeling common objects, saving a ton of labeling effort and cost. It’s impressive progress. However,

Breaking the Human Accuracy Barrier in Computer Vision Labeling Read More +

Here you’ll find a wealth of practical technical insights and expert advice to help you bring AI and visual intelligence into your products without flying blind.

Contact

Address

Berkeley Design Technology, Inc.
PO Box #4446
Walnut Creek, CA 94596

Phone
Phone: +1 (925) 954-1411
Scroll to Top