Automotive

Automotive Applications for Embedded Vision

Vision products in automotive applications can make us better and safer drivers

Vision products in automotive applications can serve to enhance the driving experience by making us better and safer drivers through both driver and road monitoring.

Driver monitoring applications use computer vision to ensure that driver remains alert and awake while operating the vehicle. These systems can monitor head movement and body language for indications that the driver is drowsy, thus posing a threat to others on the road. They can also monitor for driver distraction behaviors such as texting, eating, etc., responding with a friendly reminder that encourages the driver to focus on the road instead.

In addition to monitoring activities occurring inside the vehicle, exterior applications such as lane departure warning systems can use video with lane detection algorithms to recognize the lane markings and road edges and estimate the position of the car within the lane. The driver can then be warned in cases of unintentional lane departure. Solutions exist to read roadside warning signs and to alert the driver if they are not heeded, as well as for collision mitigation, blind spot detection, park and reverse assist, self-parking vehicles and event-data recording.

Eventually, this technology will to lead cars with self-driving capability; Google, for example, is already testing prototypes. However many automotive industry experts believe that the goal of vision in vehicles is not so much to eliminate the driving experience but to just to make it safer, at least in the near term.

Physical AI: 8 Questions Every Engineering Leader Is Asking

This blog post was originally published at Geisel Software’s website. It is reprinted here with the permission of Geisel Software. Jensen Huang called it at CES 2025: “The next frontier of AI is physical.” Since then, the phrase has been everywhere — in investor decks, conference keynotes, and vendor pitches. But for the software engineering managers, directors, VPs, and

Read More »

Everything Is Going to Be Driven by Algorithms

This blog post was originally published at Ambarella’s website. It is reprinted here with the permission of Ambarella. A security camera on a warehouse loading dock captures 86,400 seconds of video every day. A fleet telematics recorder on a long-haul truck accumulates gigabytes of road footage between fuel stops. A surgical robot’s stereo cameras generate

Read More »

Falcon Takes Flight: Connect Tech’s Vehicle AI Platform Was Built for the Worst Conditions on Earth

This blog post was originally published at Macnica America’s website. It is reprinted here with the permission of Macnica America. Dust. Vibration. Mud. Sub-zero mornings and scorching afternoons. These are the conditions that kill most compute platforms before they ever reach production, and yet it’s exactly these conditions, on farms, construction sites, and open-pit mines, where

Read More »

Bosch and Qualcomm Expand Collaboration to Strategic ADAS Solutions

Highlights: High performance solutions: Bosch and Qualcomm aim to make ADAS solutions for enhanced safety and comfort available to everyone. Continued Business Momentum: Collaboration has secured significant new business wins for both next-generation ADAS and cockpit solutions. Proven Global Success: Bosch delivers over 10 million cockpit computers powered by Snapdragon® Cockpit Platforms. Global Market Penetration:

Read More »

From ADAS to Robotaxi: How Vision Systems Must Level Up to Meet New Mobility Use Cases (Part 2)

This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. Key Takeaways How urban lighting and motion define robotaxi imaging needs Which camera features support reliable perception during day and night operation Why unified AI vision boxes reduce latency and coordination gaps How integrated vision platforms

Read More »

From ADAS to Robotaxi: How to Overcome the Major Vision Challenges

This blog post was originally published at e-con Systems’ website. It is reprinted here with the permission of e-con Systems. Key Takeaways Why does robotaxi vision need more than task-driven ADAS sensing? Impact of long-duty operation and changing lighting on perception reliability Challenges faced across vehicles, cities, and operating conditions How visual data continuity affects

Read More »

CES 2026: Physical AI moves from concept to system architecture

This market analysis was originally published at the Yole Group’s website. It is reprinted here with the permission of the Yole Group. The world’s largest consumer electronics conference demonstrated the technical synergies between automotive and robotics. At CES 2026, there was a clear cross-sector message: Physical AI is the common language across the automotive, robotaxi

Read More »

How Lenovo is scaling Level 4 autonomous robotaxis on Arm

This blog post was originally published at Arm’s website. It is reprinted here with the permission of Arm. As L4 robotaxis shift from pilot to production, Arm offers the compute foundation needed to deliver end-to-end physical AI that scales across vehicle fleets. After years of autonomous driving pilots and controlled trials, the automotive industry is moving toward the production-scale deployment of Level 4 (L4) robotaxis. This marks

Read More »

What Does a GPU Have to Do With Automotive Security?

This blog post was originally published at Imagination Technologies’ website. It is reprinted here with the permission of Imagination Technologies. The automotive industry is undergoing the most significant transformation since the advent of electronics in cars. Vehicles are becoming software-defined, connected, AI-driven, and continuously updated. This evolution brings extraordinary new capability – but it also brings

Read More »

Accelerating next-generation automotive designs with the TDA5 Virtualizer™ Development Kit

This blog post was originally published at Texas Instruments’ website. It is reprinted here with the permission of Texas Instruments. Introduction Continuous innovation in high-performance, power-efficient systems-on-a-chip (SoCs) is enabling safer, smarter and more autonomous driving experiences in even more vehicles. As another big step forward, Texas Instruments and Synopsys developed a Virtualizer Development Kit™ (VDK) for the

Read More »

Here you’ll find a wealth of practical technical insights and expert advice to help you bring AI and visual intelligence into your products without flying blind.

Contact

Address

Berkeley Design Technology, Inc.
PO Box #4446
Walnut Creek, CA 94596

Phone
Phone: +1 (925) 954-1411
Scroll to Top