Detection and Tracking
Camera sensor configuration, visual perception, lidar processing, tracking
and sensor fusion
Automated Driving Toolbox™ perception algorithms use data from cameras and lidar scans to
detect and track objects of interest and locate them in a driving scenario.
These algorithms are ideal for ADAS and autonomous driving applications,
such as automatic braking and steering.
- Camera Sensor Configuration
Monocular camera sensor calibration, image-to-vehicle coordinate
system transforms, bird’s-eye-view image transforms
- Visual Perception
Lane boundary, pedestrian, vehicle, and other object detections using
machine learning and deep learning
- Lidar Processing
Velodyne® file import, segmentation, downsampling, transformations,
visualization, and 3-D point cloud registration from lidar
- Tracking and Sensor Fusion
Object tracking and multisensor fusion, bird’s-eye plot of detections
and object tracks