Visual Perception for Self-Driving Cars
This course is part of Self-Driving Cars Specialization
Instructors: Steven Waslander +1 more
Instructor ratings
We asked all learners to give feedback on our instructors based on the quality of their teaching style.
What you'll learn
Skills you'll gain
There are 7 modules in this course
This course will introduce you to the main perception tasks in autonomous driving, static and dynamic object detection, and will survey common computer vision methods for robotic perception. By the end of this course, you will be able to work with the pinhole camera model, perform intrinsic and extrinsic camera calibration, detect, describe and match image features and design your own convolutional neural networks. You'll apply these methods to visual odometry, object detection and tracking, and semantic segmentation for drivable surface estimation. These techniques represent the main building blocks of the perception system for self-driving cars. For the final project in this course, you will develop algorithms that identify bounding boxes for objects in the scene, and define the boundaries of the drivable surface. You'll work with synthetic and real image data, and evaluate your performance on a realistic dataset. This is an advanced course, intended for learners with a background in computer vision and deep learning. To succeed in this course, you should have programming experience in Python 3.0, and familiarity with Linear Algebra (matrices, vectors, matrix multiplication, rank, Eigenvalues and vectors and inverses).
Module 1: Basics of 3D Computer Vision
Module 2: Visual Features - Detection, Description and Matching
Module 3: Feedforward Neural Networks
Module 4: 2D Object Detection
Module 5: Semantic Segmentation
Module 6: Putting it together - Perception of dynamic objects in the drivable region
Explore more from Software Development
©2025 ementorhub.com. All rights reserved