Location: Cyril Magnin II

Duration: 2:40pm - 3:20pm

Day of week: Wednesday

Share this on:

Abstract

Autonomous vehicles need to perceive their surroundings and analyze it to make decisions and act in an environment. More specifically, the autonomous vehicle detects objects on the road and maneuver through the traffic utilizing smart functional modules. In recent years, artificial intelligence, in particular, deep neural networks, have been used widely to build these smart functional modules.

While object detection (putting bounding boxes around objects), and semantic segmentation (labeling each pixel in an image) has been the focus of many researchers in autonomous driving, these methods may fall short when it comes to forming a better social understanding of pedestrian intent. In this talk, we present our approach to pedestrian intent prediction and communication, which leverages more complex computer vision algorithms that estimate human pose rather than bounding boxes or pixel labels.

The increasingly sophisticated models, like the pose estimation networks we describe, show tremendous promise as they prove to be robust at approximating complex and non-linear mapping functions from images to outputs. However, these models are typically large and have a huge number of parameters resulting in a steep cost in terms of training and inference time resource requirements. This makes the use of these networks challenging on resource and power constrained embedded systems. In this talk, we also show that the compression of neural networks results in faster predictions with smaller deep neural networks.

Speaker: Nasim Souly

Senior Engineer & Machine Learning Researcher @Volkswagen

Nasim is a Senior Machine learning engineer at Volkswagen Group of America, Electronics Research Lab in Belmont, CA, where she is a member of perception and machine learning group working on applying deep learning to advanced topics in autonomous driving.

She holds a Ph.D. of computer science from the University of Central Florida, where her research mainly focused on computer vision applications of machine learning namely saliency detection, activity recognition, and semantic segmentation.