Fwd: RI Ph.D. Thesis Proposal: Adam Villaflor - starting in 8 minutes!

Jeff Schneider jeff4 at andrew.cmu.edu
Fri Sep 16 12:53:00 EDT 2022


Hi Everyone,

Please come hear Adam tell us about self-driving cars and RL in his 
thesis proposal!

Jeff.




-------- Forwarded Message --------
Subject: 	RI Ph.D. Thesis Proposal: Adam Villaflor
Date: 	Wed, 7 Sep 2022 13:58:33 -0400
From: 	Suzanne Muth <lyonsmuth at cmu.edu>
To: 	ri-people at lists.andrew.cmu.edu



Date: 16 September 2022
Time: 1:00 p.m. (ET)
Location: GHC 4405
Zoom Link: 
https://cmu.zoom.us/j/95581369179?pwd=THdjUnVkQnBaUUFlNDdOcDBMcVhNQT09
Type: Ph.D. Thesis Proposal
Who: Adam Villaflor
Title: Combining Offline Reinforcement Learning with Stochastic 
Multi-Agent Planning for Autonomous Driving

Abstract:
Fully autonomous vehicles have the potential to greatly reduce vehicular 
accidents and revolutionize how people travel and how we transport 
goods. Many of the major challenges for autonomous driving systems 
emerge from the numerous traffic situations that require complex 
interactions with other agents. For the foreseeable future, autonomous 
vehicles will have to share the road with human-drivers and pedestrians, 
and thus cannot rely on centralized communication to address these 
interactive scenarios. Therefore, autonomous driving systems need to be 
able to negotiate and respond to unknown agents that exhibit uncertain 
behavior. To tackle these problems, most commercial autonomous driving 
stacks use a modular approach that splits perception, agent forecasting, 
and planning into separately engineered modules. By decomposing 
autonomous driving into smaller modules, it allows for simplifying 
abstractions and greater parallelization of the engineering effort.

However, fully separating prediction and planning makes it difficult to 
reason about how other vehicles will respond to the planned trajectory 
for the controlled ego-vehicle. Thus to maintain safety, many modular 
approaches have to be overly conservative when interacting with other 
agents. Ideally, we want autonomous vehicles to drive in a natural and 
confident manner, while still maintaining safety. We believe that to 
achieve this behavior we need 3 major components. First, we need an 
approach that unifies prediction and planning in a single probabilistic 
closed-loop planning framework. Second, we need to use a multi-agent 
formulation in combination with deep learning models that can scale to 
the complexities of real-world driving and effectively model the 
interactive multi-modal distributions of real-world traffic. Finally, we 
need approaches that can effectively search the space of potential 
multi-agent interactions across time efficiently in order to produce a 
suitable planned behavior. In this proposal, we will show our current 
progress in applying deep offline reinforcement learning to autonomous 
driving, and present future work to continue scaling deep learning 
approaches to more complicated and interactive autonomous driving problems.

Thesis Committee Members:
Jeff Schneider, Chair
John Dolan, Co-Chair
David Held
Philipp Krähenbühl (UT Austin)

A draft of the thesis proposal document is available at
https://drive.google.com/drive/folders/14DoKABsNlCx9AKK9O1ZY77Roau4ugEmG?usp=sharing 
<https://drive.google.com/drive/folders/14DoKABsNlCx9AKK9O1ZY77Roau4ugEmG?usp=sharing>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/autonlab-users/attachments/20220916/c068ca01/attachment.html>
-------------- next part --------------
_______________________________________________
ri-people mailing list
ri-people at lists.andrew.cmu.edu
https://lists.andrew.cmu.edu/mailman/listinfo/ri-people



More information about the Autonlab-users mailing list