We propose a novel knowledge distillation framework for effectively teaching a sensorimotor student agent to drive from the supervision of a privileged teacher agent. Current distillation for sensorimotor agents methods tend to result in suboptimal learned driving behavior by the student, which we hypothesize is due to inherent differences between the input, modeling capacity, and optimization processes of the two agents. We develop a novel distillation scheme that can address these limitations and close the gap between the sensorimotor agent and its privileged teacher. Our key insight is to design a student which learns to align their input features with the teacher’s privileged Bird’s Eye View (BEV) space. The student then can benefit from direct supervision by the teacher over the internal representation learning. To scaffold the difficult sensorimotor learning task, the student model is optimized via a student-paced coaching mechanism with various auxiliary supervision. We further propose a high-capacity imitation learned privileged agent that surpasses prior privileged agents in CARLA and ensures the student learns safe driving behavior. Our proposed sensorimotor agent results in a robust image-based behavior cloning agent in CARLA, improving over current models by over 20.6% in driving score without requiring LiDAR, historical observations, ensemble of models, on-policy data aggregation or reinforcement learning.
To ease the challenging sensorimotor agent training task, recent approaches decompose the task into stages, by first training a privileged network with complete knowledge of the world and distilling its knowledge into a less capable student network. However, current distillation methods for sensorimotor agents result in suboptimal driving behavior due to inherent differences between the inputs, modeling capacity, and optimization processes of the two agents.
Our proposed CaT framework enables highly effective knowledge transfer between a privileged teacher and a sensorimotor (i.e., image-based) student. Specifically, we first sample queries using a spatial parameterization of the BEV space and process them with a self-attention module. Then, a deformable cross-attention module is applied to populate the student’s BEV features. The residual blocks following the alignment module can consequently facilitate knowledge transfer via direct distillation of most of the teacher’s features. Our optimization objective for guiding the distillation process is a weighted sum over both distillation and auxiliary tasks, including output distillation loss, feature distillation loss, segmentation loss and command prediction loss.
@inproceedings{zhang2023coaching,
title={Coaching a Teachable Student},
author={Zhang, Jimuyang and Huang, Zanming and Ohn-Bar, Eshed},
booktitle={CVPR},
year={2023}
}