TY - JOUR
T1 - Advisable learning for self-driving vehicles by internalizing observation-to-action rules
AU - Kim, Jinkyu
AU - Moon, Suhong
AU - Rohrbach, Anna
AU - Darrell, Trevor
AU - Canny, John
N1 - Funding Information:
We thank Y. Gao, D. Wang, O. Watkins, and C. Devin at UC Berkeley for their helpful discussion. This work was supported by DARPA XAI program and Berkeley DeepDrive. J. Kim was in part supported by Samsung Scholarship.
Funding Information:
Acknowledgements. We thank Y. Gao, D. Wang, O. Watkins, and C. Devin at UC Berkeley for their helpful discussion. ThisworkwassupportedbyDARPAXAIprogram and Berkeley DeepDrive. J. Kim was in part supported by SamsungScholarship.
Publisher Copyright:
© 2020 IEEE
PY - 2020
Y1 - 2020
N2 - Humans learn to drive through both practice and theory, e.g. by studying the rules, while most self-driving systems are limited to the former. Being able to incorporate human knowledge of typical causal driving behaviour should benefit autonomous systems. We propose a new approach that learns vehicle control with the help of human advice. Specifically, our system learns to summarize its visual observations in natural language, predict an appropriate action response (e.g. “I see a pedestrian crossing, so I stop”), and predict the controls, accordingly. Moreover, to enhance interpretability of our system, we introduce a fine-grained attention mechanism which relies on semantic segmentation and object-centric RoI pooling. We show that our approach of training the autonomous system with human advice, grounded in a rich semantic representation, matches or outperforms prior work in terms of control prediction and explanation generation. Our approach also results in more interpretable visual explanations by visualizing object-centric attention maps. Code is available at https://github.com/JinkyuKimUCB/advisable-driving.
AB - Humans learn to drive through both practice and theory, e.g. by studying the rules, while most self-driving systems are limited to the former. Being able to incorporate human knowledge of typical causal driving behaviour should benefit autonomous systems. We propose a new approach that learns vehicle control with the help of human advice. Specifically, our system learns to summarize its visual observations in natural language, predict an appropriate action response (e.g. “I see a pedestrian crossing, so I stop”), and predict the controls, accordingly. Moreover, to enhance interpretability of our system, we introduce a fine-grained attention mechanism which relies on semantic segmentation and object-centric RoI pooling. We show that our approach of training the autonomous system with human advice, grounded in a rich semantic representation, matches or outperforms prior work in terms of control prediction and explanation generation. Our approach also results in more interpretable visual explanations by visualizing object-centric attention maps. Code is available at https://github.com/JinkyuKimUCB/advisable-driving.
UR - http://www.scopus.com/inward/record.url?scp=85094649940&partnerID=8YFLogxK
U2 - 10.1109/CVPR42600.2020.00968
DO - 10.1109/CVPR42600.2020.00968
M3 - Conference article
AN - SCOPUS:85094649940
SN - 1063-6919
SP - 9658
EP - 9667
JO - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
JF - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
M1 - 9157238
T2 - 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020
Y2 - 14 June 2020 through 19 June 2020
ER -