Crafted vs. Learned Representations in Predictive Models - A Case Study on Cyclist Path Prediction

Journal Article (2021)
Author(s)

E.A.I. Pool (TU Delft - Intelligent Vehicles)

J.F.P. Kooij (TU Delft - Intelligent Vehicles)

D. Gavrila (TU Delft - Intelligent Vehicles)

Research Group
Intelligent Vehicles
Copyright
© 2021 E.A.I. Pool, J.F.P. Kooij, D. Gavrila
DOI related publication
https://doi.org/10.1109/TIV.2021.3064253
More Info
expand_more
Publication Year
2021
Language
English
Copyright
© 2021 E.A.I. Pool, J.F.P. Kooij, D. Gavrila
Research Group
Intelligent Vehicles
Issue number
4
Volume number
6
Pages (from-to)
747-759
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

This paper compares two models for context-based path prediction of objects with switching dynamics: a Dynamic Bayesian Network (DBN) and a Recurrent Neural Network (RNN). These models are instances of two larger model categories, distinguished by whether expert knowledge is explicitly crafted into the state representation (and thus is interpretable) or whether the representation is learned from data, respectively. Both have shown state-of-the-art performance in previous work. In order to provide a fair comparison, we ensure that both models are treated similarly with respect to the use of context cues and parameter estimation. Specifically, we describe (1) how to integrate the context cues (used previously by the DBN) into the RNN, and (2) how to optimize the DBN with back-propagation similar to the RNN, while keeping an interpretable state representation. Experiments are performed on a scenario where a cyclist might turn left at an intersection in front of the ego-vehicle. Results show that the RNN successfully leverages the context cues, and that optimizing the DBN improves its performance with respect to existing work. While the RNN outperforms the optimized DBN in predictive log-likelihood by a significant margin, both models attain similar average Euclidean distance errors (23-39 cm for DBN and 31-34 cm for RNN, predicting 1 s ahead).

Files

Crafted_vs_Learned_Representat... (pdf)
(pdf | 1.95 Mb)
- Embargo expired in 08-09-2021
License info not available