2003 - Probabilistic Future Prediction For Video Scene Understanding
2003 - Probabilistic Future Prediction For Video Scene Understanding
Anthony Hu1,2 , Fergal Cotter1 , Nikhil Mohan1 , Corina Gurau1 , and Alex Kendall1,2
1
Wayve, London, UK. [email protected]
2
University of Cambridge.
arXiv:2003.06409v1 [cs.CV] 13 Mar 2020
1 Introduction
Building predictive cognitive models of the world is often regarded as the essence
of intelligence. It is one of the first skills that we develop as infants. We use
these models to enhance our capability at learning more complex tasks, such as
navigation or manipulating objects [48].
Unlike in humans, developing prediction models for autonomous vehicles to
anticipate the future remains hugely challenging. Road agents have to make
reliable decisions based on forward simulation to understand how relevant parts
of the scene will evolve. There are various reasons why modelling the future
is incredibly difficult: natural-scene data is rich in details, most of which are
irrelevant for the driving task, dynamic agents have complex temporal dynamics,
often controlled by unobservable variables, and the future is inherently uncertain,
as multiple futures might arise from a unique and deterministic past.
Current approaches to autonomous driving individually model each dynamic
agent by producing hand-crafted behaviours, such as trajectory forecasting, to
feed into a decision making module [8]. This largely assumes independence be-
tween agents and fails to model multi-agent interaction. Most works that holis-
tically reason about the temporal scene are limited to simple, often simulated
environments or use low dimensional input images that do not have the visual
2
complexity of real world driving scenes [47]. Some approaches tackle this prob-
lem by making simplifying assumptions to the motion model or the stochasticity
of the world [41,8]. Others avoid explicitly predicting the future scene but rather
rely on an implicit representation or Q-function (in the case of model-free rein-
forcement learning) in order to choose an action [33,27,36].
Real world future scenarios are difficult to model because of the stochasticity
and the partial observability of the world. Our work addresses this by encoding
the future state into a low-dimensional future distribution. We then allow the
model to have a privileged view of the future through the future distribution at
training time. As we cannot use the future at test time, we train a present distri-
bution (using only the current state) to match the future distribution through a
Kullback-Leibler (KL) divergence loss. We can then sample from the present dis-
tribution during inference, when we do not have access to the future. We observe
that this paradigm allows the model to learn accurate and diverse probabilistic
future prediction outputs.
In order to predict the future we need to first encode video into a motion rep-
resentation. Unlike advances in 2D convolutional architectures [60,26], learning
spatio-temporal features is more challenging due to the higher dimensionality
of video data and the complexity of modelling dynamics. State-of-the-art archi-
tectures [64,61] decompose 3D filters into spatial and temporal convolutions in
order to learn more efficiently. The model we propose further breaks down con-
volutions into many space-time combinations and context aggregation modules,
stacking them together in a more complex hierarchical representation. We show
that the learnt representation is able to jointly predict ego-motion and motion
of other dynamic agents. By explicitly modelling these dynamics we can capture
the essential features for representing causal effects for driving. Ultimately we use
this motion-aware and future-aware representation to improve an autonomous
vehicle control policy.
Our main contributions are threefold. Firstly, we present a novel deep learn-
ing framework for future video prediction. Secondly, we demonstrate that our
probabilistic model is able to generate visually diverse and plausible futures.
Thirdly, we show our future prediction representation substantially improves a
learned autonomous driving policy.
2 Related Work
This work falls in the intersection of learning scene representation from video,
probabilistic modelling of the ambiguity inherent in real-world driving data, and
using the learnt representation for control.
Visual prediction. Most works for learning dynamics from video fall under
the framework of model-based reinforcement learning [42,20,16,32] or unsuper-
vised feature learning [57,15], both regressing directly in pixel space [45,49,31]
or in a learned feature space [30,19]. For the purpose of creating good represen-
tations for driving scenes, directly predicting in the high-dimensional space of
image pixels is unnecessary, as some details about the appearance of the world
are irrelevant for planning and control. Our approach is similar to that of Luc
et al . [44] which trains a model to predict future semantic segmentation us-
ing pseudo-ground truth labels generated from a teacher model. However, our
model predicts a more complete scene representation with segmentation, depth,
and flow and is probabilistic in order to model the uncertainty of the future.
Future Prediction
zt+Nf G t+Nf
ôt
Present & Future
... Distributions ... ...
train
zt+1 ηt G ôt+1
t
Future test
Past Perception Dynamics - Y P µt,p, σt,p Control - C
xt T T T zt ĉt
xt−2 T T
xt−3 T
...
3 Model Architecture
Our model learns a spatio-temporal feature to jointly predict future scene rep-
resentation (semantic segmentation, depth, optical flow) and train a driving
policy. The architecture contains five components: Perception, an image scene
understanding model, Dynamics, which learns a spatio-temporal representation,
Present/Future Distributions, our probabilistic framework, Future Prediction,
which predicts future video scene representation, and Control, which trains a
driving policy using expert driving demonstrations. Figure 1 gives an overview
of the model and further details are described in this section and Appendix A.
3.1 Perception
The perception component of our system contains two modules: the encoder
of a scene understanding model that was trained on single image frames to
reconstruct semantic segmentation and depth [35], and the encoder of a flow
network [58], trained to predict optical flow. The combined perception features
xt ∈ RC×H×W form the input to the dynamics model. These models can also be
used as a teacher to distill the information from the future, giving pseudo-ground
truth labels for segmentation, depth and flow {st , dt , ft }. See subsection 4.1 for
more details on the teacher model.
5
3.2 Dynamics
zt = Y(xt−T +1 : xt ) (1)
We train a future prediction model that unrolls the dynamics feature, which is
a compact scene representation of the past context, into predictions about the
state of the world in the future. The future prediction model is a convolutional
recurrent network G which creates future features gtt+i that become the inputs of
individual decoders Ds , Dd , Df to decode these features to predicted segmenta-
tion ŝt+i ˆt+i ˆt+i values in the pixel space. We have introduced
t , depth dt , and flow ft
a second time superscript notation, i.e. gtt+i , represents the prediction about the
world at time t + i given the dynamics features at time t. Also note that gtt , zt .
The structure of the convolutional recurrent network G is the following: a
convolutional GRU [2] followed by three spatial residual layers, repeated D times,
6
zout ∈ RCout×T×H×W
+
1×1×1, Cout
Concat, 3C
Global Context
Compression Compression Compression Compression Avg Pool Avg Pool Avg Pool
1×1×1, C2 1×1×1, C2 1×1×1, C2 1×1×1, C2 kt×H×W kt×H2 ×W2 kt×H4 ×W4
zin ∈ RC×T×H×W
3.5 Control
ˆ
ĉt = {v̂t , v̇ˆt , θ̂t , θ̇t } = C(zt ) (7)
C compresses zt ∈ RCd ×H×W with strided convolutional layers, then stacks sev-
eral fully connected layers, compressing at each stage, to regress the four dimen-
sional output.
3.6 Losses
Nf −1
X
Ls = γfi Lt+i
s (8)
i=0
Lfuture-pred = λs Ls + λd Ld + λf Lf (9)
c −1
NX 2
Lc = γci vt+i − v̂t + iv̇ˆt +
i=0
ˆ 2
θt+i − θ̂t + iθ̇t (10)
where 0 < γc < 1 is the control discount factor penalizing less speed and steering
errors further into the future.
9
Total Loss The final loss L can be decomposed into the future prediction loss
(Lfuture-pred ), the probabilistic loss (Lprobabilistic ), and the control loss (Lc ) .
4 Experiments
Dynamics and Control The dynamics and control modules are trained using
30 hours of driving data from the urban driving dataset we collected and de-
scribed above. We address the inherent dataset bias by sampling data uniformly
across lateral and longitudinal dimensions. First, the data is split into a his-
togram of bins by steering, and subsequently by speed. We found that weighting
each data point proportionally to the width of the bin it belongs to avoids the
need for alternative approaches such as data augmentation.
4.2 Metrics
1
Mperception = (seg% increase + depth% decrease + flow% decrease ) (12)
3
10
where d is an error metric and S, S 0 , are independent samples from the present
distribution P . This metric measures performance both in terms of accuracy,
by looking at the minimum error of the samples, as well as the diversity of
the predictions by taking the expectation of the distance between N samples.
The distance d is the scale-invariant logarithmic error for depth, the average
end-point error for flow, and for segmentation d(x, y) = 1 − IoU(x, y).
To measure control performance, we report mean absolute error of speed and
steering outputs, balanced by steering histogram bins.
5 Results
We first compare our proposed spatio-temporal module to previous state-of-
the-art architectures and show that our module achieves the best performance
on future prediction metrics. Then we demonstrate that modelling the future
in a probabilistic manner further improves performance. And finally, we show
that our probabilistic future prediction representation substantially improves a
learned driving policy. All the reported results are evaluated on test routes with
no overlap with the training data.
Temporal Model Mperception (↑) Depth (↓) Flow (↓) Seg. (↑)
Repeat frame 0.0% 1.467 5.707 0.356
None -40.3% 1.980 8.573 0.229
Res. 3D Conv. [25] 6.9% 1.162 5.437 0.339
Conv. GRU [2] 7.4% 1.097 5.714 0.346
Deterministic
Sep. Inception [64] 9.6% 1.101 5.300 0.344
Ours 13.6% 1.090 5.029 0.367
Res. 3D Conv. [25] 8.1% 1.107 5.720 0.356
Conv. GRU [2] 9.0% 1.101 5.645 0.359
Probabilistic
Sep. Inception [64] 13.8% 1.040 5.242 0.371
Ours 20.0% 0.970 4.857 0.396
a (kt , 1, 1) time filter), decomposing into two subtasks helps the network learn
more efficiently. In the same spirit, we decompose the spatio-temporal convo-
lutions into all combinations of space-time convolutions: (1, ks , ks ), (kt , 1, ks ),
(kt , ks , 1), (kt , ks , ks ), and by stacking these temporal blocks together, the net-
work can learn a hierarchically more complex representation of the scene. The
second reason is that we incorporate global context in our features. By pooling
the features spatially and temporally at different scales, each individual feature
map also has information about the global scene context, which helps in am-
biguous situations. Appendix A.2 contains an ablation study of the different
component of the Temporal Block.
Table 2: Diversity Distance Metric for various temporal models evaluated on the
urban driving data, demonstrating that our model produces the most accurate
and diverse distribution.
Fig. 3: Predicted futures from our model while driving through an urban in-
tersection. From left, we show the actual past and future video sequence and
labelled semantic segmentation. Using four different noise vectors, η, we observe
the model imagining different driving manoeuvres at an intersection: being sta-
tionary, driving straight, taking a left or a right turn. We show both predicted
semantic segmentation and entropy (uncertainty) for each future. This example
demonstrates that our model is able to learn a probabilistic embedding, capable
of predicting multi-modal and plausible futures.
13
Fig. 4: Predicted futures from our model while driving through a busy urban
scene. From left, we show actual past and future video sequence and labelled
semantic segmentation, depth and optical flow. Using two different noise vectors,
η, we observe the model imagining either stopping in traffic or continuing in
motion. This illustrates our model’s efficacy at jointly predicting holistic future
behaviour of our own vehicle and other dynamic agents in the scene across all
modalities.
Further, our framework can automatically infer which scenes are unusual or
unexpected and where the model is uncertain of the future, by computing the dif-
ferential entropy of the present distribution. Simple scenes (e.g. one-way streets)
will tend to have a low entropy, corresponding to an almost deterministic future.
Any latent code sampled from the present distribution will correspond to the
same future. Conversely, complex scenes (e.g. intersections, roundabouts) will be
associated with a high-entropy. Different samples from the present distribution
will correspond to different futures, effectively modelling the stochasticity of the
future.3
Finally, to allow reproducibility, we evaluate our future prediction framework
on Cityscapes [13] and report future semantic segmentation performance in Ta-
ble 3. We compare our predictions, at resolution 256 × 512, to the ground truth
segmentation at 5 and 10 frames in the future, corresponding to respectively
0.29s and 0.59s in the future as the videos are sampled at 17Hz. Qualitative
examples on Cityscapes can be found in Appendix C.
Temporal Model IoUi=5 (↑) IoUi=10 (↑)
Repeat frame 0.393 0.331
Res. 3D Conv. [25] 0.445 0.399
Conv. GRU [2] 0.449 0.397
Probabilistic
Sep. Inception [64] 0.455 0.402
Ours 0.464 0.416
Table 4: Evaluation of the driving policy. The policy is learned from temporal
features explicitly trained to predict the future. We observe a significant perfor-
mance improvement over non-temporal and non-future-aware baselines.
All deterministic models trained with the future prediction loss outperform
the baseline, and more interestingly the temporal representation’s ability to bet-
ter predict the future (shown by Mperception ) directly translate in a control
performance gain, with our best deterministic model having, respectively, a 27%
and 38% improvement over the baseline for steering and speed.
Finally, all probabilistic models perform better than their deterministic coun-
terpart, further demonstrating that modelling the uncertainty of the future pro-
duces a more effective spatio-temporal representation. Our probabilistic model
achieves the best performance with a 33% steering and 46% speed improvement
over the baseline.
6 Conclusions
This work is the first to propose a deep learning model capable of probabilistic
future prediction of ego-motion, static scene and other dynamic agents, jointly.
We observe large performance improvements due to our proposed temporal video
encoding architecture and probabilistic modelling of present and future distri-
butions. This initial work leaves a lot of future directions to explore: leveraging
known priors and structure in the latent representation, conditioning the control
policy on future prediction and applying our future prediction architecture to
model-based reinforcement learning.
15
References
1. Amini, A., Rosman, G., Karaman, S., Rus, D.: Variational end-to-end navigation
and localization. In: Proc. International Conference on Robotics and Automation
(ICRA). IEEE (2019)
2. Ballas, N., Yao, L., Pas, C., Courville, A.: Delving deeper into convolutional net-
works for learning video representations. In: Proc. International Conference on
Learning Representations (ICLR) (2016)
3. Bellemare, M.G., Danihelka, I., Dabney, W., Mohamed, S., Lakshminarayanan,
B., Hoyer, S., Munos, R.: The cramer distance as a solution to biased wasserstein
gradients. arXiv preprint (2017)
4. Bhattacharyya, A., Schiele, B., Fritz, M.: Accurate and diverse sampling of se-
quences based on a ”best of many” sample objective. Proc. Computer Vision and
Pattern Recognition (CVPR) (2018)
5. Bilen, H., Fernando, B., Gavves, E., Vedaldi, A., Gould, S.: Dynamic image net-
works for action recognition. Proc. Computer Vision and Pattern Recognition
(CVPR) (2016)
6. Bojarski, M., Testa, D.D., Dworakowski, D., Firner, B., Flepp, B., Goyal, P., Jackel,
L.D., Monfort, M., Muller, U., Zhang, J., Zhang, X., Zhao, J., Zieba, K.: End to
End Learning for Self-Driving Cars. arXiv preprint (2016)
7. Carreira, J., Zisserman, A.: Quo vadis, action recognition? a new model and the
kinetics dataset. In: Proc. Computer Vision and Pattern Recognition (CVPR)
(2017)
8. Casas, S., Luo, W., Urtasun, R.: Intentnet: Learning to predict intention from raw
sensor data. In: Proc. Conference on Robot Learning (CoRL) (2018)
9. Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution
for semantic image segmentation. arXiv preprint (2017)
10. Chua, K., Calandra, R., McAllister, R., Levine, S.: Deep reinforcement learning
in a handful of trials using probabilistic dynamics models. Advances in Neural
Information Processing Systems (NeurIPS) (2018)
11. Clark, A., Donahue, J., Simonyan, K.: Adversarial video generation on complex
datasets. In: arXiv preprint (2019)
12. Codevilla, F., Miiller, M., López, A., Koltun, V., Dosovitskiy, A.: End-to-end
driving via conditional imitation learning. In: Proc. International Conference on
Robotics and Automation (ICRA) (2018)
13. Cordts, M., Omran, M., Ramos, S., Rehfeld, T., Enzweiler, M., Benenson, R.,
Franke, U., Roth, S., Schiele, B.: The cityscapes dataset for semantic urban
scene understanding. In: Proc. Computer Vision and Pattern Recognition (CVPR)
(2016)
14. Denton, E., Fergus, R.: Stochastic video generation with a learned prior. In: Proc.
International Conference on Machine Learning (ICML). Proceedings of Machine
Learning Research (2018)
15. Denton, E.L., Birodkar, v.: Unsupervised learning of disentangled representations
from video. In: Advances in Neural Information Processing Systems (NeurIPS)
(2017)
16. Ebert, F., Finn, C., Dasari, S., Xie, A., Lee, A.X., Levine, S.: Visual foresight:
Model-based deep reinforcement learning for vision-based robotic control. arXiv
preprint (2018)
17. Ebert, F., Finn, C., Lee, A.X., Levine, S.: Self-supervised visual planning with
temporal skip connections. In: Proc. Conference on Robot Learning (CoRL) (2017)
16
18. Feichtenhofer, C., Pinz, A., Wildes, R.P.: Spatiotemporal residual networks for
video action recognition. In: Advances in Neural Information Processing Systems
(NeurIPS) (2016)
19. Finn, C., Goodfellow, I., Levine, S.: Unsupervised learning for physical interaction
through video prediction. In: Advances in Neural Information Processing Systems
(NeurIPS) (2016)
20. Finn, C., Levine, S.: Deep visual foresight for planning robot motion. Proc. Inter-
national Conference on Robotics and Automation (ICRA) (2017)
21. Gadde, R., Jampani, V., Gehler, P.V.: Semantic video CNNs through representa-
tion warping. In: Proc. Computer Vision and Pattern Recognition (CVPR) (2017)
22. Goodfellow, I.: Nips 2016 tutorial: Generative adversarial networks (2016)
23. Ha, D., Schmidhuber, J.: World models. In: Advances in Neural Information Pro-
cessing Systems (NeurIPS) (2018)
24. Hafner, D., Lillicrap, T., Fischer, I., Villegas, R., Ha, D., Lee, H., Davidson, J.:
Learning latent dynamics for planning from pixels. In: Proc. International Confer-
ence on Machine Learning (ICML) (2019)
25. Hara, K., Kataoka, H., Satoh, Y.: Learning spatio-temporal features with 3d resid-
ual networks for action recognition. In: Proc. International Conference on Com-
puter Vision, workshop (ICCVw) (2017)
26. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition.
In: Proc. Computer Vision and Pattern Recognition (CVPR) (2016)
27. Hessel, M., Modayil, J., van Hasselt, H., Schaul, T., Ostrovski, G., Dabney, W.,
Horgan, D., Piot, B., Azar, M.G., Silver, D.: Rainbow: Combining improvements
in deep reinforcement learning. AAAI Conference on Artificial Intelligence (2018)
28. Huang, X., Cheng, X., Geng, Q., Cao, B., Zhou, D., Wang, P., Lin, Y., Yang, R.:
The apolloscape dataset for autonomous driving. In: Proc. Computer Vision and
Pattern Recognition, workshop (CVPRw) (2018)
29. Ioannou, Y., Robertson, D., Shotton, J., Cipolla, R., Criminisi, A.: Training cnns
with low-rank filters for efficient image classification. In: Proc. International Con-
ference on Learning Representations (ICLR) (2016)
30. Jaderberg, M., Mnih, V., Czarnecki, W.M., Schaul, T., Leibo, J.Z., Silver, D.,
Kavukcuoglu, K.: Reinforcement learning with unsupervised auxiliary tasks. Proc.
International Conference on Learning Representations (ICLR) (2017)
31. Jayaraman, D., Ebert, F., Efros, A., Levine, S.: Time-agnostic prediction: Predict-
ing predictable video frames (2018)
32. Kaiser, L., Babaeizadeh, M., Milos, P., Osinski, B., Campbell, R.H., Czechowski,
K., Erhan, D., Finn, C., Kozakowski, P., Levine, S., Mohiuddin, A., Sepassi, R.,
Tucker, G., Michalewski, H.: Model-based reinforcement learning for atari. In: Proc.
International Conference on Learning Representations (ICLR) (2020)
33. Kalashnikov, D., Irpan, A., Pastor, P., Ibarz, J., Herzog, A., Jang, E., Quillen,
D., Holly, E., Kalakrishnan, M., Vanhoucke, V., Levine, S.: Qt-opt: Scalable deep
reinforcement learning for vision-based robotic manipulation. Proc. International
Conference on Machine Learning (ICML) (2018)
34. Kendall, A., Gal, Y.: What uncertainties do we need in bayesian deep learn-
ing for computer vision? In: Advances in Neural Information Processing Systems
(NeurIPS) (2017)
35. Kendall, A., Gal, Y., Cipolla, R.: Multi-task learning using uncertainty to weigh
losses for scene geometry and semantics. In: Proc. Computer Vision and Pattern
Recognition (CVPR) (2018)
36. Kendall, A., Hawke, J., Janz, D., Mazur, P., Reda, D., Allen, J.M., Lam, V.D.,
Bewley, A., Shah, A.: Learning to drive in a day. In: Proc. International Conference
on Robotics and Automation (ICRA) (2019)
17
37. Kohl, S., Romera-Paredes, B., Meyer, C., Fauw, J.D., Ledsam, J.R., Maier-Hein,
K.H., Eslami, S.M.A., Rezende, D.J., Ronneberger, O.: A probabilistic u-net for
segmentation of ambiguous images. In: Advances in Neural Information Processing
Systems (NeurIPS) (2018)
38. Kurutach, T., Tamar, A., Yang, G., Russell, S.J., Abbeel, P.: Learning plannable
representations with causal infogan. In: Advances in Neural Information Processing
Systems (NeurIPS) (2018)
39. Lee, A.X., Zhang, R., Ebert, F., Abbeel, P., Finn, C., Levine, S.: Stochastic adver-
sarial video prediction. arXiv preprint (2018)
40. Lee, N., Choi, W., Vernaza, P., Choy, C.B., Torr, P.H.S., Chandraker, M.K.: DE-
SIRE: distant future prediction in dynamic scenes with interacting agents. Proc.
Computer Vision and Pattern Recognition (CVPR) (2017)
41. Levine, S., Abbeel, P.: Learning neural network policies with guided policy search
under unknown dynamics. In: Advances in Neural Information Processing Systems
27 (2014)
42. Levine, S., Finn, C., Darrell, T., Abbeel, P.: End-to-end training of deep visuomotor
policies. Journal of Machine Learning Research (2016)
43. Li, Z., Snavely, N.: MegaDepth: Learning Single-View Depth Prediction from In-
ternet Photos. Proc. Computer Vision and Pattern Recognition (CVPR) (2018)
44. Luc, P., Neverova, N., Couprie, C., Verbeek, J., LeCun, Y.: Predicting deeper
into the future of semantic segmentation. In: Proc. International Conference on
Computer Vision (ICCV) (2017)
45. Mathieu, M., Couprie, C., LeCun, Y.: Deep multi-scale video prediction beyond
mean square error. In: Proc. International Conference on Learning Representations
(ICLR) (2016)
46. Neuhold, G., Ollmann, T., Bulo, S.R., Kontschieder, P.: The Mapillary Vistas
Dataset for Semantic Understanding of Street Scenes. In: Proc. International Con-
ference on Computer Vision (ICCV)
47. Oh, J., Guo, X., Lee, H., Lewis, R., Singh, S.: Action-conditional video prediction
using deep networks in atari games. Advances in Neural Information Processing
Systems (NeurIPS) (2015)
48. Piaget, J.: Origins of intelligence in the child. London: Routledge and Kegan Paul
(1936)
49. Ranzato, M., Szlam, A., Bruna, J., Mathieu, M., Collobert, R., Chopra, S.: Video
(language) modeling: a baseline for generative models of natural videos. arXiv
preprint (2014)
50. Rhinehart, N., McAllister, R., Kitani, K.M., Levine, S.: PRECOG: prediction con-
ditioned on goals in visual multi-agent settings. Proc. International Conference on
Computer Vision (ICCV) (2019)
51. Salimans, T., Zhang, H., Radford, A., Metaxas, D.N.: Improving gans using optimal
transport. Proc. International Conference on Learning Representations (ICLR)
(2018)
52. Schuldt, Christian, L.I., Caputo, B.: Recognizing human actions: A local svm ap-
proach. In: Proc. International Conference on Pattern Recognition (2004)
53. Shi, X., Chen, Z., Wang, H., Yeung, D.Y., Wong, W.k., Woo, W.c.: Convolutional
lstm network: A machine learning approach for precipitation nowcasting. In: Ad-
vances in Neural Information Processing Systems (NeurIPS) (2015)
54. Siam, M., Valipour, S., Jagersand, M., Ray, N.: Convolutional gated recurrent net-
works for video segmentation. In: Proc. International Conference on Image Pro-
cessing (2017)
18
55. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale
image recognition. In: Proc. International Conference on Learning Representations
(ICLR) (2015)
56. Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recog-
nition in videos. In: NIPS (2014)
57. Srivastava, N., Mansimov, E., Salakhudinov, R.: Unsupervised learning of video
representations using lstms. In: ICML (2015)
58. Sun, D., Yang, X., Liu, M.Y., Kautz, J.: PWC-Net: CNNs for Optical Flow Using
Pyramid, Warping, and Cost Volume. Proc. Computer Vision and Pattern Recog-
nition (CVPR) (2018)
59. Sun, L., Jia, K., Yeung, D., Shi, B.E.: Human action recognition using factorized
spatio-temporal convolutional networks. Proc. International Conference on Com-
puter Vision (ICCV) (2015)
60. Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S.E., Anguelov, D., Erhan, D.,
Vanhoucke, V., Rabinovich, A.: Going deeper with convolutions. Proc. Computer
Vision and Pattern Recognition (CVPR) (2015)
61. Tran, D., Wang, H., Torresani, L., Ray, J., LeCun, Y., Paluri, M.: A closer look
at spatiotemporal convolutions for action recognition. In: Proc. Computer Vision
and Pattern Recognition (CVPR) (2018)
62. Wu, Z., Shen, C., van den Hengel, A.: Bridging category-level and instance-level
semantic image segmentation. arXiv preprint (2016)
63. Xie, S., Girshick, R.B., Dollár, P., Tu, Z., He, K.: Aggregated residual transforma-
tions for deep neural networks. Proc. Computer Vision and Pattern Recognition
(CVPR) (2017)
64. Xie, S., Sun, C., Huang, J., Tu, Z., Murphy, K.: Rethinking spatiotemporal feature
learning for video understanding. Proc. European Conference on Computer Vision
(ECCV) (2018)
65. Yu, F., Xian, W., Chen, Y., Liu, F., Liao, M., Madhavan, Vashisht, D.T.: Bdd100k:
A diverse driving video database with scalable annotation tooling (2018)
66. Zhu, X., Xiong, Y., Dai, J., Yuan, L., Wei, Y.: Deep feature flow for video recog-
nition. In: Proc. Computer Vision and Pattern Recognition (CVPR) (2017)
19
A Architecture
In total, our architecture has 30.4M parameters, comprising of modules:
– Perception encoder, Eperception , 25.3M parameters ;
– Dynamics module, Y, 0.8M parameters ;
– Future decoders, G, 3.5M parameters ;
– Control policy model, C, 0.7M parameters.
A.1 Perception
Semantics and Geometry. Our model is an encoder-decoder model with five
encoder blocks and three decoder blocks, followed by an atrous spatial pyramid pooling
(ASPP) module [9]. The encoders contain 2, 4, 8, 8, 8 layers respectively, downsampling
by a factor of two each time with a strided convolution. The decoders contain 3 layers
each, upsampling each time by a factor of two with a sub-strided convolution. All layers
have residual connections and many are low rank, with varying kernel and dilation sizes.
Furthermore, we employ skip connections from the encoder to decoder at each spatial
scale.
We pretrain the scene understanding encoder on a number of heterogeneous datasets
to predict semantic segmentation and depth: CityScapes [13], Mapillary Vistas [46],
ApolloScape [28] and Berkeley Deep Drive [65]. We collapse the classes to 14 seman-
tic segmentation classes shared across these datasets and sample each dataset equally
during training. We train for 200,000 gradient steps with a batch size of 32 using SGD
with an initial learning rate of 0.1 with momentum 0.9. We use cross entropy for seg-
mentation and the scale-invariant loss [43] to learn depth with a weight of 1.0 and 0.1,
respectively.
Motion. In addition to this semantics and geometry encoder, we also use a pre-
trained optical flow network, PWCNet [58]. We use the pretrained authors’ implemen-
tation.
Perception. To form our perception encoder we concatenate these two feature rep-
resentations (from the perception encoder and optical flow net) concatenated together.
We use the features two layers before the output optical flow regression as the feature
representation. The decoders of these networks are used for generating pseudo-ground
truth segmentation and depth labels to train our dynamics and future prediction mod-
ules.
B Nomenclature
Networks
Perception encoder Eperception
Temporal Block T
Dynamics module Y
Present network P
Future network F
Future prediction module G
Future decoders Ds , Dd , Df
Control module C
Tensors
Temporal context T
Future prediction horizon Nf
Future control horizon Nc
Input image it
Perception features xt = Eperception (it )
Dynamics features zt = Y(xt−T +1 : xt )
Present distribution µt,present , σt,present = P(zt )
Future distribution µt,future , σt,future = F(zt )
2
Noise vector (train) ηt ∼ N (µt,future , σt,future )
2
Noise vector (test) ηt ∼ N (µt,present , σt,present )
Future prediction inputs ut+i
t = η t
Future prediction initial hidden state gtt = zt
Future prediction output features gtt+i = G(ut+it , gt
t+i−1
)
t+i t+i ˆt+i ˆt+i
ôt = {ŝt , dt , ft }
Future perception outputs
= {Ds (gtt+i ), Dd (gtt+i ), Df (gtt+i )}
ˆ
ĉt = {v̂t , v̇ˆt , θ̂t , θ̇t }
Control outputs
= C(zt )
21
(a) Our model can correctly predict future segmentation of small classes such as poles
or traffic lights.
(b) Dynamic agents, i.e. cars and cyclists, are also accurately predicted.
(c) In this example, the bus is correctly segmented, without any class bleeding contrary
to the pseudo-ground truth segmentation, showing that our model can reason in a
holistic way.
Fig. 5: Future prediction on the CityScapes dataset, for 10 frames in the future
at 17Hz and 256 × 512 resolution.