Who doesn't know of the super cool scenes in "Minority Report": intelligent machines and innovative user interfaces with speech and gestures? In this deep dive, we will talk about how deep learning can enable such interactions using some Microsoft projects in the area of NUI (Natural User Interfaces): Kinect, Handpose, Skype Translator etc. Which predictive models are being used? What do we do if we don't have sufficient data? Finally we will dare an outlook into the future how new and innovative human-machine-interaction concepts can change our user experience with computers and in light of industry 4.0.