Predicting Physical World Destinations for Commands Given to Self-Driving Cars

Dusan Grujicic, Thierry Deruyttere, Marie-Francine Moens, Matthew Blaschko

[AAAI-22] Main Track
Abstract: In recent years, we have seen significant steps taken in the development of self-driving cars. Multiple companies are starting to roll out impressive systems that work in a variety of settings. These systems can sometimes give the impression that full self-driving is just around the corner and that we would soon build cars without even a steering wheel. However, phasing out of the traditional methods of human-vehicle interaction raises the question of the interaction between the self-driving car and the passenger, and with the increased degree of control and autonomy being given to the AI, makes it that much more critical. Recent works have tried to tackle this issue by allowing the passenger to give commands that refer to specific objects in the visual scene. Nevertheless, this is only half the task as the car should also understand the physical destination of the command, which is what we focus on in this paper. We propose an extension where we annotate the 3D end location that the car needs to reach after executing the given command, and evaluate multiple different baselines on predicting this destination location. From our results, we conclude that our dataset and proposed task are challenging. Additionally, we also propose a model that outperforms the prior works adapted for this particular setting.

Introduction Video

Sessions where this paper appears

  • Poster Session 6

    Sat, February 26 8:45 AM - 10:30 AM (+00:00)
    Red 4
    Add to Calendar

  • Poster Session 7

    Sat, February 26 4:45 PM - 6:30 PM (+00:00)
    Red 4
    Add to Calendar