CodeMachine LearningProject

LetFuser: Autonomous driving with transformers

LeTFuser: Light-weight End-to-end Transformer-Based Sensor Fusion for Autonomous Driving with Multi-Task Learning

In end-to-end autonomous driving, the utilization of existing sensor fusion techniques for imitation learning proves inadequate in challenging situations that involve numerous dynamic agents. To address this issue, we introduce LeTFuser, a transformer-based algorithm for fusing multiple RGB-D camera representations. To perform perception and control tasks simultaneously, we utilize multi-task learning. Our model comprises of two modules, the first being the perception module that is responsible for encoding the observation data obtained from the RGB-D cameras. It carries out tasks such as semantic segmentation, semantic depth cloud mapping (SDC), and traffic light state recognition. Our approach employs the Convolutional vision Transformer (CvT) \cite{wu2021cvt} to better extract and fuse features from multiple RGB cameras due to local and global feature extraction capability of convolution and transformer modules, respectively. Following this, the control module undertakes the decoding of the encoded characteristics together with supplementary data, comprising a rough simulator for static and dynamic environments, as well as various measurements, in order to anticipate the waypoints associated with a latent feature space. We use two methods to process these outputs and generate the vehicular controls (e.g. steering, throttle, and brake) levels. The first method uses a PID algorithm to follow the waypoints on the fly, whereas the second one directly predicts the control policy using the measurement features and environmental state. We evaluate the model and conduct a comparative analysis with recent models on the CARLA simulator using various scenarios, ranging from normal to adversarial conditions, to simulate real-world scenarios. Our code is available at this link to facilitate future studies.

Restricted content! log in or register for free.

Join Upaspro to get email for news in AI and Finance

4 thoughts on “LetFuser: Autonomous driving with transformers

  • Hi Pedram,
    really interesting topic. I was wondering if we can have access to the paper?
    What type of sensors you considered for the model?

    Reply
    • The paper has just been accepted to the CVPR workshop. We have not yet distribute it. We hopefully announce it after we submit our conference paper.

      Reply
  • Hello,
    Thank you for your post. The autonomous driving is a really hot topic. I was wondering what kind of problem we can get into if we use the imitation learning?

    Reply
    • Hi Roxana,
      The imitation learning can get the at most the same performance as the expert. On the other hand, if the expert does not include some adversarial scenarios, it can result to add-up error and divergence. As our future direction, we are going to consider reinforcement learning approach using sequential model learning.

      Reply

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses User Verification plugin to reduce spam. See how your comment data is processed.