to download the project base paper transformer use project.

Abstract:

Jointly processing information from multiple sensors is crucial to achieving accurate and robust perception for reliable autonomous driving systems transformer use. However, current 3D perception research follows a modality-specific paradigm, leading to additional computation overheads and inefficient collaboration between different sensor data. In this paper, we present an efficient multi-modal backbone for outdoor 3D perception named UniTR, which processes a variety of modalities with unified modeling and shared parameters. Unlike previous works, UniTR introduces a modality-agnostic transformer encoder to handle these view- discrepant sensor data for parallel modal-wise representation learning and automatic cross-modal interaction without additional fusion steps.

More importantly, to make full use of these complementary sensor types, we present a novel multi- modal integration strategy by both considering semantic-abundant 2D perspective and geometry-aware 3D sparse neighborhood relations. UniTR is also a fundamentally task-agnostic backbone that naturally supports different 3D perception tasks. It sets a new state-of-the-art performance on the nuScenes benchmark, achieving +1.1 NDS higher for 3D object detection and +12.0 higher mIoU for BEV map segmentation with lower inference latency. Data obtained from multi-sensory systems are represented in fundamentally different modalities: e.g., cameras capture visually rich perspective images, while LiDARs acquire geometry-sensitive point clouds in 3D space. Integrating these complementary sensors is an ideal solution for achieving robust 3D perception.

The code will be available Here

UNITR: A UNIFIED AND EFFICIENT MULTI-MODAL TRANSFORMER FOR BIRD’S-EYE-VIEW REPRESENTATION, transformer use, final year projects for computer science
Leave a Comment

Comments

No comments yet. Why don’t you start the discussion?

Leave a Reply

Your email address will not be published. Required fields are marked *