SwePub
Sök i LIBRIS databas

  Utökad sökning

WFRF:(Xiong Weiyi)
 

Sökning: WFRF:(Xiong Weiyi) > LXL: LiDAR Excluded...

LXL: LiDAR Excluded Lean 3D Object Detection with 4D Imaging Radar and Camera Fusion

Xiong, Weiyi (författare)
Beihang University
Liu, Jianan (författare)
Huang, Tao (författare)
James Cook University
visa fler...
Han, Qing Long (författare)
Swinburne University of Technology
Xia, Yuxuan, 1993 (författare)
Chalmers tekniska högskola,Chalmers University of Technology
Zhu, Bing (författare)
Beihang University
visa färre...
 (creator_code:org_t)
2024
2024
Engelska.
Ingår i: IEEE Transactions on Intelligent Vehicles. - 2379-8858. ; 9:1, s. 79-92
  • Tidskriftsartikel (refereegranskat)
Abstract Ämnesord
Stäng  
  • As an emerging technology and a relatively affordable device, the 4D imaging radar has already been confirmed effective in performing 3D object detection in autonomous driving. Nevertheless, the sparsity and noisiness of 4D radar point clouds hinder further performance improvement, and in-depth studies about its fusion with other modalities are lacking. On the other hand, as a new image view transformation strategy, “sampling” has been applied in a few image-based detectors and shown to outperform the widely applied “depth-based splatting” proposed in Lift-Splat-Shoot (LSS), even without image depth prediction. However, the potential of “sampling” is not fully unleashed. This paper investigates the “sampling” view transformation strategy on the camera and 4D imaging radar fusion-based 3D object detection. LiDAR Excluded Lean (LXL) model, predicted image depth distribution maps and radar 3D occupancy grids are generated from image perspective view (PV) features and radar bird's eye view (BEV) features, respectively. They are sent to the core of LXL, called “radar occupancy-assisted depth-based sampling”, to aid image view transformation. We demonstrated that more accurate view transformation can be performed by introducing image depths and radar information to enhance the “sampling” strategy. Experiments on VoD and TJ4DRadSet datasets show that the proposed method outperforms the state-of-the-art 3D object detection methods by a significant margin without bells and whistles. Ablation studies demonstrate that our method performs the best among different enhancement settings.

Ämnesord

NATURVETENSKAP  -- Fysik -- Atom- och molekylfysik och optik (hsv//swe)
NATURAL SCIENCES  -- Physical Sciences -- Atom and Molecular Physics and Optics (hsv//eng)
NATURVETENSKAP  -- Data- och informationsvetenskap -- Datorseende och robotik (hsv//swe)
NATURAL SCIENCES  -- Computer and Information Sciences -- Computer Vision and Robotics (hsv//eng)
TEKNIK OCH TEKNOLOGIER  -- Elektroteknik och elektronik -- Annan elektroteknik och elektronik (hsv//swe)
ENGINEERING AND TECHNOLOGY  -- Electrical Engineering, Electronic Engineering, Information Engineering -- Other Electrical Engineering, Electronic Engineering, Information Engineering (hsv//eng)

Nyckelord

autonomous driving
multi-modal fusion
Cameras
deep learning
Radar imaging
Three-dimensional displays
Radar detection
Feature extraction
4D imaging radar
3D object detection
Radar
camera
Object detection

Publikations- och innehållstyp

art (ämneskategori)
ref (ämneskategori)

Hitta via bibliotek

Till lärosätets databas

Sök utanför SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy