Skip to content
VuFind
  • Login
    • English
    • اللغة العربية
Advanced
  • Multi-head multi-scale self-cr...
  • Cite this
  • Text this
  • Email this
  • Print
  • Export Record
    • Export to RefWorks
    • Export to EndNoteWeb
    • Export to EndNote
  • Save to List
  • Permanent link
Multi-head multi-scale self-cross attention mechanism block.

Multi-head multi-scale self-cross attention mechanism block.

<p>Multi-head multi-scale self-cross attention mechanism block.</p>

Saved in:
Bibliographic Details
Main Author: Haixing Cheng (2240002) (author)
Other Authors: Chengyong Liu (9976282) (author), Wenzhe Gu (5802653) (author), Yuyi Wu (21807702) (author), Mengye Zhao (22184576) (author), Wentao Liu (1447222) (author), Naibang Wang (22184579) (author)
Published: 2025
Subjects:
Neuroscience
Cancer
Science Policy
Space Science
Biological Sciences not elsewhere classified
Information Systems not elsewhere classified
nuscenes validation set
experimental results show
resolution image features
provide enhanced features
better align image
approaches often fail
head adaptive cross
sparse lidar points
xlink "> multi
prior depth information
lgmmfusion achieves 71
image bev features
modal fusion framework
lidar bev features
depth information
lidar features
often lead
modal sensors
framework leverages
better match
lidar bev
head multi
novel lidar
detection head
spatial positions
small objects
scale self
point clouds
perception systems
imprecise detections
guided multi
fully exploit
feature fusion
eye view
critical role
called lgmmfusion
autonomous driving
attention mechanism
also improving
Tags: Add Tag
No Tags, Be the first to tag this record!
  • Holdings
  • Description
  • Comments
  • Similar Items
  • Staff View
Be the first to leave a comment!
You must be logged in first

Similar Items

  • Multi-head adaptive cross-attention block.
    by: Haixing Cheng (2240002)
    Published: (2025)
  • Extended ablation study results on the nuScenes validation set. This table compares the performance of LGMMfusion variants, including the impact of Image BEV (I-BEV), BEV Query (BEV-Q), attention structure modules (MHMS-SA and MHA-CA), and attention parameters (number of heads <i>H</i> and scales <i>S</i>). I-BEV variations involve the use of BatchNorm (BN) and ReLU activations.
    by: Haixing Cheng (2240002)
    Published: (2025)
  • Results on the nuScenes validation set. The results are compared across different methods using LiDAR (L) and Camera (C) modalities(Mod). The performance is evaluated both overall (mAP, NDS) and at the per-class level [75]. The classes include Car, Truck(Tru), Construction Vehicle (C.V.), Bus, Trailer(Tra), Barrier(Bar), Motorcycle (Motor), Bike, Pedestrian (Ped.), and Traffic Cone (T.C.).   Small object categories include pedestrian, traffic cone, and bicycle, which show the most notable mAP improvements.
    by: Haixing Cheng (2240002)
    Published: (2025)
  • Results on nuScenes val set. The modalities are Camera (C), LiDAR (L).
    by: Haixing Cheng (2240002)
    Published: (2025)
  • This figure shows an example of 3D annotation, each row represents a scene.
    by: Haixing Cheng (2240002)
    Published: (2025)

Find More

  • Browse the Catalog
  • Browse Alphabetically
  • Explore Channels
  • Course Reserves
  • New Items
Cannot write session to /tmp/vufind_sessions/sess_glorqk6uqdrl179v24j5n80bmr