Fine-tuning on same-model feature matching tasks. The X-FFN and Y-FFN represent the assistants of any two kinds of pre-trained different modal images in the second stage of Fig 4. The fine-tuning of the X-modal image and the fine-tuning of the Y-modal image are independent of each other.
<p>Fine-tuning on same-model feature matching tasks. The X-FFN and Y-FFN represent the assistants of any two kinds of pre-trained different modal images in the second stage of Fig 4. The fine-tuning of the X-modal image and the fine-tuning of the Y-modal image are independent of each other.<...
Saved in:
| Main Author: | Yide Di (20969124) (author) |
|---|---|
| Other Authors: | Yun Liao (160524) (author), Hao Zhou (136535) (author), Kaijun Zhu (18283913) (author), Qing Duan (541846) (author), Junhui Liu (2063140) (author), Mingyu Lu (2333083) (author) |
| Published: |
2025
|
| Subjects: | |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Fine-tuning on different-model feature matching tasks. The X-FFN and Y-FFN represent the assistants of any two kinds of pre-trained different modal images in the second stage of Fig 4. The X-Y FFN represent the assistant of the pre-trained different modal images in the third stage of Fig 4.
by: Yide Di (20969124)
Published: (2025) -
Illustration of feature matching of UFM. When working with specific data, the pre-trained backbone is frozen, and only the corresponding modal assistants need to be fine-tuned for feature matching. The multi-modal assistants contain both same-modal assistants and different-modal matching assistants.
by: Yide Di (20969124)
Published: (2025) -
Illustration of pre-training. The pre-training of NIR and SAR images is taken here as an example. Pre-training consists of 3 stages: (1) pre-train the general FFN, (2) pre-train all X-X assistants, and (3) pre-train all X-Y assistants.
by: Yide Di (20969124)
Published: (2025) -
MMA estimation on multi-modal images in SEN12MS dataset.
by: Yide Di (20969124)
Published: (2025) -
MMA estimation on different multi-modal image datasets. The modal of the images: (1) RGB-NIR, (2) Optical-SAR, (3) Optical-SAR, (4) T1-T2 (5) RGB-Depth, (6) UV-Green.
by: Yide Di (20969124)
Published: (2025)