UFM: Unified feature matching pre-training with multi-modal image assistants
UFM: Unified feature matching pre-training with multi-modal image assistants
About this item
Full title
Author / Creator
Di, Yide , Liao, Yun , Zhou, Hao , Zhu, Kaijun , Duan, Qing , Liu, Junhui and Lu, Mingyu
Publisher
United States: Public Library of Science
Journal title
Language
English
Formats
Publication information
Publisher
United States: Public Library of Science
Subjects
More information
Scope and Contents
Contents
Image feature matching, a foundational task in computer vision, remains challenging for multimodal image applications, often necessitating intricate training on specific datasets. In this paper, we introduce a Unified Feature Matching pre-trained model (UFM) designed to address feature matching challenges across a wide spectrum of modal images. We present Multimodal Image Assistant (MIA) transformers, finely tunable structures adept at handling diverse feature matching problems. UFM exhibits versatility in addressing both feature matching tasks within the same modal and those across different modals. Additionally, we propose a data augmentation algorithm and a staged pre-training strategy to effectively tackle challenges arising from sparse data in specific modals and imbalanced modal datasets. Experimental results demonstrate that UFM excels in generalization and performance across various feature matching tasks. The code will be released at: https://github.com/LiaoYun0x0/UFM ....
Alternative Titles
Full title
UFM: Unified feature matching pre-training with multi-modal image assistants
Authors, Artists and Contributors
Author / Creator
Identifiers
Primary Identifiers
Record Identifier
TN_cdi_plos_journals_3184406697
Permalink
https://devfeature-collection.sl.nsw.gov.au/record/TN_cdi_plos_journals_3184406697
Other Identifiers
ISSN
1932-6203
E-ISSN
1932-6203
DOI
10.1371/journal.pone.0319051