Log in to save to my catalogue

Super Vision Transformer

Super Vision Transformer

https://devfeature-collection.sl.nsw.gov.au/record/TN_cdi_proquest_journals_2882797984

Super Vision Transformer

About this item

Full title

Super Vision Transformer

Publisher

New York: Springer US

Journal title

International journal of computer vision, 2023-12, Vol.131 (12), p.3136-3151

Language

English

Formats

Publication information

Publisher

New York: Springer US

More information

Scope and Contents

Contents

We attempt to reduce the computational costs in vision transformers (ViTs), which increase quadratically in the token number. We present a novel training paradigm that trains only one ViT model at a time, but is capable of providing improved image recognition performance with various computational costs. Here, the trained ViT model, termed super vision transformer (SuperViT), is empowered with the versatile ability to solve incoming patches of multiple sizes as well as preserve informative tokens with multiple keeping rates (the ratio of keeping tokens) to achieve good hardware efficiency for inference, given that the available hardware resources often change from time to time. Experimental results on ImageNet demonstrate that our SuperViT can considerably reduce the computational costs of ViT models with even performance increase. For example, we reduce 2 
×
 FLOPs of DeiT-S while increasing the Top-1 accuracy by 0.2% and 0.7% for 1.5 
×
 reduction. Also, our SuperViT significantly outperforms existing studies on efficient vision transformers. For example, when consuming the same amount of FLOPs, our SuperViT surpasses the recent state-of-the-art EViT by 1.1% when using DeiT-S as their backbones. The project of this work is made publicly available at
https://github.com/lmbxmu/SuperViT
....

Alternative Titles

Full title

Super Vision Transformer

Identifiers

Primary Identifiers

Record Identifier

TN_cdi_proquest_journals_2882797984

Permalink

https://devfeature-collection.sl.nsw.gov.au/record/TN_cdi_proquest_journals_2882797984

Other Identifiers

ISSN

0920-5691

E-ISSN

1573-1405

DOI

10.1007/s11263-023-01861-3

How to access this item