A Survey on Sparsity Exploration in Transformer-Based Accelerators
A Survey on Sparsity Exploration in Transformer-Based Accelerators
About this item
Full title
Author / Creator
Publisher
Basel: MDPI AG
Journal title
Language
English
Formats
Publication information
Publisher
Basel: MDPI AG
Subjects
More information
Scope and Contents
Contents
Transformer models have emerged as the state-of-the-art in many natural language processing and computer vision applications due to their capability of attending to longer sequences of tokens and supporting parallel processing more efficiently. Nevertheless, the training and inference of transformer models are computationally expensive and memory i...
Alternative Titles
Full title
A Survey on Sparsity Exploration in Transformer-Based Accelerators
Authors, Artists and Contributors
Author / Creator
Identifiers
Primary Identifiers
Record Identifier
TN_cdi_proquest_journals_2819443357
Permalink
https://devfeature-collection.sl.nsw.gov.au/record/TN_cdi_proquest_journals_2819443357
Other Identifiers
ISSN
2079-9292
E-ISSN
2079-9292
DOI
10.3390/electronics12102299