Efficient Language Model Training through Cross-Lingual and Progressive Transfer Learning
Efficient Language Model Training through Cross-Lingual and Progressive Transfer Learning
About this item
Full title
Author / Creator
Publisher
Ithaca: Cornell University Library, arXiv.org
Journal title
Language
English
Formats
Publication information
Publisher
Ithaca: Cornell University Library, arXiv.org
Subjects
More information
Scope and Contents
Contents
Most Transformer language models are primarily pretrained on English text, limiting their use for other languages. As the model sizes grow, the performance gap between English and other languages with fewer compute and data resources increases even further. Consequently, more resource-efficient training methods are needed to bridge the gap for lang...
Alternative Titles
Full title
Efficient Language Model Training through Cross-Lingual and Progressive Transfer Learning
Authors, Artists and Contributors
Author / Creator
Identifiers
Primary Identifiers
Record Identifier
TN_cdi_proquest_journals_2768912526
Permalink
https://devfeature-collection.sl.nsw.gov.au/record/TN_cdi_proquest_journals_2768912526
Other Identifiers
E-ISSN
2331-8422