A4 Vertaisarvioitu artikkeli konferenssijulkaisussa

WikiBERT Models: Deep Transfer Learning for Many Languages




TekijätPyysalo Sampo, Kanerva Jenna, Virtanen Antti, Ginter Filip

ToimittajaSimon Dobnik, Lilja Øvrelid

Konferenssin vakiintunut nimiNordic Conference on Computational Linguistics

Julkaisuvuosi2021

JournalLinköping Electronic Conference Proceedings

Kokoomateoksen nimiProceedings of the 23rd Nordic Conference on Computational Linguistics (NoDaLiDa)

Sarjan nimiLinköping Electronic Conference Proceedings

Numero sarjassa178

Aloitussivu1

Lopetussivu10

ISBN978-91-7929-614-8

ISSN1650-3686

Verkko-osoitehttps://ep.liu.se/en/conference-article.aspx?series=ecp&issue=178&Article_No=1

Rinnakkaistallenteen osoitehttps://research.utu.fi/converis/portal/detail/Publication/56908534


Tiivistelmä
Deep neural language models such as BERT have enabled substantial recent advances in many natural language processing tasks. However, due to the effort and computational cost involved in their pre-training, such models are typically introduced only for a small number of high-resource languages such as English. While multilingual models covering large numbers of languages are available, recent work suggests monolingual training can produce better models, and our understanding of the tradeoffs between mono- and multilingual training is incomplete. In this paper, we introduce a simple, fully automated pipeline for creating language-specific BERT models from Wikipedia data and introduce 42 new such models, most for languages up to now lacking dedicated deep neural language models. We assess the merits of these models using cloze tests and the state-of-the-art UDify parser on Universal Dependencies data, contrasting performance with results using the multilingual BERT (mBERT) model. We find that the newly introduced WikiBERT models outperform mBERT in cloze tests for nearly all languages, and that UDify using WikiBERT models outperforms the parser using mBERT on average, with the language-specific models showing substantially improved performance for some languages, yet limited improvement or a decrease in performance for others. All of the methods and models introduced in this work are available under open licenses from https://github.com/turkunlp/wikibert

Ladattava julkaisu

This is an electronic reprint of the original article.
This reprint may differ from the original in pagination and typographic detail. Please cite the original version.





Last updated on 2024-26-11 at 23:16