D3 Artikkeli ammatillisessa konferenssijulkaisussa

Register Always Matters: Analysis of LLM Pretraining Data Through the Lens of Language Variation




TekijätMyntti, Amanda; Henriksson, Erik; Laippala,Veronika; Pyysalo, Sampo

ToimittajaN/A

Konferenssin vakiintunut nimiConference on Language Modeling

Julkaisuvuosi2025

Kokoomateoksen nimiProceedings of the Second Conference on Language Modeling, COLM 2025

Julkaisun avoimuus kirjaamishetkelläAvoimesti saatavilla

Julkaisukanavan avoimuus Kokonaan avoin julkaisukanava

Verkko-osoitehttps://openreview.net/forum?id=FqXXtSZWEZ

Rinnakkaistallenteen osoitehttps://research.utu.fi/converis/portal/detail/Publication/506457520

Rinnakkaistallenteen lisenssiCC BY

Rinnakkaistallennetun julkaisun versioKustantajan versio


Tiivistelmä

Pretraining data curation is a cornerstone in Large Language Model (LLM) development, leading to growing research on quality filtering of large web corpora. From statistical quality flags to LLM-based labelling systems, datasets are divided into categories, frequently reducing to a binary: those passing the filters are deemed as valuable examples, others are discarded as useless or detrimental. However, a more detailed understanding of the contribution of different kinds of texts to model performance is still largely lacking. In this article, we present the first study utilising registers or genres—a widely used standard in corpus linguistics to model linguistic variation—to curate pretraining datasets and investigate the effect of register on the performance of LLMs. We train small generative models with register classified data and evaluate them using standard benchmarks, and show that the register of pretraining data substantially affects model performance. We uncover surprising relationships between the pretraining material and the resulting models: using the News register results in subpar performance, and on the contrary, including the Opinion class, covering texts such as reviews and opinion blogs, is highly beneficial. While a model trained on the entire unfiltered dataset outperforms those trained on datasets limited to a single register, combining well-performing registers such as How-to-InstructionsInformational Description, and Opinion leads to major improvements. Furthermore, analysis of individual benchmark results reveals key differences in the strengths and drawbacks of specific register classes as pretraining data: How-to-Instructions excels at physical reasoning and sentence completion while barely crossing random baselines on world-knowledge benchmarks, while Narrative boosts performance on social interaction tasks but struggles with scientific questions. These findings show that register is an important explainer of model variation and can facilitate more deliberate and detailed future data selection practices.


Ladattava julkaisu

This is an electronic reprint of the original article.
This reprint may differ from the original in pagination and typographic detail. Please cite the original version.




Julkaisussa olevat rahoitustiedot
This project has received funding from the Finnish Doctoral Program Network in Artificial
Intelligence, AI-DOC under decision number VN/3137/2024-OKM-6, the European Union’s
Horizon Europe research and innovation programme under grant agreement No. 101070350, the
Digital Europe Programme under grant agreement No. 101195233, and the Research Council of
Finland under grant No. 362459.


Last updated on