Scaling Data-Constrained Language Models

dc.contributor.authorMuennighoff, Niklas
dc.contributor.authorRush, Alexander M.
dc.contributor.authorBarak, Boaz
dc.contributor.authorLe Scao, Teven
dc.contributor.authorPiktus, Aleksandra
dc.contributor.authorTazi, Nouamane
dc.contributor.authorPyysalo, Sampo
dc.contributor.authorWolf, Thomas
dc.contributor.authorRaffel, Colin
dc.contributor.organizationfi=data-analytiikka|en=Data-analytiikka|
dc.contributor.organization-code1.2.246.10.2458963.20.68940835793
dc.converis.publication-id492253404
dc.converis.urlhttps://research.utu.fi/converis/portal/Publication/492253404
dc.date.accessioned2025-08-27T22:19:52Z
dc.date.available2025-08-27T22:19:52Z
dc.description.abstractThe current trend of scaling language models involves increasing both parameter count and training data set size. Extrapolating this trend suggests that training data set size may soon be limited by the amount of text data available on the internet. Motivated by this limit, we investigate scaling language models in data-constrained regimes. Specifically, we run a large set of experiments varying the extent of data repetition and compute budget, ranging up to 900 billion training tokens and 9 billion parameter models. We find that with constrained data for a fixed compute budget, training with up to 4 epochs of repeated data yields negligible changes to loss compared to having unique data. However, with more repetition, the value of adding compute eventually decays to zero. We propose and empirically validate a scaling law for compute optimality that accounts for the decreasing value of repeated tokens and excess parameters. Finally, we experiment with approach esmitigating data scarcity, including augmenting the training data set with code data or removing commonly used filters. Models and data sets from our 400 training runs are freely available athttps://github.com/huggingface/datablations.
dc.identifier.eissn1533-7928
dc.identifier.jour-issn1532-4435
dc.identifier.olddbid201995
dc.identifier.oldhandle10024/185022
dc.identifier.urihttps://www.utupub.fi/handle/11111/40921
dc.identifier.urlhttps://www.jmlr.org/papers/v26/24-1000.html
dc.identifier.urnURN:NBN:fi-fe2025082789636
dc.language.isoen
dc.okm.affiliatedauthorPyysalo, Sampo
dc.okm.discipline113 Computer and information sciencesen_GB
dc.okm.discipline113 Tietojenkäsittely ja informaatiotieteetfi_FI
dc.okm.internationalcopublicationinternational co-publication
dc.okm.internationalityInternational publication
dc.okm.typeA1 ScientificArticle
dc.publisherMICROTOME PUBL
dc.publisher.countryUnited Statesen_GB
dc.publisher.countryYhdysvallat (USA)fi_FI
dc.publisher.country-codeUS
dc.publisher.placeBROOKLINE
dc.relation.articlenumber53
dc.relation.ispartofjournalJournal of Machine Learning Research
dc.relation.volume26
dc.source.identifierhttps://www.utupub.fi/handle/10024/185022
dc.titleScaling Data-Constrained Language Models
dc.year.issued2025

Tiedostot

Näytetään 1 - 1 / 1
Ladataan...
Name:
24-1000.pdf
Size:
2.08 MB
Format:
Adobe Portable Document Format