TULUN: Transparent and Adaptable Low-resource Machine Translation

dc.contributor.authorMerx, Raphael
dc.contributor.authorSuominen, Hanna
dc.contributor.authorHong, Lois Yinghui
dc.contributor.authorThieberger, Nick
dc.contributor.authorCohn, Trevor
dc.contributor.authorVylomova, Ekaterina
dc.contributor.organizationfi=tietotekniikan laitos|en=Department of Computing|
dc.contributor.organization-code1.2.246.10.2458963.20.85312822902
dc.converis.publication-id506057677
dc.converis.urlhttps://research.utu.fi/converis/portal/Publication/506057677
dc.date.accessioned2026-01-21T13:35:18Z
dc.date.available2026-01-21T13:35:18Z
dc.description.abstractMachine translation (MT) systems that support low-resource languages often struggle on specialized domains. While researchers have proposed various techniques for domain adaptation, these approaches typically require model fine-tuning, making them impractical for non-technical users and small organizations. To address this gap, we propose TULUN,(1) a versatile solution for terminology-aware translation, combining neural MT with large language model (LLM)-based post-editing guided by existing glossaries and translation memories. Our open-source web-based platform enables users to easily create, edit, and leverage terminology resources, fostering a collaborative human-machine translation process that respects and incorporates domain expertise while increasing MT accuracy. Evaluations show effectiveness in both real-world and benchmark scenarios: on medical and disaster relief translation tasks for Tetun and Bislama, our system achieves improvements of 16.90-22.41 ChrF++ points over baseline MT systems. Across six low-resource languages on the FLORES dataset, TULUN outperforms both standalone MT and LLM approaches, achieving an average improvement of 2.8 ChrF++ points over NLLB-54B. TULUN is publicly accessible at bislama-trans.rapha.dev.
dc.format.pagerange129
dc.format.pagerange139
dc.identifier.isbn979-8-89176-253-4
dc.identifier.issn0736-587X
dc.identifier.jour-issn0736-587X
dc.identifier.olddbid213129
dc.identifier.oldhandle10024/196147
dc.identifier.urihttps://www.utupub.fi/handle/11111/54807
dc.identifier.urlhttps://aclanthology.org/2025.acl-demo.13/
dc.identifier.urnURN:NBN:fi-fe202601217172
dc.language.isoen
dc.okm.affiliatedauthorSuominen, Hanna
dc.okm.discipline113 Computer and information sciencesen_GB
dc.okm.discipline113 Tietojenkäsittely ja informaatiotieteetfi_FI
dc.okm.internationalcopublicationinternational co-publication
dc.okm.internationalityInternational publication
dc.okm.typeA4 Conference Article
dc.publisher.countryUnited Statesen_GB
dc.publisher.countryYhdysvallat (USA)fi_FI
dc.publisher.country-codeUS
dc.relation.conferenceAnnual Meeting of the Association for Computational Linguistics
dc.relation.ispartofjournalAnnual Meeting of the Association for Computational Linguistics
dc.relation.volume63
dc.source.identifierhttps://www.utupub.fi/handle/10024/196147
dc.titleTULUN: Transparent and Adaptable Low-resource Machine Translation
dc.title.bookProceedings of the 63rd Annual Meeting of the Association for Computational Linguistics : (Volume 3: System Demonstrations)
dc.year.issued2025

Tiedostot

Näytetään 1 - 1 / 1
Ladataan...
Name:
2025.acl-demo.13.pdf
Size:
680.73 KB
Format:
Adobe Portable Document Format