Algebraic shortcuts for leave-one-out cross-validation in supervised network inference

dc.contributor.authorStock M
dc.contributor.authorPahikkala T
dc.contributor.authorAirola A
dc.contributor.authorWaegeman W
dc.contributor.authorDe Baets B
dc.contributor.organizationfi=tietojenkäsittelytiede|en=Computer Science|
dc.contributor.organization-code1.2.246.10.2458963.20.23479734818
dc.converis.publication-id37642437
dc.converis.urlhttps://research.utu.fi/converis/portal/Publication/37642437
dc.date.accessioned2025-08-27T23:34:20Z
dc.date.available2025-08-27T23:34:20Z
dc.description.abstractSupervised machine learning techniques have traditionally been very successful at reconstructing biological networks, such as protein-ligand interaction, protein-protein interaction and gene regulatory networks. Many supervised techniques for network prediction use linear models on a possibly nonlinear pairwise feature representation of edges. Recently, much emphasis has been placed on the correct evaluation of such supervised models. It is vital to distinguish between using a model to either predict new interactions in a given network or to predict interactions for a new vertex not present in the original network. This distinction matters because (i) the performance might dramatically differ between the prediction settings and (ii) tuning the model hyperparameters to obtain the best possible model depends on the setting of interest. Specific cross-validation schemes need to be used to assess the performance in such different prediction settings.In this work we discuss a state-of-the-art kernel-based network inference technique called two-step kernel ridge regression. We show that this regression model can be trained efficiently, with a time complexity scaling with the number of vertices rather than the number of edges. Furthermore, this framework leads to a series of cross-validation shortcuts that allow one to rapidly estimate the model performance for any relevant network prediction setting. This allows computational biologists to fully assess the capabilities of their models. The machine learning techniques with the algebraic shortcuts are implemented in the RLScore software package: https://github.com/aatapa/RLScore.
dc.format.pagerange262
dc.format.pagerange271
dc.identifier.eissn1477-4054
dc.identifier.jour-issn1467-5463
dc.identifier.olddbid204217
dc.identifier.oldhandle10024/187244
dc.identifier.urihttps://www.utupub.fi/handle/11111/52379
dc.identifier.urnURN:NBN:fi-fe2021042824507
dc.language.isoen
dc.okm.affiliatedauthorPahikkala, Tapio
dc.okm.affiliatedauthorAirola, Antti
dc.okm.discipline113 Computer and information sciencesen_GB
dc.okm.discipline113 Tietojenkäsittely ja informaatiotieteetfi_FI
dc.okm.internationalcopublicationinternational co-publication
dc.okm.internationalityInternational publication
dc.okm.typeA1 ScientificArticle
dc.publisherOxford University Press
dc.publisher.countryUnited Kingdomen_GB
dc.publisher.countryBritanniafi_FI
dc.publisher.country-codeGB
dc.relation.doi10.1093/bib/bby095
dc.relation.ispartofjournalBriefings in Bioinformatics
dc.relation.issue1
dc.relation.volume21
dc.source.identifierhttps://www.utupub.fi/handle/10024/187244
dc.titleAlgebraic shortcuts for leave-one-out cross-validation in supervised network inference
dc.year.issued2020

Tiedostot

Näytetään 1 - 1 / 1
Ladataan...
Name:
242321.1.full.pdf
Size:
744.4 KB
Format:
Adobe Portable Document Format
Description:
Final draft