How to run a world record? A Reinforcement Learning approach

dc.contributor.authorShahsavari Sajad
dc.contributor.authorImmonen Eero
dc.contributor.authorKarami Masoomeh
dc.contributor.authorHaghbayan Mohammadhashem
dc.contributor.authorPlosila Juha
dc.contributor.organizationfi=robotiikka ja autonomiset järjestelmät|en=Robotics and Autonomous Systems|
dc.contributor.organization-code1.2.246.10.2458963.20.72785230805
dc.converis.publication-id175657877
dc.converis.urlhttps://research.utu.fi/converis/portal/Publication/175657877
dc.date.accessioned2022-10-28T13:41:40Z
dc.date.available2022-10-28T13:41:40Z
dc.description.abstract<p>Finding the optimal distribution of exerted effort by an athlete in competitive sports has been widely investigated in the fields of sport science, applied mathematics and optimal control. In this article, we propose a reinforcement learning-based solution to the optimal control problem in the running race application. Well-known mathematical model of Keller is used for numerically simulating the dynamics in runner's energy storage and motion. A feed-forward neural network is employed as the probabilistic controller model in continuous action space which transforms the current state (position, velocity and available energy) of the runner to the predicted optimal propulsive force that the runner should apply in the next time step. A logarithmic barrier reward function is designed to evaluate performance of simulated races as a continuous smooth function of runner's position and time. The neural network parameters, then, are identified by maximizing the expected reward using on-policy actor-critic policy-gradient RL algorithm. We trained the controller model for three race lengths: 400, 1500 and 10000 meters and found the force and velocity profiles that produce a near-optimal solution for the runner's problem. Results conform with Keller's theoretical findings with relative percent error of 0.59% and are comparable to real world records with relative percent error of 2.38%, while the same error for Keller's findings is 2.82%.<br></p>
dc.format.pagerange159
dc.format.pagerange166
dc.identifier.isbn978-3-937436-77-7
dc.identifier.issn2522-2414
dc.identifier.jour-issn2522-2414
dc.identifier.olddbid183674
dc.identifier.oldhandle10024/166768
dc.identifier.urihttps://www.utupub.fi/handle/11111/40958
dc.identifier.urlhttps://www.scs-europe.net/dlib/2022/2022-0159.html
dc.identifier.urnURN:NBN:fi-fe2022081154615
dc.language.isoen
dc.okm.affiliatedauthorKarami, Masoomeh
dc.okm.affiliatedauthorHaghbayan, Hashem
dc.okm.affiliatedauthorPlosila, Juha
dc.okm.discipline113 Computer and information sciencesen_GB
dc.okm.discipline213 Electronic, automation and communications engineering, electronicsen_GB
dc.okm.discipline113 Tietojenkäsittely ja informaatiotieteetfi_FI
dc.okm.discipline213 Sähkö-, automaatio- ja tietoliikennetekniikka, elektroniikkafi_FI
dc.okm.internationalcopublicationnot an international co-publication
dc.okm.internationalityInternational publication
dc.okm.typeA4 Conference Article
dc.publisher.countryUnited Kingdomen_GB
dc.publisher.countryBritanniafi_FI
dc.publisher.country-codeGB
dc.relation.conferenceEuropean Conference on Modelling and Simulation
dc.relation.ispartofjournalProceedings: European Conference for Modelling and Simulation
dc.relation.ispartofseriesProceedings : European Conference for Modelling and Simulation
dc.relation.volume1
dc.relation.volume36
dc.source.identifierhttps://www.utupub.fi/handle/10024/166768
dc.titleHow to run a world record? A Reinforcement Learning approach
dc.title.bookProceedings of the 36th ECMS International Conference on Modelling and Simulation ECMS 2022 May 30th – June 3rd, 2022, Ålesund, Norway
dc.year.issued2022

Tiedostot

Näytetään 1 - 1 / 1
Ladataan...
Name:
0159_simo_ecms2022_0049.pdf
Size:
1.05 MB
Format:
Adobe Portable Document Format