The mathematics of neural networks behind image recognition

dc.contributor.authorJäppinen, Jarno
dc.contributor.departmentfi=Matematiikan ja tilastotieteen laitos|en=Department of Mathematics and Statistics|
dc.contributor.facultyfi=Matemaattis-luonnontieteellinen tiedekunta|en=Faculty of Science|
dc.contributor.studysubjectfi=Matematiikka|en=Mathematics|
dc.date.accessioned2024-06-10T21:02:45Z
dc.date.available2024-06-10T21:02:45Z
dc.date.issued2024-05-24
dc.description.abstractThis thesis aims to present a special case of neural network for image recognition: the convolutional neural network. Before that, the neural network in general, its components, architecture and operation will be presented. The idea of a convolutional neural network is to reduce the mathematical computation by focusing more on extracting the information that is necessary for recognition. With the convolution neural network technique, networks can be made quite large and complex without the computational needs increasing proportionally. In this paper, we will introduce the mathematics needed to train a neural network. The computation is based on optimizing the weight coefficients of the network by minimizing the error function using a backpropagation algorithm. The work also includes a simple fully connected neural network and a more complex convolutional neural network trained on two different datasets. The CIFAR-10 and Fashion-MNIST datasets have been used. The CIFAR-10 dataset consists of small coloured images which were more challenging to recognize with the models compared to the Fashion-MNIST dataset where the images are greyscale and slightly smaller. With the CIFAR-10 data, the convolutional neural network learned with a fairly good accuracy in recognition while the fully connected neural network did not achieve very good learning results. With the Fashion-MNIST dataset, both models, fully connected and convolutional, learned to recognize objects quite well. I conclude that the learning of recognition is particularly influenced by the content of the material. If the images are clear and contain only one object, classification is easy, as it is for humans. But if there are many different things in the picture it is much harder for a human as well as a machine to tell what is in the picture. The amount of computation required is also greatly affected by the size of the images and the number of different classification categories.
dc.format.extent54
dc.identifier.olddbid195166
dc.identifier.oldhandle10024/178219
dc.identifier.urihttps://www.utupub.fi/handle/11111/19546
dc.identifier.urnURN:NBN:fi-fe2024061049172
dc.language.isoeng
dc.rightsfi=Julkaisu on tekijänoikeussäännösten alainen. Teosta voi lukea ja tulostaa henkilökohtaista käyttöä varten. Käyttö kaupallisiin tarkoituksiin on kielletty.|en=This publication is copyrighted. You may download, display and print it for Your own personal use. Commercial use is prohibited.|
dc.rights.accessrightsavoin
dc.source.identifierhttps://www.utupub.fi/handle/10024/178219
dc.subjectneural network, fully connected neural network, convolutional neural network, backpropagation
dc.titleThe mathematics of neural networks behind image recognition
dc.type.ontasotfi=Pro gradu -tutkielma|en=Master's thesis|

Tiedostot

Näytetään 1 - 1 / 1
Ladataan...
Name:
Jappinen_Jarno_opinnayte.pdf
Size:
1.79 MB
Format:
Adobe Portable Document Format