dc.contributor |
Lacasa Saiz de Arce, Lucas Jaime |
|
dc.contributor |
Soriano, Miguel |
|
dc.contributor.author |
Danovski, Kaloyan Martinov |
|
dc.date |
2023 |
|
dc.date.accessioned |
2024-10-07T08:17:28Z |
|
dc.date.issued |
2023-10-13 |
|
dc.identifier.uri |
http://hdl.handle.net/11201/166277 |
|
dc.description.abstract |
[eng] The process of training an artificial neural network involves iteratively adapting its weight parameters so as to minimize the error of the
network’s prediction, when confronted with a learning task. This iterative change can be naturally interpreted as a trajectory in network
space, and thus the training algorithm (e.g. Gradient Descent optimization of a suitable loss function) can be interpreted as a dynamical
system in graph space, and the whole training can be characterized
by a time series of networks. In this work, we study the dynamical
properties of this system and focus on its dynamical stability. We do
this mostly by studying how the distance between initially close neural
networks evolves over time during training, in a form of “orbital stability”. We find that such distance evolution qualitatively depends on the
specific learning rate of the gradient descent scheme, and study a few
of the di↵erent regions, finding hints of regular and irregular—possibly
chaotic—behavior. Our findings are put in contrast to common wisdom
on convergence properties of neural networks and dynamical systems
theory. The research of this thesis—and thereby, the presentation of its
results—is exploratory in nature, and its exposition follows by stating
several questions bridging ideas from machine learning and dynamical
systems that we subsequently address within a vanilla neural network
model, hence shaping up a logical narrative. This thesis, we hope, not
only will shed light on the training mechanisms and inner workings of
neural networks, but will provide motivation and justification for further research at the interface between dynamical systems theory and
learning algorithms. |
ca |
dc.format |
application/pdf |
|
dc.language.iso |
eng |
ca |
dc.publisher |
Universitat de les Illes Balears |
|
dc.rights |
info:eu-repo/semantics/openAccess |
|
dc.rights |
all rights reserved |
|
dc.subject |
53 - Física |
ca |
dc.subject |
530.1 - Principis generals de la física |
ca |
dc.subject.other |
Artificial neural networks |
ca |
dc.subject.other |
Machine learning |
ca |
dc.subject.other |
Dynamical systems |
ca |
dc.subject.other |
Chaos |
ca |
dc.title |
Artificial neural networks through the lens of dynamical systems theory |
ca |
dc.type |
info:eu-repo/semantics/masterThesis |
ca |
dc.type |
info:eu-repo/semantics/publishedVersion |
|
dc.date.updated |
2024-06-03T11:22:38Z |
|
dc.date.embargoEndDate |
info:eu-repo/date/embargoEnd/2050-01-01 |
|
dc.embargo |
2050-01-01 |
|
dc.rights.accessRights |
info:eu-repo/semantics/embargoedAccess |
|