Growing Complex Networks for Better Learning of Chaotic Dynamical Systems

This thesis advances the theory of network specialization by characterizing the effect of network specialization on the eigenvectors of a network. We prove and provide explicit formulas for the eigenvectors of specialized graphs based on the eigenvectors of their parent graphs. The second portion of...

Full description

Bibliographic Details
Main Author: Passey Jr., David Joseph
Format: Others
Published: BYU ScholarsArchive 2020
Subjects:
Online Access:https://scholarsarchive.byu.edu/etd/8146
https://scholarsarchive.byu.edu/cgi/viewcontent.cgi?article=9146&context=etd
Description
Summary:This thesis advances the theory of network specialization by characterizing the effect of network specialization on the eigenvectors of a network. We prove and provide explicit formulas for the eigenvectors of specialized graphs based on the eigenvectors of their parent graphs. The second portion of this thesis applies network specialization to learning problems. Our work focuses on training reservoir computers to mimic the Lorentz equations. We experiment with random graph, preferential attachment and small world topologies and demonstrate that the random removal of directed edges increases predictive capability of a reservoir topology. We then create a new network model by growing networks via targeted application of the specialization model. This is accomplished iteratively by selecting top preforming nodes within the reservoir computer and specializing them. Our generated topology out-preforms all other topologies on average.