Growing Complex Networks for Better Learning of Chaotic Dynamical Systems
This thesis advances the theory of network specialization by characterizing the effect of network specialization on the eigenvectors of a network. We prove and provide explicit formulas for the eigenvectors of specialized graphs based on the eigenvectors of their parent graphs. The second portion of...
Main Author: | |
---|---|
Format: | Others |
Published: |
BYU ScholarsArchive
2020
|
Subjects: | |
Online Access: | https://scholarsarchive.byu.edu/etd/8146 https://scholarsarchive.byu.edu/cgi/viewcontent.cgi?article=9146&context=etd |
Summary: | This thesis advances the theory of network specialization by characterizing the effect of network specialization on the eigenvectors of a network. We prove and provide explicit formulas for the eigenvectors of specialized graphs based on the eigenvectors of their parent graphs. The second portion of this thesis applies network specialization to learning problems. Our work focuses on training reservoir computers to mimic the Lorentz equations. We experiment with random graph, preferential attachment and small world topologies and demonstrate that the random removal of directed edges increases predictive capability of a reservoir topology. We then create a new network model by growing networks via targeted application of the specialization model. This is accomplished iteratively by selecting top preforming nodes within the reservoir computer and specializing them. Our generated topology out-preforms all other topologies on average. |
---|