How important is weight symmetry in backpropagation?

Gradient backpropagation (BP) requires symmetric feedforward and feedback connections-the same weights must be used for forward and backward passes. This "weight transport problem" (Grossberg 1987) is thought to be one of the main reasons to doubt BP's biologically plausibility. Using...

Full description

Bibliographic Details
Main Authors: Liao, Qianli (Contributor), Leibo, Joel Z (Contributor), Poggio, Tomaso A (Contributor)
Other Authors: Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences (Contributor), Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science (Contributor), McGovern Institute for Brain Research at MIT (Contributor)
Format: Article
Language:English
Published: Association for the Advancement of Artificial Intelligence, 2017-11-28T18:13:54Z.
Subjects:
Online Access:Get fulltext
LEADER 02188 am a22002413u 4500
001 112304
042 |a dc 
100 1 0 |a Liao, Qianli  |e author 
100 1 0 |a Massachusetts Institute of Technology. Department of Brain and Cognitive Sciences  |e contributor 
100 1 0 |a Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science  |e contributor 
100 1 0 |a McGovern Institute for Brain Research at MIT  |e contributor 
100 1 0 |a Liao, Qianli  |e contributor 
100 1 0 |a Leibo, Joel Z  |e contributor 
100 1 0 |a Poggio, Tomaso A  |e contributor 
700 1 0 |a Leibo, Joel Z  |e author 
700 1 0 |a Poggio, Tomaso A  |e author 
245 0 0 |a How important is weight symmetry in backpropagation? 
260 |b Association for the Advancement of Artificial Intelligence,   |c 2017-11-28T18:13:54Z. 
856 |z Get fulltext  |u http://hdl.handle.net/1721.1/112304 
520 |a Gradient backpropagation (BP) requires symmetric feedforward and feedback connections-the same weights must be used for forward and backward passes. This "weight transport problem" (Grossberg 1987) is thought to be one of the main reasons to doubt BP's biologically plausibility. Using 15 different classification datasets, we systematically investigate to what extent BP really depends on weight symmetry. In a study that turned out to be surprisingly similar in spirit to Lillicrap et al.'s demonstration (Lillicrap et al. 2014) but orthogonal in its results, our experiments indicate that: (1) the magnitudes of feedback weights do not matter to performance (2) the signs of feedback weights do matter-the more concordant signs between feedforward and their corresponding feedback connections, the better (3) with feedback weights having random magnitudes and 100% concordant signs, we were able to achieve the same or even better performance than SGD. (4) some normalizations/stabilizations are indispensable for such asymmetric BP to work, namely Batch Normalization (BN) (Ioffe and Szegedy 2015) and/or a "Batch Manhattan" (BM) update rule. 
520 |a National Science Foundation (U.S.) (STC Award CCF 1231216) 
655 7 |a Article 
773 |t Thirtieth AAAI Conference on Artificial Intelligence