Resumo – Publicações

Improving deep neural network random initialization through neuronal rewiring.
SCABINI, Leonardo Felipe dos Santos; BAETS, Bernard De; BRUNO, Odemir Martinez.
Abstract: The deep learning literature is continuously updated with new architectures and training techniques. However, weight initialization is overlooked by most recent research, despite some intriguing findings regarding random weights. On the other hand, recent works have been approaching Network Science to understand the structure and dynamics of Artificial Neural Networks (ANNs) after training. Therefore, in this work, we analyze the centrality of neurons in randomly initialized networks. We show that a higher neuronal strength variance may decrease performance, while a lower neuronal strength variance usually improves it. A new method is then proposed to rewire neuronal connections according to a preferential attachment (PA) rule based on their strength, which significantly reduces the strength variance of layers initialized by common methods. In this sense, PA rewiring only reorganizes connections, while preserving the magnitude and distribution of the weights. We show through an extensive statistical analysis on image classification tasks that performance is improved in most cases, both during training and testing, when using both simple and complex architectures and learning schedules. Our results show that, aside from the magnitude, the organization of the weights is also relevant for better initialization of deep ANNs.
Neurocomputing
v. 599, p. 128130-1-128130-13 + supplementary data - Ano: 2024
Fator de Impacto: 5,5
    @article={003206967,author = {SCABINI, Leonardo Felipe dos Santos; BAETS, Bernard De; BRUNO, Odemir Martinez.},title={Improving deep neural network random initialization through neuronal rewiring},journal={Neurocomputing},note={v. 599, p. 128130-1-128130-13 + supplementary data},year={2024}}