NWU Institutional Repository

Impact of batch normalization on convolutional network representations

Loading...
Thumbnail Image

Researcher ID

Supervisors

Journal Title

Journal ISSN

Volume Title

Publisher

Record Identifier

Abstract

Batch normalization (BatchNorm) is a popular layer normalization technique used when training deep neural networks. It has been shown to enhance the training speed and accuracy of deep learning models. However, the mechanics by which BatchNorm achieves these benefits is an active area of research, and different perspectives have been proposed. In this paper, we investigate the effect of BatchNorm on the resulting hidden representations, that is, the vectors of activation valuesformed as samples are processed at each hidden layer. Specifically, we consider the sparsity of these representations, as well as their implicit clustering – the creation of groups of representations that are similar to some extent. We contrast image classification models trained with and without batch normalization and highlight consistent differences observed. These findings highlight that BatchNorm’s effect on representational sparsity is not a significant factor affecting generalization, while the representations of models trained with BatchNorm tend to show more advantageous clustering characteristics.

Sustainable Development Goals

Description

Keywords

Citation

Potgieter, H.L. et al. Impact of batch normalization on convolutional network representations. arXiv:2501.14441v2 [cs.LG] 13 Feb 2025

Collections

Endorsement

Review

Supplemented By

Referenced By