Stochastic Weight Matrix-based Regularization Methods for Deep Neural Networks

Published in 5th International Conference on Learning, Optimization and Data Science, 2019

Recommended citation: Reizinger P., Gyires-Tóth B. (2019) "Stochastic Weight Matrix-Based Regularization Methods for Deep Neural Networks.." Springer LNCS 1. 11943. https://arxiv.org/pdf/1909.11977

Abstract

The aim of this paper is to introduce two widely applicable regularization methods based on the direct modification of weight matrices. The first method, Weight Reinitialization, utilizes a simplified Bayesian assumption with partially resetting a sparse subset of the parameters. The second one, Weight Shuffling, introduces an entropy- and weight distribution-invariant non-white noise to the parameters. The latter can also be interpreted as an ensemble approach. The proposed methods are evaluated on benchmark datasets, such as MNIST, CIFAR-10 or the JSB Chorales database, and also on time series modeling tasks. We report gains both regarding performance and entropy of the analyzed networks. We also made our code available as a GitHub repository.

Citation

Reizinger P., Gyires-Tóth B. (2019) Stochastic Weight Matrix-Based Regularization Methods for Deep Neural Networks. In: Nicosia G., Pardalos P., Umeton R., Giuffrida G., Sciacca V. (eds) Machine Learning, Optimization, and Data Science. LOD 2019. Lecture Notes in Computer Science, vol 11943. Springer, Cham. https://doi.org/10.1007/978-3-030-37599-7_5

@article{reizinger2019wmm,
   title={Stochastic Weight Matrix-Based Regularization Methods for Deep Neural Networks},
   ISSN={1611-3349},
   url={http://dx.doi.org/10.1007/978-3-030-37599-7_5},
   DOI={10.1007/978-3-030-37599-7_5},
   journal={Lecture Notes in Computer Science},
   publisher={Springer International Publishing},
   author={Reizinger, Patrik and Gyires-Tóth, Bálint},
   year={2019},
   pages={45–57}
}