Stochastic Weight Matrix-based Regularization Methods for Deep Neural Networks
Published in 5th International Conference on Learning, Optimization and Data Science, 2019
Recommended citation: Reizinger P., Gyires-Tóth B. (2019) "Stochastic Weight Matrix-Based Regularization Methods for Deep Neural Networks.." Springer LNCS 1. 11943. https://arxiv.org/pdf/1909.11977
Abstract
The aim of this paper is to introduce two widely applicable regularization methods based on the direct modification of weight matrices. The first method, Weight Reinitialization, utilizes a simplified Bayesian assumption with partially resetting a sparse subset of the parameters. The second one, Weight Shuffling, introduces an entropy- and weight distribution-invariant non-white noise to the parameters. The latter can also be interpreted as an ensemble approach. The proposed methods are evaluated on benchmark datasets, such as MNIST, CIFAR-10 or the JSB Chorales database, and also on time series modeling tasks. We report gains both regarding performance and entropy of the analyzed networks. We also made our code available as a GitHub repository.
Citation
Reizinger P., Gyires-Tóth B. (2019) Stochastic Weight Matrix-Based Regularization Methods for Deep Neural Networks. In: Nicosia G., Pardalos P., Umeton R., Giuffrida G., Sciacca V. (eds) Machine Learning, Optimization, and Data Science. LOD 2019. Lecture Notes in Computer Science, vol 11943. Springer, Cham. https://doi.org/10.1007/978-3-030-37599-7_5
@article{reizinger2019wmm,
title={Stochastic Weight Matrix-Based Regularization Methods for Deep Neural Networks},
ISSN={1611-3349},
url={http://dx.doi.org/10.1007/978-3-030-37599-7_5},
DOI={10.1007/978-3-030-37599-7_5},
journal={Lecture Notes in Computer Science},
publisher={Springer International Publishing},
author={Reizinger, Patrik and Gyires-Tóth, Bálint},
year={2019},
pages={45–57}
}