This is the official source code for the paper: Improving the Trainability of Deep Neural Networks through Layerwise Batch-Entropy Regularization.
Bibtex Entry:
@article{
peer2022improving,
title={Improving the Trainability of Deep Neural Networks through Layerwise Batch-Entropy Regularization},
author={David Peer and Bart Keulen and Sebastian Stabinger and Justus Piater and Antonio Rodriguez-Sanchez},
journal={Transactions on Machine Learning Research},
year={2022},
url={https://openreview.net/forum?id=LJohl5DnZf},
note={}
}
To setup the environment simply run the setup.sh
script. It creates a virtual env. and additionally installs all the requirements needed to run the experiments as provided in the paper.
Every experiment is self-contained i.e. can be used as a code base for future work. In case we executed a hyperparameter search, we also provide the sweep files (sweep.yaml
) which contain precise hyperparameter setups that were used. Otherwise, a run.sh
script is provided. For further information how to run a sweep together with an agent we politely refer to the official wandb documentation: https://docs.wandb.ai/guides/sweeps