NeurIPS 2020 [Paper]
Pipeline:
- Sparsity Training
- Pruning
- Fine-tuning
We test our code on Python 3.6. Our code is incompatible with Python 2.x.
Install packages:
pip install -r requirements.txtWe recommend to run the code on PyTorch 1.2 and CUDA 10.0. The project is incompatible with PyTorch <= 1.0.
See README in ./imagenet or ./cifar for guidelines on running experiments on ImageNet (ILSVRC-12) or CIFAR10/100 datasets.
We upload the the pruned checkpoints on OneDrive.
We introduce a novel pruning method in our paper (Fig. 2). We have implemented multiple pruning methods in our code (option --pruning-strategy).
grad: The method introduced in our paper (Section 3.3).fixed: Use a global pruning threshold for all layers (0.01 as default).percent: Determine the threshold by a global pruning percent (as Network Slimming).search: Deprecated. Not recommend to use.
original: There is no any sparse regularization on the loss function, i.e., baseline model.sr: Apply L1 regularization on the scaling factors, i.e., Network Slimming.zol: Polarization regularization. See equation 2 in the paper.
We build our code based on rethinking-network-pruning. We'd like to thank their contribution to the research on structured pruning.