This repository contains the code for the paper:
Dongyoon Han*, Jiwhan Kim*, and Junmo Kim, "Deep Pyramidal Residual Networks", CVPR 2017 (* equal contribution).
Arxiv: https://arxiv.org/abs/1610.02915.
The code is based on Facebook's implementation of ResNet (https://github.com/facebook/fb.resnet.torch).
Please refer to this site for Caffe implementation of PyramidNet.
Deep convolutional neural networks (DCNNs) have shown remarkable performance in image classification tasks in recent years. Generally, deep neural network architectures are stacks consisting of a large number of convolution layers, and they perform downsampling along the spatial dimension via pooling to reduce memory usage. At the same time, the feature map dimension (i.e., the number of channels) is sharply increased at downsampling locations, which is essential to ensure effective performance because it increases the capability of high-level attributes. Moreover, this also applies to residual networks and is very closely related to their performance. In this research, instead of using downsampling to achieve a sharp increase at each residual unit, we gradually increase the feature map dimension at all the units to involve as many locations as possible. This is discussed in depth together with our new insights as it has proven to be an effective design to improve the generalization ability. Furthermore, we propose a novel residual unit capable of further improving the classification accuracy with our new network architecture. Experiments on benchmark CIFAR datasets have shown that our network architecture has a superior generalization ability compared to the original residual networks.
Figure 1: Schematic illustration of (a) basic residual units, (b) bottleneck, (c) wide residual units, and (d) our pyramidal residual units.
Figure 2: Visual illustrations of (a) additive PyramidNet, (b) multiplicative PyramidNet, and (c) comparison of (a) and (b).
- Install Torch (http://torch.ch) and ResNet (https://github.com/facebook/fb.resnet.torch).
- Add the files addpyramidnet.lua and mulpyramidnet.lua to the folder "models".
- Manually set the parameter "alpha" in the files addpyramidnet.lua and mulpyramidnet.lua (Line 28).
- Change the learning rate schedule in the file train.lua: "decay = epoch >= 122 and 2 or epoch >= 81 and 1 or 0" to "decay = epoch >= 225 and 2 or epoch >= 150 and 1 or 0".
- Train our PyramidNet, by running main.lua as below:
To train additive PyramidNet-164 (alpha=48) on CIFAR-10 dataset:
th main.lua -dataset cifar10 -depth 164 -nEpochs 300 -LR 0.1 -netType addpyramidnet -batchSize 128 -shareGradInput true
To train additive PyramidNet-164 (alpha=48) with 4 GPUs on CIFAR-100 dataset:
th main.lua -dataset cifar100 -depth 164 -nEpochs 300 -LR 0.5 -nGPU 4 -nThreads 8 -netType addpyramidNet -batchSize 128 -shareGradInput true
Top-1 error rates on CIFAR-10 and CIFAR-100 datasets. "alpha" denotes the widening factor; "add" and "mul" denote the results obtained with additive and multiplicative pyramidal networks, respectively.
Network | # of parameters | Output feat. dimension | CIFAR-10 | CIFAR-100 |
---|---|---|---|---|
PyramidNet-110 (mul), alpha=4.75 | 1.7M | 76 | 4.62 | 23.16 |
PyramidNet-110 (add), alpha=48 | 1.7M | 64 | 4.62 | 23.31 |
PyramidNet-110 (mul), alpha=8 | 3.8M | 128 | 4.50 | 20.94 |
PyramidNet-110 (add), alpha=84 | 3.8M | 100 | 4.27 | 20.21 |
PyramidNet-110 (mul), alpha=27 | 28.3M | 432 | 4.06 | 18.79 |
PyramidNet-110 (add), alpha=270 | 28.3M | 286 | 3.73 | 18.25 |
Top-1 error rates of our model with the bottleneck architecture on CIFAR-10 and CIFAR-100 datasets. We use the additive pyramidal networks.
Network | # of parameters | Output feat. dimension | CIFAR-10 | CIFAR-100 |
---|---|---|---|---|
PyramidNet-164 (add), alpha=48 | 1.7M | 256 | 4.21 | 19.52 |
PyramidNet-164 (add), alpha=84 | 3.8M | 400 | 3.96 | 18.32 |
PyramidNet-164 (add), alpha=270 | 27.0M | 1144 | 3.48 | 17.01 |
PyramidNet-200 (add), alpha=240 | 26.6M | 1024 | 3.44 | 16.51 |
PyramidNet-236 (add), alpha=220 | 26.8M | 944 | 3.40 | 16.37 |
PyramidNet-272 (add), alpha=200 | 26.0M | 864 | 3.31 | 16.35 |
Figure 3: Performance distribution according to number of parameters on CIFAR-10 (left) and CIFAR-100 (right).
Top-1 and Top-5 error rates of single-model, single-crop (224*224) on ImageNet dataset. We use the additive PyramidNet for our results.
Network | # of parameters | Output feat. dimension | Top-1 error | Top-5 error |
---|---|---|---|---|
PreResNet-200 | 64.5M | 2048 | 21.66 | 5.79 |
PyramidNet-200, alpha=300 | 62.1M | 1456 | 20.47 | 5.29 |
PyramidNet-200, alpha=450, Dropout (0.5) | 116.4M | 2056 | 20.11 | 5.43 |
Model files download: link
- The parameter "alpha" can only be changed in the files addpyramidnet.lua and mulpyramidnet.lua (Line 28).
- We recommend to use multi-GPU when training additive PyramidNet with alpha=270 or multiplicative PyramidNet with alpha=27. Otherwise you may get "out of memory" error.
- We are currently testing our code in the ImageNet dataset. We will upload the result when the training is completed.
07/17/2017:
- Caffe implementation of PyramidNet is released.
02/12/2017:
- Results of the bottleneck architecture on CIFAR datasets are updated.
01/23/2017:
- Added Imagenet pretrained models.
Jiwhan Kim ([email protected]), Dongyoon Han ([email protected]), Junmo Kim ([email protected])