A new deep neural network structure specially optimized for high inference speed on modern GPU. It uses full convolutions in low-level stage and depth-wises convolutions in high-level stages.
More details could be found in our arXiv preprint:
Ming Lin, Hesen Chen, Xiuyu Sun, Qi Qian, Hao Li, Rong Jin. Neural Architecture Design for GPU-Efficient Networks. arXiv:2006.14090 [cs.CV]. June 2020. [arXiv]
If you find it useful, please help us by citing this work as following:
@misc{lin2020neural,
title={Neural Architecture Design for GPU-Efficient Networks},
author={Ming Lin and Hesen Chen and Xiuyu Sun and Qi Qian and Hao Li and Rong Jin},
year={2020},
eprint={2006.14090},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
We provided three pre-trained models, GENet-large/normal/small.
- Install PyTorch and NVIDIA Apex (optional).
- Prepare ImageNet dataset. By default the data root directory is assumed to be ~/data/imagenet.
That is, ~/data/imagenet/train/ contains the training set and ~/data/imagenet/val contains the validation set.
- The ImageNet downloading script could be found here.
- Clone this repository:
git clone https://github.com/idstcv/GPU-Efficient-Networks.git
cd GPU-Efficient-Networks
- Download pre-trained parameters to ./GENet_params. [Google Drive download link]
python val.py --data ~/data/imagenet --arch GENet_large --params_dir ./GENet_params/ --use_apex
If you don't have Apex installed, turn-off the --use_apex option. Use '--arch GENet_large', '--arch GENet_normal', '--arch GENet_small' to choose different GENets.
Warning: Without Apex installed, the accuracies of the pre-trained models will be slightly different due to the difference in the FP16 quantization.
First, load pre-trained GENets:
import GENet
import torch
from torch import nn
arch = 'GENet_large'
if arch == 'GENet_large':
input_image_size = 256
model = GENet.genet_large(pretrained=True, root='./GENet_params/')
if arch == 'GENet_normal':
input_image_size = 192
model = GENet.genet_normal(pretrained=True, root='./GENet_params/')
if arch == 'GENet_small':
input_image_size = 192
model = GENet.genet_small(pretrained=True, root='./GENet_params/')
Note GENet-large/normal/small use different input image resolutions.
Then, create your own FC-head:
# number of classes in your target dataset
num_classes = 10
model.fc_linear = nn.Linear(model.last_channels, num_classes, bias=True)
If you use GPU 0 for fine-tuning:
gpu = 0
torch.cuda.set_device(gpu)
model = model.cuda(0)
To inference,
x = get_one_image_from_your_dataset(input_image_size=input_image_size)
x = x.cuda(gpu)
output = model(x)
Copyright (C) 2010-2020 Alibaba Group Holding Limited. Released under the Apache License.