Computer Vision models (include classification models and detection models) implemented by PyTorch
Architecure
pip install -r requirements.txt
simply run the cmd for the training:
## 1 GPU for vgg
CUDA_VISIBLE_DEVICES=0 python -u train_classifier.py --work-path ./experiments/vgg/cifar10
## resume from ckpt
CUDA_VISIBLE_DEVICES=0 python -u train_classifier.py --work-path ./experiments/vgg/cifar10 --resume
## 2 GPUs for inception_v1
CUDA_VISIBLE_DEVICES=0,1 python -u train_classifier.py --work-path ./experiments/inception_v1/cifar10
## 4 GPUs for yolo_v2
CUDA_VISIBLE_DEVICES=0,1,2,3 python -u train_detector.py --work-path ./experience/yolo_v2/voc2007
We use yaml file config.yaml
to save the parameters, check any files in ./experimets
for more details.
You can see the training curve via tensorboard, tensorboard --logdir path-to-event --port your-port
. (Not verified)
The training log will be dumped via logging, check log.txt
in your work path.
architecture | params | batch size | epoch | C10 test acc (%) | C100 test acc (%) |
---|---|---|---|---|---|
alexnet | 24.7M | 128 | 250 | 82.70 | - |
inception_v1 | 7.2M | 128 | 250 | 91.52 | - |
vgg19_bn | 38.9M | 128 | 250 | 93.37 | - |
resnet | 1.7M | 128 | 250 | 90.47 | - |
mobilenet_v1 | 3.2M | 128 | 200 | 86.01 | - |