内容简介:Simpfly implementation of Quantization Aware Training[1][2] with MXNet-scala module.Tested on Ubuntu 14.041, compile MXNet with CUDA, then compile the scala-pkg,doc:
MXNET-Scala TrainQuantization
Simpfly implementation of Quantization Aware Training[1][2] with MXNet-scala module.
Setup
Tested on Ubuntu 14.04
Requirements
- sbt 0.13 http://www.scala-sbt.org/
- Mxnet v1.4 https://github.com/dmlc/mxnet
Build steps
1, compile MXNet with CUDA, then compile the scala-pkg,doc: https://github.com/dmlc/mxnet/tree/master/scala-package
2, under the Mxnet-Scala/TrainQuantization folder:
mkdir lib; ln -s $MXNET_HOME/scala-package/assembly/linux-x86_64-gpu/target/mxnet-full_2.11-linux-x86_64-gpu-1.5.0-SNAPSHOT.jar lib
3, run sbt
and then compile the project
Train vgg on Cifar10
Using the script train_vgg16_cifar10.sh
under the scripts folder to train vgg from scratch on Cifar10:
FINETUNE_MODEL_EPOCH=-1 FINETUNE_MODEL_PREFIX=$ROOT/models/
Or you can finetune with the provided pretrain model:
FINETUNE_MODEL_EPOCH=46 FINETUNE_MODEL_PREFIX=$ROOT/models/cifar10_vgg16_acc_0.8772035
I did not use any data augmentation and carefully tune the hyper-parameters during training, the best accuracy I got was 0.877, worse than the best accracy 0.93 reported on Cifar10.
Train vgg with fake quantization on Cifar10
Using the script train_quantize_vgg16_cifar10.sh
under the scripts folder to train vgg with fake quantization on Cifar10,
you must provide the pretrained model:
FINETUNE_MODEL_EPOCH=46 FINETUNE_MODEL_PREFIX=$ROOT/models/cifar10_vgg16_acc_0.8772035
If everything goes right, you should get almost the same accuray with pretrained model after serveral epoch.
Test vgg with simulated quantization on Cifar10
Using the script test_quantize_vgg16_cifar10.sh
under the scripts folder to test pretrained fake quantization vgg with simulated quantization on Cifar10, you must provide the pretrained model:
FINETUNE_MODEL_EPOCH=57 FINETUNE_MODEL_PREFIX=$ROOT/models/cifar10_quantize_vgg16_acc_0.877504
Warning
Currently there is memory leak some where in the code, but I can't figure out the reason. You will see the memory usage keep increasing when you run the tranining script. So remenber to stop the traning script when memory usage is too high, and you can resume the training process with saved model previously.
Reference
[1] Quantizing deep convolutional networks for efficient inference: A whitepaper. https://arxiv.org/pdf/1806.08342.pdf
[2] Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference. https://arxiv.org/pdf/1712.05877.pdf
以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持 码农网
猜你喜欢:- Pytorch实现卷积神经网络训练量化(QAT)
- TensorFlow 模型优化工具包 — 训练后整型量化
- 【邢不行|量化小讲堂系列05-Python量化入门】计算创业板平均市盈率
- 量化交易的尬舞
- 抄底的艺术:量化交易之路
- 用R语言开始量化投资
本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们。
解放战争(上)(1945年8月—1948年9月)
王树增 / 人民文学出版社 / 2009-8 / 60.00
《解放战争》为王树增非虚构文学著述中规模最大的作品。武器简陋、兵力不足的军队对抗拥有现代武器装备的兵力庞大的军队,数量不多、面积有限的解放区最终扩展成为九百六十万平方公里的共和国,解放战争在短短四年时间里演绎的是人类历史上的战争传奇。国际风云,政治智慧,时事洞察,军事谋略,军队意志,作战才能,作品具有宏阔的视角和入微的体察,包含着惊心动魄的人生沉浮和变幻莫测的战场胜负,尽展中国历史上规模最大的一场......一起来看看 《解放战争(上)(1945年8月—1948年9月)》 这本书的介绍吧!