Knowledge distillation using CIFAR 100
Paper : https://arxiv.org/abs/1503.02531
Resnet 20 (you can change with 50 or others)
MobileNet or CNN - LSTM
ImageNet Dataset : http://www.image-net.org/
Link = http://hpkim0512.blogspot.com/2017/12/sdf.html?m=1
Stop Training if overfitting(validation set accuracy doesn't increase) happens more than 50 epochs
Python 3.6 Tensorflow, keras
windows 10, ubuntu linux