-
Notifications
You must be signed in to change notification settings - Fork 172
Walkthrough: AlexNet
Contents:
- About Minerva owl.net
- About ImageNet
- About AlexNet
- Training AlexNet using Minerva
- Multi-view classification using AlexNet
- Using AlexNet to extract feature
owl.net is a DNN training framework build on Minerva's python interface owl. The main purposes of this package are: 1) Providing a simple way for Minerva users to train deep neural network for computer vision problems. 2) Providing a prototype about how to build user applications utilizing the advantages of Minerva.
We borrow Caffe's well-defined network and solver configure file format but the execution is conducted in Minerva engine. It's a showcase of Minerva's flexibile interface (building Caffe's main functionality in several hundreds of lines) and computation efficiency (Multi-GPU training).
If you are not familiar with ImageNet Large Scale Visual Recognition Challenge, please see here. The classification task contains 1.28 million images belong to 1000 classes.
To make IO efficient, we recommend transfer the original image into LMDB after you download the dataset. We could use the tool provided by Caffe to do the convertion. After converting the images, we need to compute the mean value of each pixel among the dataset. When training, mean values are subtracted from the image to produce a zero-mean input. Mean_file for ILSVRC12 can be downloaded by the script provided by Caffe.
In ILSVRC2012, AlexNet was proposed. It's the winning model of ILSVRC2012 classification task and it achieved a large accuracy margin compared with the non-DNN methods. It contains 5 convolutional layers and 3 fully-connected layers. During training, some randomness is introduced in the data augmentation process and dropout layer. Those details are shown below and should be defined in the configure file provided by Caffe. Note that currently we don't support convolutions with more than one group, for one GPU released recently has enough RAM to hold the whole model.
Layer | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 |
---|---|---|---|---|---|---|---|---|
Type | conv+max+norm | conv+max+norm | conv | conv | conv+max | full | full | full |
Channels | 96 | 256 | 384 | 384 | 256 | 4096 | 4096 | 1000 |
Filter Size | 11*11 | 5*5 | 3*3 | 3*3 | 3*3 | - | - | - |
Convolution Stride | 4*4 | 1*1 | 1*1 | 1*1 | 1*1 | - | - | - |
Pooling Size | 3*3 | 3*3 | - | - | 3*3 | - | - | - |
Pooling Stride | 2*2 | 2*2 | - | - | 2*2 | - | - | - |
Padding Size | 2*2 | 1*1 | 1*1 | 1*1 | 1*1 | - | - | - |
We implemented the DNN training logic in trainer.py. The main body of training code and updating code is below, which explicitly show how to train and update using multiple-gpu:
for iteridx in range(s.snapshot * s.owl_net.solver.snapshot, s.owl_net.solver.max_iter):
# train on multi-gpu
for gpuid in range(s.num_gpu):
owl.set_device(s.gpu[gpuid])
s.owl_net.forward('TRAIN')
s.owl_net.backward('TRAIN')
for wid in wunits:
wgrad[gpuid].append(s.owl_net.units[wid].weightgrad)
bgrad[gpuid].append(s.owl_net.units[wid].biasgrad)
# weight update
for i in range(len(wunits)):
wid = wunits[i]
upd_gpu = i * s.num_gpu / len(wunits)
owl.set_device(s.gpu[upd_gpu])
for gid in range(s.num_gpu):
if gid == upd_gpu:
continue
wgrad[upd_gpu][i] += wgrad[gid][i]
bgrad[upd_gpu][i] += bgrad[gid][i]
s.owl_net.units[wid].weightgrad = wgrad[upd_gpu][i]
s.owl_net.units[wid].biasgrad = bgrad[upd_gpu][i]
s.owl_net.update(wid)
The training procedure is controlled by solver file. The information need to be defined through solver are:
- network configuration file
- snapshot saving directory
- max iteration
- testing interval
- test iteration
- snapshot saving interval
- learning rate tuning strategy
- momentum
- weight decay
AlexNet usually need traverse the training set 70-90 times before converging and the learning rate should be tune smaller several times. The standard solver for AlexNet could be found here.
User could use following command to start training given Caffe's solver under the scripts folder.
./net_trainer.py <solver_file> [--snapshot SNAPSHOT] [-n NUM_GPU]
solver_file is the file name in Caffe's solver format. SNAPSHOT is the index of the snapshot to start with (default: 0). NUM_GPU is the number of gpu to use.
If we set NUM_GPU greater than 1, our code will slice a training batch into NUM_GPU pieces and FF/BP in parallel. The update is executed synchronously, so the training result using one gpu or NUM_GPU gpus will be the same.
The SNAPSHOT parameter will guide the system find the saved model under "snapshot saving directory" with that index. If the model can be found and the weight dimension is matched with the network configure file, OwlNet will load the model and continue training. Otherwise, it will initialize the weight according to the weight_filler parameter in configure file.
An example of call:
./net_trainer.py /path/to/solver_file --snapshot=0 -n=4
In multi-view classification, each image will be cropped out 5 patches located on left-upper, left-down, right-upper, right-down, center, and also generate the horizontal flipped version of them. Our data provider will handle this augmentation for you when you indicate the phase "MULTI_VIEW" in network configure file.
include {
phase: MULTI_VIEW
}
Each data batch will be turned to ten batches with different view. The final prediction will be made after we sum up the softmax value of different views. The main body of the code is below:
for testiteridx in range(s.owl_net.solver.test_iter[0]):
for i in range(10):
s.owl_net.forward('MULTI_VIEW')
if i == 0:
softmax_val = loss_unit.ff_y
batch_size = softmax_val.shape[1]
softmax_label = loss_unit.y
else:
softmax_val = softmax_val + loss_unit.ff_y
test_num += batch_size
predict = softmax_val.argmax(0)
truth = softmax_label.argmax(0)
correct = (predict - truth).count_zero()
acc_num += correct
Use following command to perform multi-view testing on the given trained network
./multiview_test.py <solver_file> <snapshot> [-g GPU_IDX]