This tool aims to load caffe prototxt and weights directly in pytorch without explicitly converting model from caffe to pytorch.
from caffenet import *
def load_image(imgfile):
import caffe
image = caffe.io.load_image(imgfile)
transformer = caffe.io.Transformer({'data': (1, 3, args.height, args.width)})
transformer.set_transpose('data', (2, 0, 1))
transformer.set_mean('data', np.array([args.meanB, args.meanG, args.meanR]))
transformer.set_raw_scale('data', args.scale)
transformer.set_channel_swap('data', (2, 1, 0))
image = transformer.preprocess('data', image)
image = image.reshape(1, 3, args.height, args.width)
return image
def forward_pytorch(protofile, weightfile, image):
net = CaffeNet(protofile)
print(net)
net.load_weights(weightfile)
net.eval()
image = torch.from_numpy(image)
image = Variable(image)
blobs = net(image)
return blobs, net.models
imgfile = 'data/cat.jpg'
protofile = 'resnet50/deploy.prototxt'
weightfile = 'resnet50/resnet50.caffemodel'
image = load_image(imgfile)
pytorch_blobs, pytorch_models = forward_pytorch(protofile, weightfile, image)
- support forward classification networks: AlexNet, VGGNet, GoogleNet, ResNet, ResNeXt, DenseNet
- support forward detection networks: SSD300, S3FD, FPN
Each layer in caffe will have a corresponding layer in pytorch.
- Convolution
- InnerProduct
- BatchNorm
- Scale
- ReLU
- Pooling
- Reshape
- Softmax
- SoftmaxWithLoss.
- Dropout
- Eltwise
- Normalize
- Permute
- Flatten
- Slice
- Concat
- PriorBox
- LRN : gpu version is ok, cpu version produce big difference
- DetectionOutput: support batchsize=1, num_classes=1 forward
- Crop
- Deconvolution
The script verify.py can verify the parameter and output difference between caffe and pytorch.
python verify.py --protofile resnet50/deploy.prototxt --weightfile resnet50/resnet50.caffemodel --imgfile data/cat.jpg --meanB 104.01 --meanG 116.67 --meanR 122.68 --scale 255 --height 224 --width 224 --synset_words data/synset_words.txt --cuda
Note:
- synset_words.txt contains class information, each line represents the description of a class.
- resnet50 is downloaded from BaiduYun
Outputs:
------------ Parameter Difference ------------
conv1 weight_diff: 0.000000 bias_diff: 0.000000
bn_conv1 running_mean_diff: 0.000000 running_var_diff: 0.000000
scale_conv1 weight_diff: 0.000000 bias_diff: 0.000000
res2a_branch1 weight_diff: 0.000000
bn2a_branch1 running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2a_branch1 weight_diff: 0.000000 bias_diff: 0.000000
res2a_branch2a weight_diff: 0.000000
bn2a_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2a_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res2a_branch2b weight_diff: 0.000000
bn2a_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2a_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res2a_branch2c weight_diff: 0.000000
bn2a_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2a_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res2b_branch2a weight_diff: 0.000000
bn2b_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2b_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res2b_branch2b weight_diff: 0.000000
bn2b_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2b_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res2b_branch2c weight_diff: 0.000000
bn2b_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2b_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res2c_branch2a weight_diff: 0.000000
bn2c_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2c_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res2c_branch2b weight_diff: 0.000000
bn2c_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2c_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res2c_branch2c weight_diff: 0.000000
bn2c_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale2c_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res3a_branch1 weight_diff: 0.000000
bn3a_branch1 running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3a_branch1 weight_diff: 0.000000 bias_diff: 0.000000
res3a_branch2a weight_diff: 0.000000
bn3a_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3a_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res3a_branch2b weight_diff: 0.000000
bn3a_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3a_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res3a_branch2c weight_diff: 0.000000
bn3a_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3a_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res3b_branch2a weight_diff: 0.000000
bn3b_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3b_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res3b_branch2b weight_diff: 0.000000
bn3b_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3b_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res3b_branch2c weight_diff: 0.000000
bn3b_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3b_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res3c_branch2a weight_diff: 0.000000
bn3c_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3c_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res3c_branch2b weight_diff: 0.000000
bn3c_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3c_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res3c_branch2c weight_diff: 0.000000
bn3c_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3c_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res3d_branch2a weight_diff: 0.000000
bn3d_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3d_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res3d_branch2b weight_diff: 0.000000
bn3d_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3d_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res3d_branch2c weight_diff: 0.000000
bn3d_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale3d_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res4a_branch1 weight_diff: 0.000000
bn4a_branch1 running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4a_branch1 weight_diff: 0.000000 bias_diff: 0.000000
res4a_branch2a weight_diff: 0.000000
bn4a_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4a_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res4a_branch2b weight_diff: 0.000000
bn4a_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4a_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res4a_branch2c weight_diff: 0.000000
bn4a_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4a_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res4b_branch2a weight_diff: 0.000000
bn4b_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4b_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res4b_branch2b weight_diff: 0.000000
bn4b_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4b_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res4b_branch2c weight_diff: 0.000000
bn4b_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4b_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res4c_branch2a weight_diff: 0.000000
bn4c_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4c_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res4c_branch2b weight_diff: 0.000000
bn4c_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4c_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res4c_branch2c weight_diff: 0.000000
bn4c_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4c_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res4d_branch2a weight_diff: 0.000000
bn4d_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4d_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res4d_branch2b weight_diff: 0.000000
bn4d_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4d_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res4d_branch2c weight_diff: 0.000000
bn4d_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4d_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res4e_branch2a weight_diff: 0.000000
bn4e_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4e_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res4e_branch2b weight_diff: 0.000000
bn4e_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4e_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res4e_branch2c weight_diff: 0.000000
bn4e_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4e_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res4f_branch2a weight_diff: 0.000000
bn4f_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4f_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res4f_branch2b weight_diff: 0.000000
bn4f_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4f_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res4f_branch2c weight_diff: 0.000000
bn4f_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale4f_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res5a_branch1 weight_diff: 0.000000
bn5a_branch1 running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5a_branch1 weight_diff: 0.000000 bias_diff: 0.000000
res5a_branch2a weight_diff: 0.000000
bn5a_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5a_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res5a_branch2b weight_diff: 0.000000
bn5a_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5a_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res5a_branch2c weight_diff: 0.000000
bn5a_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5a_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res5b_branch2a weight_diff: 0.000000
bn5b_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5b_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res5b_branch2b weight_diff: 0.000000
bn5b_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5b_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res5b_branch2c weight_diff: 0.000000
bn5b_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5b_branch2c weight_diff: 0.000000 bias_diff: 0.000000
res5c_branch2a weight_diff: 0.000000
bn5c_branch2a running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5c_branch2a weight_diff: 0.000000 bias_diff: 0.000000
res5c_branch2b weight_diff: 0.000000
bn5c_branch2b running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5c_branch2b weight_diff: 0.000000 bias_diff: 0.000000
res5c_branch2c weight_diff: 0.000000
bn5c_branch2c running_mean_diff: 0.000000 running_var_diff: 0.000000
scale5c_branch2c weight_diff: 0.000000 bias_diff: 0.000000
------------ Output Difference ------------
data output_diff: 0.000000
conv1 output_diff: 0.000000
pool1 output_diff: 0.000000
res2a_branch1 output_diff: 0.000000
res2a_branch2a output_diff: 0.000000
res2a_branch2b output_diff: 0.000000
res2a_branch2c output_diff: 0.000000
res2a output_diff: 0.000000
res2b_branch2a output_diff: 0.000000
res2b_branch2b output_diff: 0.000000
res2b_branch2c output_diff: 0.000001
res2b output_diff: 0.000000
res2c_branch2a output_diff: 0.000000
res2c_branch2b output_diff: 0.000001
res2c_branch2c output_diff: 0.000001
res2c output_diff: 0.000001
res3a_branch1 output_diff: 0.000001
res3a_branch2a output_diff: 0.000000
res3a_branch2b output_diff: 0.000000
res3a_branch2c output_diff: 0.000001
res3a output_diff: 0.000000
res3b_branch2a output_diff: 0.000000
res3b_branch2b output_diff: 0.000000
res3b_branch2c output_diff: 0.000000
res3b output_diff: 0.000000
res3c_branch2a output_diff: 0.000000
res3c_branch2b output_diff: 0.000000
res3c_branch2c output_diff: 0.000000
res3c output_diff: 0.000000
res3d_branch2a output_diff: 0.000000
res3d_branch2b output_diff: 0.000000
res3d_branch2c output_diff: 0.000001
res3d output_diff: 0.000001
res4a_branch1 output_diff: 0.000001
res4a_branch2a output_diff: 0.000000
res4a_branch2b output_diff: 0.000000
res4a_branch2c output_diff: 0.000001
res4a output_diff: 0.000001
res4b_branch2a output_diff: 0.000000
res4b_branch2b output_diff: 0.000000
res4b_branch2c output_diff: 0.000001
res4b output_diff: 0.000000
res4c_branch2a output_diff: 0.000000
res4c_branch2b output_diff: 0.000000
res4c_branch2c output_diff: 0.000001
res4c output_diff: 0.000000
res4d_branch2a output_diff: 0.000000
res4d_branch2b output_diff: 0.000000
res4d_branch2c output_diff: 0.000001
res4d output_diff: 0.000000
res4e_branch2a output_diff: 0.000000
res4e_branch2b output_diff: 0.000000
res4e_branch2c output_diff: 0.000001
res4e output_diff: 0.000000
res4f_branch2a output_diff: 0.000000
res4f_branch2b output_diff: 0.000000
res4f_branch2c output_diff: 0.000001
res4f output_diff: 0.000000
res5a_branch1 output_diff: 0.000002
res5a_branch2a output_diff: 0.000000
res5a_branch2b output_diff: 0.000000
res5a_branch2c output_diff: 0.000001
res5a output_diff: 0.000001
res5b_branch2a output_diff: 0.000000
res5b_branch2b output_diff: 0.000000
res5b_branch2c output_diff: 0.000001
res5b output_diff: 0.000001
res5c_branch2a output_diff: 0.000000
res5c_branch2b output_diff: 0.000000
res5c_branch2c output_diff: 0.000002
res5c output_diff: 0.000001
pool5 output_diff: 0.000000
fc1000 output_diff: 0.000001
prob output_diff: 0.000000
------------ Classification ------------
pytorch classification top1: 0.193016 n02113023 Pembroke, Pembroke Welsh corgi
caffe classification top1: 0.193018 n02113023 Pembroke, Pembroke Welsh corgi