You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I was testing several BLAS implementations to see the performance difference. I'm using the MNIST dataset as instructed in its tutorial but with max_iter set to 1000.
I just discovered that the training (using train_lenet.sh) is significantly slow compared to the master branch. I tested on two different machines. The results below are from an Intel Xeon W3530 Nehalem CPU. I'm training using CPU mode. Is this an expected slow-down caused by some implementation change?
I can't reproduce this and in fact my dev is faster than master according to caffe time. Are you sure there weren't interfering tasks increasing the CPU or disk load during the test for dev?
It could be the switch to the DataTransformer from inline data transformation in dev since you are running in purely in CPU. You could try your evaluation at aee9cd3 before it was merged in #954 to test if it makes a difference.
Hi,
I was testing several BLAS implementations to see the performance difference. I'm using the MNIST dataset as instructed in its tutorial but with
max_iter
set to1000
.I just discovered that the training (using
train_lenet.sh
) is significantly slow compared to the master branch. I tested on two different machines. The results below are from an Intel Xeon W3530 Nehalem CPU. I'm training using CPU mode. Is this an expected slow-down caused by some implementation change?The text was updated successfully, but these errors were encountered: