You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Nov 17, 2023. It is now read-only.
I have tested the three cases many times as below:
When running train_mnist.py(GPU) on the Cmake built mxnet on Ubuntu, the average epoch time cost is
0.7s. But the Make built mxnet on Ubuntu takes only 0.42s.
When running train_mnist.py(GPU) on the Cmake built mxnet on Windows, the average epoch time cost is 0.822s, even longer than running the script on MKL. The CPU(MKL) mode is about 0.6s.
The Windows MKL built mxnet version is better than the Ubuntu MKL built one. Each epoch of the former takes about 0.62s and each epoch of the later takes above 0.8s.
It's very weird. I hope the performance differences can be eliminated. I suggest that mxnet should bring a standard benchmark tool and reference performance index to measure running time such as the 'make runtest' of Caffe.
I have tested the three cases many times as below:
When running train_mnist.py(GPU) on the Cmake built mxnet on Ubuntu, the average epoch time cost is
0.7s. But the Make built mxnet on Ubuntu takes only 0.42s.
When running train_mnist.py(GPU) on the Cmake built mxnet on Windows, the average epoch time cost is 0.822s, even longer than running the script on MKL. The CPU(MKL) mode is about 0.6s.
The Windows MKL built mxnet version is better than the Ubuntu MKL built one. Each epoch of the former takes about 0.62s and each epoch of the later takes above 0.8s.
It's very weird. I hope the performance differences can be eliminated. I suggest that mxnet should bring a standard benchmark tool and reference performance index to measure running time such as the 'make runtest' of Caffe.