Comments (7)
Torch's MM based convolutions on CPU use a lot more memory, and the shapes probably are not as optimized for OpenBLAS-ARM (as it unfolds all mini-batches and does a single MM call, rather than doing per-batch unfold + gemm in caffe). I'd suggest trying out:
https://github.com/mvitez/OpenBLAS-conv
https://github.com/mvitez/thnets
This very old fork of torch also has optimized assembly NEON based convolutions in there, but only for 32-bit ARM: soumith/torch-android@af6dc1e
from convnet-benchmarks.
that being said, 20x seems hugely suspect, as they are both calling gemm.
from convnet-benchmarks.
@soumith Thanks for the pointers! I'll try out the codes you've linked and post back here. If it is indeed a difference in unfolding all batches vs. per-batch unfolds (in Caffe), then it should make a huge difference. Thanks!
from convnet-benchmarks.
@soumith I've validated AlexNet using thnets (https://github.com/mvitez/thnets) and the TX1's ARM A57 CPU is now within 18% of the Caffe implementation (4.54 FPS on thnets vs. 5.3 FPS on Caffe) for a batch_size = 4 on thnets. I attribute the speedup to assembly-level intrinsics + highly optimized openBLAS kernels for the ARM platform.
I couldn't verify your claim that Torch unfolds all batches and performs a single MM call (and that Caffe unfolds per batch and performs multiple MM calls. Running ~/tegrastats to monitor memory usage, it appears that Torch (for my original Torch benchmark) actually uses less memory than Caffe.
Anyways, you solved my problem :) thanks man.
from convnet-benchmarks.
Hi, these numbers seems very bad... Just if it helps let me say that I am developing my on toolkit for academic purposes:
https://github.com/RParedesPalacios/Layers
(i have still to upload src code)
And AlexNet with batch=100 and only forward (inference) it takes 2 secs approx. I use lowering and all the batch unfolded. I will try to upload the code to try it.
regards
from convnet-benchmarks.
@RParedesPalacios, are we talking about inference on the TX1's ARM CPU? If so, isn't 50 images per second unreasonable?
Say it's 720 MFLOP to perform a single forward pass for one image [1]. 50 images per second would roughly translate to 36 GFLOPS (720 MFLOP * 50 images/s). I suspect the peak performance of the TX1's ARM CPUs cannot surpass 10 GFLOPS even with NEON and multithreading enabled.
[1] https://groups.google.com/forum/#!topic/caffe-users/cUD3IF5NMOk
from convnet-benchmarks.
@carlodelmundo, hi i misunderstood the point.. I read, "... Intel Xeon E5-2637 CPU ..." and i thought that the following numbers refer to that.. but i read again and i understand that all the numbers refer to the ARM CPU!.
sorry for that ,regards!
from convnet-benchmarks.
Related Issues (20)
- CuDNN[R4]-fp16 (Torch) results
- Torch 7 HOT 5
- After FFT & Winograd, what next? HOT 3
- Number of kernels in alexnet_benmark HOT 1
- Use Tensorflow benchmark without GPU HOT 1
- Problem running on older GPUs HOT 2
- Add PyTorch Benchmarks HOT 5
- Enable XLA support for Tensorflow HOT 2
- Tensorflow benchmarks cause error when running run_forward_backward HOT 1
- Updating benchmarks for recent cuDNN v6 HOT 1
- Tesla GV100 results? HOT 1
- Issue running tf_cnn_benchmark on Xeon Phi
- Amd Vega results with MIOpen?
- Cudnn 7 support?
- Tensorflow benchmark files not updated after migration?
- this project has stop update?
- worse chainer convnet-benchmarks performance on cupy-2.0.0 as compared to cupy-1.0.0.1 HOT 2
- cltorch googlenet.lua: attempt to index global 'cudnn' (a nil value)
- convnet-benchmark is not working with tensorflow 1.8 on AMD or Nvidia cards HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from convnet-benchmarks.