Comments (6)
AVX2 instructions are used here to accelerate the Kmeans processing. What's the version of your gcc? Does it build if you remove "-mavx2" compiler option?
from caffemodelcompression.
I got the issue done using my other computer. However, it gave me an error;
ImportError: No module named rpn.proposal_layer
Traceback (most recent call last):
File "caffemodel_compress.py", line 96, in
caffe_model_compress(prototxt, caffemodel, output, 6, 2)
File "caffemodel_compress.py", line 13, in caffe_model_compress
net = caffe.Net(model, caffe.TEST);
SystemError: NULL result without error in PyObject_Call
In the mean time, I want to ask you @yuanyuanli85 that if the compression is beneficial for making a net faster. For instance I got 500 ms detection time on Faster RCNN on Jetson TX1 module. Would the compression help to make it faster?
from caffemodelcompression.
Hi @eonurk, I got similar questions regarding the speed-up effect of such kind of compression techniques. Did you have any progress or rough result? Personally I wonder if it can help to speed up the convolutional neural networks with only conv and no fc layers. Thanks a lot!
from caffemodelcompression.
Actually I did not have a time to strugling with this problem. But in the end, it would be very beneficial for many people to know whether compression makes it faster or not. So please let me know if you find something about the issue:)
from caffemodelcompression.
Model compression reduces the computation a lot, ie. 32-bit float ->16-bit half-float. To run the compressed network, you need a hardware to support 16-bit half-float computation and a software framework. As I know, Nvidia P4/P40 has the support for half-float and int8. if we compress the weights to less bits, say 4 bits, we probably need a specific hardware, asic or fpga to run it.
from caffemodelcompression.
@yuanyuanli85 @eonurk Thanks for replying. It is that with specific hardware and smaller data types, the reduced computation is huge. How about sticking to the 32-bit float? From what I am understanding, the saved computation mainly comes from pruning stage and the paper only benchmarks the fc layers in terms of speedup. I think the speed-up requires a really sparse weight matrix to work with cuSPARSE (as they did with fc layers and more than 95% of the parameters are pruned!) For the convolutional layers, the sparsity is only around 40%-60%. So we may only expect little speed-up or even slower computation if cuSPARSE is used? Thanks
from caffemodelcompression.
Related Issues (17)
- compress the alexnet and decompress it give wrong prediction sometimes
- can you provide c++ interface HOT 1
- some trouble about compressed network
- usage of compressed model
- Not Speed up in pvanet?
- 压缩后测试问题 HOT 4
- The meaning of the parameter "newlabel"
- illegal instruction (core dump)——compressing layer conv1
- How to use .npz file to test?
- Compression Result HOT 8
- Compress Layers whitout bias HOT 3
- How to implement acceleration in the paper? HOT 1
- Use case Instructions
- OverflowError: long int too large to convert to int HOT 1
- illegal instruction (core dump) HOT 2
- help HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from caffemodelcompression.