Code Monkey home page Code Monkey logo

nanodet_tensorrt_int8's Introduction

环境配置

ubuntu:18.04

cuda:11.0

cudnn:8.0

tensorrt:7.2.16

OpenCV:3.4.2

cuda,cudnn,tensorrt和OpenCV安装包(编译好了,也可以自己从官网下载编译)可以从链接: https://pan.baidu.com/s/1dpMRyzLivnBAca2c_DIgGw 密码: 0rct

cuda安装

如果系统有安装驱动,运行如下命令卸载

sudo apt-get purge nvidia*

禁用nouveau,运行如下命令

sudo vim /etc/modprobe.d/blacklist.conf

在末尾添加

blacklist nouveau

然后执行

sudo update-initramfs -u

chmod +x cuda_11.0.2_450.51.05_linux.run

sudo ./cuda_11.0.2_450.51.05_linux.run

是否接受协议: accept

然后选择Install

最后回车

vim ~/.bashrc 添加如下内容:

export PATH=/usr/local/cuda-11.0/bin:$PATH

export LD_LIBRARY_PATH=/usr/local/cuda-11.0/lib64:$LD_LIBRARY_PATH

source .bashrc 激活环境

cudnn 安装

tar -xzvf cudnn-11.0-linux-x64-v8.0.4.30.tgz

cd cuda/include

sudo cp *.h /usr/local/cuda-11.0/include

cd cuda/lib64

sudo cp libcudnn* /usr/local/cuda-11.0/lib64

tensorrt及OpenCV安装

定位到用户根目录

tar -xzvf TensorRT-7.2.1.6.Ubuntu-18.04.x86_64-gnu.cuda-11.0.cudnn8.0.tar.gz

cd TensorRT-7.2.1.6/python,该目录有4个python版本的tensorrt安装包

sudo pip3 install tensorrt-7.2.1.6-cp37-none-linux_x86_64.whl(根据自己的python版本安装)

pip install pycuda 安装python版本的cuda

定位到用户根目录

tar -xzvf opencv-3.4.2.zip 以备推理调用

nanodet模型转换onnx

pip install onnx

pip install onnx-simplifier

git clone https://github.com/Wulingtian/nanodet.git

cd nanodet

cd config 配置模型文件(注意激活函数要换为relu!tensorrt支持relu量化),训练模型

定位到nanodet目录,进入tools目录,打开export.py文件,配置cfg_path model_path out_path三个参数

定位到nanodet目录,运行 python tools/export.py 得到转换后的onnx模型

python3 -m onnxsim onnx模型名称 nanodet-simple.onnx 得到最终简化后的onnx模型

onnx模型转换为 int8 tensorrt引擎

git clone https://github.com/Wulingtian/nanodet_tensorrt_int8_tools.git(求star)

cd nanodet_tensorrt_int8_tools

vim convert_trt_quant.py 修改如下参数

BATCH_SIZE 模型量化一次输入多少张图片

BATCH 模型量化次数

height width 输入图片宽和高

CALIB_IMG_DIR 训练图片路径,用于量化

onnx_model_path onnx模型路径

python convert_trt_quant.py 量化后的模型存到models_save目录下

tensorrt模型推理

git clone https://github.com/Wulingtian/nanodet_tensorrt_int8.git(求star)

cd nanodet_tensorrt_int8

vim CMakeLists.txt

修改USER_DIR参数为自己的用户根目录

vim nanodet_infer.cc 修改如下参数

output_name模型有一个输出

我们可以通过netron查看模型输出名

pip install netron 安装netron

vim netron_nanodet.py 把如下内容粘贴

    import netron

    netron.start('此处填充简化后的onnx模型路径', port=3344)

python netron_nanodet.py 即可查看 模型输出名

trt_model_path 量化的的tensorrt推理引擎(models_save目录下trt后缀的文件)

test_img 测试图片路径

INPUT_W INPUT_H 输入图片宽高

NUM_CLASS 训练的模型有多少类

NMS_THRESH nms阈值

CONF_THRESH 置信度阈值

参数配置完毕

mkdir build

cd build

cmake ..

make

./NanoDetEngine 输出平均推理时间,以及保存预测图片到当前目录下,至此,部署完成!

nanodet_tensorrt_int8's People

Contributors

wulingtian avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

nanodet_tensorrt_int8's Issues

输入层tensor的Shape疑问求解(cur_in_tensor_params[0].pShape)?

感谢题主的开源代码,受益匪浅!
我有两个疑问请教一下 :

1 输入层tensor的Shape, ( nanodet_infer.cc 的235行 ):

// 第一个输入tensor
cur_in_tensor_params[0].nDims = 4;
cur_in_tensor_params[0].type = DT_FLOAT;
cur_in_tensor_params[0].pShape[0] = 1; //batch size can't set to -1
cur_in_tensor_params[0].pShape[1] = 3;
cur_in_tensor_params[0].pShape[2] = INPUT_W;
cur_in_tensor_params[0].pShape[3] = INPUT_H;

输入的tensor的shape按道理是NCHW, 即上述应该是 pShape[2] = INPUT_H, pShape[3] = INPUT_W。不知道我的理解是否正确? 是因为nanodet的输入尺寸是320x320, 所以这里即使反了,运行和结果才没出现问题 ?

2 题主在另一个Issue (https://github.com/Wulingtian/nanodet_tensorrt_int8/issues/1) 中提供的 LibMyTtrEngine-trt721.so生成代码,与github上TensorRT的模型加载运行的代码(例如 https://hub.fastgit.org/linghu8812/tensorrt_inference/tree/master/nanodet )不太相同,但题主的nanodet推理的效果和精度有明显的提升! 请问题主做了哪些相关的优化?

谢谢。 期待题主的回复!

model intput

Thank your for open this project.
**.trt is int8 model, however input data is float. code

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.