Accelerate mobileNet-ssd with tensorRT
**when I run the command: ./mobileNet then I get the following errors:** `attempting to open cache file ../../model/MobileNetSSD_deploy.caffemodel.1.tensorcache loading network profile from cache... createInference The engine plan file is incompatible with this version of TensorRT, expecting 5.0.6.3got -2100382470.0.2.1852793695, please rebuild. createInference_end Bindings after deserializing: Segmentation fault (core dumped) ` But the requirement lists shows TensorRT4 is OK. **It looks like I should install tensorRT 5 according to the error, but when I run dpkg -l | grep TensorRT it shows:** `...~/MobileNet-SSD-TensorRT$ dpkg -l | grep TensorRT ii graphsurgeon-tf 5.0.6-1+cuda10.0 arm64 GraphSurgeon for TensorRT package ii libnvinfer-dev 5.0.6-1+cuda10.0 arm64 TensorRT development libraries and headers ii libnvinfer-samples 5.0.6-1+cuda10.0 all TensorRT samples and documentation ii libnvinfer5 5.0.6-1+cuda10.0 arm64 TensorRT runtime libraries ii python-libnvinfer 5.0.6-1+cuda10.0 arm64 Python bindings for TensorRT ii python-libnvinfer-dev 5.0.6-1+cuda10.0 arm64 Python development package for TensorRT ii python3-libnvinfer 5.0.6-1+cuda10.0 arm64 Python 3 bindings for TensorRT ii python3-libnvinfer-dev 5.0.6-1+cuda10.0 arm64 Python 3 development package for TensorRT ii tensorrt 5.0.6.3-1+cuda10.0 arm64 Meta package of TensorRT ii uff-converter-tf 5.0.6-1+cuda10.0 arm64 UFF converter for TensorRT package` **My TensorRT is 5! Someone can help me out with it? Thanks.**
This issue appears to be discussing a feature request or bug report related to the repository. Based on the content, it seems to be still under discussion. The issue was opened by masteryun and has received 7 comments.