This page provides instructions for optimizing the performance of models running on DLA.
Make sure you install the Python dependencies from requirements.txt.
Below steps were run with TensorRT 8.5 on an Orin L4T platform where trtexec was present in /usr/local/tensorrt/bin/trtexec.
Original location: https://zenodo.org/record/4735647/files/resnet50_v1.onnx
Mirror: https://web.archive.org/web/20221201211434/https://zenodo.org/record/4735647/files/resnet50_v1.onnx
- Download
resnet50_v1.onnxfrom above links - Run
python3 resnet50.pyon any platform (host or Orin target) - Copy the generated
resnet50_v1_prepared.onnxto your Orin target - Run the following command:
./trtexec --useDLACore=0 --int8 --memPoolSize=dlaSRAM:1MiB --inputIOFormats=int8:dla_hwc4 --outputIOFormats=int8:chw32 --onnx=resnet50_v1_prepared.onnx --shapes=input_tensor:0:2x3x224x224
Original location: https://zenodo.org/record/3228411/files/resnet34-ssd1200.onnx
- Download
resnet34-ssd1200.onnxfrom above links - Run
python3 ssd_resnet34.pyon any platform (host or Orin target) - Copy the generated
resnet34-ssd1200_prepared.onnxto your Orin target - Run the following command on your Orin target:
./trtexec --useDLACore=0 --int8 --memPoolSize=dlaSRAM:1MiB --inputIOFormats=int8:dla_hwc4 --outputIOFormats=int8:chw32 --onnx=dla/regression/testdb/resnet34-ssd1200_prepared.onnx
Original location: https://zenodo.org/record/4735652/files/ssd_mobilenet_v1_coco_2018_01_28.onnx
-
Download
ssd_mobilenet_v1_coco_2018_01_28.onnxfrom above links -
Run
python3 ssd_mobilenetv1.pyon any platform (host or Orin target) -
Copy the generated
ssd_mobilenet_v1_coco_2018_01_28_prepared.onnxto your Orin target -
Run the following command on your Orin target:
./trtexec --useDLACore=0 --int8 --memPoolSize=dlaSRAM:1MiB --inputIOFormats=int8:dla_hwc4 --outputIOFormats=int8:chw32 --onnx=ssd_mobilenet_v1_coco_2018_01_28_prepared.onnx --shapes=Preprocessor/sub:0:2x3x300x300