Skip to content

Model inference performance degradation when converting to onnx and tensorrt engine. #1124

@alexgrabit

Description

@alexgrabit

Hi everybody,

Problem description
We converted a EfficientnetV2 trained model to ONNX format and use it to inference some test images. Results are pretty much worse than the inference using .pth.tar file directly (they have no sense for us). Results are similar when inferencing with TensorRT engine converted from ONNX model, so we suppose the problem is in .pth.tar to ONNX conversion.

Does anybody know what could we be doing wrong?

Thanks in advance.

Metadata

Metadata

Assignees

Labels

bugSomething isn't working

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions