Skip to content

result mismatch between qdq-onnx and tensorrt engine. #4708

@zzqiuzz

Description

@zzqiuzz

Description

i tried to use tensort8.6.1 to convert a onnx model with qdq nodes, however, the output of the converted engine dismatched with that of original onnx.

Environment

TensorRT Version:8.6.1 GA

NVIDIA GPU: Geforce 3060

NVIDIA Driver Version:Driver Version: 535.230.02

CUDA Version: nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2022 NVIDIA Corporation
Built on Wed_Sep_21_10:33:58_PDT_2022
Cuda compilation tools, release 11.8, V11.8.89
Build cuda_11.8.r11.8/compiler.31833905_0

CUDNN Version: 8.6.0

Operating System: ubuntu20.04

Python Version (if applicable):3.9.23

Tensorflow Version (if applicable):

PyTorch Version (if applicable):

Baremetal or Container (if so, version):

Relevant Files

Steps To Reproduce

  1. trtexec --onnx=quant_base.onnx --saveEngine=quant_base.plan --dumpProfile=true --int8 --fp16 --verbose=true > log.log 2>&1
    quant_base.onnx
  2. python _infer.py
    i use randomly generated input to forward onnx and engine, compare their results by calculating cosine similarity.
Image

Metadata

Metadata

Assignees

No one assigned

    Labels

    Module:AccuracyOutput mismatch between TensorRT and other frameworks

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions