Skip to content

Tritonserver serving Tensorrt model with cuda graph results in weird unconsistent outputs. #4651

@WingEdge777

Description

@WingEdge777

Hi TensorRT team.

I'm not sure which component should be blamed for this problem exactly. So I filed a ticket here as well as triton-inference-server/server#8550.

In conclusion, when we send requests sequentially to TritonServer, in the AAAAABBBBBAAAABBBB pattern. In every A(B)‘s requesting round, the first few A(B) requests probably return the last period round B(A)'s results, which is absurdly wrong for the current input.

Details can be found in the issue I posted above.

Thanks for your product. Looking forward to your response.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions