Skip to content

Conversation

keshavv27
Copy link
Contributor

Description

  • Use total inference time instead of the submission time for output statistics calculation

Motivation and Context

  • The min, max, and other statistics reported for inference were using device submission time instead of the inference time.

@ishwar-raut1 @gaugarg-nv @thevishalagarwal @umangb-09 @gedoensmax

@chilo-ms
Copy link
Contributor

/azp run Linux QNN CI Pipeline,Win_TRT_Minimal_CUDA_Test_CI,Windows ARM64 QNN CI Pipeline,Windows GPU Doc Gen CI Pipeline,Windows x64 QNN CI Pipeline

Copy link

Azure Pipelines successfully started running 4 pipeline(s).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants