You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I've seen #869, but this appears to be a separate issue. Maybe @fxmarty or @JingyaHuang can help me?
I am able to get significant speedup with Whisper with CUDAExecutionProvider and wav2vec2 with TensorrtExecutionProvider, but Whisper with TensorrtExecutionProvider yields very poor results.
I'm using trt_engine_cache and warming up the model, but Whisper with TensorrtExecutionProvider consistently is over 2x slower than a vanilla pipeline.
Ex, running the script below yields me:
Benchmarking baseline pipeline...
Run 1: 2.48 seconds
Run 2: 1.87 seconds
Run 3: 1.85 seconds
Benchmarking TensorRT pipeline...
Run 1: 5.61 seconds
Run 2: 4.99 seconds
Run 3: 5.00 seconds
Information
The official example scripts
My own modified scripts
Tasks
An officially supported task in the examples folder (such as GLUE/SQuAD, ...)
System Info
Who can help?
I've seen #869, but this appears to be a separate issue. Maybe @fxmarty or @JingyaHuang can help me?
I am able to get significant speedup with Whisper with
CUDAExecutionProvider
and wav2vec2 withTensorrtExecutionProvider
, but Whisper withTensorrtExecutionProvider
yields very poor results.I'm using
trt_engine_cache
and warming up the model, but Whisper withTensorrtExecutionProvider
consistently is over 2x slower than a vanilla pipeline.Ex, running the script below yields me:
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction (minimal, reproducible, runnable)
Expected behavior
TensorrtExecutionProvider
yields a significant speedup.The text was updated successfully, but these errors were encountered: