You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have noticed that the several graph optimizations like operator fusions, etc are occured by default while inferencing in TensorRT.
Is there any way to disable them or does TRT have levels of graph ops just as in ONNXRuntime?
Because I want to compare the performaces between before opt and after opt, for ex.
The text was updated successfully, but these errors were encountered:
@banu-narayana-dnjp Unfortunately, TensorRT API doesn't support disabling fusions at the moment.
However, you can experiment with the builder optimization level to achieve different fusion patterns. Please refer to doc on optimization level for usage
Thank you @asfiyab-nvidia and @lix19937 for your clarifications. Just one question, does MARK_ALL with the polygraphy tool result in any change in the final score(considering it prevents the fusions)?
I have noticed that the several graph optimizations like operator fusions, etc are occured by default while inferencing in TensorRT.
Is there any way to disable them or does TRT have levels of graph ops just as in ONNXRuntime?
Because I want to compare the performaces between before opt and after opt, for ex.
The text was updated successfully, but these errors were encountered: