guides/triton-inference-server/ #8241
Replies: 6 comments 23 replies
-
i am following the same approach and able to complete inference with above code, however i want to run inference with different model size (i.e. 1024 or 800 or 1088), as i want model to be more accuratly work with certain size of images the config.pbxt file i am using are also working fine |
Beta Was this translation helpful? Give feedback.
-
Hi , I am following the same thing Triton inference server using ultralytics . But what I understood is the ultralytics library is typically used to run inference locally rather than interacting with a remote server like Triton. For running inference on a model hosted on Triton, we need to use a client that can communicate with Triton's REST or GRPC APIs. I am using yolov8n model and deployed it to triton and this is my config file name: Is there any script to perform inference using the tritonclient library? |
Beta Was this translation helpful? Give feedback.
-
I followed the outline in this post to successfully infer on Triton using the onnx model as instructed. However, the resulting images are labeling with class numbers instead of the names. I can easily translate the numbers to names after the fact, but is there a way to insert the names ahead of inference so images are labeled with names? |
Beta Was this translation helpful? Give feedback.
-
Hi, I am following the same thing Triton inference server using ultralytics, and using wrk to request the server. However, the GPU utilization keeps low(using 319M, 8G in total) when running inference, is there any advice to improve the GPU utilization? thanks. |
Beta Was this translation helpful? Give feedback.
-
I followed the steps and I am encountering this error: can someone help? |
Beta Was this translation helpful? Give feedback.
-
hello, i am trying to use .engine (tensorrt) format when using triton inference server but i get error, because triton does not recognize the backend needed for .engine file. |
Beta Was this translation helpful? Give feedback.
-
guides/triton-inference-server/
A step-by-step guide on integrating Ultralytics YOLOv8 with Triton Inference Server for scalable and high-performance deep learning inference deployments.
https://docs.ultralytics.com/guides/triton-inference-server/
Beta Was this translation helpful? Give feedback.
All reactions