-
Notifications
You must be signed in to change notification settings - Fork 184
Slow inference time #170
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Hi! Times in Probably the times that you are measuring are based on PyTorch model in FP32 precision, so it's ok that they are higher. Am I right? |
One more thing, how to use TensorRT to accelerate the inference of RFDETR? is there any tutorial? |
Hi @programath 👋🏻, unfortunately, we don't have a tutorial for this. Do you think it would be helpful? |
Good afternoon @SkalskiP and @programath, I've converted a custom RT-DETR-B and RT-DETR-L model to TensorRT for deployment on a Jetson Orin Nano 8GB. Would it be beneficial if I wrote a tutorial for this? |
Hi @DatSplit! That would be amazing — really appreciate it! 🔥 Here’s what I’m thinking: we should start breaking out the documentation into separate Could you open a PR that creates a
Something like this for structure:
Later, we’ll link to Let me know what you think! |
Good evening @SkalskiP, That sounds like a solid plan! |
@DatSplit the sooner the better. It can be really rough. I can help you out with structuring it properly. I mostly care about the steps to follow. |
Good afternoon @SkalskiP, I have quite a busy weekend. |
Hi @DatSplit , how has the real-time performance (in FPS) been for the base and large models on your Jetson so far? |
@DatSplit that sounds perfect! Thank you! 🙏🏻 |
Hi, I am testing the latency of the model RFDETRBase at different platform. (Quadro GV100 and A800), the latency of inference time are 0.25s and 0.14s. which is much slower then the reported speed. Is that normal speed of rf-detr?
the timer is calcuated as follows:
The text was updated successfully, but these errors were encountered: