You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Dec 23, 2022. It is now read-only.
Describe the bug
After fixing lr, I ran the DETR training but it seems loss didn't decrease at all.
I know DETR convergence is so slow, but is this loss behavior natural?
I cannot tell you since I have no GPUs at my disposal. What I can tell you is that the loss is decreasing during an overfit on a single image. You can try the detr overfit notebook.
If I had GPUs I would correct the possible bugs but currently I cannot :s. I runned the overfit with the same inputs on the official codebase and kerod and I ended up with the same results.
For same inputs same loss (I just modified a bit the loss of Detr to instead having the background classes at pos 80 + 1 have it at pos 0 like in kerod)
Small overfit on a single batch
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Describe the bug
After fixing lr, I ran the DETR training but it seems loss didn't decrease at all.
I know DETR convergence is so slow, but is this loss behavior natural?
To Reproduce
run this notebook
https://colab.research.google.com/github/Emgarr/kerod/blob/master/notebooks/detr_coco_training_multi_gpu.ipynb
Expected behavior
A clear and concise description of what you expected to happen.
Screenshots
Desktop (please complete the following information):
colab notebook
Additional context
Add any other context about the problem here.
The text was updated successfully, but these errors were encountered: