-
Notifications
You must be signed in to change notification settings - Fork 10
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training process abnormal. #4
Comments
I haven't encountered this issue before. Please paste your complete settings and startup command here, and I'll find some time in the next few days to test it. |
I got an similar issue and the training configuration and the log is below. Training Configuration
The Log
|
also environmental information
|
I also encounter the same problem.... |
No decline in loss when iou becomes 0. Maybe it is caused by gradient exploration. |
@Liuliuliudalu
|
Sorry for my typo, it is 1e-4. I'm still training the segrefiner_lr. Maybe I can reply you later after evaluation. |
Thank you all for your attention to this work and sorry for the delayed response. I believe the issue is due to training instability caused by a too-small batch size. The |
And through testing, I have confirmed that for the initial learning rate of 4e-4 in HR-SegRefiner, the issue of IOU dropping to 0 does not occur when the global batch size reaches 4. |
@Liuliuliudalu @MengyuWang826
|
This is a bit strange. When I have enough free GPUs available, I will check and retrain it. |
@MengyuWang826 Thank you! |
Hello, I also encountered the same problem, may I ask you to solve it? |
@MengyuWang826 |
I have tried many different settings, but during the training of segrefiner, the IoU becomes zero after only a few steps. Is this a known issue?
Thanks.
The text was updated successfully, but these errors were encountered: