-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
咨询META训练时超参数设置问题 #2
Comments
Hello! I also encountered the problem what you said when using the default BASE_LR: 0.04,IMS_PER_BATCH: 64. But when I set BASE_LR to 0.0003 and IMS_PER_BATCH=64, I reported another error when finishing epoch/iter: 2/1999. Do you or the author knows what the problem is or have you ever encountered it? -- Process 0 terminated with the following error: |
and often occur the gradient overflow: |
Hello, I haven't met the second question. I trained with four graphics cards. However, after my successful training, the performance still lags behind that in the paper. I sent an email to the author of the paper, but he didn't reply to me. |
I have met the same problem, have you addressed ? Thanks a lot. |
In fact, I kind of forgot about this. If memory serves, the number of classes in config.yml should be +1. |
我从您论文摘要所给出的github链接下载了META工程,用四张显存各为12G的2080Ti,按照您配置文件中默认的参数,BASE_LR: 0.04,IMS_PER_BATCH: 64,训练时回报错“FloatingPointError: Loss became infinite or NaN at iteration=510!”,查阅资料后好像时是学习率有点大。我就根据您论文实验部分的参数设置,设置BASE_LR: 0.0003,IMS_PER_BATCH: 64,这样能训练完,但是rank1为39.16,mAP为14.83,达不到论文中的性能。请问是我训练过程有问题吗,能否分享一下您的超参数设置呢?
The text was updated successfully, but these errors were encountered: