Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LR hyperparameters tuning method #14

Closed
Cadene opened this issue Oct 28, 2018 · 1 comment
Closed

LR hyperparameters tuning method #14

Cadene opened this issue Oct 28, 2018 · 1 comment

Comments

@Cadene
Copy link

Cadene commented Oct 28, 2018

Hi,

Thanks again for your code. Unfortunately, I ran into a little issue. I can't reproduce some of your results because I am obliged to reduce my batch size (from 512 (yours) to 75). Thus I need to change the hyperparameters related to the learning rate.

Finding the right learning rate can be easily done by a little grid search. However, I would like to know how did you tune the hyperparameters related to the scheduler.
Especially:

  • __C.training_parameters.wu_factor = 0.2
  • __C.training_parameters.wu_iters = 1000
  • __C.training_parameters.lr_steps = [5000, 7000, 9000, 11000]
  • __C.training_parameters.lr_ratio = 0.1

Sharing your method would be awesome :)

Thanks for your help!
Remi

@YuJiang01
Copy link
Contributor

we roughly adjusted based on the number of epochs when changing batch size for wu_iters and lr_steps

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants