Is there a way to set learning rate annealing with tune.run()?
The grid search below will run two training: one LR at 1e-5 and second at 1e-6. How do I, for example, have a LR schedule where at the end of 1000 iterations is LR is reduced from 1e-5 to 1e-6 and from 1e-6 to 1e-7? I’ve gone through the docs and can’t seem to find a solution.
The 1000000 is the (sampled) timesteps at which you would like the 1e-6 to be reached. RLlib will linearly decrease the learning rat from 1e-5 to 1e-6 and after 1Mio ts, stick with the 1e-6 as final value.
Yes, you are right, the very first lr used is the “lr” one ignoring the schedule, after that, we correctly switch to using the schedule. This is a small bug, that we should fix for clarity.