Use dynamic learning rate decay for convergence#39
Use dynamic learning rate decay for convergence#39begeekmyfriend wants to merge 2 commits intoseungwonpark:masterfrom
Conversation
Signed-off-by: begeekmyfriend <begeekmyfriend@gmail.com>
|
Hi, your code looks great, and thanks for kindly sending PR! |
|
melgan_eval_mandarin.zip |
Signed-off-by: begeekmyfriend <begeekmyfriend@gmail.com>
|
|
||
|
|
||
| def cosine_decay(init_val, final_val, step, decay_steps): | ||
| alpha = final_val / init_val |
There was a problem hiding this comment.
The learning rate decays. You might write a demo for testing.
init_val = 1e-4
final_val = 1e-5There was a problem hiding this comment.
According to the following source it's "Minimum learning rate value as a fraction of learning_rate."
https://docs.w3cub.com/tensorflow~python/tf/train/cosine_decay/
Given the values, it looks like it's correct. The naming is just off - it should be the smallest value in the numerator and largest value in the denominator.
|
Is this different from pytorch's built-in CosineAnnealingLR? |
|
@begeekmyfriend |
|
It is just a preference. Pick it or other as you like. |
|
@begeekmyfriend Thank you for your quick reply. I used your branch of tacotron and found it is one of the best among a lot of code branchs. I will try the cos-lr as well as the apex in tfgan. |
The evaluation sounds better than that with fixed learning rate.
Signed-off-by: begeekmyfriend begeekmyfriend@gmail.com