Skip to content

Commit a84c741

Browse files
authored
update with new records, new fixes
1 parent 8077b9b commit a84c741

File tree

1 file changed

+6
-3
lines changed

1 file changed

+6
-3
lines changed

README.md

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,12 +5,15 @@ Medium article with more info:
55
https://medium.com/@lessw/new-deep-learning-optimizer-ranger-synergistic-combination-of-radam-lookahead-for-the-best-of-2dc83f79a48d
66

77
Multiple updates:
8+
1 - Ranger is the optimizer we used to beat the high scores for 8 different categories on the FastAI leaderboards! (Previous records all held with AdamW optimizer).
89

9-
1 - We used Ranger to beat the FastAI leaderboard score by nearly 20% (19.77%). The trick was to combine Ranger with: Mish activation function, and flat+ cosine anneal training curve.
10+
2 - Highly recommend combining Ranger with: Mish activation function, and flat+ cosine anneal training curve.
1011

11-
2 - Based on that, also found .95 is better than .90 for beta1 (momentum) param (ala betas=(0.95, 0.999)).
12+
3 - Based on that, also found .95 is better than .90 for beta1 (momentum) param (ala betas=(0.95, 0.999)).
1213

13-
3 - Verified no load/save issues in our codebase here. It was an issue for people that were using LookAhead/RAdam as seperate components.
14+
Fixes:
15+
1 - Differential Group learning rates now supported. This was fix in RAdam and ported here thanks to @sholderbach.
16+
2 - In progress fix - save and then load may leave first run weights stranded in memory, slowing down future runs...trying to investigate and fix now.
1417

1518

1619
Usage and notebook to test are available here:

0 commit comments

Comments
 (0)