You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@@ -32,8 +33,8 @@ If this repository has been useful to you in your research, please cite it using
32
33
33
34
| Symbol | Meaning | Count |
34
35
|:--------------|:--------|:------|
35
-
| None | Paper |11|
36
-
|:outbox_tray:| Summary |2|
36
+
| None | Paper |17|
37
+
|:outbox_tray:| Summary |3|
37
38
|:computer:| Code | 0 |
38
39
39
40
@@ -53,36 +54,55 @@ If this repository has been useful to you in your research, please cite it using
53
54
4.[KOALA: A Kalman Optimization Algorithm with Loss Adaptivity](https://arxiv.org/abs/2107.03331)[:outbox_tray:]()[:computer:]()
54
55
Aram Davtyan, Sepehr Sameni, Llukman Cerkezi, Givi Meishvilli, Adam Bielski, Paolo Favaro; 2021
55
56
57
+
## Momentum based Optimizers
58
+
59
+
5.[On the Momentum Term in Gradient Descent Learning Algorithms](https://reader.elsevier.com/reader/sd/pii/S0893608098001166?token=3147494EED9FE670AF728F3408B795675246C9934481200C4E86611D7FE34FAEDDFF1E9BD5C6AE9455320BF21F3FEA3B&originRegion=eu-west-1&originCreation=20230223114928)[:outbox_tray:]()[:computer:]()
60
+
Ning Qian; 1999
61
+
62
+
6.[Symbolic Discovery of Optimization Algorithms](https://arxiv.org/abs/2302.06675)[:outbox_tray:]()[:computer:]() Xiangning Chen, Chen Liang, Da Huang; 2023
6.[Adam: A Method for Stochastic Optimization](https://arxiv.org/abs/1412.6980)[:outbox_tray:]()[:computer:]()
77
+
10.[Adam: A Method for Stochastic Optimization](https://arxiv.org/abs/1412.6980)[:outbox_tray:]()[:computer:]()
64
78
Diederik P. Kingma, Jimmy Ba; 2014
65
79
80
+
8.[AdamP: Slowing Down the Slowdown for Momentum Optimizers on Scale-invariant Weights](https://arxiv.org/abs/2006.08217)[:outbox_tray:]()[:computer:]()
8.[Gradient Centralization: A New Optimization Technique for Deep Neural Networks](https://arxiv.org/abs/2004.01461)[:outbox_tray:](survey/gradient-centralization.md)[:computer:]()
95
+
14.[Gradient Centralization: A New Optimization Technique for Deep Neural Networks](https://arxiv.org/abs/2004.01461)[:outbox_tray:](survey/gradient-centralization.md)[:computer:]()
76
96
Hongwei Yong, Jianqiang Huang, Xiansheng Hua, Lei Zhang; 2020
77
97
78
98
79
99
## Optimizer Analysis and Meta-research
80
-
9.[On Empirical Comparisons of Optimizers for Deep Learning](https://arxiv.org/abs/1910.05446)[:outbox_tray:]()
100
+
15.[On Empirical Comparisons of Optimizers for Deep Learning](https://arxiv.org/abs/1910.05446)[:outbox_tray:]()
81
101
Dami Choi, Christopher J. Shallue, Zachary Nado, Jaehoon Lee, Chris J. Maddison, George E. Dahl; 2019
82
102
83
-
10.[Adam Can Converge Without Any Modification on Update Rules](https://arxiv.org/abs/2208.09632)[:outbox_tray:](survey/adam-can-converge.md)
103
+
11.[Adam Can Converge Without Any Modification on Update Rules](https://arxiv.org/abs/2208.09632)[:outbox_tray:](survey/adam-can-converge.md)
0 commit comments