Skip to content

Commit 590d6fa

Browse files
authored
Merge pull request #17 from nithinkr2000/main
Added 3 papers to the list
2 parents d658ec6 + af1942a commit 590d6fa

File tree

1 file changed

+17
-13
lines changed

1 file changed

+17
-13
lines changed

README.md

Lines changed: 17 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ If this repository has been useful to you in your research, please cite it using
3333

3434
| Symbol | Meaning | Count |
3535
|:--------------|:--------|:------|
36-
| None | Paper | 17 |
36+
| :page_facing_up: | Paper | 20 |
3737
| :outbox_tray: | Summary | 3 |
3838
| :computer: | Code | 0 |
3939

@@ -61,48 +61,52 @@ If this repository has been useful to you in your research, please cite it using
6161

6262
6. [Symbolic Discovery of Optimization Algorithms](https://arxiv.org/abs/2302.06675) [:outbox_tray:]() [:computer:]() Xiangning Chen, Chen Liang, Da Huang; 2023
6363

64-
64+
7. [Demon: Improved Neural Network Training with Momentum Decay](https://arxiv.org/abs/1910.04952) John Chen, Cameron Wolfe, Zhao Li, Anastasios Kyrillidis ; 2021
6565
## Adaptive Optimizers
6666

67-
7. [Adaptive Subgradient Methods for Online Learning and Stochastic Optimization](https://dl.acm.org/doi/10.5555/1953048.2021068) [:outbox_tray:]() [:computer:]() John Duchi, Elad Hazan, Yoram Singer; 2011
67+
8. [Adaptive Subgradient Methods for Online Learning and Stochastic Optimization](https://dl.acm.org/doi/10.5555/1953048.2021068) [:outbox_tray:]() [:computer:]() John Duchi, Elad Hazan, Yoram Singer; 2011
6868

69-
8. [ADADELTA: An Adaptive Learning Rate Method](https://arxiv.org/abs/1212.5701) [:outbox_tray:]() [:computer:]()
69+
9. [ADADELTA: An Adaptive Learning Rate Method](https://arxiv.org/abs/1212.5701) [:outbox_tray:]() [:computer:]()
7070
Matthew D. Zeiler; 2012
7171

72-
6. [RMSProp](http://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf) [:outbox_tray:]() [:computer:]()
72+
10. [RMSProp](http://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf) [:outbox_tray:]() [:computer:]()
7373
Geoffrey Hinton; 2013
7474

7575
## Adam Family of Optimizers
7676

77-
10. [Adam: A Method for Stochastic Optimization](https://arxiv.org/abs/1412.6980) [:outbox_tray:]() [:computer:]()
77+
11. [Adam: A Method for Stochastic Optimization](https://arxiv.org/abs/1412.6980) [:outbox_tray:]() [:computer:]()
7878
Diederik P. Kingma, Jimmy Ba; 2014
7979

80-
8. [AdamP: Slowing Down the Slowdown for Momentum Optimizers on Scale-invariant Weights](https://arxiv.org/abs/2006.08217) [:outbox_tray:]() [:computer:]()
80+
12. [AdamP: Slowing Down the Slowdown for Momentum Optimizers on Scale-invariant Weights](https://arxiv.org/abs/2006.08217) [:outbox_tray:]() [:computer:]()
8181
Byeongho Heo, Sanghyuk Chun, Seong Joon Oh, Dongyoon Han; 2020
8282

83-
9. [On the Variance of the Adaptive Learning Rate and Beyond](https://arxiv.org/abs/1908.03265) [:outbox_tray:]() [:computer:]()
83+
13. [On the Variance of the Adaptive Learning Rate and Beyond](https://arxiv.org/abs/1908.03265) [:outbox_tray:]() [:computer:]()
8484
Liyuan Liu, Haoming Jiang, Pengcheng He; 2021
8585

86+
14. [AdaBelief Optimizer: Adapting Stepsizes by the Belief in Observed Gradients](https://arxiv.org/abs/2010.07468) Juntang Zhuang, Tommy Tang, Yifan Ding, Sekhar Tatikonda, Nicha Dvornek, Xenophon Papademetris, James S. Duncan ; 2020
87+
88+
15. [Momentum Centering and Asynchronous Update for Adaptive Gradient Methods](https://arxiv.org/abs/2110.05454) Juntang Zhuang, Yifan Ding, Tommy Tang, Nicha Dvornek, Sekhar Tatikonda, James S. Duncan ; 2021
89+
8690
# Second-order Optimizers
8791

88-
13. [Shampoo: Preconditioned Stochastic Tensor Optimization](https://arxiv.org/abs/1802.09568) [:outbox_tray:]() [:computer:]()
92+
16. [Shampoo: Preconditioned Stochastic Tensor Optimization](https://arxiv.org/abs/1802.09568) [:outbox_tray:]() [:computer:]()
8993
Vineet Gupta, Tomer Koren, Yoram Singer; 2018
9094

9195

9296
# Other Optimisation-Related Research
9397

9498
## General Improvements
95-
14. [Gradient Centralization: A New Optimization Technique for Deep Neural Networks](https://arxiv.org/abs/2004.01461) [:outbox_tray:](survey/gradient-centralization.md) [:computer:]()
99+
17. [Gradient Centralization: A New Optimization Technique for Deep Neural Networks](https://arxiv.org/abs/2004.01461) [:outbox_tray:](survey/gradient-centralization.md) [:computer:]()
96100
Hongwei Yong, Jianqiang Huang, Xiansheng Hua, Lei Zhang; 2020
97101

98102

99103
## Optimizer Analysis and Meta-research
100-
15. [On Empirical Comparisons of Optimizers for Deep Learning](https://arxiv.org/abs/1910.05446) [:outbox_tray:]()
104+
18. [On Empirical Comparisons of Optimizers for Deep Learning](https://arxiv.org/abs/1910.05446) [:outbox_tray:]()
101105
Dami Choi, Christopher J. Shallue, Zachary Nado, Jaehoon Lee, Chris J. Maddison, George E. Dahl; 2019
102106

103-
11. [Adam Can Converge Without Any Modification on Update Rules](https://arxiv.org/abs/2208.09632) [:outbox_tray:](survey/adam-can-converge.md)
107+
19. [Adam Can Converge Without Any Modification on Update Rules](https://arxiv.org/abs/2208.09632) [:outbox_tray:](survey/adam-can-converge.md)
104108
Yushun Zhang, Congliang Chen, Naichen Shi, Ruoyu Sun, Zhi-Quan Luo; 2022
105109

106110
## Hyperparameter Tuning
107-
17. [Gradient Descent: The Ultimate Optimizer](https://arxiv.org/abs/1909.13371) [:outbox_tray:]() [:computer:]()
111+
20. [Gradient Descent: The Ultimate Optimizer](https://arxiv.org/abs/1909.13371) [:outbox_tray:]() [:computer:]()
108112
Kartik Chandra, Audrey Xie, Jonathan Ragan-Kelley, Erik Meijer; 2019

0 commit comments

Comments
 (0)