Skip to content
This repository was archived by the owner on Apr 29, 2021. It is now read-only.

Conversation

@julianmack
Copy link
Contributor

@julianmack julianmack commented Jan 15, 2020

Adds Gradient accumulation to enable arbitrary batch size.

Note that the configs do not change when using accumulation. This is because accumulation over multiple steps (or gpus in future?) will be identical to no accumulation with larger batch size per step. E.g. accumulation=2, batch/step = 32 is identical to no accumulation at batch=64.

This is blocked by PRs:

@julianmack julianmack changed the title Adds Gradient Accumulation Gradient Accumulation Jan 15, 2020
@julianmack julianmack changed the base branch from master to rnnt January 28, 2020 15:34
@julianmack julianmack changed the base branch from rnnt to rnnt_lr_warm January 28, 2020 15:34
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants