Training on GPU #7975
Phat-Loc
started this conversation in
Help: Best practices
Training on GPU
#7975
Replies: 1 comment 3 replies
-
Assuming your training data is large enough to fill your GPU memory, I would try increasing That said you don't seem to have a Transformer in your pipeline. You can train on GPU without a Transformer, but is that what you intend to do? If you have a good GPU usually using Transformers is worth it. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
What settings should I use in the training section of config the config to max out my gpu usage? Most of the time it is my gpu runs between 0 - 10% during training. Here is my settings
[system]
gpu_allocator = "pytorch"
#gpu_allocator = null
seed = 0
[nlp]
lang = "en"
pipeline = ["tok2vec", "senter"]
tokenizer = {"@Tokenizers":"spacy.Tokenizer.v1"}
before_creation = null
after_creation = null
after_pipeline_creation = null
disabled = []
batch_size = 1000
[training]
train_corpus = "corpora.train"
dev_corpus = "corpora.dev"
seed = ${system.seed}
gpu_allocator = ${system.gpu_allocator}
dropout = 0.1
accumulate_gradient = 1
patience = 1600
max_epochs = 0
max_steps = 20000
eval_frequency = 200
frozen_components = []
before_to_disk = null
[training.batcher]
@batchers = "spacy.batch_by_words.v1"
discard_oversize = false
tolerance = 0.2
get_length = null
[training.batcher.size]
@schedules = "compounding.v1"
start = 100
stop = 1000
compound = 1.001
t = 0.0
[training.logger]
@Loggers = "spacy.ConsoleLogger.v1"
progress_bar = false
[training.optimizer]
@optimizers = "Adam.v1"
beta1 = 0.9
beta2 = 0.999
L2_is_weight_decay = true
L2 = 0.01
grad_clip = 1.0
use_averages = false
eps = 0.00000001
[training.optimizer.learn_rate]
@schedules = "warmup_linear.v1"
warmup_steps = 250
total_steps = 20000
initial_rate = 0.00005
[training.score_weights]
ents_per_type = null
ents_f = 1.0
ents_p = 0.0
ents_r = 0.0
Beta Was this translation helpful? Give feedback.
All reactions