File tree Expand file tree Collapse file tree 3 files changed +4
-29
lines changed Expand file tree Collapse file tree 3 files changed +4
-29
lines changed Original file line number Diff line number Diff line change 3030)
3131from transformers .trainer_utils import EvalPrediction
3232from transformers .training_args import OptimizerNames
33- from transformers .utils import is_apex_available , is_peft_available
33+ from transformers .utils import is_peft_available
3434
3535from ..data_utils import is_conversational , maybe_apply_chat_template
3636from ..models .modeling_base import GeometricMixtureWrapper
4747)
4848
4949
50- if is_apex_available ():
51- from apex import amp
52-
53-
5450if is_peft_available ():
5551 from peft import PeftModel
5652
@@ -501,10 +497,6 @@ def training_step(
501497 if self .args .n_gpu > 1 :
502498 loss = loss .mean () # mean() to average on multi-gpu parallel training
503499
504- if self .use_apex :
505- with amp .scale_loss (loss , self .optimizer ) as scaled_loss :
506- scaled_loss .backward ()
507- else :
508- self .accelerator .backward (loss , ** kwargs )
500+ self .accelerator .backward (loss , ** kwargs )
509501
510502 return loss .detach () / self .args .gradient_accumulation_steps
Original file line number Diff line number Diff line change 4343 ProcessorMixin ,
4444 Trainer ,
4545 TrainerCallback ,
46- is_apex_available ,
4746)
4847from transformers .models .auto .modeling_auto import MODEL_FOR_IMAGE_TEXT_TO_TEXT_MAPPING_NAMES
4948from transformers .trainer_utils import EvalPrediction , seed_worker
7877if is_peft_available ():
7978 from peft import PeftConfig , PeftModel
8079
81- if is_apex_available ():
82- from apex import amp
83-
8480
8581if is_sagemaker_mp_enabled ():
8682 from smdistributed .modelparallel import __version__ as SMP_VERSION
@@ -1457,11 +1453,7 @@ def training_step(
14571453 if self .args .n_gpu > 1 :
14581454 loss = loss .mean () # mean() to average on multi-gpu parallel training
14591455
1460- if self .use_apex :
1461- with amp .scale_loss (loss , self .optimizer ) as scaled_loss :
1462- scaled_loss .backward ()
1463- else :
1464- self .accelerator .backward (loss , ** kwargs )
1456+ self .accelerator .backward (loss , ** kwargs )
14651457
14661458 return loss .detach () / self .args .gradient_accumulation_steps
14671459
Original file line number Diff line number Diff line change 2727 PreTrainedTokenizerBase ,
2828 ProcessorMixin ,
2929 TrainerCallback ,
30- is_apex_available ,
3130)
3231from transformers .trainer_utils import EvalPrediction
3332from transformers .training_args import OptimizerNames
4746from .xpo_config import XPOConfig
4847
4948
50- if is_apex_available ():
51- from apex import amp
52-
53-
5449if is_peft_available ():
5550 from peft import PeftModel
5651
@@ -545,10 +540,6 @@ def training_step(
545540 if self .args .n_gpu > 1 :
546541 loss = loss .mean () # mean() to average on multi-gpu parallel training
547542
548- if self .use_apex :
549- with amp .scale_loss (loss , self .optimizer ) as scaled_loss :
550- scaled_loss .backward ()
551- else :
552- self .accelerator .backward (loss , ** kwargs )
543+ self .accelerator .backward (loss , ** kwargs )
553544
554545 return loss .detach () / self .args .gradient_accumulation_steps
You can’t perform that action at this time.
0 commit comments