File tree Expand file tree Collapse file tree 1 file changed +3
-0
lines changed Expand file tree Collapse file tree 1 file changed +3
-0
lines changed Original file line number Diff line number Diff line change @@ -92,6 +92,7 @@ def test_packing(
92
92
93
93
94
94
@pytest .mark .skip_missing_tokenizer
95
+ @patch ("llama_recipes.utils.train_utils.torch.cuda.is_bf16_supported" )
95
96
@patch ("llama_recipes.finetuning.torch.cuda.is_available" )
96
97
@patch ('llama_recipes.finetuning.train' )
97
98
@patch ('llama_recipes.finetuning.AutoTokenizer' )
@@ -119,6 +120,7 @@ def test_distributed_packing(
119
120
tokenizer ,
120
121
train ,
121
122
cuda_is_available ,
123
+ cuda_is_bf16_supported ,
122
124
setup_tokenizer ,
123
125
setup_processor ,
124
126
llama_version ,
@@ -133,6 +135,7 @@ def test_distributed_packing(
133
135
get_mmodel .return_value .get_input_embeddings .return_value .weight .shape = [0 ]
134
136
get_config .return_value = Config (model_type = model_type )
135
137
cuda_is_available .return_value = False
138
+ cuda_is_bf16_supported .return_value = False
136
139
137
140
rank = 1
138
141
os .environ ['LOCAL_RANK' ] = f'{ rank } '
You can’t perform that action at this time.
0 commit comments