We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 27368ed commit 7ab243aCopy full SHA for 7ab243a
tests/common/base_training_pipeline_test.py
@@ -22,7 +22,8 @@ def mock_pipeline(mocker):
22
"train_ratio": 0.7,
23
"val_ratio": 0.1,
24
"downsample_rate": 1,
25
- "do_anomaly": False
+ "do_anomaly": False,
26
+ "hf_hub_model": None
27
}
28
config.model_config = mocker.MagicMock()
29
config.train_params["lora"] = False
@@ -50,7 +51,8 @@ def test_create_model_lora_enabled(mocker):
50
51
"tmax": 10,
52
"learning_rate": 1e-3,
53
"model": "LTSM",
- "local_pretrain": "None"
54
+ "local_pretrain": "None",
55
56
57
58
config.train_params["lora"] = True
0 commit comments