Skip to content

Commit 0e074e4

Browse files
authored
Merge branch 'main' into kylesayrs/tracing-testing
2 parents 71c5575 + e5780c5 commit 0e074e4

File tree

4 files changed

+16
-16
lines changed

4 files changed

+16
-16
lines changed

src/llmcompressor/logger.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -76,7 +76,7 @@ def configure_logger(config: Optional[LoggerConfig] = None) -> None:
7676
if (log_file := os.getenv("LLM_COMPRESSOR_LOG_FILE")) is not None:
7777
logger_config.log_file = log_file
7878
if (log_file_level := os.getenv("LLM_COMPRESSOR_LOG_FILE_LEVEL")) is not None:
79-
logger_config.log_file_level = log_file_level
79+
logger_config.log_file_level = log_file_level.upper()
8080

8181
if logger_config.disabled:
8282
logger.disable("llmcompressor")

tests/e2e/vLLM/run_tests.sh

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ SUCCESS=0
44

55
while getopts "c:t:" OPT; do
66
case ${OPT} in
7-
c )
7+
c )
88
CONFIG="$OPTARG"
99
;;
1010
t )
@@ -25,9 +25,7 @@ do
2525

2626
export TEST_DATA_FILE="$MODEL_CONFIG"
2727
pytest \
28-
-r a \
2928
--capture=tee-sys \
30-
--junitxml="test-results/e2e-$(date +%s).xml" \
3129
"$TEST" || LOCAL_SUCCESS=$?
3230

3331
if [[ $LOCAL_SUCCESS == 0 ]]; then

tests/e2e/vLLM/test_vllm.py

Lines changed: 7 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,6 @@
88
import yaml
99
from huggingface_hub import HfApi
1010
from loguru import logger
11-
from parameterized import parameterized_class
1211

1312
from llmcompressor.core import active_session
1413
from tests.e2e.e2e_utils import run_oneshot_for_e2e_testing
@@ -43,7 +42,9 @@
4342
# Will run each test case in its own process through run_tests.sh
4443
# emulating vLLM CI testing
4544
@requires_gpu_count(1)
46-
@parameterized_class("test_data_file", [(TEST_DATA_FILE,)])
45+
@pytest.mark.parametrize(
46+
"test_data_file", [pytest.param(TEST_DATA_FILE, id=TEST_DATA_FILE)]
47+
)
4748
@pytest.mark.skipif(not vllm_installed, reason="vLLM is not installed, skipping test")
4849
class TestvLLM:
4950
"""
@@ -62,10 +63,8 @@ class TestvLLM:
6263
be used for quantization. Otherwise, the recipe will always be used if given.
6364
""" # noqa: E501
6465

65-
def set_up(self):
66-
eval_config = yaml.safe_load(
67-
Path(self.test_data_file).read_text(encoding="utf-8")
68-
)
66+
def set_up(self, test_data_file: str):
67+
eval_config = yaml.safe_load(Path(test_data_file).read_text(encoding="utf-8"))
6968

7069
if os.environ.get("CADENCE", "commit") != eval_config.get("cadence"):
7170
pytest.skip("Skipping test; cadence mismatch")
@@ -97,10 +96,10 @@ def set_up(self):
9796
]
9897
self.api = HfApi()
9998

100-
def test_vllm(self):
99+
def test_vllm(self, test_data_file: str):
101100
# Run vLLM with saved model
102101

103-
self.set_up()
102+
self.set_up(test_data_file)
104103
if not self.save_dir:
105104
self.save_dir = self.model.split("/")[1] + f"-{self.scheme}"
106105
oneshot_model, tokenizer = run_oneshot_for_e2e_testing(

tests/lmeval/test_lmeval.py

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -39,6 +39,9 @@ class LmEvalConfig(BaseModel):
3939
# Will run each test case in its own process through run_tests.sh
4040
# emulating vLLM CI testing
4141
@requires_gpu_count(1)
42+
@pytest.mark.parametrize(
43+
"test_data_file", [pytest.param(TEST_DATA_FILE, id=TEST_DATA_FILE)]
44+
)
4245
@pytest.mark.skipif(
4346
not lm_eval_installed, reason="lm eval is not installed, skipping test"
4447
)
@@ -58,8 +61,8 @@ class TestLMEval:
5861
be used for quantization. Otherwise, the recipe will always be used if given.
5962
""" # noqa: E501
6063

61-
def set_up(self):
62-
eval_config = yaml.safe_load(Path(TEST_DATA_FILE).read_text(encoding="utf-8"))
64+
def set_up(self, test_data_file: str):
65+
eval_config = yaml.safe_load(Path(test_data_file).read_text(encoding="utf-8"))
6366

6467
if os.environ.get("CADENCE", "commit") != eval_config.get("cadence"):
6568
pytest.skip("Skipping test; cadence mismatch")
@@ -88,9 +91,9 @@ def set_up(self):
8891
self.num_calibration_samples = 512
8992
self.max_seq_length = 2048
9093

91-
def test_lm_eval(self):
94+
def test_lm_eval(self, test_data_file: str):
9295
# Run vLLM with saved model
93-
self.set_up()
96+
self.set_up(test_data_file)
9497
if not self.save_dir:
9598
self.save_dir = self.model.split("/")[1] + f"-{self.scheme}"
9699
oneshot_model, processor = run_oneshot_for_e2e_testing(

0 commit comments

Comments
 (0)