Skip to content
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
38 changes: 18 additions & 20 deletions src/llmcompressor/entrypoints/oneshot.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@
import os
from datetime import datetime
from pathlib import Path
from typing import TYPE_CHECKING, Callable
from typing import TYPE_CHECKING, Any, Callable

from loguru import logger
from torch.utils.data import DataLoader
Expand Down Expand Up @@ -281,25 +281,15 @@ def oneshot(
dataloader_num_workers: int = 0,
min_tokens_per_module: float | None = None,
moe_calibrate_all_experts: bool = True,
pipeline: str | None = "independent",
tracing_ignore: list[str] = [
"_update_causal_mask",
"create_causal_mask",
"_update_mamba_mask",
"make_causal_mask",
"get_causal_mask",
"mask_interface",
"mask_function",
"_prepare_4d_causal_attention_mask",
"_prepare_fsmt_decoder_inputs",
"_prepare_4d_causal_attention_mask_with_cache_position",
"_update_linear_attn_mask",
"project_per_layer_inputs",
],
sequential_targets: list[str] | None = None,
sequential_offload_device: str = "cpu",
pipeline: str = "independent",
tracing_ignore: list[str] | None = None,
raw_kwargs: dict[str, Any] | None = None,
preprocessing_func: Callable | None = None,
remove_columns: list[str] | None = None,
dvc_data_repository: str | None = None,
quantization_aware_calibration: bool = True,
# Miscellaneous arguments
sequential_targets: list[str] | None = None,
# Miscellaneous arguments
output_dir: str | None = None,
log_dir: str | None = None,
**kwargs,
Expand Down Expand Up @@ -398,10 +388,18 @@ def oneshot(
:return: The calibrated PreTrainedModel
"""

# pass all args directly into Oneshot
if sequential_targets and pipeline == "independent":
raise ValueError(
"Invalid configuration: "
"'sequential_targets' cannot be used with 'independent' pipeline. "
"Please use 'sequential' or 'layer_sequential' pipeline when specifying "
"sequential_targets."
)

local_args = {
k: v for k, v in locals().items() if k not in ("local_args", "kwargs")
}

one_shot = Oneshot(**local_args, **kwargs)
one_shot()

Expand Down
15 changes: 15 additions & 0 deletions tests/llmcompressor/transformers/oneshot/test_api_inputs.py
Original file line number Diff line number Diff line change
@@ -1,10 +1,15 @@
import logging

import pytest
from transformers import AutoModelForCausalLM, AutoTokenizer

from llmcompressor import oneshot
from tests.llmcompressor.transformers.oneshot.dataset_processing import get_data_utils
from tests.testing_utils import parse_params

logging.basicConfig(level=logging.INFO)
logger = logging.getLogger(__name__)

CONFIGS_DIRECTORY = "tests/llmcompressor/transformers/oneshot/oneshot_configs"

# TODO: Seems better to mark test type (smoke, sanity, regression) as a marker as
Expand Down Expand Up @@ -42,6 +47,16 @@ def wrapped_preprocess_func(sample):
dataset_config_name=config.get("dataset_config_name"),
)

args["pipeline"] = config.get("pipeline", "independent")
args["sequential_targets"] = config.get("sequential_targets", None)
args["tracing_ignore"] = config.get("tracing_ignore", [])
args["raw_kwargs"] = config.get("raw_kwargs", {})
args["preprocessing_func"] = config.get("preprocessing_func", None)
args["remove_columns"] = config.get("remove_columns", None)
args["dvc_data_repository"] = config.get("dvc_data_repository", None)
args["splits"] = config.get("splits", {"calibration": "train[:50]"})
args["log_dir"] = config.get("log_dir", "sparse_logs")

return args


Expand Down
Loading