Skip to content

Commit 0e8fe08

Browse files
committed
Cleanup
1 parent d77e518 commit 0e8fe08

File tree

3 files changed

+6
-4
lines changed

3 files changed

+6
-4
lines changed

auto_fp8/modeling.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -140,8 +140,6 @@ def get_layers_to_ignore(model, ignore_patterns) -> List[str]:
140140
if ignore_pattern.startswith(regex_prefix):
141141
# check if name matches regex and add to set if true
142142
regex_pattern = ignore_pattern[len(regex_prefix) :]
143-
print(regex_pattern)
144-
print(name)
145143
if re.search(regex_pattern, name):
146144
ignored_layers.add(name)
147145
else:

example_dataset.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,8 @@
77
quantized_model_dir = "Meta-Llama-3-8B-Instruct-FP8"
88

99
tokenizer = AutoTokenizer.from_pretrained(pretrained_model_dir, use_fast=True)
10+
tokenizer.pad_token = tokenizer.eos_token
11+
1012
ds = load_dataset("mgoin/ultrachat_2k", split="train_sft").select(512)
1113
examples = [tokenizer.apply_chat_template(batch["messages"], tokenize=False) for batch in ds]
1214
examples = tokenizer(examples, padding=True, truncation=True, return_tensors="pt").to("cuda")

examples/example_mixtral.py

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,9 @@
77
quantized_model_dir = "Mixtral-8x7B-Instruct-v0.1-FP8"
88

99
tokenizer = AutoTokenizer.from_pretrained(pretrained_model_dir, use_fast=True)
10-
ds = load_dataset("mgoin/ultrachat_2k", split="train_sft").select(10)
10+
tokenizer.pad_token = tokenizer.eos_token
11+
12+
ds = load_dataset("mgoin/ultrachat_2k", split="train_sft").select(range(10))
1113
examples = [tokenizer.apply_chat_template(batch["messages"], tokenize=False) for batch in ds]
1214
examples = tokenizer(examples, padding=True, truncation=True, return_tensors="pt").to("cuda")
1315

@@ -21,4 +23,4 @@
2123
pretrained_model_dir, quantize_config=quantize_config
2224
)
2325
model.quantize(examples)
24-
model.save_quantized(quantized_model_dir)
26+
model.save_quantized(quantized_model_dir)

0 commit comments

Comments
 (0)