Skip to content
This repository was archived by the owner on Sep 10, 2025. It is now read-only.

Commit 1df7101

Browse files
authored
Merge branch 'main' into patch-10
2 parents 616104d + d7b681a commit 1df7101

File tree

4 files changed

+26
-10
lines changed

4 files changed

+26
-10
lines changed

install/install_requirements.sh

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -14,19 +14,21 @@ then
1414
if [[ -z ${CONDA_DEFAULT_ENV:-} ]] || [[ ${CONDA_DEFAULT_ENV:-} == "base" ]] || [[ ! -x "$(command -v python)" ]];
1515
then
1616
PYTHON_EXECUTABLE=python3
17+
else
18+
PYTHON_EXECUTABLE=python
1719
fi
1820
fi
1921
echo "Using python executable: $PYTHON_EXECUTABLE"
2022

2123
PYTHON_SYS_VERSION="$($PYTHON_EXECUTABLE -c "import sys; print(f'{sys.version_info.major}.{sys.version_info.minor}')")"
22-
# Check python version. Expect 3.10.x or 3.11.x
24+
# Check python version. Expect at least 3.10.x
2325
if ! $PYTHON_EXECUTABLE -c "
2426
import sys
25-
if sys.version_info < (3, 10) or sys.version_info >= (3, 12):
27+
if sys.version_info < (3, 10):
2628
sys.exit(1)
2729
";
2830
then
29-
echo "Python version must be 3.10.x or 3.11.x. Detected version: $PYTHON_SYS_VERSION"
31+
echo "Python version must be at least 3.10.x. Detected version: $PYTHON_SYS_VERSION"
3032
exit 1
3133
fi
3234

install/requirements.txt

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,6 @@ snakeviz
1414
sentencepiece
1515
# numpy version range required by GGUF util
1616
numpy >= 1.17, < 2.0
17-
gguf
1817
blobfile
1918
tomli >= 1.1.0 ; python_version < "3.11"
2019
openai

torchchat/cli/cli.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -533,7 +533,7 @@ def arg_init(args):
533533
# Localized import to minimize expensive imports
534534
from torchchat.utils.build_utils import get_device_str
535535

536-
if args.device is None:
536+
if args.device is None or args.device == "fast":
537537
args.device = get_device_str(
538538
args.quantize.get("executor", {}).get("accelerator", default_device)
539539
)

torchchat/generate.py

Lines changed: 20 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1189,12 +1189,27 @@ def callback(x, *, done_generating=False):
11891189
f"Mean Accepted: {sum([idx * i for idx, i in enumerate(counts_aggregated)])/sum(counts_aggregated)}"
11901190
)
11911191

1192-
print(
1193-
f"\n Average tokens/sec (total): {torch.mean(torch.tensor(aggregate_metrics['tokens_per_sec'])).item():.2f} \
1194-
\nAverage tokens/sec (first token): {torch.mean(torch.tensor(aggregate_metrics['first_token_per_sec'])).item():.2f} \
1195-
\nAverage tokens/sec (next tokens): {torch.mean(torch.tensor(aggregate_metrics['next_tokens_per_sec'])).item():.2f} \n\
1192+
avg_tokens_sec = torch.mean(
1193+
torch.tensor(aggregate_metrics["tokens_per_sec"])
1194+
).item()
1195+
avg_first_token_sec = torch.mean(
1196+
torch.tensor(aggregate_metrics["first_token_per_sec"])
1197+
).item()
1198+
avg_next_tokens_sec = torch.mean(
1199+
torch.tensor(aggregate_metrics["next_tokens_per_sec"])
1200+
).item()
1201+
1202+
if not (
1203+
torch.isnan(torch.tensor(avg_tokens_sec))
1204+
or torch.isnan(torch.tensor(avg_first_token_sec))
1205+
or torch.isnan(torch.tensor(avg_next_tokens_sec))
1206+
):
1207+
print(
1208+
f"\n Average tokens/sec (total): {avg_tokens_sec:.2f} \
1209+
\nAverage tokens/sec (first token): {avg_first_token_sec:.2f} \
1210+
\nAverage tokens/sec (next tokens): {avg_next_tokens_sec:.2f} \n\
11961211
"
1197-
)
1212+
)
11981213
if torch.cuda.is_available():
11991214
print(f"Memory used: {torch.cuda.max_memory_reserved() / 1e9:.02f} GB")
12001215

0 commit comments

Comments
 (0)