File tree Expand file tree Collapse file tree 1 file changed +11
-1
lines changed
examples/llm_qat/notebooks Expand file tree Collapse file tree 1 file changed +11
-1
lines changed Original file line number Diff line number Diff line change 608608 "metadata" : {},
609609 "outputs" : [],
610610 "source" : [
611+ " %%sh # [run in command line outside notebook]\n " ,
612+ " \n " ,
611613 " docker run --rm --ipc=host -it \\\n " ,
612614 " --ulimit stack=67108864 --ulimit memlock=-1 \\\n " ,
613615 " --gpus all -p 8000:8000 -e TRTLLM_ENABLE_PDL=1 \\\n " ,
640642 "metadata" : {},
641643 "outputs" : [],
642644 "source" : [
645+ " %%sh # [run in TensorRT-LLM container]\n " ,
646+ " \n " ,
643647 " git clone https://github.com/NVIDIA/TensorRT-Model-Optimizer.git"
644648 ]
645649 },
658662 "metadata" : {},
659663 "outputs" : [],
660664 "source" : [
665+ " %%sh # [run in TensorRT-LLM container]\n " ,
666+ " \n " ,
661667 " cd TensorRT-Model-Optimizer/\n " ,
662668 " pip install -e ."
663669 ]
679685 },
680686 "outputs" : [],
681687 "source" : [
688+ " %%sh # [run in TensorRT-LLM container]\n " ,
689+ " \n " ,
682690 " # set export path for converted checkpoints. The script saves the converted checkpoint in ${ROOT_SAVE_PATH}/saved_models_${MODEL_FULL_NAME}\n " ,
683691 " export ROOT_SAVE_PATH=/app/tensorrt_llm\n " ,
684692 " \n " ,
710718 "metadata" : {},
711719 "outputs" : [],
712720 "source" : [
721+ " %%sh # [run in TensorRT-LLM container]\n " ,
722+ " \n " ,
713723 " trtllm-serve /app/tensorrt_llm/saved_models_checkpoint-450_nvfp4_hf/ \\\n " ,
714724 " --max_batch_size 1 --max_num_tokens 1024 \\\n " ,
715725 " --max_seq_len 4096 --tp_size 8 --pp_size 1 \\\n " ,
803813 },
804814 {
805815 "cell_type" : " code" ,
806- "execution_count" : 13 ,
816+ "execution_count" : null ,
807817 "id" : " fb78741b-30cb-46f2-a292-c5192cbca9ed" ,
808818 "metadata" : {},
809819 "outputs" : [
You can’t perform that action at this time.
0 commit comments