File tree Expand file tree Collapse file tree 1 file changed +4
-4
lines changed
Expand file tree Collapse file tree 1 file changed +4
-4
lines changed Original file line number Diff line number Diff line change @@ -767,9 +767,9 @@ jobs:
767767 pushd optimum-executorch
768768 # There is no release yet, for CI stability, always test from the same commit on main
769769 git checkout $OPTIMUM_ET_COMMIT
770- python install_dev.py --skip_override_torch
770+ ${CONDA_RUN} python install_dev.py --skip_override_torch
771771 popd
772- pip list
772+ ${CONDA_RUN} pip list
773773 echo "::endgroup::"
774774
775775 echo "::group::Export to ExecuTorch"
@@ -787,15 +787,15 @@ jobs:
787787 "--output_dir" "${OUTPUT_DIR}"
788788 )
789789
790- optimum-cli export executorch "${ARGS[@]}"
790+ ${CONDA_RUN} optimum-cli export executorch "${ARGS[@]}"
791791
792792 ls -FlAGhp ${OUTPUT_DIR}
793793 popd
794794 echo "::endgroup::"
795795
796796 echo "::group::Inference using python API"
797797 pushd optimum-executorch
798- python -c "
798+ ${CONDA_RUN} python -c "
799799 import os
800800 from optimum.executorch import ExecuTorchModelForCausalLM
801801 from transformers import AutoTokenizer
You can’t perform that action at this time.
0 commit comments