File tree Expand file tree Collapse file tree 2 files changed +22
-8
lines changed
fluid/inference/api/demo_ci Expand file tree Collapse file tree 2 files changed +22
-8
lines changed Original file line number Diff line number Diff line change 2
2
PADDLE_ROOT=$1
3
3
TURN_ON_MKL=$2 # use MKL or Openblas
4
4
TEST_GPU_CPU=$3 # test both GPU/CPU mode or only CPU mode
5
+ DATA_DIR=$4 # dataset
6
+ cd ` dirname $0 `
7
+ current_dir=` pwd`
5
8
if [ $2 == ON ]; then
6
9
# You can export yourself if move the install path
7
10
MKL_LIB=${PADDLE_ROOT} /build/fluid_install_dir/third_party/install/mklml/lib
@@ -29,15 +32,15 @@ function download() {
29
32
fi
30
33
cd ..
31
34
}
32
- mkdir -p data
33
- cd data
35
+ mkdir -p $DATA_DIR
36
+ cd $DATA_DIR
34
37
vis_demo_list=' se_resnext50 ocr mobilenet'
35
38
for vis_demo_name in $vis_demo_list ; do
36
39
download $vis_demo_name
37
40
done
38
- cd ..
39
41
40
42
# compile and test the demo
43
+ cd $current_dir
41
44
mkdir -p build
42
45
cd build
43
46
@@ -73,9 +76,9 @@ for WITH_STATIC_LIB in ON OFF; do
73
76
for use_gpu in $use_gpu_list ; do
74
77
for vis_demo_name in $vis_demo_list ; do
75
78
./vis_demo \
76
- --modeldir=../data /$vis_demo_name /model \
77
- --data=../data /$vis_demo_name /data.txt \
78
- --refer=../data /$vis_demo_name /result.txt \
79
+ --modeldir=$DATA_DIR /$vis_demo_name /model \
80
+ --data=$DATA_DIR /$vis_demo_name /data.txt \
81
+ --refer=$DATA_DIR /$vis_demo_name /result.txt \
79
82
--use_gpu=$use_gpu
80
83
if [ $? -ne 0 ]; then
81
84
echo " vis demo $vis_demo_name runs fail."
Original file line number Diff line number Diff line change @@ -654,11 +654,21 @@ function gen_fluid_inference_lib() {
654
654
if [[ ${WITH_C_API:- OFF} == " OFF" && ${WITH_INFERENCE:- ON} == " ON" ]] ; then
655
655
cat << EOF
656
656
========================================
657
- Deploying fluid inference library ...
657
+ Generating fluid inference library ...
658
658
========================================
659
659
EOF
660
660
cmake .. -DWITH_DISTRIBUTE=OFF
661
661
make -j ` nproc` inference_lib_dist
662
+ fi
663
+ }
664
+
665
+ function tar_fluid_inference_lib() {
666
+ if [[ ${WITH_C_API:- OFF} == " OFF" && ${WITH_INFERENCE:- ON} == " ON" ]] ; then
667
+ cat << EOF
668
+ ========================================
669
+ Taring fluid inference library ...
670
+ ========================================
671
+ EOF
662
672
cd ${PADDLE_ROOT} /build
663
673
cp -r fluid_install_dir fluid
664
674
tar -czf fluid.tgz fluid
@@ -673,7 +683,7 @@ function test_fluid_inference_lib() {
673
683
========================================
674
684
EOF
675
685
cd ${PADDLE_ROOT} /paddle/fluid/inference/api/demo_ci
676
- ./run.sh ${PADDLE_ROOT} ${WITH_MKL:- ON} ${WITH_GPU:- OFF}
686
+ ./run.sh ${PADDLE_ROOT} ${WITH_MKL:- ON} ${WITH_GPU:- OFF} ${INFERENCE_DEMO_INSTALL_DIR}
677
687
./clean.sh
678
688
fi
679
689
}
@@ -722,6 +732,7 @@ function main() {
722
732
fluid_inference_lib)
723
733
cmake_gen ${PYTHON_ABI:- " " }
724
734
gen_fluid_inference_lib
735
+ tar_fluid_inference_lib
725
736
test_fluid_inference_lib
726
737
;;
727
738
check_style)
You can’t perform that action at this time.
0 commit comments