@@ -14,6 +14,7 @@ concurrency:
14
14
15
15
env :
16
16
l_ov_link : https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2024.3.0-15945-a349dc82f9a/l_openvino_toolkit_ubuntu20_2024.3.0.dev20240708_x86_64.tgz
17
+ m_ov_link : https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2024.3.0-15945-a349dc82f9a/m_openvino_toolkit_macos_12_6_2024.3.0.dev20240708_x86_64.tgz
17
18
w_ov_link : https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2024.3.0-15945-a349dc82f9a/w_openvino_toolkit_windows_2024.3.0.dev20240708_x86_64.zip
18
19
jobs :
19
20
cpp-multinomial-greedy_causal_lm-ubuntu :
@@ -584,3 +585,119 @@ jobs:
584
585
timeout 30s ./samples/python/chat_sample/chat_sample.py ./TinyLlama-1.1B-Chat-v1.0/ < input.txt > ./pred2.txt
585
586
diff pred2.txt ref.txt
586
587
echo "Chat sample python" passed
588
+
589
+ cpp-continuous-batching-ubuntu :
590
+ runs-on : ubuntu-20.04-8-cores
591
+ steps :
592
+ - uses : actions/checkout@v4
593
+ with :
594
+ submodules : recursive
595
+ - uses : actions/setup-python@v4
596
+ with :
597
+ python-version : 3.8
598
+ - name : Install OpenVINO
599
+ run : |
600
+ mkdir ./ov/
601
+ curl ${{ env.l_ov_link }} | tar --directory ./ov/ --strip-components 1 -xz
602
+ sudo ./ov/install_dependencies/install_openvino_dependencies.sh
603
+ - name : Download, convert and build
604
+ run : |
605
+ source ./ov/setupvars.sh
606
+ python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
607
+ python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
608
+ optimum-cli export openvino --trust-remote-code --weight-format fp16 --model TinyLlama/TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v1.0
609
+ cmake -DCMAKE_BUILD_TYPE=Release -S ./ -B ./build/
610
+ cmake --build ./build/ --config Release -j
611
+ - name : Run gtests
612
+ run : |
613
+ source ./ov/setupvars.sh
614
+ ./build/tests/cpp/tests_continuous_batching
615
+ - name : Run accuracy_sample
616
+ run : |
617
+ source ./ov/setupvars.sh
618
+ timeout 50s ./build/samples/cpp/continuous_batching_accuracy/continuous_batching_accuracy -m ./TinyLlama-1.1B-Chat-v1.0/ -n 5
619
+ - name : Run throughput_benchmark
620
+ run : |
621
+ wget -q https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/resolve/main/ShareGPT_V3_unfiltered_cleaned_split.json
622
+ source ./ov/setupvars.sh
623
+ timeout 200s ./build/samples/cpp/continuous_batching_benchmark/continuous_batching_benchmark -n 10 --dynamic_split_fuse -m ./TinyLlama-1.1B-Chat-v1.0/ --dataset ./ShareGPT_V3_unfiltered_cleaned_split.json --cache_size 1
624
+
625
+
626
+ cpp-continuous-batching-windows :
627
+ runs-on : windows-latest
628
+ defaults :
629
+ run :
630
+ shell : cmd
631
+ steps :
632
+ - uses : actions/checkout@v4
633
+ with :
634
+ submodules : recursive
635
+ - uses : actions/setup-python@v4
636
+ with :
637
+ python-version : 3.8
638
+ - name : Install OpenVINO
639
+ run : |
640
+ curl --output ov.zip ${{ env.w_ov_link }}
641
+ unzip -d ov ov.zip
642
+ dirs=(ov/*) && mv ov/*/* ov && rmdir "${dirs[@]}"
643
+ shell : bash
644
+ - name : Install dependencies and build
645
+ run : |
646
+ call .\ov\setupvars.bat
647
+ python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
648
+ python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
649
+ optimum-cli export openvino --trust-remote-code --weight-format fp16 --model TinyLlama/TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v1.0
650
+ cmake -DCMAKE_BUILD_TYPE=Releas -S ./ -B ./build/
651
+ cmake --build ./build/ --config Release -j
652
+ - name : Run gtests
653
+ run : |
654
+ set PATH=.\build\openvino_genai\;%PATH%
655
+ call .\ov\setupvars.bat
656
+ .\build\tests\cpp\Release\tests_continuous_batching.exe
657
+ - name : Run accuracy_sample
658
+ run : |
659
+ set PATH=.\build\openvino_genai\;%PATH%
660
+ call .\ov\setupvars.bat
661
+ .\build\samples\cpp\continuous_batching_accuracy\Release\continuous_batching_accuracy.exe -m .\TinyLlama-1.1B-Chat-v1.0\ -n 5
662
+ - name : Run throughput_benchmark
663
+ run : |
664
+ curl -o .\ShareGPT_V3_unfiltered_cleaned_split.json -s -L "https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/resolve/main/ShareGPT_V3_unfiltered_cleaned_split.json"
665
+ set PATH=.\build\openvino_genai\;%PATH%
666
+ call .\ov\setupvars.bat
667
+ .\build\samples\cpp\continuous_batching_benchmark\Release\continuous_batching_benchmark.exe -n 2 --dynamic_split_fuse -m .\TinyLlama-1.1B-Chat-v1.0\ --dataset .\ShareGPT_V3_unfiltered_cleaned_split.json --cache_size 1
668
+
669
+ cpp-continuous-batching-macos :
670
+ runs-on : macos-12
671
+ steps :
672
+ - uses : actions/checkout@v4
673
+ with :
674
+ submodules : recursive
675
+ - uses : actions/setup-python@v4
676
+ with :
677
+ python-version : 3.8
678
+ - name : Install OpenVINO
679
+ run : |
680
+ mkdir ./ov/
681
+ curl ${{ env.m_ov_link }} | tar --directory ./ov/ --strip-components 1 -xz
682
+ brew install coreutils scons
683
+ - name : Download, convert and build
684
+ run : |
685
+ source ./ov/setupvars.sh
686
+ python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
687
+ python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
688
+ optimum-cli export openvino --trust-remote-code --weight-format fp16 --model TinyLlama/TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v1.0
689
+ cmake -DCMAKE_BUILD_TYPE=Release -S ./ -B ./build/
690
+ cmake --build ./build/ --config Release -j
691
+ - name : Run gtests
692
+ run : |
693
+ source ./ov/setupvars.sh
694
+ ./build/tests/cpp/tests_continuous_batching
695
+ - name : Run accuracy_sample
696
+ run : |
697
+ source ./ov/setupvars.sh
698
+ timeout 120s ./build/samples/cpp/continuous_batching_accuracy/continuous_batching_accuracy -m ./TinyLlama-1.1B-Chat-v1.0/ -n 5
699
+ - name : Run throughput_benchmark
700
+ run : |
701
+ wget -q https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/resolve/main/ShareGPT_V3_unfiltered_cleaned_split.json
702
+ source ./ov/setupvars.sh
703
+ ./build/samples/cpp/continuous_batching_benchmark/continuous_batching_benchmark -n 5 --dynamic_split_fuse -m ./TinyLlama-1.1B-Chat-v1.0/ --dataset ./ShareGPT_V3_unfiltered_cleaned_split.json --cache_size 1
0 commit comments