@@ -713,7 +713,7 @@ jobs:
713
713
echo "Chat sample python" passed
714
714
715
715
716
- visual_language_chat_sample-ubuntu :
716
+ visual_language_chat_sample-ubuntu-minicpm_v2_6 :
717
717
runs-on : ubuntu-22.04-16-cores
718
718
steps :
719
719
- uses : actions/checkout@v4
@@ -722,21 +722,13 @@ jobs:
722
722
- uses : actions/setup-python@v4
723
723
with :
724
724
python-version : 3.11
725
- - name : Install OpenVINO
726
- run : |
727
- mkdir ./ov/
728
- curl ${{ env.l_u22_ov_link }} | tar --directory ./ov/ --strip-components 1 -xz
729
- sudo ./ov/install_dependencies/install_openvino_dependencies.sh
730
- - name : Build app
731
- run : |
732
- source ./ov/setupvars.sh
733
- cmake -DCMAKE_BUILD_TYPE=Release -S ./ -B ./build/
734
- cmake --build ./build/ --config Release --target visual_language_chat py_openvino_genai -j
735
- - name : Install dependencies
736
- run : |
737
- source ./ov/setupvars.sh
738
- python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
739
- python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt opencv-python
725
+ - uses : ./.github/actions/install_openvino
726
+ with :
727
+ ov_link : ${{ env.l_u22_ov_link }}
728
+ - uses : ./.github/actions/build_app
729
+ with :
730
+ build_target : ' visual_language_chat py_openvino_genai'
731
+ - uses : ./.github/actions/install_python_deps
740
732
- name : Download and convert tiny-random-minicpmv-2_6 model and an image
741
733
run : |
742
734
python -m pip install git+https://github.com/eaidova/optimum-intel.git@ea/minicpmv
@@ -764,13 +756,6 @@ jobs:
764
756
&& ./build/samples/cpp/visual_language_chat/visual_language_chat ./tiny-random-minicpmv-2_6/ ./images/
765
757
<<< $'Describe the images?' | tee cpp.txt
766
758
timeout-minutes : 2
767
- - name : Encode cpp.txt with Python encoding instead of terminal one
768
- shell : python
769
- run : |
770
- with open("cpp.txt", "rb") as f:
771
- content = f.read().decode("utf-8", "replace")
772
- with open("cpp.txt", "wb") as f:
773
- f.write(content.encode("utf-8"))
774
759
- name : Run visual_language_chat Python sample - tiny-random-minicpmv-2_6
775
760
run : >
776
761
set -o pipefail
@@ -779,6 +764,13 @@ jobs:
779
764
<<< $'Describe the images?' | tee py.txt
780
765
env :
781
766
PYTHONPATH : " ./build/"
767
+ - name : Encode cpp.txt with Python encoding instead of terminal one
768
+ shell : python
769
+ run : |
770
+ with open("cpp.txt", "rb") as f:
771
+ content = f.read().decode("utf-8", "replace")
772
+ with open("cpp.txt", "wb") as f:
773
+ f.write(content.encode("utf-8"))
782
774
- run : diff cpp.txt py.txt
783
775
- name : Run visual_language_chat C++ sample with 2 prompts - tiny-random-minicpmv-2_6
784
776
run : >
@@ -803,39 +795,51 @@ jobs:
803
795
with open("cpp2.txt", "wb") as f:
804
796
f.write(content.encode("utf-8"))
805
797
- run : diff cpp2.txt py2.txt
806
- - name : Download and convert LLaVa 1.5 model and an image
807
- run : |
808
- source ./ov/setupvars.sh
809
- optimum-cli export openvino --model llava-hf/llava-1.5-7b-hf ./llava_1_5_7b_ov/
810
- wget https://llava-vl.github.io/static/images/monalisa.jpg
811
- - name : Run visual_language_chat C++ sample - LLaVa 1.5
812
- run : >
813
- source ./ov/setupvars.sh
814
- && ./build/samples/cpp/visual_language_chat/visual_language_chat ./llava_1_5_7b_ov/ monalisa.jpg
815
- <<< $'Who drew this painting?\nWhen did the painter live?'
816
- timeout-minutes : 4
817
- - name : Download and convert LLaVa-Next model
818
- run : |
819
- source ./ov/setupvars.sh
820
- optimum-cli export openvino --model llava-hf/llava-v1.6-mistral-7b-hf ./llava_v1_6_mistral_7b_ov/
821
- - name : Run visual_language_chat C++ sample - LLaVa-Next
822
- run : >
823
- source ./ov/setupvars.sh
824
- && ./build/samples/cpp/visual_language_chat/visual_language_chat ./llava_v1_6_mistral_7b_ov/ monalisa.jpg
825
- <<< $'Who drew this painting?\nWhen did the painter live?'
826
- timeout-minutes : 4
798
+
799
+ visual_language_chat_sample-ubuntu-llava_1_5 :
800
+ uses : ./.github/workflows/job_vlm_sample_llava.yml
801
+ with :
802
+ model_id : llava-hf/llava-1.5-7b-hf
803
+ model_dir : llava_1_5_7b_ov
804
+
805
+ visual_language_chat_sample-ubuntu-llava_next :
806
+ uses : ./.github/workflows/job_vlm_sample_llava.yml
807
+ with :
808
+ model_id : llava-hf/llava-v1.6-mistral-7b-hf
809
+ model_dir : llava_v1_6_mistral_7b_ov
810
+
811
+ visual_language_chat_sample-ubuntu-internvl2 :
812
+ runs-on : ubuntu-22.04-16-cores
813
+ steps :
814
+ - uses : actions/checkout@v4
815
+ with :
816
+ submodules : recursive
817
+ - uses : actions/setup-python@v4
818
+ with :
819
+ python-version : 3.11
820
+ - uses : ./.github/actions/install_openvino
821
+ with :
822
+ ov_link : ${{ env.l_u22_ov_link }}
823
+ - uses : ./.github/actions/build_app
824
+ with :
825
+ build_target : ' visual_language_chat py_openvino_genai'
826
+ - uses : ./.github/actions/install_python_deps
827
827
- name : Download and convert InternVL2 model
828
828
run : |
829
+ # Lowering transformers version, workaround for https://huggingface.co/OpenGVLab/InternVL2-1B/discussions/7
830
+ python -m pip install -U "transformers<4.45.0"
829
831
source ./ov/setupvars.sh
830
832
optimum-cli export openvino --model OpenGVLab/InternVL2-4B ./internvl2_4b_ov/ --trust-remote-code
833
+ - name : Download images
834
+ run : |
835
+ wget https://llava-vl.github.io/static/images/monalisa.jpg
831
836
- name : Run visual_language_chat C++ sample - InternVL2
832
837
run : >
833
838
source ./ov/setupvars.sh
834
839
&& ./build/samples/cpp/visual_language_chat/visual_language_chat ./internvl2_4b_ov/ monalisa.jpg
835
840
<<< $'Who drew this painting?\nWhen did the painter live?'
836
841
timeout-minutes : 4
837
842
838
-
839
843
cpp-continuous-batching-ubuntu :
840
844
runs-on : ubuntu-20.04-8-cores
841
845
defaults :
@@ -975,7 +979,7 @@ jobs:
975
979
cpp-greedy_causal_lm-Qwen-7B-Chat, cpp-beam_search_causal_lm-Qwen1_5-7B-Chat, cpp-beam_search_causal_lm-Phi-2,
976
980
cpp-beam_search_causal_lm-notus-7b-v1, cpp-speculative_decoding_lm-ubuntu, cpp-prompt_lookup_decoding_lm-ubuntu,
977
981
cpp-Phi-1_5, cpp-greedy_causal_lm-redpajama-3b-chat, cpp-chat_sample-ubuntu, cpp-continuous-batching-ubuntu,
978
- visual_language_chat_sample-ubuntu,
982
+ visual_language_chat_sample-ubuntu-minicpm_v2_6, visual_language_chat_sample-ubuntu-llava_1_5, visual_language_chat_sample-ubuntu-llava_next, visual_language_chat_sample-ubuntu-internvl2 ,
979
983
cpp-continuous-batching-windows, cpp-continuous-batching-macos]
980
984
if : ${{ always() }}
981
985
runs-on : ubuntu-latest
0 commit comments