38
38
- name : CMake - build
39
39
run : cmake --build build -j`nproc` -- llama_cpp_plugin llama_cpp_e2e_tests
40
40
41
+
41
42
- name : Upload build artifacts
42
- uses : alehechka /upload-tartifact@v2
43
+ uses : actions /upload-artifact@v4
43
44
with :
44
45
name : build_artifacts
45
46
path : ${{ github.workspace }}/openvino/bin/intel64/Release/
@@ -49,20 +50,23 @@ jobs:
49
50
runs-on : ubuntu-20.04
50
51
steps :
51
52
- name : Download build artifacts
52
- uses : alehechka /download-tartifact@v2
53
+ uses : actions /download-artifact@v4
53
54
with :
54
55
name : build_artifacts
55
- path : binaries
56
+ path : ${{ github.workspace }}/ binaries
56
57
57
- - name : Prepare test data
58
+ - name : Prepare test data - checkout llama.cpp repo
58
59
uses : actions/checkout@v3
59
60
with :
60
61
repository : ggerganov/llama.cpp
61
62
path : llama.cpp
62
- run : pip install llama.cpp/requirements/requirements-convert-hf-to-gguf.txt
63
- run : huggingface-cli huggingface-cli download gpt2 model.safetensors tokenizer.json tokenizer_config.json vocab.json config.json merges.txt --local-dir hf_gpt2
64
- run : mkdir -p ${{ github.workspace }}/test_data
65
- run : python3 llama.cpp/convert-hf-to-gguf.py hf_gpt2 --outtype f32 --outfile ${{ github.workspace }}/test_data/gpt2.gguf
63
+
64
+ - name : Prepare test data - convert test model files
65
+ run : |
66
+ pip install -r llama.cpp/requirements/requirements-convert-hf-to-gguf.txt
67
+ huggingface-cli download gpt2 model.safetensors tokenizer.json tokenizer_config.json vocab.json config.json merges.txt --local-dir hf_gpt2
68
+ mkdir -p ${{ github.workspace }}/test_data
69
+ python3 llama.cpp/convert-hf-to-gguf.py hf_gpt2 --outtype f32 --outfile ${{ github.workspace }}/test_data/gpt2.gguf
66
70
67
71
- name : Run E2E tests
68
72
run : ${{ github.workspace }}/binaries/llama_cpp_e2e_tests
0 commit comments