You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: image_generation/lcm_dreamshaper_v7/cpp/README.md
+3-8
Original file line number
Diff line number
Diff line change
@@ -18,7 +18,7 @@ Prepare a python environment and install dependencies:
18
18
conda create -n openvino_lcm_cpp python==3.10
19
19
conda activate openvino_lcm_cpp
20
20
conda update -c conda-forge --all
21
-
conda install -c conda-forge openvino=2024.1.0 c-compiler cxx-compiler make cmake
21
+
conda install -c conda-forge openvino=2024.1.0 c-compiler cxx-compiler git make cmake
22
22
# Ensure that Conda standard libraries are used
23
23
conda env config vars set LD_LIBRARY_PATH=$CONDA_PREFIX/lib:$LD_LIBRARY_PATH
24
24
```
@@ -37,13 +37,8 @@ conda env config vars set LD_LIBRARY_PATH=$CONDA_PREFIX/lib:$LD_LIBRARY_PATH
37
37
```
38
38
39
39
2. Download the model from Huggingface and convert it to OpenVINO IR via [optimum-intel CLI](https://github.com/huggingface/optimum-intel). Example commandfor downloading and exporting FP16 model:
# Converting tokenizer manually (`--convert-tokenizer` flag of `optimum-cli` results in "OpenVINO Tokenizer export for CLIPTokenizer is not supported.")
Copy file name to clipboardexpand all lines: text_generation/causal_lm/cpp/README.md
+2-4
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
# Text generation C++ samples that support most popular models like LLaMA 2
2
2
3
-
These examples showcase inference of text-generation Large Language Models (LLMs): `chatglm`, `LLaMA`, `Qwen` and other models with the same signature. The applications don't have many configuration options to encourage the reader to explore and modify the source code. Loading `openvino_tokenizers` to `ov::Core` enables tokenization. Run `convert_tokenizer` to generate IRs for the samples. [group_beam_searcher.hpp](group_beam_searcher.hpp) implements the algorithm of the same name, which is used by `beam_search_causal_lm`. There is also a Jupyter [notebook](https://github.com/openvinotoolkit/openvino_notebooks/tree/main/notebooks/254-llm-chatbot) which provides an example of LLM-powered Chatbot in Python.
3
+
These examples showcase inference of text-generation Large Language Models (LLMs): `chatglm`, `LLaMA`, `Qwen` and other models with the same signature. The applications don't have many configuration options to encourage the reader to explore and modify the source code. Loading `openvino_tokenizers` to `ov::Core` enables tokenization. Run `optimum-cli` to generate IRs for the samples. [group_beam_searcher.hpp](group_beam_searcher.hpp) implements the algorithm of the same name, which is used by `beam_search_causal_lm`. There is also a Jupyter [notebook](https://github.com/openvinotoolkit/openvino_notebooks/tree/main/notebooks/254-llm-chatbot) which provides an example of LLM-powered Chatbot in Python.
4
4
5
5
## How it works
6
6
@@ -49,7 +49,7 @@ This approach reduces the need for multiple infer requests to the main model, en
49
49
50
50
## Install OpenVINO
51
51
52
-
Install [OpenVINO Archives >= 2024.0](docs.openvino.ai/install). `master` and possibly the latest `releases/*` branch correspond to not yet released OpenVINO versions. https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/ can be used for these branches early testing. `<INSTALL_DIR>` below refers to the extraction location.
52
+
Install [OpenVINO Archives >= 2024.1](docs.openvino.ai/install). `master` and possibly the latest `releases/*` branch correspond to not yet released OpenVINO versions. https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/ can be used for these branches early testing. `<INSTALL_DIR>` below refers to the extraction location.
53
53
54
54
## Build `greedy_causal_lm`, `beam_search_causal_lm` and `openvino_tokenizers`
0 commit comments