Skip to content

Commit 95591b6

Browse files
authored
Merge branch 'main' into sramakr1/refactor_xpu_layers
2 parents e2e26e9 + 1ae2b58 commit 95591b6

21 files changed

+79
-22
lines changed

.github/workflows/container-ci.yaml

+4
Original file line numberDiff line numberDiff line change
@@ -158,6 +158,10 @@ jobs:
158158
# password: ${{ secrets.REGISTRY_TOKEN }}
159159
- name: Pull Image
160160
run: docker pull ${{ secrets.REGISTRY }}/${{ secrets.REPO }}:${{ matrix.container }}
161+
- uses: actions/cache@3624ceb22c1c5a301c8db4169662070a689d9ea8 # v4.1.1
162+
with:
163+
path: ${{ github.workspace }}/.cache/trivy
164+
key: ${{ github.head_ref || github.ref_name }}-trivy
161165
- name: Scan Container
162166
uses: aquasecurity/trivy-action@915b19bbe73b92a6cf82a1bc12b087c9a19a5fe2 # v0.28.0
163167
with:

jax/README.md

+7-7
Original file line numberDiff line numberDiff line change
@@ -8,13 +8,13 @@ The images below include [JAX\*] and [Intel® Extension for OpenXLA\*].
88

99
| Tag(s) | [JAX\*] | [Intel® Extension for OpenXLA\*] | [Flax] | Dockerfile |
1010
| -------------------------- | --------- | -------------------------------- | -------- | --------------- |
11-
| `0.4.0-pip-base`, `latest` | [v0.4.32] | [v0.4.0-jax] | [v0.9.0] | [v0.4.0] |
11+
| `0.4.0-pip-base`, `latest` | [v0.4.26] | [v0.4.0-jax] | [v0.8.2] | [v0.4.1] |
1212

1313
The images below additionally include [Jupyter Notebook](https://jupyter.org/) server:
1414

1515
| Tag(s) | [JAX\*] | [Intel® Extension for OpenXLA\*] | [Flax] | Dockerfile |
1616
| ------------------- | --------- | ----------------- | -------- | --------------- |
17-
| `0.4.0-pip-jupyter` | [v0.4.32] | [v0.4.0-jax] | [v0.9.0] | [v0.4.0] |
17+
| `0.4.0-pip-jupyter` | [v0.4.26] | [v0.4.0-jax] | [v0.8.2] | [v0.4.1] |
1818

1919
### Run the Jupyter Container
2020

@@ -35,13 +35,13 @@ The images below include [Intel® Distribution for Python*]:
3535

3636
| Tag(s) | [JAX\*] | [Intel® Extension for OpenXLA\*] | [Flax] | Dockerfile |
3737
| ---------------- | --------- | ----------------- | -------- | --------------- |
38-
| `0.4.0-idp-base` | [v0.4.32] | [v0.4.0-jax] | [v0.9.0] | [v0.4.0] |
38+
| `0.4.0-idp-base` | [v0.4.26] | [v0.4.0-jax] | [v0.8.2] | [v0.4.1] |
3939

4040
The images below additionally include [Jupyter Notebook](https://jupyter.org/) server:
4141

4242
| Tag(s) | [JAX\*] | [Intel® Extension for OpenXLA\*] | [Flax] | Dockerfile |
4343
| ------------------- | --------- | ----------------- | -------- | --------------- |
44-
| `0.4.0-idp-jupyter` | [v0.4.32] | [v0.4.0-jax] | [v0.9.0] | [v0.4.0] |
44+
| `0.4.0-idp-jupyter` | [v0.4.26] | [v0.4.0-jax] | [v0.8.2] | [v0.4.1] |
4545

4646
## Build from Source
4747

@@ -77,10 +77,10 @@ It is the image user's responsibility to ensure that any use of The images below
7777
[JAX\*]: https://github.com/google/jax
7878
[Flax]: https://github.com/google/flax
7979

80-
[v0.4.32]: https://github.com/google/jax/releases/tag/jax-v0.4.32
80+
[v0.4.26]: https://github.com/google/jax/releases/tag/jax-v0.4.26
8181

8282
[v0.4.0-jax]: https://github.com/intel/intel-extension-for-openxla/releases/tag/0.4.0
8383

84-
[v0.9.0]: https://github.com/google/Flax/releases/tag/v0.9.0
84+
[v0.8.2]: https://github.com/google/Flax/releases/tag/v0.8.2
8585

86-
[v0.4.0]: https://github.com/intel/ai-containers/blob/v0.4.0/jax/Dockerfile
86+
[v0.4.1]: https://github.com/intel/ai-containers/blob/main/jax/Dockerfile

workflows/charts/tensorflow-serving/Chart.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
# limitations under the License.
1414

1515
apiVersion: v2
16-
name: tensorflow-serving-on-intel
16+
name: tensorflow-serving
1717
description: TensorFlow Serving is a flexible, high-performance serving system for machine learning models, designed for production environments. TensorFlow Serving makes it easy to deploy new algorithms and experiments, while keeping the same server architecture and APIs. TensorFlow Serving provides out-of-the-box integration with TensorFlow models, but can be easily extended to serve other types of models and data.
1818

1919
# A chart can be either an 'application' or a 'library' chart.

workflows/charts/tensorflow-serving/README.md

+9-6
Original file line numberDiff line numberDiff line change
@@ -1,14 +1,17 @@
1-
# tensorflow-serving-on-intel
1+
# TensorFlow Serving on Intel GPU
2+
3+
TensorFlow Serving is a flexible, high-performance serving system for machine learning models, designed for production environments. TensorFlow Serving makes it easy to deploy new algorithms and experiments, while keeping the same server architecture and APIs. TensorFlow Serving provides out-of-the-box integration with TensorFlow models, but can be easily extended to serve other types of models and data.
24

35
![Version: 0.1.0](https://img.shields.io/badge/Version-0.1.0-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) ![AppVersion: 1.16.0](https://img.shields.io/badge/AppVersion-1.16.0-informational?style=flat-square)
46

5-
TensorFlow Serving is a flexible, high-performance serving system for machine learning models, designed for production environments. TensorFlow Serving makes it easy to deploy new algorithms and experiments, while keeping the same server architecture and APIs. TensorFlow Serving provides out-of-the-box integration with TensorFlow models, but can be easily extended to serve other types of models and data.
7+
## Installation
68

7-
## Maintainers
9+
```bash
10+
helm repo add intel https://intel.github.io/helm-charts
11+
helm install my-tgi intel/text-generation-inference
12+
```
813

9-
| Name | Email | Url |
10-
| ---- | ------ | --- |
11-
| tylertitsworth | <tyler.titsworth@intel.com> | <https://github.com/tylertitsworth> |
14+
Then, follow the installation notes to test the deployment
1215

1316
## Values
1417

Original file line numberDiff line numberDiff line change
@@ -0,0 +1,18 @@
1+
# TensorFlow Serving on Intel GPU
2+
3+
{{ template "chart.description" . }}
4+
5+
{{ template "chart.versionBadge" . }}{{ template "chart.typeBadge" . }}{{ template "chart.appVersionBadge" . }}
6+
7+
## Installation
8+
9+
```bash
10+
helm repo add intel https://intel.github.io/helm-charts
11+
helm install my-tgi intel/text-generation-inference
12+
```
13+
14+
Then, follow the installation notes to test the deployment
15+
16+
{{ template "chart.valuesSection" . }}
17+
18+
{{ template "helm-docs.versionFooter" . }}

workflows/charts/tgi/Chart.yaml workflows/charts/text-generation-inference/Chart.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
# limitations under the License.
1414

1515
apiVersion: v2
16-
name: tgi-on-intel
16+
name: text-generation-inference
1717
description: A Rust, Python and gRPC server for text generation inference by huggingface on Intel GPUs.
1818

1919
# A chart can be either an 'application' or a 'library' chart.

workflows/charts/tgi/README.md workflows/charts/text-generation-inference/README.md

+9
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,15 @@ For more information about how to use Huggingface text-generation-inference with
99
1010
![Version: 0.2.1](https://img.shields.io/badge/Version-0.2.1-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) ![AppVersion: 1.16.0](https://img.shields.io/badge/AppVersion-1.16.0-informational?style=flat-square)
1111

12+
## Installation
13+
14+
```bash
15+
helm repo add intel https://intel.github.io/helm-charts
16+
helm install my-tgi intel/text-generation-inference
17+
```
18+
19+
Then, follow the installation notes to test the deployment
20+
1221
## Values
1322

1423
| Key | Type | Default | Description |

workflows/charts/tgi/README.md.gotmpl workflows/charts/text-generation-inference/README.md.gotmpl

+8-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,14 @@ For more information about how to use Huggingface text-generation-inference with
99

1010
{{ template "chart.versionBadge" . }}{{ template "chart.typeBadge" . }}{{ template "chart.appVersionBadge" . }}
1111

12-
{{ template "chart.requirementsSection" . }}
12+
## Installation
13+
14+
```bash
15+
helm repo add intel https://intel.github.io/helm-charts
16+
helm install my-tgi intel/text-generation-inference
17+
```
18+
19+
Then, follow the installation notes to test the deployment
1320

1421
{{ template "chart.valuesSection" . }}
1522

workflows/charts/torchserve/Chart.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,8 @@
1313
# limitations under the License.
1414

1515
apiVersion: v2
16-
name: torchserve-on-intel
17-
description: TorchServe on Intel is a performant, flexible and easy to use tool for serving PyTorch models in production.
16+
name: torchserve
17+
description: TorchServe is a performant, flexible and easy to use tool for serving PyTorch models in production on Intel GPUs.
1818

1919
# A chart can be either an 'application' or a 'library' chart.
2020
#

workflows/charts/torchserve/README.md

+11-2
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,20 @@
11
# TorchServe with Intel Optimizations
22

3-
TorchServe on Intel is a performant, flexible and easy to use tool for serving PyTorch models in production.
3+
TorchServe is a performant, flexible and easy to use tool for serving PyTorch models in production on Intel GPUs.
44

5-
For more information about how to use TorchServe with Intel Optimizations, check out the [container documentation](../../../pytorch/serving/README.md).
5+
For more information about how to use TorchServe with Intel Optimizations, check out the [container documentation](https://github.com/intel/ai-containers/blob/main/pytorch/serving/README.md).
66

77
![Version: 0.1.0](https://img.shields.io/badge/Version-0.1.0-informational?style=flat-square) ![Type: application](https://img.shields.io/badge/Type-application-informational?style=flat-square) ![AppVersion: 1.16.0](https://img.shields.io/badge/AppVersion-1.16.0-informational?style=flat-square)
88

9+
## Installation
10+
11+
```bash
12+
helm repo add intel https://intel.github.io/helm-charts
13+
helm install my-tgi intel/text-generation-inference
14+
```
15+
16+
Then, follow the installation notes to test the deployment
17+
918
## Values
1019

1120
| Key | Type | Default | Description |

workflows/charts/torchserve/README.md.gotmpl

+9-2
Original file line numberDiff line numberDiff line change
@@ -2,11 +2,18 @@
22

33
{{ template "chart.description" . }}
44

5-
For more information about how to use TorchServe with Intel Optimizations, check out the [container documentation](../../../pytorch/serving/README.md).
5+
For more information about how to use TorchServe with Intel Optimizations, check out the [container documentation](https://github.com/intel/ai-containers/blob/main/pytorch/serving/README.md).
66

77
{{ template "chart.versionBadge" . }}{{ template "chart.typeBadge" . }}{{ template "chart.appVersionBadge" . }}
88

9-
{{ template "chart.requirementsSection" . }}
9+
## Installation
10+
11+
```bash
12+
helm repo add intel https://intel.github.io/helm-charts
13+
helm install my-tgi intel/text-generation-inference
14+
```
15+
16+
Then, follow the installation notes to test the deployment
1017

1118
{{ template "chart.valuesSection" . }}
1219

0 commit comments

Comments
 (0)