Skip to content

Commit 731c640

Browse files
committed
modified the workbench image names
1 parent 493d235 commit 731c640

22 files changed

Lines changed: 94 additions & 79 deletions

docs/openshift-ai/data-science-project/using-projects-the-rhoai.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -131,11 +131,11 @@ On the Create workbench page, complete the following information.
131131

132132
Once you have entered the information for your workbench, click **Create**.
133133

134-
![Fill Workbench Information](images/tensorflow-workbench.png)
134+
![Fill Workbench Information](images/tensor flow-workbench.png)
135135

136136
For our example project, let's name it "Tensorflow Workbench". We'll select the
137-
**TensorFlow** image, choose a **Deployment size** of **Small**,
138-
**Accelerator** of **NVIDIA A100 GPU**, **Number of accelerators**
137+
**Jupyter | TensorFlow | CUDA | Python 3.12** image, choose a
138+
of **Small**, **Accelerator** of **NVIDIA A100 GPU**, **Number of accelerators**
139139
as **1** and allocate a **Cluster storage** space of **20GB** (Selected By Default).
140140

141141
After creating the workbench, you will return to your project page. It shows the

docs/openshift-ai/other-projects/RAG-talk-with-your-pdf.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -71,10 +71,10 @@ be populated with documents.
7171
![RAG Workbench Information](images/RAG-Jupyter-Notebook-Workbench.png)
7272

7373
For our example project, let's name it "RAG Workbench". We'll select the
74-
**TensorFlow** image with Recommended Version (selected by default), choose
75-
a **Deployment size** of **Medium**, **Accelerator** as **None** (no GPU is
76-
needed for this setup) and allocate a **Cluster storage** space of **20GB**
77-
(Selected By Default).
74+
**Jupyter | TensorFlow | CUDA | Python 3.12** image with Recommended Version
75+
(selected by default), choose a **Deployment size** of **Medium**, **Accelerator**
76+
as **None** (no GPU is needed for this setup) and allocate a **Cluster storage**
77+
space of **20GB** (Selected By Default).
7878

7979
!!! tip "Tip"
8080

docs/openshift-ai/other-projects/configure-jupyter-notebook-use-gpus-aiml-modeling.md

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -21,9 +21,10 @@ and Number of accelerators (GPUs).
2121
![PyTorch Workbench Information](images/pytorch-workbench.png)
2222

2323
For our example project, let's name it "PyTorch Workbench". We'll select the
24-
**PyTorch** image, choose a **Deployment size** of **Small**, choose **Accelerator**
25-
of **NVIDIA V100 GPU**, **Number of accelerators** as **1**, and allocate a
26-
**Cluster storage** space of **20GB** (Selected By Default).
24+
**Jupyter | PyTorch | CUDA | Python 3.12** image, choose a **Deployment size**
25+
of **Small**, choose **Accelerator** of **NVIDIA V100 GPU**,
26+
**Number of accelerators** as **1**, and allocate a **Cluster storage** space of
27+
**20GB** (Selected By Default).
2728

2829
!!! tip "Hardware Acceleration using GPU"
2930

docs/openshift-ai/other-projects/connect-vscode-to-rhoai-wb.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ Use the following steps to connect your local VS Code to RHOAI Workbench kernel:
7373
You can find this port from the service in your project with name same
7474
as your workbench.
7575

76-
- Open the Jupyter notebook in your local VS Code
76+
- Open the Jupyter notebook in your local VS Code:
7777

7878
![Jupyter Notebook](images/jupyter-nb.png)
7979

docs/openshift-ai/other-projects/deploying-a-llama-model-with-kserve.md

Lines changed: 69 additions & 55 deletions
Original file line numberDiff line numberDiff line change
@@ -381,64 +381,78 @@ and use `Llama 3.2 3B Modelcar` as the connection name, as shown below:
381381

382382
![Create Connection](images/create-connection-using-uri.png)
383383

384-
!!! note "Important Note: ModelCar Requirements & Guidance"
385-
386-
You have several options for deploying models to your OpenShift AI cluster.
387-
We recommend using **[ModelCar](https://kserve.github.io/website/docs/model-serving/storage/providers/oci#using-modelcars)**
388-
because it removes the need to manually download models from Hugging Face Hub,
389-
upload them to S3, or manage access permissions. With ModelCar, you can package
390-
models as OCI images and pull them at runtime or precache them. This simplifies
391-
versioning, improves traceability, and integrates cleanly into CI/CD workflows.
392-
ModelCar images also ensure reproducibility and maintain versioned model releases.
393-
394-
You can deploy our own model using a ModelCar container, which packages all
395-
model files into an OCI container image. To learn more about ModelCar containers,
396-
read this article **[Build and deploy a ModelCar container in OpenShift AI](https://developers.redhat.com/articles/2025/01/30/build-and-deploy-modelcar-container-openshift-ai)**.
397-
It explains the benefits of ModelCar containers, how to build a ModelCar image,
398-
and how to deploy it with OpenShift AI.
399-
400-
For additional patterns and prebuilt ModelCar images, explore the Red Hat AI
401-
Services **[ModelCar Catalog repository](https://github.com/redhat-ai-services/modelcar-catalog)**
402-
on GitHub. Prebuilt images from this catalog are also available in the
403-
**[ModelCar Catalog registry](https://quay.io/repository/redhat-ai-services/modelcar-catalog)**
404-
on *Quay.io*.
405-
406-
!!! tip "Use Any Other Available Model from the ModelCar Catalog registry."
407-
408-
**You can use any model from the ModelCar Catalog registry in a similar way.**
409-
For example, for the `Granite-3.3-8B-Instruct` model, you can use the publicly
410-
available container image from the **Quay.io** registry: **[quay.io/redhat-ai-services/modelcar-catalog:granite-3.3-8b-instruct](https://quay.io/repository/redhat-ai-services/modelcar-catalog?tag=granite-3.3-8b-instruct)**.
411-
412-
The **[Granite-3.3-8B-Instruct](https://huggingface.co/ibm-granite/granite-3.3-8b-instruct)**
413-
model is an 8-billion-parameter, 128K context-length language model fine-tuned
414-
for improved reasoning and instruction-following capabilities. It is built
415-
on top of the `Granite-3.3-8B-Base` model.
416-
417-
To create a connection for the `Granite-3.3-8B-Instruct` model, use the
418-
following URI:
419-
420-
```sh
421-
oci://quay.io/redhat-ai-services/modelcar-catalog:granite-3.3-8b-instruct
422-
```
384+
##### Using Publicly Available ModelCar Catalog registry
385+
386+
You have several options for deploying models to your OpenShift AI cluster.
387+
We recommend using **[ModelCar](https://kserve.github.io/website/docs/model-serving/storage/providers/oci#using-modelcars)**
388+
because it removes the need to manually download models from Hugging Face Hub,
389+
upload them to S3, or manage access permissions. With ModelCar, you can package
390+
models as OCI images and pull them at runtime or precache them. This simplifies
391+
versioning, improves traceability, and integrates cleanly into CI/CD workflows.
392+
ModelCar images also ensure reproducibility and maintain versioned model releases.
393+
394+
You can deploy your own model using a ModelCar container, which packages all
395+
model files into an OCI container image. To learn more about ModelCar containers,
396+
read this article **[Build and deploy a ModelCar container in OpenShift AI](https://developers.redhat.com/articles/2025/01/30/build-and-deploy-modelcar-container-openshift-ai)**.
397+
398+
It explains the benefits of ModelCar containers, how to build a ModelCar image,
399+
and how to deploy it with OpenShift AI.
400+
401+
For additional patterns and prebuilt ModelCar images, explore the Red Hat AI
402+
Services **[ModelCar Catalog repository](https://github.com/redhat-ai-services/modelcar-catalog)**
403+
on GitHub. Prebuilt images from this catalog are also available in the
404+
**[ModelCar Catalog registry](https://quay.io/repository/redhat-ai-services/modelcar-catalog)**
405+
on *Quay.io*.
423406

424-
In the **Additional serving runtime arguments** field under **Configuration
425-
parameters** section, specify the following recommended arguments:
426-
427-
```yaml
428-
--dtype=half
429-
--max-model-len=20000
430-
--gpu-memory-utilization=0.95
431-
--enable-chunked-prefill
432-
--enable-auto-tool-choice
433-
--tool-call-parser=granite
434-
--chat-template=/app/data/template/tool_chat_template_granite.jinja
435-
```
407+
!!! tip "Use Any Other Available Model from the ModelCar Catalog registry."
408+
409+
**You can use any model from the ModelCar Catalog registry in a similar way.**
410+
For example, for the `Granite-3.3-8B-Instruct` model, you can use the publicly
411+
available container image from the **Quay.io** registry: **[quay.io/redhat-ai-services/modelcar-catalog:granite-33-8b-instruct](https://quay.io/repository/redhat-ai-services/modelcar-catalog?tag=granite-3.3-8b-instruct)**.
412+
413+
The **[Granite-3.3-8B-Instruct](https://huggingface.co/ibm-granite/granite-3.3-8b-instruct)**
414+
model is an 8-billion-parameter, 128K context-length language model fine-tuned
415+
for improved reasoning and instruction-following capabilities. It is built
416+
on top of the `Granite-3.3-8B-Base` model.
417+
418+
To create a connection for the `Granite-3.3-8B-Instruct` model, use the
419+
following URI:
420+
421+
```sh
422+
oci://quay.io/redhat-ai-services/modelcar-catalog:granite-3.3-8b-instruct
423+
```
424+
425+
In the **Additional serving runtime arguments** field under **Configuration
426+
parameters** section, specify the following recommended arguments:
427+
428+
```yaml
429+
--dtype=half
430+
--max-model-len=20000
431+
--gpu-memory-utilization=0.95
432+
--enable-chunked-prefill
433+
--enable-auto-tool-choice
434+
--tool-call-parser=granite
435+
--chat-template=/app/data/template/tool_chat_template_granite.jinja
436+
```
437+
438+
Additionally, you may find it helpful to read **[Optimize and deploy LLMs for
439+
production with OpenShift AI](https://developers.redhat.com/articles/2025/10/06/optimize-and-deploy-llms-production-openshift-ai)**.
440+
441+
##### Using Model Catalog
442+
443+
Recent version of RHOAI include support for the **Model Catalog**, enabling users
444+
to easily discover, evaluate, and deploy generative AI models from a centralized
445+
interface. This feature provides access to models from multiple providers such
446+
as Red Hat, IBM, Meta, NVIDIA, Mistral AI, and Google, with built-in benchmarking
447+
based on open-source evaluation datasets to compare performance and quality. It
448+
simplifies the workflow by allowing data scientists and AI engineers to select
449+
suitable models, register them in a model registry, and deploy them directly to
450+
a serving runtime.
436451

437-
However, note that all these images are compiled for the **x86 architecture**.
438-
If you're targeting ARM, you'll need to rebuild these images on an ARM machine,
439-
as demonstrated in **[this guide](https://pandeybk.medium.com/serving-vllm-and-granite-models-on-arm-with-red-hat-openshift-ai-0178adba550e)**.
452+
![Model Catalog](images/model-catalog.png)
440453

441-
Additionally, you may find it helpful to read **[Optimize and deploy LLMs for production with OpenShift AI](https://developers.redhat.com/articles/2025/10/06/optimize-and-deploy-llms-production-openshift-ai)**.
454+
Models can be deployed directly from the model catalog to streamline the deployment
455+
process. For more details, refer to the [**official documentation**](https://docs.redhat.com/en/documentation/red_hat_openshift_ai_self-managed/3.0/html-single/working_with_model_registries/index#deploying-a-model-from-the-model-catalog_model-registry).
442456

443457
## Setting up Single-model Server and Deploy the model
444458

docs/openshift-ai/other-projects/fraud-detection-predictive-ai-app.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -647,10 +647,10 @@ look like this:
647647
![Fraud detection Workbench Information](images/fraud-detection-workbench.png)
648648

649649
For our example project, let's name it "Fraud detection". We'll select the
650-
**TensorFlow** image with Recommended Version (selected by default), choose
651-
a **Deployment size** of **Small**, choose **Accelerator** of
652-
**NVIDIA V100 GPU**, **Number of accelerators** as **1**, and allocate
653-
a **Cluster storage** space of **20GB** (Selected By Default).
650+
****Jupyter | TensorFlow | CUDA | Python 3.12** image with Recommended Version
651+
(selected by default), choose a **Deployment size** of **Small**, choose
652+
**Accelerator** of **NVIDIA V100 GPU**, **Number of accelerators** as **1**, and
653+
allocate a **Cluster storage** space of **20GB** (Selected By Default).
654654

655655
!!! info "Running Workbench without GPU"
656656

docs/openshift-ai/other-projects/how-access-s3-data-then-download-and-analyze-it.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,8 +16,8 @@ and Number of accelerators (GPUs).
1616
![Standard Data Science Workbech Information](images/standard-data-science-workbench.png)
1717

1818
For our example project, let's name it "Standard Data Science Workbench". We'll
19-
select the **Standard Data Science** image with Recommended Version
20-
(selected by default), choose a **Deployment size** of **Small**,
19+
select the **Jupyter | Data Science | CPU | Python 3.12** image with Recommended
20+
Version (selected by default), choose a **Deployment size** of **Small**,
2121
**Accelerator** as **None** (no GPU is needed for this setup), and allocate a
2222
**Cluster storage** space of **1GB**.
2323

28.5 KB
Loading
2.63 KB
Loading
12.7 KB
Loading

0 commit comments

Comments
 (0)