Pre-requisites for Deploying NAI
In this part of the lab we will prepare pre-requisites for LLM application on GPU nodes.
The following is the flow of the applications lab:
stateDiagram-v2
direction LR
state PreRequisites {
[*] --> CreateFilesShare
CreateFilesShare --> PrepareHuggingFace
PrepareHuggingFace --> [*]
}
state CreateOfflineHelmContainers {
[*] --> PrepareNAIHelmCharts
PrepareNAIHelmCharts --> PrepareNAIContainerImages
PrepareNAIContainerImages --> [*]
}
[*] --> PreRequisites
PreRequisites --> CreateOfflineHelmContainers
CreateOfflineHelmContainers --> DeployNAI : next section
DeployNAI --> TestNAI
TestNAI --> [*]
Prepare the following pre-requisites needed to deploy NAI on target kubernetes cluster.
Create Nutanix Files Storage Class
We will create Nutanix Files storage class which will be used to create a pvc that will store the LLama-3-8B model files.
- In Prism Central, choose Files from the menu
- Choose the file server (e.g. labFS)
- Click on Shares & Exports
- Click on +New Share or Export
-
Fill the details of the Share
- Name - model_share
- Description - for NAI model store
- Share path - leave blank
- Max Size - 10 GiB (adjust to the model file size)
- Primary Protocol Access - NFS
-
Click Next and make sure Enable compression in checked
- Click Next
-
In NFS Protocol Access, choose the following:
- Authentication - System
- Default Access (for all clients) - Read-Write
- Squash - Root Squash
Note
Consider changing access options for Production environment
-
Click Next
- Confirm the share details and click on Create
Create the Files Storage Class
-
Run the following command to check K8S status of the
nkpdevcluster -
In VSC Explorer, click on New File and create a config file with the following name:
Add the following content and replace the
nfsServerNamewith the name of the Nutanix Files server name .
-
Create the storage class
-
Check storage classes in the cluster for the Nutanix Files storage class
kubectl get storageclass NAME PROVISIONER RECLAIMPOLICY VOLUMEBINDINGMODE ALLOWVOLUMEEXPANSION AGE dkp-object-store kommander.ceph.rook.io/bucket Delete Immediate false 28h nai-nfs-storage csi.nutanix.com Delete Immediate true 24h nutanix-volume (default) csi.nutanix.com Delete WaitForFirstConsumer false 28h
Request Access to Model on Hugging Face
Follow these steps to request access to the meta-llama/Meta-Llama-3.1-8B-Instruct model:
LLM Recommendation
From testing google/gemma-2-2b-it model is quicker to download and obtain download rights, than meta-llama/Meta-Llama-3.1-8B-Instruct model.
Feel free to use the google/gemma-2-2b-it model if necessary. The procedure to request access to the model is the same.
-
Sign in to your Hugging Face account:
- Visit Hugging Face and log in to your account.
-
Navigate to the model page:
- Go to the Meta-Llama-3.1-8B-Instruct model page.
-
Request access:
- On the model page, you will see a section or button labeled Request Access (this is usually near the top of the page or near the "Files and versions" section).
- Click Request Access.
-
Complete the form:
- You may be prompted to fill out a form or provide additional details about your intended use of the model.
- Complete the required fields and submit the request.
-
Wait for approval:
- After submitting your request, you will receive a notification or email once your access is granted.
- This process can take some time depending on the approval workflow.
Once access is granted, there will be an email notification.
Note
Email from Hugging Face can take a few minutes or hours before it arrives.
Create a Hugging Face Token with Read Permissions
Follow these steps to create a Hugging Face token with read permissions:
-
Sign in to your Hugging Face account:
- Visit Hugging Face and log in to your account.
-
Access your account settings:
- Click on your profile picture in the top-right corner.
- From the dropdown, select Settings.
-
Navigate to the "Access Tokens" section:
- In the sidebar, click on Access Tokens.
- You will see a page where you can create and manage tokens.
-
Create a new token:
- Click the New token button.
- Enter a name for your token (i.e.,
read-only-token).
-
Set token permissions:
- Under the permissions dropdown, select Read. For Example:

- Under the permissions dropdown, select Read. For Example:
-
Create and copy the token:
- After selecting the permissions, click Create.
- Your token will be generated and displayed only once, so make sure to copy it and store it securely.
Use this token for accessing Hugging Face resources with read-only permissions.
Prepare Helm Charts
In this section we will prepare the helm charts necessary for NAI and pre-requisite applications install
- NAI
- Envoy Gateway
- Kserve
- OpenTelemetry Operator
The procedure will be done on the jumphost VM.
-
Login to Nutanix Portal using your credentials
-
Go to Downloads > NAI Airgapped Bundle
-
Download and extract the NAI air-gap
helmbundle from the link you copied earlier- Open new
VSCodewindow on your jumphost VM
- Open new
-
In
VSCodeExplorer pane, click on existing$HOMEfolder -
Click on New Folder name it:
airgap-nai -
On
VSCodeExplorer plane, click the$HOME/airgap-naifolder -
On
VSCodemenu, selectTerminal>New Terminal -
Browse to
airgap-naidirectory -
In
VSC, under the newly createdairgap-naifolder, click on New File and create file with the following name: -
Add (append) the following environment variables and save it
-
Source the
.envfile to import environment variables -
Download the NAI
2.5.0helm chart bundle from Nutanix Portal -
Login to Harbor registry on the command line (if not done so)
-
Upload the downloaded and prepared helm charts to Harbor
for chart in $(ls *.tgz); do echo $chart;helm push $chart oci://$REGISTRY_HOST;done gateway-crds-helm-v1.5.0.tgz Pushed: harbor.10.x.x.134.nip.io/nkp/gateway-crds-helm:v1.5.0 Digest: sha256:d8f06d0d28bbcb40188f211be0fc3f0dee20376c5ad59935bf68bc5cc67149fc gateway-helm-v1.5.0.tgz Pushed: harbor.10.x.x.134.nip.io/nkp/gateway-helm:v1.5.0 Digest: sha256:4e49511296e23e3d1400c92cfb38a5c26030501ec7353883e4ccad9fd7cc4c2c kserve-crd-v0.15.0.tgz Pushed: harbor.10.x.x.134.nip.io/nkp/kserve-crd:v0.15.0 Digest: sha256:d540337971ae91fbd7593764eb90d8689012bc835f5d2f8ee42af47543a18403 kserve-v0.15.0.tgz Pushed: harbor.10.x.x.134.nip.io/nkp/kserve:v0.15.0 Digest: sha256:cafd90ab1d91a54a28c1ff2761d976bdda0bb173675ef392a16ac250b044d15f nai-core-2.5.0.tgz Pushed: harbor.10.x.x.134.nip.io/nkp/nai-core:2.5.0 Digest: sha256:6326c6348f31c2ee0de93d03b164863addb82f8054283a4d418c8d2910b7935d nai-operators-2.5.0.tgz Pushed: harbor.10.x.x.134.nip.io/nkp/nai-operators:2.5.0 Digest: sha256:09775842c1bae10a094f54a54bc3f82177f0a9b92f0e52a20219d1cc7bf41d96 opentelemetry-operator-0.93.0.tgz Pushed: harbor.10.x.x.134.nip.io/nkp/opentelemetry-operator:0.93.0 Digest: sha256:01e050311f4f335fb2b4ae9c3dabe6efc7017cfba37cc3dc8c9acce756e52732 Now the charts are available in the OCI compatible container/chart registry.
Prepare Container Images
The Jumphost VM will be used as a medium to download the NAI container images and upload them to the internal Harbor container registry.
stateDiagram-v2
direction LR
state LoginToNutanixPortal {
[*] --> CreateDockerIDandAccessToken
CreateDockerIDandAccessToken --> LoginToDockerCLI
LoginToDockerCLI --> [*]
}
state PrepareNAIDockerImages {
[*] --> DownloadUploadImagesToHarbor
DownloadUploadImagesToHarbor --> [*]
}
[*] --> LoginToNutanixPortal
LoginToNutanixPortal --> PrepareNAIDockerImages
PrepareNAIDockerImages --> [*]
Upload NAI Docker Images to Harbor
Info
The download and upload of the container images will be done in one docker push command which will use the internal Harbor container registry details.
This will be a two-step process.
- Upload the container images from the downloaded
nai-2.x.x.tarto the jumphost VM local docker images store - Upload it to the internal Harbor container registry
-
Download the NAI air-gap bundles (NAI container images) from the Nutanix Portal > Downloads > Nutanix Enterprise AI
-
Since we will be using the same internal Harbor container registry to upload container images, make sure the following environment variables are set (these were already set during air-gap NKP preparation)
-
(Optional) - To view the container images loaded in your local docker container registry, run the following command:
nutanix/nai-api:v2.5.0 nutanix/nai-inference-ui:v2.5.0 nutanix/nai-model-processor:v2.5.0 nutanix/nai-iep-operator:v2.5.0 nutanix/nai-tgi:3.3.4-b2485c9 nutanix/nai-kserve-huggingfaceserver:v0.15.2 nutanix/nai-kserve-huggingfaceserver:v0.15.2 nutanix/nai-kserve-huggingfaceserver:v0.15.2-gpu nutanix/nai-kserve-controller:v0.15.0 nutanix/nai-postgres:16.1-alpine -
Push the images to the jumphost VM local docker images store
-
Tag and push all the NAI images to refer to the internal harbor registry
Now we are ready to deploy our NAI workloads.