Getting Started
In this part of the lab we will deploy LLM on GPU nodes.
We will also deploy a Kubernetes cluster so far as per the NVD design requirements.
Darksite NKP cluster: to host the dev LLM and ChatBot application - this will use GPU passed through to the kubernetes worker node.
Deploy the kubernetes cluster with the following components:
- 3 x Control plane nodes
- 4 x Worker nodes
- 1 x GPU node (with a minimum of 40GB of RAM and 16 vCPUs based on
llama3-8B
LLM model)
We will deploy the GPT-in-a-Box v2 NVD Reference App - backed by llama3-8B
model.
The following is the flow of the NAI lab:
stateDiagram-v2
direction LR
state DeployHarborRegistry {
[*] --> IncreaseJumphostResources
IncreaseJumphostResources --> DeployHarbor
DeployHarbor --> [*]
}
state DeployNKP {
[*] --> CreateNkpMachineImage
CreateNkpMachineImage --> CreateNkpSelfManagedCluster
CreateNkpSelfManagedCluster --> DeployGPUNodePool
DeployGPUNodePool --> [*]
}
state NAIPreRequisites {
[*] --> ReserveIPs
ReserveIPs --> CreateFilesShare
CreateFilesShare --> [*]
}
state DeployNAI {
[*] --> BootStrapDevCluster
BootStrapDevCluster --> MonitorResourcesDeployment
MonitorResourcesDeployment --> [*]
}
state TestLLMApp {
[*] --> TestQueryLLM
TestQueryLLM --> TestChatApp
TestChatApp --> [*]
}
[*] --> DeployHarborRegistry
DeployHarborRegistry --> DeployNKP
DeployNKP --> NAIPreRequisites
NAIPreRequisites --> DeployNAI
DeployNAI --> TestLLMApp
TestLLMApp --> [*]