Install on GKE¶
TIP: Make sure you have enough quota in your GCP project.
Open the cloud console quotas page: https://console.cloud.google.com/iam-admin/quotas. Make sure your project is selected in the top left.
You will need to verify that you have enough quota for the accelerators you want to use. Below is table of common quotas you will have to increase depending on your needs.
Quota | Location | Min Value |
---|---|---|
Preemptible TPU v5 Lite Podslice chips | <your-region> |
8 |
Preemptible NVIDIA L4 GPUs | <your-region> |
2 |
GPUs (all regions) | - | 2 |
CPUs (all regions) | - | 24 |
See the following screenshot examples of how these quotas appear in the console:
1. Create a cluster¶
Option: GKE Autopilot¶
Create an Autopilot cluster (replace us-central1
with a region that you have quota).
gcloud container clusters create-auto cluster-1 \
--location=us-central1
Option: GKE Standard¶
TODO: Reference gcloud commands for creating a GKE standard cluster.
2. Install KubeAI¶
Add KubeAI Helm repository.
helm repo add kubeai https://www.kubeai.org
helm repo update
Make sure you have a HuggingFace Hub token set in your environment (HUGGING_FACE_HUB_TOKEN
).
Install KubeAI with Helm.
curl -L -O https://raw.githubusercontent.com/substratusai/kubeai/refs/heads/main/charts/kubeai/values-gke.yaml
helm upgrade --install kubeai kubeai/kubeai \
-f values-gke.yaml \
--set secrets.huggingface.token=$HUGGING_FACE_HUB_TOKEN \
--wait
3. Optionally configure models¶
Optionally install preconfigured models.
cat <<EOF > kubeai-models.yaml
catalog:
llama-3.1-8b-instruct-fp8-l4:
enabled: true
EOF
helm install kubeai-models kubeai/models \
-f ./kubeai-models.yaml