On-prem
Helmcharts
This document provides information about the Helm chart for deploying Rime Labs services on Kubernetes.
Chart Overview
The Helm chart deploys a two-tier application consisting of an API service and a model service. The API service communicates with the model service for inference operations.
Prerequisites
Kubernetes 1.19+ Helm 3.0+ NVIDIA GPU Operator installed (for GPU support) PV provisioner support in the underlying infrastructure (if using persistent storage)
Chart Structure
Installation
Example values.yaml
Example Deployment Templates
Here’s a simplified example of what the deployment templates might look like:
API Deployment
Model Deployment
Troubleshooting
Common Issues
- GPU not recognized: Ensure the NVIDIA GPU Operator is installed correctly in your cluster.
- Services cannot communicate: Verify that service names are correctly referenced in environment variables.
- Resource constraints: If pods are in a pending state, check if you have sufficient resources (CPU, memory, GPUs) in your cluster.