# Alternative Installation ### Installing Both Ollama and Open WebUI Using Kustomize For cpu-only pod ```bash kubectl apply -f ./kubernetes/manifest/base ``` For gpu-enabled pod ```bash kubectl apply -k ./kubernetes/manifest ``` ### Installing Both Ollama and Open WebUI Using Helm Package Helm file first ```bash helm package ./kubernetes/helm/ ``` For cpu-only pod ```bash helm install open-webui ./open-webui-*.tgz ``` For gpu-enabled pod ```bash helm install open-webui ./open-webui-*.tgz --set ollama.resources.limits.nvidia.com/gpu="1" ``` Check the `kubernetes/helm/values.yaml` file to know which parameters are available for customization