How-tos for the Administrator of Liz: The Rancher AI Assistant

Configure OpenAI provider

Select OpenAI via the UI

Navigate to the ‘Global Settings’ → ‘AI Assistant’ tab.

  1. Select OpenAI, provide an OpenAI API Key. Head to platform.openai.com to sign up to OpenAI and generate an API key.

  2. Select which model to use.

  3. Click on Apply, the agent will restart which may take a few seconds.

    OpenAI configuration screen

Select OpenAI via the helm chart

Use the following helm values to configure OpenAI from the Agent helm chart:

llmModel: "gpt-4o"
openaiLlmModel: "gpt-4o"
openaiApiKey: "xxxxxxxxx"

Update the chart:

helm upgrade --install --namespace cattle-ai-agent-system --create-namespace -f values.yaml rancher-ai-agent rancher-ai/agent

Configure an OpenAI like endpoint

From the UI or via the helm chart you can set an openAI like endpoint.

  • On the UI: Click on the Advanced settings section. Enter a valid endpoint, and click on apply.

  • In the helm chart: Set the openaiUrl value.

    OpenAI like endpoint configuration screen

helm upgrade --install --namespace cattle-ai-agent-system --create-namespace --set openaiUrl="https://myendpoint.example" rancher-ai-agent rancher-ai/agent

Configure Gemini provider

Select Gemini via the UI

Navigate to the ‘Global Settings’ → ‘AI Assistant’ tab.

  1. Select Gemini, provide a Google API Key via Google AI Studio or create an API key credential in GCP portal.

  2. Select which model to use.

  3. Click on Apply, the agent will restart which may take a few seconds.

    Gemini configuration screen

Select Gemini via the helm chart

Use the following helm values to configure Gemini from the Agent helm chart:

llmModel: "gemini-2.0-flash"
geminiLlmModel: "gemini-2.0-flash"
googleApiKey: "xxxxxxxxx"

Update the chart:

helm upgrade --install --namespace cattle-ai-agent-system --create-namespace -f values.yaml rancher-ai-agent rancher-ai/agent

Configure AWS Bedrock provider

Select AWS Bedrock via the UI

Navigate to the ‘Global Settings’ → ‘AI Assistant’ tab.

  1. Select Bedrock, provide a Bedrock Bearer Token (recommended). Follow AWS procedure to generate a Bedrock API Key.

  2. Enter a valid AWS Region.

  3. Select which model to use by going to AWS Bedrock Cross Region Inference tab, copy the Inference profile ID.

Choose a model that supports Tools call. Currently the Anthropic Claude Opus model has been tested.

  1. Click on Apply, the agent will restart which may take a few seconds.

    AWS Bedrock configuration screen

Select AWS Bedrock via the helm chart

Use the following helm values to configure AWS Bedrock from the Agent helm chart:

llmModel: "global.anthropic.claude-opus-4-5-20251101-v1:0"
bedrockLlmModel: "global.anthropic.claude-opus-4-5-20251101-v1:0"
awsBedrock:
  bearerToken: "xxxxxxxx"
  region: "us-east-1"

Update the chart:

helm upgrade --install --namespace cattle-ai-agent-system --create-namespace -f values.yaml rancher-ai-agent rancher-ai/agent

Control Access (RBAC)

To use the AI agent a user needs to get permission to the llm-config secret and the http:rancher-ai-agent:80 services/proxy.

  • Global Role

You can create the following Global Role on the local cluster or using Rancher UI on Users & Authentication → Role Templates.

apiVersion: management.cattle.io/v3
displayName: ai
kind: GlobalRole
metadata:
  name: ai-agent
namespacedRules:
  cattle-ai-agent-system:
  - apiGroups:
    - ''
    resourceNames:
    - http:rancher-ai-agent:80
    resources:
    - services/proxy
    verbs:
    - get
  - apiGroups:
    - ''
    resourceNames:
    - llm-config
    resources:
    - secrets
    verbs:
    - get

Assign this role to any user that needs to access Liz on the local cluster. This will be improved in the future with a native role.

Configure RAG

This feature is still in alpha stage and heavy development and improvement. The Rancher AI Assistant utilizes an internal Retrieval Augmented Generation (RAG) system, which is continually supplied with current Rancher documentation.

Select an Embeddings

  • Ollama: We recommend using qwen3-embedding:4b.

  • OpenAI: Consult the list of supported models.

  • Gemini: Consult the list of supported models.

  • AWS Bedrock: Consult the list of supported models.

Use the following helm values to configure RAG from the Agent helm chart:

# Enable RAG with embedded rancher documentation
rag:
  enabled: true
  embeddings_model: "qwen3-embedding:4b"
  pvc:

You can configure an existing Persistent Volume Claim using pvc. This will mount a volume to the Rancher AI Agent and significantly improve startup time for the RAG if the Agent pod restarts.

Update the chart:

helm upgrade --install --namespace cattle-ai-agent-system --create-namespace -f values.yaml rancher-ai-agent rancher-ai/agent

Configure RAG via the UI

  1. Navigate to the ‘Global Settings’ → ‘AI Assistant’ tab.

  2. Click on Show Advanced.

    Configure RAG via the UI
  3. Click on Enable RAG for this assistant.

  4. Select an Embeddings Model to use for the RAG.

  5. Click on Apply and the agent will restart which may take a few minutes to initiate the RAG.

Please note that during the initialisation, Liz is not available.