Custom toolsetsΒΆ
By adding custom toolsets, users can extend HolmesGPT's investigation capabilities to address unique use cases, specific infrastructure setups, or organization-specific requirements.
For example, custom toolsets might include specialized log analysis patterns or integration with external monitoring systems.
ExamplesΒΆ
Below are examples of custom toolsets and how to add them to Holmes:
Example 1: Grafana ToolsetΒΆ
This toolset lets Holmes view Grafana dashboards and suggest relevant dashboards to the user:
Prerequisites:
Grafana URL (e.g. http://localhost:3000 or https://grafana.example.com)
Grafana service account token with Basic role -> Viewer and Data sources -> Reader permissions. Check out this video on creating a Grafana service account token.
Configuration:
Helm Values:
holmes:
# provide environment variables the toolset needs - can be pulled from secrets or provided in plaintext
additionalEnvVars:
- name: GRAFANA_API_KEY
value: <token_goes_here>
- name: GRAFANA_URL
value: <grafana_url>
# define the toolset
toolsets:
grafana:
# this tool can only be enabled if these prerequisites are met
prerequisites:
# we need the GRAFANA_URL and GRAFANA_API_KEY environment variables to be set
- env:
- "GRAFANA_URL"
- "GRAFANA_API_KEY"
# curl must be installed - we check by running `curl --version` (if it's not installed, the command will fail)
- command: "curl --version"
# human-readable description of the toolset (this is not seen by the AI model - its just for users)
description: "Grafana tools"
# tools (capabilities) that will be provided to HolmesGPT when this toolset is enabled
tools:
- name: "grafana_get_dashboard"
# the LLM sees this description and uses it to decide when to use this tool
description: "Get list of grafana dashboards"
# the command that will be executed when this tool is used
# environment variables like GRAFANA_URL and GRAFANA_API_KEY can be used in the command
# they will not be exposed to the AI model, as the AI model doesn't see the command that was run
command: "curl \"${GRAFANA_URL}/api/search\" -H \"Authorization: Bearer ${GRAFANA_API_KEY}\""
- name: "grafana_get_url"
description: "Get the URL of a Grafana dashboard by UID, including the real URL of Grafana"
# in this command we use a variable called `{{ dashboard_uid }}`
# unlike enviroment variables that were provided by the user, variables like `{{ dashboard_uid }}` are provided by the AI model
# the AI model sees the tool description, decides to use this tool, and then provides a value for all {{ template_variables }} to invoke the tool
command: "echo \"${GRAFANA_URL}/d/{{ dashboard_uid }}\""
Update your Helm values with the provided YAML configuration, then apply the changes with Helm upgrade:
helm upgrade robusta robusta/robusta --values=generated_values.yaml --set clusterName=<YOUR_CLUSTER_NAME>
After the deployment is complete, you can open the HolmesGPT chat in the Robusta SaaS UI and ask questions like what grafana dashboard should I look at to investigate high pod cpu?.
Suggesting relevant dashboards during alert investigations: Add runbook instructions to your alert in the Robusta UI, instructing Holmes to search for related Grafana dashboards.

grafana_toolset.yaml:
toolsets:
grafana:
# this tool can only be enabled if these prerequisites are met
prerequisites:
# we need the GRAFANA_URL and GRAFANA_API_KEY environment variables to be set
- env:
- "GRAFANA_URL"
- "GRAFANA_API_KEY"
# curl must be installed - we check by running `curl --version` (if it's not installed, the command will fail)
- command: "curl --version"
# human-readable description of the toolset (this is not seen by the AI model - its just for users)
description: "Grafana tools"
# tools (capabilities) that will be provided to HolmesGPT when this toolset is enabled
tools:
- name: "grafana_get_dashboard"
# the LLM sees this description and uses it to decide when to use this tool
description: "Get list of grafana dashboards"
# the command that will be executed when this tool is used
# environment variables like GRAFANA_URL and GRAFANA_API_KEY can be used in the command
# they will not be exposed to the AI model, as the AI model doesn't see the command that was run
command: "curl \"${GRAFANA_URL}/api/search\" -H \"Authorization: Bearer ${GRAFANA_API_KEY}\""
- name: "grafana_get_url"
description: "Get the URL of a Grafana dashboard by UID, including the real url of grafana"
# in this command we use a variable called `{{ dashboard_uid }}`
# unlike enviroment variables that were provided by the user, variables like `{{ dashboard_uid }}` are provided by the AI model
# the AI model sees the tool description, decides to use this tool, and then provides a value for all {{ template_variables }} to invoke the tool
command: "echo \"${GRAFANA_URL}/d/{{ dashboard_uid }}\""
Set the appropriate environment variables and run Holmes:
export GRAFANA_API_KEY="<grafana_api_key>"
export GRAFANA_URL="<grafana_url>"
To test, run:
holmes ask -t grafana_toolset.yaml "what grafana dashboard should I look at to investigate high pod cpu?"
Example 2: Kubernetes Diagnostics ToolsetΒΆ
This toolset provides diagnostics for Kubernetes clusters, helping developers identify and resolve issues.
holmes:
toolsets:
kubernetes/diagnostics:
description: "Advanced diagnostics and troubleshooting tools for Kubernetes clusters"
docs_url: "https://kubernetes.io/docs/home/"
icon_url: "https://encrypted-tbn0.gstatic.com/images?q=tbn:ANd9GcRPKA-U9m5BxYQDF1O7atMfj9EMMXEoGu4t0Q&s"
tags:
- core
- cluster
prerequisites:
- command: "kubectl version --client"
tools:
- name: "kubectl_node_health"
description: "Check the health status of all nodes in the cluster."
command: "kubectl get nodes -o wide"
- name: "kubectl_check_resource_quota"
description: "Fetch the resource quota for a specific namespace."
command: "kubectl get resourcequota -n {{ namespace }} -o yaml"
- name: "kubectl_find_evicted_pods"
description: "List all evicted pods in a specific namespace."
command: "kubectl get pods -n {{ namespace }} --field-selector=status.phase=Failed | grep Evicted"
Update the generated_values.yaml
file with the provided YAML configuration, then apply the changes by executing the Helm upgrade command:
helm upgrade robusta robusta/robusta --values=generated_values.yaml --set clusterName=<YOUR_CLUSTER_NAME>
Once deployed, Holmes will have access to advanced diagnostic tools for Kubernetes clusters. For example, you can ask Holmes, "Can you do a node health check?"
and it will automatically use the newly added tools to provide you the answer.
Example 3: GitHub ToolsetΒΆ
This toolset enables Holmes to fetch information from GitHub repositories.
First create a GitHub Personal Access Token with fine-grained permissions. For this example, you can leave the default permissions.
Helm Values:
holmes:
# provide environment variables the toolset needs
additionalEnvVars:
- name: GITHUB_TOKEN
value: <token_goes_here>
# define the toolset itself
toolsets:
github_tools:
description: "Tools for managing GitHub repositories"
tags:
- cli
prerequisites:
- env:
- "GITHUB_TOKEN"
- command: "curl --version"
tools:
- name: "get_recent_commits"
description: "Fetches the most recent commits for a repository"
command: "curl -H 'Authorization: token ${GITHUB_TOKEN}' https://api.github.com/repos/{{ owner }}/{{ repo }}/commits?per_page={{ limit }} "
- name: "get_repo_details"
description: "Fetches details of a specific repository"
command: "curl -H 'Authorization: token ${GITHUB_TOKEN}' https://api.github.com/repos/{{ owner }}/{{ repo }}"
# In the above examples, LLM-provided parameters like {{ owner }} are inferred automatically from the command
# you can also define them explicitly - this is useful if:
# - You want to enforce parameter requirements (e.g., `owner` and `repo` are required).
# - You want to define provide a default value for optional parameters.
parameters:
owner:
type: "string"
description: "Owner of the repository."
required: true
repo:
type: "string"
description: "Name of the repository."
required: true
Update your Helm values with the provided YAML configuration, then apply the changes with Helm upgrade:
helm upgrade robusta robusta/robusta --values=generated_values.yaml --set clusterName=<YOUR_CLUSTER_NAME>
After the deployment is complete, the GitHub toolset will be available. HolmesGPT will be able to use it to interact with GitHub repositories. For example, you can now open the HolmesGPT chat in the Robusta SaaS UI and ask, who made the last commit to the robusta-dev/holmesgpt repo on github?.

First, add the following environment vairables:
export GITHUB_TOKEN="<token_goes_here>"
Then, add the following to ~/.holmes/config.yaml, creating the file if it doesn't exist:
toolsets:
github_tools:
description: "Tools for managing GitHub repositories"
tags:
- cli
prerequisites:
- env:
- "GITHUB_TOKEN"
- command: "curl --version"
tools:
- name: "get_recent_commits"
description: "Fetches the most recent commits for a repository"
command: "curl -H 'Authorization: token ${GITHUB_TOKEN}' https://api.github.com/repos/{{ owner }}/{{ repo }}/commits?per_page={{ limit }} "
# In the above examples, LLM-provided parameters like {{ owner }} are inferred automatically from the command
# you can also define them explicitly - this is useful if:
# - You want to enforce parameter requirements (e.g., `owner` and `repo` are required).
# - You want to define provide a default value for optional parameters.
parameters:
owner:
type: "string"
description: "Owner of the repository."
required: true
repo:
type: "string"
description: "Name of the repository."
required: true
To test, run:
holmes ask -t github_toolset.yaml "who made the last commit to the robusta-dev/holmesgpt repo on github?"
ReferenceΒΆ
A toolset is defined in your Helm values (generated_values.yaml
). Each toolset has a unique name and has to contain tools.
toolsets:
<toolset_name>:
enabled: <true|false>
name: "<string>"
description: "<string>"
docs_url: "<string>"
icon_url: "<string>"
tags:
- <cli|cluster|core>
installation_instructions: "<string>"
prerequisites:
- command: "<shell_command>"
expected_output: "<expected output of the command>"
- env:
- "<environment variable>"
additional_instructions: "<string>"
tools:
- name: "<string>"
description: "<string>"
command: "<shell command template>"
script: "<script content>"
parameters:
<parameter_name>:
type: "<string>"
description: "<string>"
required: <true|false>
Toolset FieldsΒΆ
Parameter |
Type |
Description |
Required |
---|---|---|---|
|
boolean |
Indicates whether the toolset is enabled. Defaults to |
No |
|
string |
A unique identifier for the toolset. Used for informational purposes and logging. |
Yes |
|
string |
A summary of the toolset's purpose. This description is visible to the LLM and helps it decide when to use the toolset for specific tasks. |
No |
|
string |
A URL pointing to documentation related to the toolset. |
No |
|
string |
A URL to an icon representing the toolset. |
No |
|
list |
Tags for categorizing toolsets, |
No |
|
string |
Instructions on how to install prerequisites required by the toolset. |
No |
|
list |
A list of conditions that must be met for the toolset to be enabled. Prerequisites can include commands or environment variables, or both. |
No |
|
string |
Additional shell commands or processing instructions applied to the output of tools in this toolset. |
No |
|
list |
A list of tools defined within the toolset. Each tool is an object with its own set of fields. |
Yes |
Tool Fields
Parameter |
Type |
Description |
Required |
---|---|---|---|
|
string |
A unique identifier for the tool within the toolset. |
Yes |
|
string |
A brief description of the tool's purpose. Helps Holmes decide when to use this tool. |
Yes |
|
string |
A shell command template that the tool will execute. Can include environment variables using |
Either |
|
string |
The content of a script that the tool will execute. Use this if your tool requires a multi-line script. |
Either |
|
dictionary |
Specifying parameters is optional, as they can be inferred by the LLM from the prompt context. When defined, parameters specify the inputs required for the tool, allowing dynamic customization of its execution. Each parameter has its own fields, such as type, description, and whether it is required. |
No |
|
string |
Additional shell commands or processing instructions applied to the output of this tool. This is can be useful for post-processing the results of a command, such as filtering, formatting, or transforming the data before it is returned to the user. For example, you could use |
No |
Parameter Fields (Within `parameters`, if missing we infer it)
Parameter |
Type |
Description |
Required |
---|---|---|---|
|
string |
The data type of the parameter (e.g., |
No (defaults to |
|
string |
A description of the parameter. |
No |
|
boolean |
Indicates whether the parameter is required. Defaults to |
No |
Variable Syntax in CommandsΒΆ
In toolset commands, variables can be defined using two syntaxes: {{ }}
and ${ }
.
Variables written as {{ variable_name }}
are placeholders that are inferred by Holmes and dynamically filled by the LLM based on the context or user prompts. These variables are visible to the LLM and allow flexible, context-aware execution. For example:
command: "kubectl describe pod {{ pod_name }} -n {{ namespace }}"
Here, {{ pod_name }}
and {{ namespace }}`
are inferred and dynamically filled during execution.
Variables written as ${VARIABLE_NAME}
are static or environment-specific values, such as API keys or configuration parameters. These are not visible to the LLM and are expanded directly by the shell at runtime. For example:
command: "curl -H 'Authorization: token ${GITHUB_TOKEN}' https://api.github.com/repos/{{ owner }}/{{ repo }}"
In this case, ${GITHUB_TOKEN}
is an environment variable, while {{ owner }}
and {{ repo }}
are dynamically inferred by Holmes.
Best Practices for Variable Usage:
Use
${}
for sensitive or static environment variables, such as API keys and credentials.Use
{{}}
for parameters that the LLM can dynamically infer and fill based on the context or user inputs.
Adding a tool that requires a new binaryΒΆ
In some cases, adding a new tool to Holmes might require installing additional packages that are not included in the base Holmes Docker image. This guide explains how to create a custom Docker image that includes the new binaries and update your Helm deployment to use the custom image.
As an example, we'll add a new HolmesGPT tool that uses the jq
binary, which isn't present in the original image:
Example Dockerfile to add jq:
FROM python:3.11-slim
ENV PYTHONUNBUFFERED=1
ENV PATH="/venv/bin:$PATH"
ENV PYTHONPATH=$PYTHONPATH:.:/app/holmes
WORKDIR /app
COPY --from=builder /app/venv /venv
COPY . /app
# We're installing here libexpat1, to upgrade the package to include a fix to 3 high CVEs. CVE-2024-45491,CVE-2024-45490,CVE-2024-45492
RUN apt-get update \
&& apt-get install -y \
git \
apt-transport-https \
gnupg2 \
&& apt-get purge -y --auto-remove \
&& apt-get install -y --no-install-recommends libexpat1 \
&& rm -rf /var/lib/apt/lists/*
# Example of installing jq
RUN apt-get install -y jq
Now, you will need to build and push the Docker image to your container registry.
Abstracted Instructions for Building and Pushing the Docker Image:
Build the Docker Image: Depending on the tools and binaries you need, build the custom Docker image with the appropriate tag.
docker build -t <your-registry>/<your-project>/holmes-custom:<tag> .
Replace: -
<your-registry>
: Your Docker registry (e.g.,us-central1-docker.pkg.dev
for Google Artifact Registry). -<your-project>
: Your project or repository name. -<tag>
: The desired tag for the image (e.g.,latest
,v1.0
).Push the Image to Your Registry: After building the image, push it to your container registry:
docker push <your-registry>/<your-project>/holmes-custom:<tag>
This ensures that the image is available for your Kubernetes deployment.
After pushing your custom Docker image, update your generated_values.yaml
to use this custom image for Holmes.
enableHolmesGPT: true
holmes:
registry: <your-registry>/<your-project> # Use your custom registry
image: <image>:<tag> # Specify the image with the tag you used when pushing the image
additionalEnvVars:
- name: ROBUSTA_AI
value: "true"
toolsets:
json_processor:
description: "A toolset for processing JSON data using jq"
prerequisites:
- command: "jq --version" # Ensure jq is installed
tools:
- name: "process_json"
description: "A tool that uses jq to process JSON input"
command: "echo '{{ json_input }}' | jq '.'" # Example jq command to format JSON
Finally, after updating your generated_values.yaml
, apply the changes to your Helm deployment:
helm upgrade robusta robusta/robusta --values=generated_values.yaml --set clusterName=<YOUR_CLUSTER_NAME>
This will update the deployment to use the custom Docker image, which includes the new binaries. The toolsets
defined in the configuration will now be available for Holmes to use, including any new binaries like jq
.