Prometheus¶
Connect HolmesGPT to Prometheus for metrics analysis and query generation. This integration enables detection of memory leaks, CPU throttling, queue backlogs, and performance issues.
Prerequisites¶
- A running and accessible Prometheus server
- Ensure HolmesGPT can connect to the Prometheus endpoint
Configuration¶
Add the following to ~/.holmes/config.yaml. Create the file if it doesn't exist:
toolsets:
prometheus/metrics:
enabled: true
config:
prometheus_url: http://<your-prometheus-service>:9090
# Optional:
#headers:
# Authorization: "Basic <base_64_encoded_string>"
When using the standalone Holmes Helm Chart, update your values.yaml
:
toolsets:
prometheus/metrics:
enabled: true
config:
prometheus_url: http://<your-prometheus-service>:9090
# Optional:
#headers:
# Authorization: "Basic <base_64_encoded_string>"
Apply the configuration:
helm upgrade holmes holmes/holmes --values=values.yaml
When using the Robusta Helm Chart (which includes HolmesGPT), update your generated_values.yaml
:
holmes:
toolsets:
prometheus/metrics:
enabled: true
config:
prometheus_url: http://<your-prometheus-service>:9090
# Optional:
#headers:
# Authorization: "Basic <base_64_encoded_string>"
Apply the configuration:
helm upgrade robusta robusta/robusta --values=generated_values.yaml --set clusterName=<YOUR_CLUSTER_NAME>
💡 Alternative: Set environment variables instead of using the config file:
- PROMETHEUS_URL
: The Prometheus server URL
- PROMETHEUS_AUTH_HEADER
: Optional authorization header value (e.g., "Bearer token123")
Validation¶
To test your connection, run:
Troubleshooting¶
Finding your Prometheus URL¶
There are several ways to find your Prometheus URL:
Option 1: Simple method (port-forwarding)
# Find Prometheus services
kubectl get svc -A | grep prometheus
# Port forward for testing
kubectl port-forward svc/<your-prometheus-service> 9090:9090 -n <namespace>
# Then access Prometheus at: http://localhost:9090
Option 2: Advanced method (get full cluster DNS URL)
If you want to find the full internal DNS URL for Prometheus, run:
kubectl get svc --all-namespaces -o jsonpath='{range .items[*]}{.metadata.name}{"."}{.metadata.namespace}{".svc.cluster.local:"}{.spec.ports[0].port}{"\n"}{end}' | grep prometheus | grep -Ev 'operat|alertmanager|node|coredns|kubelet|kube-scheduler|etcd|controller' | awk '{print "http://"$1}'
This will print all possible Prometheus service URLs in your cluster. Pick the one that matches your deployment.
Common Issues¶
- Connection refused: Check if the Prometheus URL is accessible from HolmesGPT.
- Authentication errors: Verify the headers configuration for secured Prometheus endpoints.
- No metrics returned: Ensure that Prometheus is scraping your targets.
Advanced Configuration¶
You can further customize the Prometheus toolset with the following options:
toolsets:
prometheus/metrics:
enabled: true
config:
prometheus_url: http://<prometheus-host>:9090
healthcheck: "-/healthy" # Path for health checking (default: -/healthy)
headers:
Authorization: "Basic <base_64_encoded_string>"
# Time windows and limits
default_metadata_time_window_hrs: 1 # Time window for metadata APIs (default: 1 hour)
query_response_size_limit: 20000 # Max characters in query response (default: 20000)
# Timeout configuration
default_query_timeout_seconds: 20 # Default timeout for PromQL queries (default: 20)
max_query_timeout_seconds: 180 # Maximum allowed timeout for PromQL queries (default: 180)
default_metadata_timeout_seconds: 20 # Default timeout for metadata/discovery APIs (default: 20)
max_metadata_timeout_seconds: 60 # Maximum allowed timeout for metadata APIs (default: 60)
# Other options
rules_cache_duration_seconds: 1800 # Cache duration for Prometheus rules (default: 30 minutes)
prometheus_ssl_enabled: true # Enable SSL verification (default: true)
tool_calls_return_data: true # If false, disables returning Prometheus data (default: true)
additional_labels: # Additional labels to add to all queries
cluster: "production"
Config option explanations:
prometheus_url
: The base URL for Prometheus. Should include protocol and port.healthcheck
: Path used for health checking Prometheus or Mimir/Cortex endpoint. Defaults to-/healthy
for Prometheus, use/ready
for Grafana Mimir.headers
: Extra headers for all Prometheus HTTP requests (e.g., for authentication).default_metadata_time_window_hrs
: Time window (in hours) for metadata/discovery APIs to look for active metrics. Default: 1 hour.query_response_size_limit
: Maximum number of characters in a query response before truncation. Set tonull
to disable. Default: 20000.default_query_timeout_seconds
: Default timeout for PromQL queries. Can be overridden per query. Default: 20.max_query_timeout_seconds
: Maximum allowed timeout for PromQL queries. Default: 180.default_metadata_timeout_seconds
: Default timeout for metadata/discovery API calls. Default: 20.max_metadata_timeout_seconds
: Maximum allowed timeout for metadata API calls. Default: 60.rules_cache_duration_seconds
: How long to cache Prometheus rules. Set tonull
to disable caching. Default: 1800 (30 minutes).prometheus_ssl_enabled
: Enable SSL certificate verification. Default: true.tool_calls_return_data
: Iffalse
, disables returning Prometheus data to HolmesGPT (useful if you hit token limits). Default: true.additional_labels
: Dictionary of labels to add to all queries (currently only implemented for AWS/AMP).
Capabilities¶
Tool Name | Description |
---|---|
list_prometheus_rules | List all defined Prometheus rules with descriptions and annotations |
get_metric_names | Get list of metric names (fastest discovery method) - requires match filter |
get_label_values | Get all values for a specific label (e.g., pod names, namespaces) |
get_all_labels | Get list of all label names available in Prometheus |
get_series | Get time series matching a selector (returns full label sets) |
get_metric_metadata | Get metadata (type, description, unit) for metrics |
execute_prometheus_instant_query | Execute an instant PromQL query (single point in time) |
execute_prometheus_range_query | Execute a range PromQL query for time series data with graph generation |
Coralogix Prometheus Configuration¶
To use a Coralogix PromQL endpoint with HolmesGPT:
- Go to Coralogix Documentation and choose the relevant PromQL endpoint for your region.
- In Coralogix, create an API key with permissions to query metrics (Data Flow → API Keys).
-
Create a Kubernetes secret for the API key and expose it as an environment variable in your Helm values:
-
Add the following under your toolsets in the Helm chart:
holmes: toolsets: prometheus/metrics: enabled: true config: healthcheck: "/api/v1/query?query=up" # This is important for Coralogix prometheus_url: "https://prom-api.eu2.coralogix.com" # Use your region's endpoint headers: token: "{{ env.CORALOGIX_API_KEY }}" default_metadata_time_window_hrs: 72 # Look back 72 hours for metrics tool_calls_return_data: true
AWS Managed Prometheus (AMP) Configuration¶
To connect HolmesGPT to AWS Managed Prometheus:
holmes:
toolsets:
prometheus/metrics:
enabled: true
config:
prometheus_url: https://aps-workspaces.us-east-1.amazonaws.com/workspaces/ws-xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx/
aws_region: us-east-1
aws_service_name: aps # Default value, can be omitted
# Optional: Specify credentials (otherwise uses default AWS credential chain)
aws_access_key: "{{ env.AWS_ACCESS_KEY_ID }}"
aws_secret_access_key: "{{ env.AWS_SECRET_ACCESS_KEY }}"
# Optional: Assume a role for cross-account access
assume_role_arn: "arn:aws:iam::123456789012:role/PrometheusReadRole"
refresh_interval_seconds: 900 # Refresh AWS credentials every 15 minutes (default)
Notes:
- The toolset automatically detects AWS configuration when aws_region
is present
- Uses SigV4 authentication for all requests
- Supports IAM roles and cross-account access via assume_role_arn
- Credentials refresh automatically based on refresh_interval_seconds
Grafana Cloud (Mimir) Configuration¶
To connect HolmesGPT to Grafana Cloud's Prometheus/Mimir endpoint:
- Create a service account token in Grafana Cloud:
- Navigate to "Administration → Service accounts"
- Create a new service account
-
Generate a service account token (starts with
glsa_
) -
Find your Prometheus datasource UID:
-
Configure HolmesGPT:
Important notes:
- Use the proxy endpoint URL format
/api/datasources/proxy/uid/
- this handles authentication and routing to Mimir automatically - The toolset automatically detects and uses the most appropriate APIs for discovery