Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Your 24/7 On-Call AI Agent - Solve Alerts Faster with Automatic Correlations, Investigations, and More

License

NotificationsYou must be signed in to change notification settings

robusta-dev/holmesgpt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

HolmesGPT is an AI agent for investigating problems in your cloud, finding the root cause, and suggesting remediations. It has dozens of built-in integrations for cloud providers, observability tools, and on-call systems.

🎉HolmesGPT is now a CNCF Sandbox Project! We're thrilled to be part of the Cloud Native Computing Foundation.Learn more about our journey.

Find more about HolmesGPT's maintainers and adoptershere.

📚Read the full documentation at holmesgpt.dev for installation guides, tutorials, API reference, and more.

How it Works |Installation |LLM Providers |YouTube Demo |Ask DeepWiki

HolmesGPT Investigation Demo

How it Works

HolmesGPT connects AI models with live observability data and organizational knowledge. It uses anagentic loop to analyze data from multiple sources and identify possible root causes.

holmesgpt-architecture-diagram

🔗 Data Sources

HolmesGPT integrates with popular observability and cloud platforms. The following data sources ("toolsets") are built-in.Add your own.

Data SourceStatusNotes
ArgoCDArgoCDGet status, history and manifests and more of apps, projects and clusters
AWS RDSAWS RDSFetch events, instances, slow query logs and more
ConfluenceConfluencePrivate runbooks and documentation
Coralogix LogsCoralogix LogsRetrieve logs for any resource
DatetimeDatetimeDate and time-related operations
DockerDockerGet images, logs, events, history and more
GitHubGitHub🟡 BetaRemediate alerts by opening pull requests with fixes
DataDogDataDog🟡 BetaFetches log data from datadog
LokiLokiQuery logs for Kubernetes resources or any query
TempoTempoFetch trace info, debug issues like high latency in application.
HelmHelmRelease status, chart metadata, and values
InternetInternetPublic runbooks, community docs etc
KafkaKafkaFetch metadata, list consumers and topics or find lagging consumer groups
KubernetesKubernetesPod logs, K8s events, and resource status (kubectl describe)
NewRelicNewRelic🟡 BetaInvestigate alerts, query tracing data
OpenSearchOpenSearchQuery health, shard, and settings related info of one or more clusters
PrometheusPrometheusInvestigate alerts, query metrics and generate PromQL queries
RabbitMQRabbitMQInfo about partitions, memory/disk alerts to troubleshoot split-brain scenarios and more
RobustaRobustaMulti-cluster monitoring, historical change data, user-configured runbooks, PromQL graphs and more
SlabSlabTeam knowledge base and runbooks on demand

🚀 End-to-End Automation

HolmesGPT can fetch alerts/tickets to investigate from external systems, then write the analysis back to the source or Slack.

IntegrationStatusNotes
Slack🟡 BetaDemo. Tag HolmesGPT bot in any Slack message
Prometheus/AlertManagerRobusta SaaS or HolmesGPT CLI
PagerDutyHolmesGPT CLI only
OpsGenieHolmesGPT CLI only
JiraHolmesGPT CLI only
GitHubHolmesGPT CLI only

Installation

All Installation Methods

Read theinstallation documentation to learn how to install HolmesGPT.

Supported LLM Providers

All Integration Providers

Read theLLM Providers documentation to learn how to set up your LLM API key.

Using HolmesGPT

holmes ask"what pods are unhealthy and why?"

You can also provide files as context:

holmes ask"summarize the key points in this document" -f ./mydocument.txt

You can also load the prompt from a file using the--prompt-file option:

holmes ask --prompt-file~/long-prompt.txtEnter interactive mode to ask follow-up questions:```bashholmes ask"what pods are unhealthy and why?" --interactive# orholmes ask"what pods are unhealthy and why?" -i

Also supported:

HolmesGPT CLI: investigate Prometheus alerts

Pull alerts from AlertManager and investigate them with HolmesGPT:

holmes investigate alertmanager --alertmanager-url http://localhost:9093# if on Mac OS and using the Holmes Docker image👇#  holmes investigate alertmanager --alertmanager-url http://docker.for.mac.localhost:9093

To investigate alerts in your browser, sign up for a free trial ofRobusta SaaS.

Optional: port-forward to AlertManager before running the command mentioned above (if running Prometheus inside Kubernetes)

kubectl port-forward alertmanager-robusta-kube-prometheus-st-alertmanager-0 9093:9093&
HolmesGPT CLI: investigate PagerDuty and OpsGenie alerts
holmes investigate opsgenie --opsgenie-api-key<OPSGENIE_API_KEY>holmes investigate pagerduty --pagerduty-api-key<PAGERDUTY_API_KEY># to write the analysis back to the incident as a commentholmes investigate pagerduty --pagerduty-api-key<PAGERDUTY_API_KEY> --update

For more details, runholmes investigate <source> --help

Customizing HolmesGPT

HolmesGPT can investigate many issues out of the box, with no customization or training. Optionally, you can extend Holmes to improve results:

Custom Data Sources: Add data sources (toolsets) to improve investigations

Custom Runbooks: Give HolmesGPT instructions for known alerts:

  • If using Robusta SaaS: Use the Robusta UI to add runbooks
  • If using the CLI: Use-r flag withcustom runbook files or add to~/.holmes/config.yaml

You can save common settings and API Keys in a config file to avoid passing them from the CLI each time:

Reading settings from a config file

You can save common settings and API keys in config file for re-use. Place the config file in~/.holmes/config.yaml` or pass it using the --config

You can view an example config file with all available settingshere.

Tool Output Transformers

HolmesGPT supportstransformers to process large tool outputs before sending them to your primary LLM. This feature helps manage context window limits while preserving essential information.

The most common transformer isllm_summarize, which uses a fast secondary model to summarize lengthy outputs from tools likekubectl describe, log queries, or metrics collection.

📖Learn more:Tool Output Transformers Documentation

🔐 Data Privacy

By design, HolmesGPT hasread-only access and respects RBAC permissions. It is safe to run in production environments.

We donot train HolmesGPT on your data. Data sent to Robusta SaaS is private to your account.

For extra privacy,bring an API key for your own AI model.

Evals

Because HolmesGPT relies on LLMs, it relies ona suite of pytest based evaluations to ensure the prompt and HolmesGPT's default set of tools work as expected with LLMs.

License

Distributed under the Apache 2.0 License. SeeLICENSE for more information.

Community

Join our community to discuss the HolmesGPT roadmap and share feedback:

📹First Community Meetup Recording:Watch on YouTube

Support

If you have any questions, feel free to message us onrobustacommunity.slack.com

How to Contribute

Please read ourCONTRIBUTING.md for guidelines and instructions.

For help, contact us onSlack or askDeepWiki AI your questions.

Please make sure to follow the CNCF code of conduct -details here.Ask DeepWiki

Packages

No packages published

[8]ページ先頭

©2009-2025 Movatter.jp