Observability AI Assistantedit

This functionality is in technical preview and may be changed or removed in a future release. Elastic will work to fix any issues, but features in technical preview are not subject to the support SLA of official GA features.

The Observability AI Assistant is a large language model (LLM) integration that helps Elastic Observability users by adding context, explaining errors and messages, and suggesting remediation in the Observability interface. You can connect the Observability AI Assistant to either the OpenAI or Microsoft Azure LLM service.

You can find Observability AI Assistant prompts throughout Observability:

Observability AI assistant preview

Observability AI Assistant is in technical preview, and its capabilities are still developing. Users should leverage it sensibly as the reliability of its responses might vary. Always cross-verify any returned advice for accurate threat detection and response, insights, and query generation.

Also, the data you provide to the Observability AI assistant is not anonymized, and is stored and processed by the third-party AI provider. This includes any data used in conversations for analysis or context, such as alert or event data, detection rule configurations, and queries. Therefore, be careful about sharing any confidential or sensitive details while using this feature.

Requirementsedit

You need following to use the Observability AI Assistant:

  • Elastic Stack version 8.9 and later.
  • An account with a third-party generative AI provider, which the Observability AI Assistant uses to generate responses. The Observability AI Assistant supports the gpt-3.5-turbo or gpt-4 models of OpenAI and Azure OpenAI Service.

Set up the Observability AI Assistantedit

Complete the following steps to use the Observability AI Assistant:

  1. Create an API key with your AI provider to authenticate requests from the Observability AI Assistant. You’ll use this in the next step. Refer to your provider’s documentation for generating API keys:

  2. Update Kibana’s configuration settings to include the Observability AI Assistant feature flag and additional information about your AI set up.

    1. Navigate to the Kibana configuration settings according to your deployment:

      • Self-managed (on-premises) deployments: Add the configuration to the kibana.yml file, which is used to configure Kibana.
      • Elastic Cloud deployments: Use the YAML editor in the Elastic Cloud console to add the configuration to Kibana user settings.
    2. Add one of the following configurations that corresponds to the AI service you’re using:

      • OpenAI:

        In the following example, xpack.observability.aiAssistant.provider.openAI.apiKey is the API key you generated in the first step.

        xpack.observability.aiAssistant.enabled: true
        xpack.observability.aiAssistant.provider.openAI.apiKey: <insert API key>
        xpack.observability.aiAssistant.provider.openAI.model: <insert model name, e.g. gpt-4>
      • Azure OpenAI Service:

        In the following example, xpack.observability.aiAssistant.provider.azureOpenAI.apiKey is the API key you generated in the first step, and xpack.observability.aiAssistant.provider.azureOpenAI.resourceName is the name of the Azure OpenAI resource used to deploy your model.

        xpack.observability.aiAssistant.enabled: true
        xpack.observability.aiAssistant.provider.azureOpenAI.deploymentId: <insert deployment ID>
        xpack.observability.aiAssistant.provider.azureOpenAI.resourceName: <insert resource name>
        xpack.observability.aiAssistant.provider.azureOpenAI.apiKey: <insert API key>
  3. If you’re using a self-managed deployment, restart Kibana.

Interact with the Observability AI Assistantedit

You can find Observability AI Assistant prompts throughout Observability that provide the following information:

  • Universal Profiling – explains the most expensive libraries and functions in your fleet and provides optimization suggestions.
  • Application performance monitoring (APM) – explains APM errors and provides remediation suggestions.
  • Infrastructure Observability – explains the processes running on a host.
  • Logs – explains log messages and generates search patterns to find similar issues.
  • Alerting – provides possible log spike causes and remediation suggestions.

For example, in the log details, you’ll see prompts for What’s this message? and How do I find similar log messages?:

obs ai logs prompts

Clicking a prompt generates a message specific to that log entry:

Observability AI assistant example