Set up a data streamedit

To set up a data stream, follow these steps:

You can also convert an existing index alias to a data stream.

Optional: Configure an ILM lifecycle policyedit

While optional, we recommend you configure an index lifecycle management (ILM) policy to automate the management of your data stream’s backing indices.

In Kibana, open the menu and go to Stack Management > Index Lifecycle Policies. Click Create policy.

Create Policy page
API example

Use the create lifecycle policy API to configure a policy:

PUT /_ilm/policy/my-data-stream-policy
{
  "policy": {
    "phases": {
      "hot": {
        "actions": {
          "rollover": {
            "max_size": "25GB"
          }
        }
      },
      "delete": {
        "min_age": "30d",
        "actions": {
          "delete": {}
        }
      }
    }
  }
}

Create an index templateedit

  1. In Kibana, open the menu and go to Stack Management > Index Management.
  2. In the Index Templates tab, click Create template.
  3. In the Create template wizard, use the Data stream toggle to indicate the template is used for data streams.
  4. Use the wizard to finish defining your template. Specify:

    • One or more index patterns that match the data stream’s name.
      We recommend using the Elastic data stream naming scheme. Data stream names must meet the following criteria:

      • Lowercase only
      • Cannot include \, /, *, ?, ", <, >, |, ,, #, :, or a space character
      • Cannot start with -, _, +, or .ds-
      • Cannot be . or ..
      • Cannot be longer than 255 bytes. Multi-byte characters count towards this limit faster.
    • Mappings and settings for the stream’s backing indices.
    • A priority for the index template

      Elasticsearch has built-in index templates, each with a priority of 100, for the following index patterns:

      • logs-*-*
      • metrics-*-*
      • synthetics-*-*

      The Elastic Agent uses these templates to create data streams. If you use the Elastic Agent, assign your index templates a priority lower than 100 to avoid overriding the built-in templates. Otherwise, to avoid accidentally applying the built-in templates, do one or more of the following:

      • To disable all built-in index and component templates, set stack.templates.enabled to false using the cluster update settings API.
      • Use a non-overlapping index pattern.
      • Assign templates with an overlapping pattern a priority higher than 100. For example, if you don’t use the Elastic Agent and want to create a template for the logs-* index pattern, assign your template a priority of 200. This ensures your template is applied instead of the built-in template for logs-*-*.

Every document indexed to a data stream must contain a @timestamp field, mapped as a date or date_nanos field type. If the index template doesn’t specify a mapping for the @timestamp field, Elasticsearch maps @timestamp as a date field with default options.

If using ILM, specify your lifecycle policy in the index.lifecycle.name setting.

Carefully consider your template’s mappings and settings. Later changes may require reindexing. See Change mappings and settings for a data stream.

Create template page
API example

Use the put index template API to create an index template. The template must include a data_stream object, indicating it’s used for data streams.

PUT /_index_template/my-data-stream-template
{
  "index_patterns": [ "my-data-stream*" ],
  "data_stream": { },
  "priority": 200,
  "template": {
    "settings": {
      "index.lifecycle.name": "my-data-stream-policy"
    }
  }
}

Create the data streamedit

To automatically create the data stream, submit an indexing request to the stream. The stream’s name must match one of your template’s index patterns.

POST /my-data-stream/_doc/
{
  "@timestamp": "2099-03-07T11:04:05.000Z",
  "user": {
    "id": "vlb44hny"
  },
  "message": "Login attempt failed"
}

You can also use the create data stream API to manually create the data stream. The stream’s name must match one of your template’s index patterns.

PUT /_data_stream/my-data-stream

When you create a data stream, Elasticsearch automatically creates a backing index for the stream. This index also acts as the stream’s first write index.

Convert an index alias to a data streamedit

Prior to Elasticsearch 7.9, you would typically use an index alias with a write index to manage time series data. Data streams replace most of this functionality and usually require less maintenance.

To convert an index alias with a write index to a new data stream with the same name, use the migrate to data stream API. During conversion, the alias’s indices become hidden backing indices for the stream. The alias’s write index becomes the stream’s write index. Note the data stream still requires a matching index template.

POST /_data_stream/_migrate/my-time-series-data

Secure the data streamedit

To control access to the data stream and its data, use Elasticsearch’s security features.

Get information about a data streamedit

In Kibana, open the menu and go to Stack Management > Index Management. In the Data Streams tab, click the data stream’s name.

Data Streams tab
API example

Use the get data stream API to retrieve information about one or more data streams:

GET /_data_stream/my-data-stream

Delete a data streamedit

To delete a data stream and its backing indices, open the Kibana menu and go to Stack Management > Index Management. In the Data Streams tab, click the trash icon. The trash icon only displays if you have the delete_index security privilege for the data stream.

Data Streams tab
API example

Use the delete data stream API to delete a data stream and its backing indices:

DELETE /_data_stream/my-data-stream