HTTP JSON inputedit

Use the httpjson input to read messages from an HTTP API with JSON payloads.

This input supports:

  • Auth

    • Basic
    • OAuth2
  • Retrieval at a configurable interval
  • Pagination
  • Retries
  • Rate limiting
  • Proxying
  • Request transformations
  • Response transformations

Example configurations:

filebeat.inputs:
# Fetch your public IP every minute.
- type: httpjson
  interval: 1m
  request.url: https://api.ipify.org/?format=json
  processors:
    - decode_json_fields:
        fields: ["message"]
        target: "json"
filebeat.inputs:
- type: httpjson
  request.url: http://localhost:9200/_search?scroll=5m
  request.method: POST
  response.split:
    target: body.hits.hits
  response.pagination:
    - set:
        target: url.value
        value: http://localhost:9200/_search/scroll
    - set:
        target: url.params.scroll_id
        value: '[[.last_response.body._scroll_id]]'
    - set:
        target: body.scroll
        value: 5m

Additionally, it supports authentication via Basic auth, HTTP Headers or oauth2.

Example configurations with authentication:

filebeat.inputs:
- type: httpjson
  request.url: http://localhost
  request.transforms:
    - set:
        target: header.Authorization
        value: 'Basic aGVsbG86d29ybGQ='
filebeat.inputs:
- type: httpjson
  auth.oauth2:
    client.id: 12345678901234567890abcdef
    client.secret: abcdef12345678901234567890
    token_url: http://localhost/oauth2/token
  request.url: http://localhost
filebeat.inputs:
- type: httpjson
  auth.oauth2:
    client.id: 12345678901234567890abcdef
    client.secret: abcdef12345678901234567890
    token_url: http://localhost/oauth2/token
    user: user@domain.tld
    password: P@$$W0₹D
  request.url: http://localhost

Input stateedit

The httpjson input keeps a runtime state between requests. This state can be accessed by some configuration options and transforms.

The state has the following elements:

  • last_response.url.value: The full URL with params and fragments from the last request with a successful response.
  • last_response.url.params: A url.Values of the params from the URL in last_response.url.value. Can be queried with the Get function.
  • last_response.header: A map containing the headers from the last successful response.
  • last_response.body: A map containing the parsed JSON body from the last successful response. This is the response as it comes from the remote server.
  • last_response.page: A number indicating the page number of the last response. It starts with the value 0 at every interval.
  • first_event: A map representing the first event sent to the output (result from applying transforms to last_response.body).
  • last_event: A map representing the last event of the current request in the requests chain (result from applying transforms to last_response.body).
  • url: The last requested URL as a raw url.URL Go type.
  • header: A map containing the headers. References the next request headers when used in request.rate_limit.early_limit or response.pagination configuration sections, and to the last response headers when used in response.transforms, response.split, or request.rate_limit.limit configuration sections.
  • body: A map containing the body. References the next request body when used in request.rate_limit.early_limit or response.pagination configuration sections, and to the last response body when used in response.transforms or response.split configuration sections.
  • cursor: A map containing any data the user configured to be stored between restarts (See cursor).

All of the mentioned objects are only stored at runtime, except cursor, which has values that are persisted between restarts.

Transformsedit

A transform is an action that lets the user modify the input state. Depending on where the transform is defined, it will have access for reading or writing different elements of the state.

The access limitations are described in the corresponding configuration sections.

appendedit

Appends a value to an array. If the field does not exist, the first entry will create a new array. If the field exists, the value is appended to the existing field and converted to a list.

- append:
    target: body.foo.bar
    value: '[[.cursor.baz]]'
    default: "a default value"
  • target defines the destination field where the value is stored.
  • value defines the value that will be stored and it is a value template.
  • default defines the fallback value whenever value is empty or the template parsing fails. Default templates do not have access to any state, only to functions.
  • value_type defines the type of the resulting value. Possible values are: string, json, and int. Default is string.
  • fail_on_template_error if set to true an error will be returned and the request will be aborted when the template evaluation fails. Default is false.

deleteedit

Deletes the target field.

- delete:
    target: body.foo.bar
  • target defines the destination field to delete. If target is a list and not a single element, the complete list will be deleted.

setedit

Sets a value.

- set:
    target: body.foo.bar
    value: '[[.cursor.baz]]'
    default: "a default value"
  • target defines the destination field where the value is stored.
  • value defines the value that will be stored and it is a value template.
  • default defines the fallback value whenever value is empty or the template parsing fails. Default templates do not have access to any state, only to functions.
  • value_type defines how the resulting value will be treated. Possible values are: string, json, and int. Default is string.
  • fail_on_template_error if set to true an error will be returned and the request will be aborted when the template evaluation fails. Default is false.

Value templatesedit

Some configuration options and transforms can use value templates. Value templates are Go templates with access to the input state and to some built-in functions. Please note that delimiters are changed from the default {{ }} to [[ ]] to improve interoperability with other templating mechanisms.

To see which state elements and operations are available, see the documentation for the option or transform where you want to use a value template.

A value template looks like:

- set:
    target: body.foo.bar
    value: '[[.cursor.baz]] more data'
    default: "a default value"

The content inside the brackets [[ ]] is evaluated. For more information on Go templates please refer to the Go docs.

Some built-in helper functions are provided to work with the input state inside value templates:

  • add: adds a list of integers and returns their sum.
  • base64DecodeNoPad: Decodes the base64 string without padding. Any binary output will be converted to a UTF8 string.
  • base64Decode: Decodes the base64 string. Any binary output will be converted to a UTF8 string.
  • base64EncodeNoPad: Joins and base64 encodes all supplied strings without padding. Example [[base64EncodeNoPad "string1" "string2"]]
  • base64Encode: Joins and base64 encodes all supplied strings. Example [[base64Encode "string1" "string2"]]
  • beatInfo: returns a map containing information about the Beat. Available keys in the map are goos (running operating system), goarch (running system architecture), commit (git commit of current build), buildtime (compile time of current build), version (version of current build). Example: [[ beatInfo.version ]] returns {version}.
  • div: does the integer division of two integer values.
  • formatDate: formats a time.Time. By default the format layout is RFC3339 but optionally can accept any of the Golang predefined layouts or a custom one. It will default to UTC timezone when formatting, but you can specify a different timezone. If the timezone is incorrect, it will default to UTC. Example: [[ formatDate (now) "UnixDate" ]], [[ formatDate (now) "UnixDate" "America/New_York" ]].
  • getRFC5988Link: extracts a specific relation from a list of RFC5988 links. It is useful when parsing header values for pagination. Example: [[ getRFC5988Link "next" .last_response.header.Link ]].
  • hashBase64: calculates the hash of a list of strings concatenated together. Returns a base64 encoded hash. Supports sha1 or sha256. Example [[hash "sha256" "string1" "string2" (formatDate (now) "RFC1123")]]
  • hash: calculates the hash of a list of strings concatenated together. Returns a hex encoded hash. Supports sha1 or sha256. Example [[hash "sha256" "string1" "string2" (formatDate (now) "RFC1123")]]
  • hexDecode: Decodes the hexadecimal string. Any hexadecimal string will be converted to its bytes representation. Example [[hexDecode "b0a92a08a9b4883aa3aa2d0957be12a678cbdbb32dc5db09fe68239a09872f96"]]; Expected Output: "\xb0\xa9*\b\xa9\xb4\x88:\xa3\xaa-\tW\xbe\x12\xa6x\xcb۳-\xc5\xdb\t\xfeh#\x9a\t\x87/\x96"
  • hmacBase64: calculates the hmac signature of a list of strings concatenated together. Returns a base64 encoded signature. Supports sha1 or sha256. Example [[hmac "sha256" "secret" "string1" "string2" (formatDate (now) "RFC1123")]]
  • hmac: calculates the hmac signature of a list of strings concatenated together. Returns a hex encoded signature. Supports sha1 or sha256. Example [[hmac "sha256" "secret" "string1" "string2" (formatDate (now) "RFC1123")]]
  • join: joins a list using the specified separator. Example: [[join .body.arr ","]]
  • mul: multiplies two integers.
  • now: returns the current time.Time object in UTC. Optionally, it can receive a time.Duration as a parameter. Example: [[now (parseDuration "-1h")]] returns the time at 1 hour before now.
  • parseDate: parses a date string and returns a time.Time in UTC. By default the expected layout is RFC3339 but optionally can accept any of the Golang predefined layouts or a custom one. Example: [[ parseDate "2020-11-05T12:25:32Z" ]], [[ parseDate "2020-11-05T12:25:32.1234567Z" "RFC3339Nano" ]], [[ (parseDate "Thu Nov 5 12:25:32 +0000 2020" "Mon Jan _2 15:04:05 -0700 2006").UTC ]].
  • parseDuration: parses duration strings and returns time.Duration. Example: [[parseDuration "1h"]].
  • parseTimestampMilli: parses a timestamp in milliseconds and returns a time.Time in UTC. Example: [[parseTimestamp 1604582732000]] returns 2020-11-05 13:25:32 +0000 UTC.
  • parseTimestampNano: parses a timestamp in nanoseconds and returns a time.Time in UTC. Example: [[parseTimestamp 1604582732000000000]] returns 2020-11-05 13:25:32 +0000 UTC.
  • parseTimestamp: parses a timestamp in seconds and returns a time.Time in UTC. Example: [[parseTimestamp 1604582732]] returns 2020-11-05 13:25:32 +0000 UTC.
  • replaceAll(old, new, s): replaces all non-overlapping instances of old with new in s. Example: [[ replaceAll "some" "my" "some value" ]] returns my value.
  • sprintf: formats according to a format specifier and returns the resulting string. Refer to the Go docs for usage. Example: [[sprintf "%d:%q" 34 "quote this"]]
  • toInt: converts a value of any type to an integer when possible. Returns 0 if the conversion fails.
  • toJSON: converts a value to a JSON string. This can be used with value_type: json to create an object from a template. Example: [[ toJSON .last_response.body.pagingIdentifiers ]].
  • urlEncode: URL encodes the supplied string. Example [[urlEncode "string1"]]. Example [[urlEncode "<string1>"]] will return %3Cstring1%3E.
  • userAgent: generates the User Agent with optional additional values. If no arguments are provided, it will generate the default User Agent that is added to all requests by default. It is recommended to delete the existing User-Agent header before setting a new one. Example: [[ userAgent "integration/1.2.3" ]] would generate Elastic-Filebeat/8.1.0 (darwin; amd64; 9b893e88cfe109e64638d65c58fd75c2ff695402; 2021-12-15 13:20:00 +0000 UTC; integration_name/1.2.3)
  • uuid: returns a random UUID such as a11e8780-e3e7-46d0-8e76-f66e75acf019. Example: [[ uuid ]]

In addition to the provided functions, any of the native functions for time.Time, http.Header, and url.Values types can be used on the corresponding objects. Examples: [[(now).Day]], [[.last_response.header.Get "key"]]

Configuration optionsedit

The httpjson input supports the following configuration options plus the Common options described later.

intervaledit

Duration between repeated requests. It may make additional pagination requests in response to the initial request if pagination is enabled. Default: 60s.

auth.basic.enablededit

When set to false, disables the basic auth configuration. Default: true.

Basic auth settings are disabled if either enabled is set to false or the auth.basic section is missing.

auth.basic.useredit

The user to authenticate with.

auth.basic.passwordedit

The password to use.

auth.oauth2.enablededit

When set to false, disables the oauth2 configuration. Default: true.

OAuth2 settings are disabled if either enabled is set to false or the auth.oauth2 section is missing.

auth.oauth2.provideredit

Used to configure supported oauth2 providers. Each supported provider will require specific settings. It is not set by default. Supported providers are: azure, google.

auth.oauth2.client.idedit

The client ID used as part of the authentication flow. It is always required except if using google as provider. Required for providers: default, azure.

auth.oauth2.client.secretedit

The client secret used as part of the authentication flow. It is always required except if using google as provider. Required for providers: default, azure.

auth.oauth2.useredit

The user used as part of the authentication flow. It is required for authentication - grant type password. It is only available for provider default.

auth.oauth2.passwordedit

The password used as part of the authentication flow. It is required for authentication - grant type password. It is only available for provider default.

user and password are required for grant_type password. If user and password is not used then it will automatically use the token_url and client credential method.

auth.oauth2.scopesedit

A list of scopes that will be requested during the oauth2 flow. It is optional for all providers.

auth.oauth2.token_urledit

The endpoint that will be used to generate the tokens during the oauth2 flow. It is required if no provider is specified.

For azure provider either token_url or azure.tenant_id is required.

auth.oauth2.endpoint_paramsedit

Set of values that will be sent on each request to the token_url. Each param key can have multiple values. Can be set for all providers except google.

- type: httpjson
  auth.oauth2:
    endpoint_params:
      Param1:
        - ValueA
        - ValueB
      Param2:
        - Value

auth.oauth2.azure.tenant_idedit

Used for authentication when using azure provider. Since it is used in the process to generate the token_url, it can’t be used in combination with it. It is not required.

For information about where to find it, you can refer to https://docs.microsoft.com/en-us/azure/active-directory/develop/howto-create-service-principal-portal.

auth.oauth2.azure.resourceedit

The accessed WebAPI resource when using azure provider. It is not required.

auth.oauth2.google.credentials_fileedit

The credentials file for Google.

Only one of the credentials settings can be set at once. If none is provided, loading default credentials from the environment will be attempted via ADC. For more information about how to provide Google credentials, please refer to https://cloud.google.com/docs/authentication.

auth.oauth2.google.credentials_jsonedit

Your credentials information as raw JSON.

Only one of the credentials settings can be set at once. If none is provided, loading default credentials from the environment will be attempted via ADC. For more information about how to provide Google credentials, please refer to https://cloud.google.com/docs/authentication.

auth.oauth2.google.jwt_fileedit

The JWT Account Key file for Google.

Only one of the credentials settings can be set at once. If none is provided, loading default credentials from the environment will be attempted via ADC. For more information about how to provide Google credentials, please refer to https://cloud.google.com/docs/authentication.

auth.oauth2.google.jwt_jsonedit

The JWT Account Key file as raw JSON.

Only one of the credentials settings can be set at once. If none is provided, loading default credentials from the environment will be attempted via ADC. For more information about how to provide Google credentials, please refer to https://cloud.google.com/docs/authentication.

auth.oauth2.google.delegated_accountedit

Email of the delegated account used to create the credentials (usually an admin). Used in combination with auth.oauth2.google.jwt_file or auth.oauth2.google.jwt_json.

request.urledit

The URL of the HTTP API. Required.

request.methodedit

HTTP method to use when making requests. GET or POST are the options. Default: GET.

request.encode_asedit

ContentType used for encoding the request body. If set it will force the encoding in the specified format regardless of the Content-Type header value, otherwise it will honor it if possible or fallback to application/json. By default the requests are sent with Content-Type: application/json. Supported values: application/json and application/x-www-form-urlencoded. application/x-www-form-urlencoded will url encode the url.params and set them as the body. It is not set by default.

request.bodyedit

An optional HTTP POST body. The configuration value must be an object, and it will be encoded to JSON. This is only valid when request.method is POST. Defaults to null (no HTTP body).

- type: httpjson
  request.method: POST
  request.body:
    query:
      bool:
        filter:
          term:
            type: authentication

request.timeoutedit

Duration before declaring that the HTTP client connection has timed out. Valid time units are ns, us, ms, s, m, h. Default: 30s.

request.ssledit

This specifies SSL/TLS configuration. If the ssl section is missing, the host’s CAs are used for HTTPS connections. See SSL for more information.

request.proxy_urledit

This specifies proxy configuration in the form of http[s]://<user>:<password>@<server name/ip>:<port>

filebeat.inputs:
# Fetch your public IP every minute.
- type: httpjson
  interval: 1m
  request.url: https://api.ipify.org/?format=json
  request.proxy_url: http://proxy.example:8080

request.keep_alive.disableedit

This specifies whether to disable keep-alives for HTTP end-points. Default: true.

request.keep_alive.max_idle_connectionsedit

The maximum number of idle connections across all hosts. Zero means no limit. Default: 0.

request.keep_alive.max_idle_connections_per_hostedit

The maximum idle connections to keep per-host. If zero, defaults to two. Default: 0.

request.keep_alive.idle_connection_timeoutedit

The maximum amount of time an idle connection will remain idle before closing itself. Valid time units are ns, us, ms, s, m, h. Zero means no limit. Default: 0s.

request.retry.max_attemptsedit

The maximum number of retries for the HTTP client. Default: 5.

request.retry.wait_minedit

The minimum time to wait before a retry is attempted. Default: 1s.

request.retry.wait_maxedit

The maximum time to wait before a retry is attempted. Default: 60s.

request.redirect.forward_headersedit

When set to true request headers are forwarded in case of a redirect. Default: false.

request.redirect.headers_ban_listedit

When redirect.forward_headers is set to true, all headers except the ones defined in this list will be forwarded. Default: [].

request.redirect.max_redirectsedit

The maximum number of redirects to follow for a request. Default: 10.

request.rate_limit.limitedit

The value of the response that specifies the total limit. It is defined with a Go template value. Can read state from: [.last_response.header]

request.rate_limit.remainingedit

The value of the response that specifies the remaining quota of the rate limit. It is defined with a Go template value. Can read state from: [.last_response.header] If the remaining header is missing from the Response, no rate-limiting will occur.

request.rate_limit.resetedit

The value of the response that specifies the epoch time when the rate limit will reset. It is defined with a Go template value. Can read state from: [.last_response.header]

request.rate_limit.early_limitedit

Optionally start rate-limiting prior to the value specified in the Response.

Under the default behavior, Requests will continue while the remaining value is non-zero. Specifying an early_limit will mean that rate-limiting will occur prior to reaching 0.

  • If the value specified for early_limit is less than 1, the value is treated as a percentage of the Response provided limit. e.g. specifying 0.9 will mean that Requests will continue until reaching 90% of the rate-limit — for a limit value of 120, the rate-limit starts when the remaining reaches 12. If the limit header is missing from the Response, default rate-limiting will occur (when remaining reaches 0).
  • If the value specified for early_limit is greater than or equal to 1, the value is treated as the target value for remaining. e.g. instead of rate-limiting when remaining hits 0, rate-limiting will occur when remaining hits the value specified.

It is not set by default (by default the rate-limiting as specified in the Response is followed).

request.transformsedit

List of transforms to apply to the request before each execution.

Available transforms for request: [append, delete, set].

Can read state from: [.first_response.*,.last_response.*, .parent_last_response.* .last_event.*, .cursor.*, .header.*, .url.*, .body.*].

Can write state to: [body.*, header.*, url.*].

filebeat.inputs:
- type: httpjson
  request.url: http://localhost:9200/_search?scroll=5m
  request.method: POST
  request.transforms:
    - set:
        target: body.from
        value: '[[now (parseDuration "-1h")]]'

The clause .parent_last_response. should only be used from within chain steps and when pagination exists at the root request level. If pagination does not exist at the root level, please use the clause .first_response. to access parent response object from within chains. You can look at this example below for a better idea.

Example Config:

  filebeat.inputs:
  - type: httpjson
    enabled: true
    id: my-httpjson-id
    request.url: http://xyz.com/services/data/v1.0/export_ids/page
    request.method: POST
    interval: 1h
    request.retry.max_attempts: 2
    request.retry.wait_min: 5s
    request.transforms:
    - set:
        target: body.page
        value: 0
    response.request_body_on_pagination: true
    response.pagination:
      - set:
          target: body.page
          value: '[[ .last_response.body.page ]]'
          fail_on_template_error: true
    chain:
    - step:
          request.url: http://xyz.com/services/data/v1.0/$.exportId/export_ids/$.files[:].id/info
          request.method: POST
          request.transforms:
          - set:
              target: body.exportId
              value: '[[ .parent_last_response.body.exportId ]]'
          replace: $.files[:].id
          replace_with: '$.exportId,.parent_last_response.body.exportId'

Here we can see that the chain step uses .parent_last_response.body.exportId only because response.pagination is present for the parent (root) request. However if response.pagination was not present in the parent (root) request, replace_with clause should have used .first_response.body.exportId. This is because when pagination does not exist at the parent level parent_last_response object is not populated with required values for performance reasons, but the first_response object always stores the very first response in the process chain.

request.tracer.filenameedit

It is possible to log httpjson requests and responses to a local file-system for debugging configurations. This option is enabled by setting the request.tracer.filename value. Additional options are available to tune log rotation behavior.

Enabling this option compromises security and should only be used for debugging.

request.tracer.maxsizeedit

This value sets the maximum size, in megabytes, the log file will reach before it is rotated. By default logs are allowed to reach 1MB before rotation.

request.tracer.maxageedit

This specifies the number days to retain rotated log files. If it is not set, log files are retained indefinitely.

request.tracer.maxbackupsedit

The number of old logs to retain. If it is not set all old logs are retained subject to the request.tracer.maxage setting.

request.tracer.localtimeedit

Whether to use the host’s local time rather that UTC for timestamping rotated log file names.

request.tracer.compressedit

This determines whether rotated logs should be gzip compressed.

response.decode_asedit

ContentType used for decoding the response body. If set it will force the decoding in the specified format regardless of the Content-Type header value, otherwise it will honor it if possible or fallback to application/json. Supported values: application/json, application/x-ndjson, text/csv, application/zip. It is not set by default.

For text/csv, one event for each line will be created, using the header values as the object keys. For this reason is always assumed that a header exists.

For application/zip, the zip file is expected to contain one or more .json or .ndjson files. The contents of all of them will be merged into a single list of JSON objects.

response.transformsedit

List of transforms to apply to the response once it is received.

Available transforms for response: [append, delete, set].

Can read state from: [.last_response.*, .last_event.*, .cursor.*, .header.*, .url.*].

Can write state to: [body.*].

filebeat.inputs:
- type: httpjson
  request.url: http://localhost:9200/_search?scroll=5m
  request.method: POST
  response.transforms:
    - delete:
        target: body.very_confidential
  response.split:
    target: body.hits.hits
  response.pagination:
    - set:
        target: url.value
        value: http://localhost:9200/_search/scroll
    - set:
        target: url.params.scroll_id
        value: '[[.last_response.body._scroll_id]]'
    - set:
        target: body.scroll
        value: 5m

response.splitedit

Split operation to apply to the response once it is received. A split can convert a map, array, or string into multiple events. If the split target is empty the parent document will be kept. If documents with empty splits should be dropped, the ignore_empty_value option should be set to true.

response.split[].targetedit

Defines the target field upon the split operation will be performed.

response.split[].typeedit

Defines the field type of the target. Allowed values: array, map, string. string requires the use of the delimiter options to specify what characters to split the string on. delimiter always behaves as if keep_parent is set to true. Default: array.

response.split[].transformsedit

A set of transforms can be defined. This list will be applied after response.transforms and after the object has been modified based on response.split[].keep_parent and response.split[].key_field.

Available transforms for response: [append, delete, set].

Can read state from: [.last_response.*, .first_event.*, .last_event.*, .cursor.*, .header.*, .url.*].

Can write state to: [body.*].

in this context, body.* will be the result of all the previous transformations.

response.split[].keep_parentedit

If set to true, the fields from the parent document (at the same level as target) will be kept. Otherwise a new document will be created using target as the root. Default: false.

response.split[].delimiteredit

Required if using split type of string. This is the sub string used to split the string. For example if delimiter was "\n" and the string was "line 1\nline 2", then the split would result in "line 1" and "line 2".

response.split[].key_fieldedit

Valid when used with type: map. When not empty, defines a new field where the original key value will be stored.

response.split[].ignore_empty_valueedit

If set to true, empty or missing value will be ignored and processing will pass on to the next nested split operation instead of failing with an error. Default: false.

response.split[].splitedit

Nested split operation. Split operations can be nested at will. An event won’t be created until the deepest split operation is applied.

response.request_body_on_paginationedit

If set to true, the values in request.body are sent for pagination requests. Default: false.

response.paginationedit

List of transforms that will be applied to the response to every new page request. All the transforms from request.transform will be executed and then response.pagination will be added to modify the next request as needed. For subsequent responses, the usual response.transforms and response.split will be executed normally.

Available transforms for pagination: [append, delete, set].

Can read state from: [.last_response.*, .first_event.*, .last_event.*, .cursor.*, .header.*, .url.*, .body.*].

Can write state to: [body.*, header.*, url.*].

Examples using split:

  • We have a response with two nested arrays, and we want a document for each of the elements of the inner array:

    {
      "this": "is kept",
      "alerts": [
        {
          "this_is": "also kept",
          "entities": [
            {
              "something": "something"
            },
            {
              "else": "else"
            }
          ]
        },
        {
          "this_is": "also kept 2",
          "entities": [
            {
              "something": "something 2"
            },
            {
              "else": "else 2"
            }
          ]
        }
      ]
    }

    The config will look like:

    filebeat.inputs:
    - type: httpjson
      interval: 1m
      request.url: https://example.com
      response.split:
        target: body.alerts
        type: array
        keep_parent: true
        split:
          # paths in nested splits need to represent the state of body, not only their current level of nesting
          target: body.alerts.entities
          type: array
          keep_parent: true

    This will output:

    [
      {
        "this": "is kept",
        "alerts": {
          "this_is": "also kept",
          "entities": {
            "something": "something"
          }
        }
      },
      {
        "this": "is kept",
        "alerts": {
          "this_is": "also kept",
          "entities": {
            "else": "else"
          }
        }
      },
      {
        "this": "is kept",
        "alerts": {
          "this_is": "also kept 2",
          "entities": {
            "something": "something 2"
          }
        }
      },
      {
        "this": "is kept",
        "alerts": {
          "this_is": "also kept 2",
          "entities": {
            "else": "else 2"
          }
        }
      }
    ]
  • We have a response with an array with two objects, and we want a document for each of the object keys while keeping the keys values:

    {
      "this": "is not kept",
      "alerts": [
        {
          "this_is": "kept",
          "entities": {
            "id1": {
              "something": "something"
            }
          }
        },
        {
          "this_is": "kept 2",
          "entities": {
            "id2": {
              "something": "something 2"
            }
          }
        }
      ]
    }

    The config will look like:

    filebeat.inputs:
    - type: httpjson
      interval: 1m
      request.url: https://example.com
      response.split:
        target: body.alerts
        type: array
        keep_parent: false
        split:
          # this time alerts will not exist because previous keep_parent is false
          target: body.entities
          type: map
          keep_parent: true
          key_field: id

    This will output:

    [
      {
        "this_is": "kept",
        "entities": {
          "id": "id1",
          "something": "something"
        }
      },
      {
        "this_is": "kept 2",
        "entities": {
          "id": "id2",
          "something": "something 2"
        }
      }
    ]
  • We have a response with an array with two objects, and we want a document for each of the object keys while applying a transform to each:

    {
      "this": "is not kept",
      "alerts": [
        {
          "this_is": "also not kept",
          "entities": {
            "id1": {
              "something": "something"
            }
          }
        },
        {
          "this_is": "also not kept",
          "entities": {
            "id2": {
              "something": "something 2"
            }
          }
        }
      ]
    }

    The config will look like:

    filebeat.inputs:
    - type: httpjson
      interval: 1m
      request.url: https://example.com
      response.split:
        target: body.alerts
        type: array
        split:
          transforms:
            - set:
                target: body.new
                value: will be added to each
          target: body.entities
          type: map

    This will output:

    [
      {
        "something": "something",
        "new": "will be added for each"
      },
      {
        "something": "something 2",
        "new": "will be added for each"
      }
    ]
  • We have a response with a keys whose value is a string. We want the string to be split on a delimiter and a document for each sub strings.

    {
      "this": "is kept",
      "lines": "Line 1\nLine 2\nLine 3"
    }

    The config will look like:

    filebeat.inputs:
    - type: httpjson
      interval: 1m
      request.url: https://example.com
      response.split:
        target: body.lines
        type: string
        delimiter: "\n"

    This will output:

    [
      {
        "this": "is kept",
        "lines": "Line 1"
      },
      {
        "this": "is kept",
        "lines": "Line 2"
      },
      {
        "this": "is kept",
        "lines": "Line 3"
      }
    ]

chainedit

A chain is a list of requests to be made after the first one.

chain[].stepedit

Contains basic request and response configuration for chained calls.

chain[].step.requestedit

See request parameters. Required.

Example:

First call: https://example.com/services/data/v1.0/

Second call: https://example.com/services/data/v1.0/1/export_ids

Third call: https://example.com/services/data/v1.0/export_ids/file_1/info

chain[].step.response.splitedit

See response split parameter.

+

chain[].step.replaceedit

A JSONPath string to parse values from responses JSON, collected from previous chain steps. Place same replace string in url where collected values from previous call should be placed. Required.

Example:

  • First call: https://example.com/services/data/v1.0/

    response

  • Second call: https://example.com/services/data/v1.0/$.records[:].id/export_ids

    response

  • Third call: https://example.com/services/data/v1.0/export_ids/$.file_name/info
filebeat.inputs:
- type: httpjson
  enabled: true
  # first call
  request.url: https://example.com/services/data/v1.0/records
  interval: 1h
  chain:
    # second call
    - step:
      request.url: https://example.com/services/data/v1.0/$.records[:].id/export_ids
      request.method: GET
      replace: $.records[:].id
    # third call
    - step:
      request.url: https://example.com/services/data/v1.0/export_ids/$.file_name/info
      request.method: GET
      replace: $.file_name

Example:

  • First call to collect record ids

    request_url: https://example.com/services/data/v1.0/records

    response_json:

    {
        "records": [
            {
                "id": 1,
            },
            {
                "id": 2,
            },
            {
                "id": 3,
            },
        ]
    }
  • Second call to collect file_name using collected ids from first call.

    request_url using id as 1: https://example.com/services/data/v1.0/1/export_ids

    response_json using id as 1:

    {
        "file_name": "file_1"
    }

    request_url using id as 2: https://example.com/services/data/v1.0/2/export_ids

    response_json using id as 2:

    {
        "file_name": "file_2"
    }
  • Third call to collect files using collected file_name from second call.

    request_url using file_name as file_1: https://example.com/services/data/v1.0/export_ids/file_1/info

    request_url using file_name as file_2: https://example.com/services/data/v1.0/export_ids/file_2/info

    Collect and make events from response in any format supported by httpjson for all calls.

    chain[].step.replace_withedit

The replace_with: "pattern,value" clause is used to replace a fixed pattern string defined in request.url with the given value. The fixed pattern must have a $. prefix, for example: $.xyz. The value may be hard coded or extracted from context variables like [.last_response.*, .first_response.*, .parent_last_response.*] etc. The replace_with clause can be used in combination with the replace clause thus providing a lot of flexibility in the logic of chain requests.

Example:

  • First call: https://example.com/services/data/v1.0/exports

    response

  • Second call: https://example.com/services/data/v1.0/$.exportId/files

    response

filebeat.inputs:
- type: httpjson
  enabled: true
  # first call
  request.url: https://example.com/services/data/v1.0/exports
  interval: 1h
  chain:
    # second call
    - step:
      request.url: https://example.com/services/data/v1.0/$.exportId/files
      request.method: GET
      replace_with: '$.exportId,.first_response.body.exportId'

Example:

  • First call to fetch exportId

    request_url: https://example.com/services/data/v1.0/exports

    response_json:

    {
        "exportId" : "2212"
    }
  • Second call to fetch file ids using exportId from first call.

    request_url using exportId as 2212: https://example.com/services/data/v1.0/2212/files

    response_json using exportId as 2212:

    {
        "files": [
            {
                "id": 1,
            },
            {
                "id": 2,
            },
            {
                "id": 3,
            },
        ]
    }

    This behaviour of targeted fixed pattern replacement in the url helps solve various use cases.

Some useful points to remember:-

  1. If you want the value to be treated as an expression to be evaluated for data extraction from context variables, it should always have a single . (dot) prefix. Example: replace_with: '$.exportId,.first_response.body.exportId'. Anything more or less will have the internal processor treat it as a hard coded value, replace_with: '$.exportId,..first_response.body.exportId' (more than one . (dot) as prefix) or replace_with:'$.exportId,first_response.body.exportId' (no . dot as prefix)
  2. Incomplete value expressions will cause an error while processing. Example: replace_with: '$.exportId,.first_response.', replace_with: '$.exportId,.last_response.' etc. These expressions are incomplete because they do not evaluate down to a valid key that can be extracted from the context variables. The value expression: .first_response., on processing, will result in an array [first_response ""] where the key to be extrated becomes "" (an empty string), which has no definition within any context variable.

Fixed patterns must not contain commas in their definition. String replacement patterns are matched by the replace_with processor with exact string matching.

chain[].whileedit

Contains basic request and response configuration for chained while calls. Chained while calls will keep making the requests for a given number of times until a condition is met or the maximum number of attempts gets exhausted. While chain has an attribute until which holds the expression to be evaluated. Ideally the until field should always be used together with the attributes request.retry.max_attempts and request.retry.wait_min which specifies the maximum number of attempts to evaluate until before giving up and the maximum wait time in between such requests. If request.retry.max_attempts is not specified, it will only try to evaluate the expression once and give up if it fails. If request.retry.wait_min is not specified the default wait time will always be 0 as in successive calls will be made immediately.

chain[].while.requestedit

See request parameters .

Example:

First call: http://example.com/services/data/v1.0/exports

Second call: http://example.com/services/data/v1.0/9ef0e6a5/export_ids/status

Third call: http://example.com/services/data/v1.0/export_ids/1/info

chain[].while.response.splitedit

See response split parameter .

chain[].while.replaceedit

See chain[].step.replace .

Example:

  • First call: http://example.com/services/data/v1.0/exports

    response

  • Second call: http://example.com/services/data/v1.0/$.exportId/export_ids/status

    response

  • Third call: http://example.com/services/data/v1.0/export_ids/$.files[:].id/info

    response 1, response 2

filebeat.inputs:
- type: httpjson
  enabled: true
  # first call
  id: my-httpjson-id
  request.url: http://example.com/services/data/v1.0/exports
  interval: 1h
  chain:
    # second call
    - while:
        request.url: http://example.com/services/data/v1.0/$.exportId/export_ids/status
        request.method: GET
        replace: $.exportId
        until: '[[ eq .last_response.body.status "completed" ]]'
        request.retry.max_attempts: 5
        request.retry.wait_min: 5s
    # third call
    - step:
        request.url: http://example.com/services/data/v1.0/export_ids/$.files[:].id/info
        request.method: GET
        replace: $.files[:].id

Example:

  • First call to collect export ids

    request_url: https://example.com/services/data/v1.0/exports

    response_json:

    {
        "exportId": "9ef0e6a5"
    }
  • Second call to collect file_ids using collected id from first call when response.body.sataus == "completed". This call continues until the condition is satisfied or the maximum number of attempts gets exhausted.

    request_url using id as 9ef0e6a5: https://example.com/services/data/v1.0/9ef0e6a5/export_ids/status

    response_json using id as 9ef0e6a5:

    {
        "status": "completed",
        "files": [
          {
            "id": 1
          },
          {
            "id": 2
          },
          {
            "id": 3
          }
        ]
    }
  • Third call to collect files using collected file_id from second call.

    request_url using file_id as 1: https://example.com/services/data/v1.0/export_ids/1/info

    request_url using file_id as 2: https://example.com/services/data/v1.0/export_ids/2/info

    response_json using id as 1:

    {
        "file_name": "file_1",
        "file_data": "some data"
    }

    response_json using id as 2:

    {
        "file_name": "file_2",
        "file_data": "some data"
    }

    Collect and make events from response in any format supported by httpjson for all calls.

httpjson chain will only create and ingest events from last call on chained configurations. Also, the current chain only supports the following: all request parameters, response.transforms and response.split.

chain[].while.replace_withedit

See chain[].step.replace_with .

cursoredit

Cursor is a list of key value objects where arbitrary values are defined. The values are interpreted as value templates and a default template can be set. Cursor state is kept between input restarts and updated once all the events for a request are published.

Each cursor entry is formed by:

  • A value template, which will define the value to store when evaluated.
  • A default template, which will define the value to store when the value template fails or is empty.
  • An ignore_empty_value flag. When set to true, will not store empty values, preserving the previous one, if any. Default: true.

Can read state from: [.last_response.*, .first_event.*, .last_event.*].

Default templates do not have access to any state, only to functions.

filebeat.inputs:
- type: httpjson
  interval: 1m
  request.url: https://api.ipify.org/?format=json
  response.transforms:
    - set:
        target: body.last_requested_at
        value: '[[.cursor.last_requested_at]]'
        default: "[[now]]"
  cursor:
    last_requested_at:
      value: '[[now]]'
  processors:
    - decode_json_fields:
        fields: ["message"]
        target: "json"

Request life cycleedit

Request lifecycle

  1. At every defined interval a new request is created.
  2. The request is transformed using the configured request.transforms.
  3. The resulting transformed request is executed.
  4. The server responds (here is where any retry or rate limit policy takes place when configured).
  5. The response is transformed using the configured response.transforms and response.split.
  6. If a chain step is configured. Each step will generate new requests based on collected IDs from responses. The requests will be transformed using configured request.transforms and the resulting generated transformed requests will be executed. This process will happen for all the steps mentioned in the chain.
  7. Each resulting event is published to the output.
  8. If a response.pagination is configured and there are more pages, a new request is created using it, otherwise the process ends until the next interval.

Chain Request lifecycle

  1. Response from regular call will be processed.
  2. Extract data from response and generate new requests from responses.
  3. Process generated requests and collect responses from server.
  4. Go back to step-2 for the next step.
  5. Publish collected responses from the last chain step.

Common optionsedit

The following configuration options are supported by all inputs.

enablededit

Use the enabled option to enable and disable inputs. By default, enabled is set to true.

tagsedit

A list of tags that Filebeat includes in the tags field of each published event. Tags make it easy to select specific events in Kibana or apply conditional filtering in Logstash. These tags will be appended to the list of tags specified in the general configuration.

Example:

filebeat.inputs:
- type: httpjson
  . . .
  tags: ["json"]
fieldsedit

Optional fields that you can specify to add additional information to the output. For example, you might add fields that you can use for filtering log data. Fields can be scalar values, arrays, dictionaries, or any nested combination of these. By default, the fields that you specify here will be grouped under a fields sub-dictionary in the output document. To store the custom fields as top-level fields, set the fields_under_root option to true. If a duplicate field is declared in the general configuration, then its value will be overwritten by the value declared here.

filebeat.inputs:
- type: httpjson
  . . .
  fields:
    app_id: query_engine_12
fields_under_rootedit

If this option is set to true, the custom fields are stored as top-level fields in the output document instead of being grouped under a fields sub-dictionary. If the custom field names conflict with other field names added by Filebeat, then the custom fields overwrite the other fields.

processorsedit

A list of processors to apply to the input data.

See Processors for information about specifying processors in your config.

pipelineedit

The ingest pipeline ID to set for the events generated by this input.

The pipeline ID can also be configured in the Elasticsearch output, but this option usually results in simpler configuration files. If the pipeline is configured both in the input and output, the option from the input is used.

keep_nulledit

If this option is set to true, fields with null values will be published in the output document. By default, keep_null is set to false.

indexedit

If present, this formatted string overrides the index for events from this input (for elasticsearch outputs), or sets the raw_index field of the event’s metadata (for other outputs). This string can only refer to the agent name and version and the event timestamp; for access to dynamic fields, use output.elasticsearch.index or a processor.

Example value: "%{[agent.name]}-myindex-%{+yyyy.MM.dd}" might expand to "filebeat-myindex-2019.11.01".

publisher_pipeline.disable_hostedit

By default, all events contain host.name. This option can be set to true to disable the addition of this field to all events. The default value is false.