API Referenceedit
bulkedit
Allows to perform multiple index/update/delete operations in a single request.
client.bulk({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string): Default index for items which don’t provide one -
operations
(Optional, { index, create, update, delete } | { detect_noop, doc, doc_as_upsert, script, scripted_upsert, _source, upsert } | object[]) -
pipeline
(Optional, string): The pipeline id to preprocess incoming documents with -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
(the default) then do nothing with refreshes. -
routing
(Optional, string): Specific routing value -
_source
(Optional, boolean | string | string[]): True or false to return the _source field or not, or default list of fields to return, can be overridden on each sub-request -
_source_excludes
(Optional, string | string[]): Default list of fields to exclude from the returned _source field, can be overridden on each sub-request -
_source_includes
(Optional, string | string[]): Default list of fields to extract and return from the _source field, can be overridden on each sub-request -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of shard copies that must be active before proceeding with the bulk operation. Defaults to 1, meaning the primary shard only. Set toall
for all shard copies, otherwise set to any non-negative value less than or equal to the total number of copies for the shard (number of replicas + 1) -
require_alias
(Optional, boolean): Sets require_alias for all incoming documents. Defaults to unset (false)
-
clear_scrolledit
Explicitly clears the search context for a scroll.
client.clearScroll({ ... })
Argumentsedit
-
Request (object):
-
scroll_id
(Optional, string | string[]): A list of scroll IDs to clear
-
close_point_in_timeedit
Close a point in time
client.closePointInTime({ id })
Argumentsedit
-
Request (object):
-
id
(string)
-
countedit
Returns number of documents matching a query.
client.count({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of indices to restrict the results -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
analyzer
(Optional, string): The analyzer to use for the query string -
analyze_wildcard
(Optional, boolean): Specify whether wildcard and prefix queries should be analyzed (default: false) -
default_operator
(Optional, Enum("and" | "or")): The default operator for query string query (AND or OR) -
df
(Optional, string): The field to use as default where no field prefix is given in the query string -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_throttled
(Optional, boolean): Whether specified concrete, expanded or aliased indices should be ignored when throttled -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
lenient
(Optional, boolean): Specify whether format-based query failures (such as providing text to a numeric field) should be ignored -
min_score
(Optional, number): Include only documents with a specific_score
value in the result -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
routing
(Optional, string): A list of specific routing values -
terminate_after
(Optional, number): The maximum count for each shard, upon reaching which the query execution will terminate early -
q
(Optional, string): Query in the Lucene query string syntax
-
createedit
Creates a new document in the index.
Returns a 409 response when a document with a same ID already exists in the index.
client.create({ id, index })
Argumentsedit
-
Request (object):
-
id
(string): Document ID -
index
(string): The name of the index -
document
(Optional, object): A document. -
pipeline
(Optional, string): The pipeline id to preprocess incoming documents with -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
(the default) then do nothing with refreshes. -
routing
(Optional, string): Specific routing value -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
version
(Optional, number): Explicit version number for concurrency control -
version_type
(Optional, Enum("internal" | "external" | "external_gte" | "force")): Specific version type -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of shard copies that must be active before proceeding with the index operation. Defaults to 1, meaning the primary shard only. Set toall
for all shard copies, otherwise set to any non-negative value less than or equal to the total number of copies for the shard (number of replicas + 1)
-
deleteedit
Removes a document from the index.
client.delete({ id, index })
Argumentsedit
-
Request (object):
-
id
(string): The document ID -
index
(string): The name of the index -
if_primary_term
(Optional, number): only perform the delete operation if the last operation that has changed the document has the specified primary term -
if_seq_no
(Optional, number): only perform the delete operation if the last operation that has changed the document has the specified sequence number -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
(the default) then do nothing with refreshes. -
routing
(Optional, string): Specific routing value -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
version
(Optional, number): Explicit version number for concurrency control -
version_type
(Optional, Enum("internal" | "external" | "external_gte" | "force")): Specific version type -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of shard copies that must be active before proceeding with the delete operation. Defaults to 1, meaning the primary shard only. Set toall
for all shard copies, otherwise set to any non-negative value less than or equal to the total number of copies for the shard (number of replicas + 1)
-
delete_by_queryedit
Deletes documents matching the provided query.
client.deleteByQuery({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names to search; use_all
or empty string to perform the operation on all indices -
max_docs
(Optional, number) -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
slice
(Optional, { field, id, max }) -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
analyzer
(Optional, string): The analyzer to use for the query string -
analyze_wildcard
(Optional, boolean): Specify whether wildcard and prefix queries should be analyzed (default: false) -
conflicts
(Optional, Enum("abort" | "proceed")): What to do when the delete by query hits version conflicts? -
default_operator
(Optional, Enum("and" | "or")): The default operator for query string query (AND or OR) -
df
(Optional, string): The field to use as default where no field prefix is given in the query string -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
from
(Optional, number): Starting offset (default: 0) -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
lenient
(Optional, boolean): Specify whether format-based query failures (such as providing text to a numeric field) should be ignored -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
refresh
(Optional, boolean): Should the affected indexes be refreshed? -
request_cache
(Optional, boolean): Specify if request cache should be used for this request or not, defaults to index level setting -
requests_per_second
(Optional, float): The throttle for this request in sub-requests per second. -1 means no throttle. -
routing
(Optional, string): A list of specific routing values -
q
(Optional, string): Query in the Lucene query string syntax -
scroll
(Optional, string | -1 | 0): Specify how long a consistent view of the index should be maintained for scrolled search -
scroll_size
(Optional, number): Size on the scroll request powering the delete by query -
search_timeout
(Optional, string | -1 | 0): Explicit timeout for each search request. Defaults to no timeout. -
search_type
(Optional, Enum("query_then_fetch" | "dfs_query_then_fetch")): Search operation type -
slices
(Optional, number | Enum("auto")): The number of slices this task should be divided into. Defaults to 1, meaning the task isn’t sliced into subtasks. Can be set toauto
. -
sort
(Optional, string[]): A list of <field>:<direction> pairs -
stats
(Optional, string[]): Specific tag of the request for logging and statistical purposes -
terminate_after
(Optional, number): The maximum number of documents to collect for each shard, upon reaching which the query execution will terminate early. -
timeout
(Optional, string | -1 | 0): Time each individual bulk request should wait for shards that are unavailable. -
version
(Optional, boolean): Specify whether to return document version as part of a hit -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of shard copies that must be active before proceeding with the delete by query operation. Defaults to 1, meaning the primary shard only. Set toall
for all shard copies, otherwise set to any non-negative value less than or equal to the total number of copies for the shard (number of replicas + 1) -
wait_for_completion
(Optional, boolean): Should the request should block until the delete by query is complete.
-
delete_by_query_rethrottleedit
Changes the number of requests per second for a particular Delete By Query operation.
client.deleteByQueryRethrottle({ task_id })
Argumentsedit
-
Request (object):
-
task_id
(string | number): The task id to rethrottle -
requests_per_second
(Optional, float): The throttle to set on this request in floating sub-requests per second. -1 means set no throttle.
-
delete_scriptedit
Deletes a script.
client.deleteScript({ id })
Argumentsedit
-
Request (object):
-
id
(string): Script ID -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
existsedit
Returns information about whether a document exists in an index.
client.exists({ id, index })
Argumentsedit
-
Request (object):
-
id
(string): The document ID -
index
(string): The name of the index -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
realtime
(Optional, boolean): Specify whether to perform the operation in realtime or search mode -
refresh
(Optional, boolean): Refresh the shard containing the document before performing the operation -
routing
(Optional, string): Specific routing value -
_source
(Optional, boolean | string | string[]): True or false to return the _source field or not, or a list of fields to return -
_source_excludes
(Optional, string | string[]): A list of fields to exclude from the returned _source field -
_source_includes
(Optional, string | string[]): A list of fields to extract and return from the _source field -
stored_fields
(Optional, string | string[]): A list of stored fields to return in the response -
version
(Optional, number): Explicit version number for concurrency control -
version_type
(Optional, Enum("internal" | "external" | "external_gte" | "force")): Specific version type
-
exists_sourceedit
Returns information about whether a document source exists in an index.
client.existsSource({ id, index })
Argumentsedit
-
Request (object):
-
id
(string): The document ID -
index
(string): The name of the index -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
realtime
(Optional, boolean): Specify whether to perform the operation in realtime or search mode -
refresh
(Optional, boolean): Refresh the shard containing the document before performing the operation -
routing
(Optional, string): Specific routing value -
_source
(Optional, boolean | string | string[]): True or false to return the _source field or not, or a list of fields to return -
_source_excludes
(Optional, string | string[]): A list of fields to exclude from the returned _source field -
_source_includes
(Optional, string | string[]): A list of fields to extract and return from the _source field -
version
(Optional, number): Explicit version number for concurrency control -
version_type
(Optional, Enum("internal" | "external" | "external_gte" | "force")): Specific version type
-
explainedit
Returns information about why a specific matches (or doesn’t match) a query.
client.explain({ id, index })
Argumentsedit
-
Request (object):
-
id
(string): The document ID -
index
(string): The name of the index -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
analyzer
(Optional, string): The analyzer for the query string query -
analyze_wildcard
(Optional, boolean): Specify whether wildcards and prefix queries in the query string query should be analyzed (default: false) -
default_operator
(Optional, Enum("and" | "or")): The default operator for query string query (AND or OR) -
df
(Optional, string): The default field for query string query (default: _all) -
lenient
(Optional, boolean): Specify whether format-based query failures (such as providing text to a numeric field) should be ignored -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
routing
(Optional, string): Specific routing value -
_source
(Optional, boolean | string | string[]): True or false to return the _source field or not, or a list of fields to return -
_source_excludes
(Optional, string | string[]): A list of fields to exclude from the returned _source field -
_source_includes
(Optional, string | string[]): A list of fields to extract and return from the _source field -
stored_fields
(Optional, string | string[]): A list of stored fields to return in the response -
q
(Optional, string): Query in the Lucene query string syntax
-
field_capsedit
Returns the information about the capabilities of fields among multiple indices.
client.fieldCaps({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): List of data streams, indices, and aliases used to limit the request. Supports wildcards (*). To target all data streams and indices, omit this parameter or use * or _all. -
fields
(Optional, string | string[]): List of fields to retrieve capabilities for. Wildcard (*
) expressions are supported. -
index_filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): Allows to filter indices if the provided query rewrites to match_none on every shard. -
runtime_mappings
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>): Defines ad-hoc runtime fields in the request similar to the way it is done in search requests. These fields exist only as part of the query and take precedence over fields defined with the same name in the index mappings. -
allow_no_indices
(Optional, boolean): If false, the request returns an error if any wildcard expression, index alias, or_all
value targets only missing or closed indices. This behavior applies even if the request targets other open indices. For example, a request targetingfoo*,bar*
returns an error if an index starts with foo but no index starts with bar. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard patterns can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. Supports a list of values, such asopen,hidden
. -
ignore_unavailable
(Optional, boolean): Iftrue
, missing or closed indices are not included in the response. -
include_unmapped
(Optional, boolean): If true, unmapped fields are included in the response. -
filters
(Optional, string): An optional set of filters: can include +metadata,-metadata,-nested,-multifield,-parent -
types
(Optional, string[]): Only return results for fields that have one of the types in the list
-
getedit
Returns a document.
client.get({ id, index })
Argumentsedit
-
Request (object):
-
id
(string): Unique identifier of the document. -
index
(string): Name of the index that contains the document. -
preference
(Optional, string): Specifies the node or shard the operation should be performed on. Random by default. -
realtime
(Optional, boolean): Boolean) If true, the request is real-time as opposed to near-real-time. -
refresh
(Optional, boolean): If true, Elasticsearch refreshes the affected shards to make this operation visible to search. If false, do nothing with refreshes. -
routing
(Optional, string): Target the specified primary shard. -
_source
(Optional, boolean | string | string[]): True or false to return the _source field or not, or a list of fields to return. -
_source_excludes
(Optional, string | string[]): A list of source fields to exclude in the response. -
_source_includes
(Optional, string | string[]): A list of source fields to include in the response. -
stored_fields
(Optional, string | string[]): A list of stored fields to return in the response -
version
(Optional, number): Explicit version number for concurrency control. The specified version must match the current version of the document for the request to succeed. -
version_type
(Optional, Enum("internal" | "external" | "external_gte" | "force")): Specific version type: internal, external, external_gte.
-
get_scriptedit
Returns a script.
client.getScript({ id })
Argumentsedit
-
Request (object):
-
id
(string): Script ID -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master
-
get_script_contextedit
Returns all script contexts.
client.getScriptContext()
get_script_languagesedit
Returns available script types, languages and contexts
client.getScriptLanguages()
get_sourceedit
Returns the source of a document.
client.getSource({ id, index })
Argumentsedit
-
Request (object):
-
id
(string): Unique identifier of the document. -
index
(string): Name of the index that contains the document. -
preference
(Optional, string): Specifies the node or shard the operation should be performed on. Random by default. -
realtime
(Optional, boolean): Boolean) If true, the request is real-time as opposed to near-real-time. -
refresh
(Optional, boolean): If true, Elasticsearch refreshes the affected shards to make this operation visible to search. If false, do nothing with refreshes. -
routing
(Optional, string): Target the specified primary shard. -
_source
(Optional, boolean | string | string[]): True or false to return the _source field or not, or a list of fields to return. -
_source_excludes
(Optional, string | string[]): A list of source fields to exclude in the response. -
_source_includes
(Optional, string | string[]): A list of source fields to include in the response. -
stored_fields
(Optional, string | string[]) -
version
(Optional, number): Explicit version number for concurrency control. The specified version must match the current version of the document for the request to succeed. -
version_type
(Optional, Enum("internal" | "external" | "external_gte" | "force")): Specific version type: internal, external, external_gte.
-
health_reportedit
Returns the health of the cluster.
client.healthReport({ ... })
Argumentsedit
-
Request (object):
-
feature
(Optional, string | string[]): A feature of the cluster, as returned by the top-level health report API. -
timeout
(Optional, string | -1 | 0): Explicit operation timeout. -
verbose
(Optional, boolean): Opt-in for more information about the health of the system. -
size
(Optional, number): Limit the number of affected resources the health report API returns.
-
indexedit
Creates or updates a document in an index.
client.index({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the index -
id
(Optional, string): Document ID -
document
(Optional, object): A document. -
if_primary_term
(Optional, number): only perform the index operation if the last operation that has changed the document has the specified primary term -
if_seq_no
(Optional, number): only perform the index operation if the last operation that has changed the document has the specified sequence number -
op_type
(Optional, Enum("index" | "create")): Explicit operation type. Defaults toindex
for requests with an explicit document ID, and to `create`for requests without an explicit document ID -
pipeline
(Optional, string): The pipeline id to preprocess incoming documents with -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
(the default) then do nothing with refreshes. -
routing
(Optional, string): Specific routing value -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
version
(Optional, number): Explicit version number for concurrency control -
version_type
(Optional, Enum("internal" | "external" | "external_gte" | "force")): Specific version type -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of shard copies that must be active before proceeding with the index operation. Defaults to 1, meaning the primary shard only. Set toall
for all shard copies, otherwise set to any non-negative value less than or equal to the total number of copies for the shard (number of replicas + 1) -
require_alias
(Optional, boolean): When true, requires destination to be an alias. Default is false
-
infoedit
Returns basic information about the cluster.
client.info()
knn_searchedit
Performs a kNN search.
client.knnSearch({ index, knn })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names to search; use_all
or to perform the operation on all indices -
knn
({ field, query_vector, k, num_candidates }): kNN query to execute -
_source
(Optional, boolean | { excludes, includes }): Indicates which source fields are returned for matching documents. These fields are returned in the hits._source property of the search response. -
docvalue_fields
(Optional, { field, format, include_unmapped }[]): The request returns doc values for field names matching these patterns in the hits.fields property of the response. Accepts wildcard (*) patterns. -
stored_fields
(Optional, string | string[]): List of stored fields to return as part of a hit. If no fields are specified, no stored fields are included in the response. If this field is specified, the _source parameter defaults to false. You can pass _source: true to return both source fields and stored fields in the search response. -
fields
(Optional, string | string[]): The request returns values for field names matching these patterns in the hits.fields property of the response. Accepts wildcard (*) patterns. -
filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type } | { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }[]): Query to filter the documents that can match. The kNN search will return the topk
documents that also match this filter. The value can be a single query or a list of queries. Iffilter
isn’t provided, all documents are allowed to match. -
routing
(Optional, string): A list of specific routing values
-
mgetedit
Allows to get multiple documents in one request.
client.mget({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string): Name of the index to retrieve documents from whenids
are specified, or when a document in thedocs
array does not specify an index. -
docs
(Optional, { _id, _index, routing, _source, stored_fields, version, version_type }[]): The documents you want to retrieve. Required if no index is specified in the request URI. -
ids
(Optional, string | string[]): The IDs of the documents you want to retrieve. Allowed when the index is specified in the request URI. -
preference
(Optional, string): Specifies the node or shard the operation should be performed on. Random by default. -
realtime
(Optional, boolean): Iftrue
, the request is real-time as opposed to near-real-time. -
refresh
(Optional, boolean): Iftrue
, the request refreshes relevant shards before retrieving documents. -
routing
(Optional, string): Custom value used to route operations to a specific shard. -
_source
(Optional, boolean | string | string[]): True or false to return the_source
field or not, or a list of fields to return. -
_source_excludes
(Optional, string | string[]): A list of source fields to exclude from the response. You can also use this parameter to exclude fields from the subset specified in_source_includes
query parameter. -
_source_includes
(Optional, string | string[]): A list of source fields to include in the response. If this parameter is specified, only these source fields are returned. You can exclude fields from this subset using the_source_excludes
query parameter. If the_source
parameter isfalse
, this parameter is ignored. -
stored_fields
(Optional, string | string[]): Iftrue
, retrieves the document fields stored in the index rather than the document_source
.
-
msearchedit
Allows to execute several search operations in one request.
client.msearch({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): List of data streams, indices, and index aliases to search. -
searches
(Optional, { allow_no_indices, expand_wildcards, ignore_unavailable, index, preference, request_cache, routing, search_type, ccs_minimize_roundtrips, allow_partial_search_results, ignore_throttled } | { aggregations, collapse, query, explain, ext, stored_fields, docvalue_fields, knn, from, highlight, indices_boost, min_score, post_filter, profile, rescore, script_fields, search_after, size, sort, _source, fields, terminate_after, stats, timeout, track_scores, track_total_hits, version, runtime_mappings, seq_no_primary_term, pit, suggest }[]) -
allow_no_indices
(Optional, boolean): If false, the request returns an error if any wildcard expression, index alias, or _all value targets only missing or closed indices. This behavior applies even if the request targets other open indices. For example, a request targeting foo*,bar* returns an error if an index starts with foo but no index starts with bar. -
ccs_minimize_roundtrips
(Optional, boolean): If true, network roundtrips between the coordinating node and remote clusters are minimized for cross-cluster search requests. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard expressions can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. -
ignore_throttled
(Optional, boolean): If true, concrete, expanded or aliased indices are ignored when frozen. -
ignore_unavailable
(Optional, boolean): If true, missing or closed indices are not included in the response. -
max_concurrent_searches
(Optional, number): Maximum number of concurrent searches the multi search API can execute. -
max_concurrent_shard_requests
(Optional, number): Maximum number of concurrent shard requests that each sub-search request executes per node. -
pre_filter_shard_size
(Optional, number): Defines a threshold that enforces a pre-filter roundtrip to prefilter search shards based on query rewriting if the number of shards the search request expands to exceeds the threshold. This filter roundtrip can limit the number of shards significantly if for instance a shard can not match any documents based on its rewrite method i.e., if date filters are mandatory to match but the shard bounds and the query are disjoint. -
rest_total_hits_as_int
(Optional, boolean): If true, hits.total are returned as an integer in the response. Defaults to false, which returns an object. -
routing
(Optional, string): Custom routing value used to route search operations to a specific shard. -
search_type
(Optional, Enum("query_then_fetch" | "dfs_query_then_fetch")): Indicates whether global term and document frequencies should be used when scoring returned documents. -
typed_keys
(Optional, boolean): Specifies whether aggregation and suggester names should be prefixed by their respective types in the response.
-
msearch_templateedit
Allows to execute several search template operations in one request.
client.msearchTemplate({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names to use as default -
search_templates
(Optional, { allow_no_indices, expand_wildcards, ignore_unavailable, index, preference, request_cache, routing, search_type, ccs_minimize_roundtrips, allow_partial_search_results, ignore_throttled } | { aggregations, collapse, query, explain, ext, stored_fields, docvalue_fields, knn, from, highlight, indices_boost, min_score, post_filter, profile, rescore, script_fields, search_after, size, sort, _source, fields, terminate_after, stats, timeout, track_scores, track_total_hits, version, runtime_mappings, seq_no_primary_term, pit, suggest }[]) -
ccs_minimize_roundtrips
(Optional, boolean): Indicates whether network round-trips should be minimized as part of cross-cluster search requests execution -
max_concurrent_searches
(Optional, number): Controls the maximum number of concurrent searches the multi search api will execute -
search_type
(Optional, Enum("query_then_fetch" | "dfs_query_then_fetch")): Search operation type -
rest_total_hits_as_int
(Optional, boolean): Indicates whether hits.total should be rendered as an integer or an object in the rest search response -
typed_keys
(Optional, boolean): Specify whether aggregation and suggester names should be prefixed by their respective types in the response
-
mtermvectorsedit
Returns multiple termvectors in one request.
client.mtermvectors({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string): The index in which the document resides. -
docs
(Optional, { _id, _index, routing, _source, stored_fields, version, version_type }[]) -
ids
(Optional, string[]) -
fields
(Optional, string | string[]): A list of fields to return. Applies to all returned documents unless otherwise specified in body "params" or "docs". -
field_statistics
(Optional, boolean): Specifies if document count, sum of document frequencies and sum of total term frequencies should be returned. Applies to all returned documents unless otherwise specified in body "params" or "docs". -
offsets
(Optional, boolean): Specifies if term offsets should be returned. Applies to all returned documents unless otherwise specified in body "params" or "docs". -
payloads
(Optional, boolean): Specifies if term payloads should be returned. Applies to all returned documents unless otherwise specified in body "params" or "docs". -
positions
(Optional, boolean): Specifies if term positions should be returned. Applies to all returned documents unless otherwise specified in body "params" or "docs". -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) .Applies to all returned documents unless otherwise specified in body "params" or "docs". -
realtime
(Optional, boolean): Specifies if requests are real-time as opposed to near-real-time (default: true). -
routing
(Optional, string): Specific routing value. Applies to all returned documents unless otherwise specified in body "params" or "docs". -
term_statistics
(Optional, boolean): Specifies if total term frequency and document frequency should be returned. Applies to all returned documents unless otherwise specified in body "params" or "docs". -
version
(Optional, number): Explicit version number for concurrency control -
version_type
(Optional, Enum("internal" | "external" | "external_gte" | "force")): Specific version type
-
open_point_in_timeedit
Open a point in time that can be used in subsequent searches
client.openPointInTime({ index, keep_alive })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names to open point in time; use_all
or empty string to perform the operation on all indices -
keep_alive
(string | -1 | 0): Specific the time to live for the point in time -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
routing
(Optional, string): Specific routing value -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both.
-
pingedit
Returns whether the cluster is running.
client.ping()
put_scriptedit
Creates or updates a script.
client.putScript({ id, script })
Argumentsedit
-
Request (object):
-
id
(string): Script ID -
script
({ lang, options, source }) -
context
(Optional, string): Script context -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
rank_evaledit
Allows to evaluate the quality of ranked search results over a set of typical search queries
client.rankEval({ requests })
Argumentsedit
-
Request (object):
-
requests
({ id, request, ratings, template_id, params }[]): A set of typical search requests, together with their provided ratings. -
index
(Optional, string | string[]): List of data streams, indices, and index aliases used to limit the request. Wildcard (*
) expressions are supported. To target all data streams and indices in a cluster, omit this parameter or use_all
or*
. -
metric
(Optional, { precision, recall, mean_reciprocal_rank, dcg, expected_reciprocal_rank }): Definition of the evaluation metric to calculate. -
allow_no_indices
(Optional, boolean): Iffalse
, the request returns an error if any wildcard expression, index alias, or_all
value targets only missing or closed indices. This behavior applies even if the request targets other open indices. For example, a request targetingfoo*,bar*
returns an error if an index starts withfoo
but no index starts withbar
. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Iftrue
, missing or closed indices are not included in the response. -
search_type
(Optional, string): Search operation type
-
reindexedit
Allows to copy documents from one index to another, optionally filtering the source documents by a query, changing the destination index settings, or fetching the documents from a remote cluster.
client.reindex({ dest, source })
Argumentsedit
-
Request (object):
-
dest
({ index, op_type, pipeline, routing, version_type }) -
source
({ index, query, remote, size, slice, sort, _source, runtime_mappings }) -
conflicts
(Optional, Enum("abort" | "proceed")) -
max_docs
(Optional, number) -
script
(Optional, { lang, options, source } | { id }) -
size
(Optional, number) -
refresh
(Optional, boolean): Should the affected indexes be refreshed? -
requests_per_second
(Optional, float): The throttle to set on this request in sub-requests per second. -1 means no throttle. -
scroll
(Optional, string | -1 | 0): Control how long to keep the search context alive -
slices
(Optional, number | Enum("auto")): The number of slices this task should be divided into. Defaults to 1, meaning the task isn’t sliced into subtasks. Can be set toauto
. -
timeout
(Optional, string | -1 | 0): Time each individual bulk request should wait for shards that are unavailable. -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of shard copies that must be active before proceeding with the reindex operation. Defaults to 1, meaning the primary shard only. Set toall
for all shard copies, otherwise set to any non-negative value less than or equal to the total number of copies for the shard (number of replicas + 1) -
wait_for_completion
(Optional, boolean): Should the request should block until the reindex is complete. -
require_alias
(Optional, boolean)
-
reindex_rethrottleedit
Changes the number of requests per second for a particular Reindex operation.
client.reindexRethrottle({ task_id })
Argumentsedit
-
Request (object):
-
task_id
(string): The task id to rethrottle -
requests_per_second
(Optional, float): The throttle to set on this request in floating sub-requests per second. -1 means set no throttle.
-
render_search_templateedit
Allows to use the Mustache language to pre-render a search definition.
client.renderSearchTemplate({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): The id of the stored search template -
file
(Optional, string) -
params
(Optional, Record<string, User-defined value>) -
source
(Optional, string)
-
scripts_painless_executeedit
Allows an arbitrary script to be executed and a result to be returned
client.scriptsPainlessExecute({ ... })
Argumentsedit
-
Request (object):
-
context
(Optional, string) -
context_setup
(Optional, { document, index, query }) -
script
(Optional, { lang, options, source })
-
scrolledit
Allows to retrieve a large numbers of results from a single search request.
client.scroll({ scroll_id })
Argumentsedit
-
Request (object):
-
scroll_id
(string): Scroll ID of the search. -
scroll
(Optional, string | -1 | 0): Period to retain the search context for scrolling. -
rest_total_hits_as_int
(Optional, boolean): If true, the API response’s hit.total property is returned as an integer. If false, the API response’s hit.total property is returned as an object.
-
searchedit
Returns results matching a query.
client.search({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names to search; use_all
or empty string to perform the operation on all indices -
aggregations
(Optional, Record<string, { aggregations, meta, adjacency_matrix, auto_date_histogram, avg, avg_bucket, boxplot, bucket_script, bucket_selector, bucket_sort, bucket_count_ks_test, bucket_correlation, cardinality, categorize_text, children, composite, cumulative_cardinality, cumulative_sum, date_histogram, date_range, derivative, diversified_sampler, extended_stats, extended_stats_bucket, frequent_item_sets, filter, filters, geo_bounds, geo_centroid, geo_distance, geohash_grid, geo_line, geotile_grid, geohex_grid, global, histogram, ip_range, ip_prefix, inference, line, matrix_stats, max, max_bucket, median_absolute_deviation, min, min_bucket, missing, moving_avg, moving_percentiles, moving_fn, multi_terms, nested, normalize, parent, percentile_ranks, percentiles, percentiles_bucket, range, rare_terms, rate, reverse_nested, sampler, scripted_metric, serial_diff, significant_terms, significant_text, stats, stats_bucket, string_stats, sum, sum_bucket, terms, top_hits, t_test, top_metrics, value_count, weighted_avg, variable_width_histogram }>) -
collapse
(Optional, { field, inner_hits, max_concurrent_group_searches, collapse }) -
explain
(Optional, boolean): If true, returns detailed information about score computation as part of a hit. -
ext
(Optional, Record<string, User-defined value>): Configuration of search extensions defined by Elasticsearch plugins. -
from
(Optional, number): Starting document offset. By default, you cannot page through more than 10,000 hits using the from and size parameters. To page through more hits, use the search_after parameter. -
highlight
(Optional, { encoder, fields }) -
track_total_hits
(Optional, boolean | number): Number of hits matching the query to count accurately. If true, the exact number of hits is returned at the cost of some performance. If false, the response does not include the total number of hits matching the query. Defaults to 10,000 hits. -
indices_boost
(Optional, Record<string, number>[]): Boosts the _score of documents from specified indices. -
docvalue_fields
(Optional, { field, format, include_unmapped }[]): Array of wildcard (*) patterns. The request returns doc values for field names matching these patterns in the hits.fields property of the response. -
knn
(Optional, { field, query_vector, query_vector_builder, k, num_candidates, boost, filter, similarity } | { field, query_vector, query_vector_builder, k, num_candidates, boost, filter, similarity }[]): Defines the approximate kNN search to run. -
rank
(Optional, { rrf }): Defines the Reciprocal Rank Fusion (RRF) to use -
min_score
(Optional, number): Minimum _score for matching documents. Documents with a lower _score are not included in the search results. -
post_filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
profile
(Optional, boolean) -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): Defines the search definition using the Query DSL. -
rescore
(Optional, { query, window_size } | { query, window_size }[]) -
script_fields
(Optional, Record<string, { script, ignore_failure }>): Retrieve a script evaluation (based on different fields) for each hit. -
search_after
(Optional, number | number | string | boolean | null | User-defined value[]) -
size
(Optional, number): The number of hits to return. By default, you cannot page through more than 10,000 hits using the from and size parameters. To page through more hits, use the search_after parameter. -
slice
(Optional, { field, id, max }) -
sort
(Optional, string | { _score, _doc, _geo_distance, _script } | string | { _score, _doc, _geo_distance, _script }[]) -
_source
(Optional, boolean | { excludes, includes }): Indicates which source fields are returned for matching documents. These fields are returned in the hits._source property of the search response. -
fields
(Optional, { field, format, include_unmapped }[]): Array of wildcard (*) patterns. The request returns values for field names matching these patterns in the hits.fields property of the response. -
suggest
(Optional, { text }) -
terminate_after
(Optional, number): Maximum number of documents to collect for each shard. If a query reaches this limit, Elasticsearch terminates the query early. Elasticsearch collects documents before sorting. Defaults to 0, which does not terminate query execution early. -
timeout
(Optional, string): Specifies the period of time to wait for a response from each shard. If no response is received before the timeout expires, the request fails and returns an error. Defaults to no timeout. -
track_scores
(Optional, boolean): If true, calculate and return document scores, even if the scores are not used for sorting. -
version
(Optional, boolean): If true, returns document version as part of a hit. -
seq_no_primary_term
(Optional, boolean): If true, returns sequence number and primary term of the last modification of each hit. See Optimistic concurrency control. -
stored_fields
(Optional, string | string[]): List of stored fields to return as part of a hit. If no fields are specified, no stored fields are included in the response. If this field is specified, the _source parameter defaults to false. You can pass _source: true to return both source fields and stored fields in the search response. -
pit
(Optional, { id, keep_alive }): Limits the search to a point in time (PIT). If you provide a PIT, you cannot specify an <index> in the request path. -
runtime_mappings
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>): Defines one or more runtime fields in the search request. These fields take precedence over mapped fields with the same name. -
stats
(Optional, string[]): Stats groups to associate with the search. Each group maintains a statistics aggregation for its associated searches. You can retrieve these stats using the indices stats API. -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
allow_partial_search_results
(Optional, boolean): Indicate if an error should be returned if there is a partial search failure or timeout -
analyzer
(Optional, string): The analyzer to use for the query string -
analyze_wildcard
(Optional, boolean): Specify whether wildcard and prefix queries should be analyzed (default: false) -
batched_reduce_size
(Optional, number): The number of shard results that should be reduced at once on the coordinating node. This value should be used as a protection mechanism to reduce the memory overhead per search request if the potential number of shards in the request can be large. -
ccs_minimize_roundtrips
(Optional, boolean): Indicates whether network round-trips should be minimized as part of cross-cluster search requests execution -
default_operator
(Optional, Enum("and" | "or")): The default operator for query string query (AND or OR) -
df
(Optional, string): The field to use as default where no field prefix is given in the query string -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_throttled
(Optional, boolean): Whether specified concrete, expanded or aliased indices should be ignored when throttled -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
lenient
(Optional, boolean): Specify whether format-based query failures (such as providing text to a numeric field) should be ignored -
max_concurrent_shard_requests
(Optional, number): The number of concurrent shard requests per node this search executes concurrently. This value should be used to limit the impact of the search on the cluster in order to limit the number of concurrent shard requests -
min_compatible_shard_node
(Optional, string): The minimum compatible version that all shards involved in search should have for this request to be successful -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
pre_filter_shard_size
(Optional, number): A threshold that enforces a pre-filter roundtrip to prefilter search shards based on query rewriting if the number of shards the search request expands to exceeds the threshold. This filter roundtrip can limit the number of shards significantly if for instance a shard can not match any documents based on its rewrite method ie. if date filters are mandatory to match but the shard bounds and the query are disjoint. -
request_cache
(Optional, boolean): Specify if request cache should be used for this request or not, defaults to index level setting -
routing
(Optional, string): A list of specific routing values -
scroll
(Optional, string | -1 | 0): Specify how long a consistent view of the index should be maintained for scrolled search -
search_type
(Optional, Enum("query_then_fetch" | "dfs_query_then_fetch")): Search operation type -
suggest_field
(Optional, string): Specifies which field to use for suggestions. -
suggest_mode
(Optional, Enum("missing" | "popular" | "always")): Specify suggest mode -
suggest_size
(Optional, number): How many suggestions to return in response -
suggest_text
(Optional, string): The source text for which the suggestions should be returned. -
typed_keys
(Optional, boolean): Specify whether aggregation and suggester names should be prefixed by their respective types in the response -
rest_total_hits_as_int
(Optional, boolean): Indicates whether hits.total should be rendered as an integer or an object in the rest search response -
_source_excludes
(Optional, string | string[]): A list of fields to exclude from the returned _source field -
_source_includes
(Optional, string | string[]): A list of fields to extract and return from the _source field -
q
(Optional, string): Query in the Lucene query string syntax
-
search_mvtedit
Searches a vector tile for geospatial values. Returns results as a binary Mapbox vector tile.
client.searchMvt({ index, field, zoom, x, y })
Argumentsedit
-
Request (object):
-
index
(string | string[]): List of data streams, indices, or aliases to search -
field
(string): Field containing geospatial data to return -
zoom
(number): Zoom level for the vector tile to search -
x
(number): X coordinate for the vector tile to search -
y
(number): Y coordinate for the vector tile to search -
aggs
(Optional, Record<string, { aggregations, meta, adjacency_matrix, auto_date_histogram, avg, avg_bucket, boxplot, bucket_script, bucket_selector, bucket_sort, bucket_count_ks_test, bucket_correlation, cardinality, categorize_text, children, composite, cumulative_cardinality, cumulative_sum, date_histogram, date_range, derivative, diversified_sampler, extended_stats, extended_stats_bucket, frequent_item_sets, filter, filters, geo_bounds, geo_centroid, geo_distance, geohash_grid, geo_line, geotile_grid, geohex_grid, global, histogram, ip_range, ip_prefix, inference, line, matrix_stats, max, max_bucket, median_absolute_deviation, min, min_bucket, missing, moving_avg, moving_percentiles, moving_fn, multi_terms, nested, normalize, parent, percentile_ranks, percentiles, percentiles_bucket, range, rare_terms, rate, reverse_nested, sampler, scripted_metric, serial_diff, significant_terms, significant_text, stats, stats_bucket, string_stats, sum, sum_bucket, terms, top_hits, t_test, top_metrics, value_count, weighted_avg, variable_width_histogram }>): Sub-aggregations for the geotile_grid.
-
Supports the following aggregation types:
- avg
- cardinality
- max
- min
- sum
buffer
(Optional, number): Size, in pixels, of a clipping buffer outside the tile. This allows renderers
to avoid outline artifacts from geometries that extend past the extent of the tile.
exact_bounds
(Optional, boolean): If false, the meta layer’s feature is the bounding box of the tile.
If true, the meta layer’s feature is a bounding box resulting from a
geo_bounds aggregation. The aggregation runs on <field> values that intersect
the <zoom>/<x>/<y> tile with wrap_longitude set to false. The resulting
bounding box may be larger than the vector tile.
extent
(Optional, number): Size, in pixels, of a side of the tile. Vector tiles are square with equal sides.
fields
(Optional, string | string[]): Fields to return in the hits
layer. Supports wildcards (*
).
This parameter does not support fields with array values. Fields with array
values may return inconsistent results.
grid_agg
(Optional, Enum("geotile" | "geohex")): Aggregation used to create a grid for the field
.
grid_precision
(Optional, number): Additional zoom levels available through the aggs layer. For example, if <zoom> is 7
and grid_precision is 8, you can zoom in up to level 15. Accepts 0-8. If 0, results
don’t include the aggs layer.
grid_type
(Optional, Enum("grid" | "point" | "centroid")): Determines the geometry type for features in the aggs layer. In the aggs layer,
each feature represents a geotile_grid cell. If grid each feature is a Polygon
of the cells bounding box. If point each feature is a Point that is the centroid
of the cell.
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): Query DSL used to filter documents for the search.
runtime_mappings
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>): Defines one or more runtime fields in the search request. These fields take
precedence over mapped fields with the same name.
size
(Optional, number): Maximum number of features to return in the hits layer. Accepts 0-10000.
If 0, results don’t include the hits layer.
sort
(Optional, string | { _score, _doc, _geo_distance, _script } | string | { _score, _doc, _geo_distance, _script }[]): Sorts features in the hits layer. By default, the API calculates a bounding
box for each feature. It sorts features based on this box’s diagonal length,
from longest to shortest.
track_total_hits
(Optional, boolean | number): Number of hits matching the query to count accurately. If true
, the exact number
of hits is returned at the cost of some performance. If false
, the response does
not include the total number of hits matching the query.
* *with_labels
(Optional, boolean): If true
, the hits and aggs layers will contain additional point features representing
suggested label positions for the original features.
search_shardsedit
Returns information about the indices and shards that a search request would be executed against.
client.searchShards({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names to search; use_all
or empty string to perform the operation on all indices -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false) -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
routing
(Optional, string): Specific routing value
-
search_templateedit
Allows to use the Mustache language to pre-render a search definition.
client.searchTemplate({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): List of data streams, indices, and aliases to search. Supports wildcards (*). -
explain
(Optional, boolean) -
id
(Optional, string): ID of the search template to use. If no source is specified, this parameter is required. -
params
(Optional, Record<string, User-defined value>) -
profile
(Optional, boolean) -
source
(Optional, string): An inline search template. Supports the same parameters as the search API’s request body. Also supports Mustache variables. If no id is specified, this parameter is required. -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
ccs_minimize_roundtrips
(Optional, boolean): Indicates whether network round-trips should be minimized as part of cross-cluster search requests execution -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_throttled
(Optional, boolean): Whether specified concrete, expanded or aliased indices should be ignored when throttled -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
routing
(Optional, string): Custom value used to route operations to a specific shard. -
scroll
(Optional, string | -1 | 0): Specifies how long a consistent view of the index should be maintained for scrolled search. -
search_type
(Optional, Enum("query_then_fetch" | "dfs_query_then_fetch")): The type of the search operation. -
rest_total_hits_as_int
(Optional, boolean): If true, hits.total are rendered as an integer in the response. -
typed_keys
(Optional, boolean): Specify whether aggregation and suggester names should be prefixed by their respective types in the response
-
terms_enumedit
The terms enum API can be used to discover terms in the index that begin with the provided string. It is designed for low-latency look-ups used in auto-complete scenarios.
client.termsEnum({ index, field })
Argumentsedit
-
Request (object):
-
index
(string): List of data streams, indices, and index aliases to search. Wildcard (*) expressions are supported. -
field
(string): The string to match at the start of indexed terms. If not provided, all terms in the field are considered. -
size
(Optional, number): How many matching terms to return. -
timeout
(Optional, string | -1 | 0): The maximum length of time to spend collecting results. Defaults to "1s" (one second). If the timeout is exceeded the complete flag set to false in the response and the results may be partial or empty. -
case_insensitive
(Optional, boolean): When true the provided search string is matched against index terms without case sensitivity. -
index_filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): Allows to filter an index shard if the provided query rewrites to match_none. -
string
(Optional, string): The string after which terms in the index should be returned. Allows for a form of pagination if the last result from one request is passed as the search_after parameter for a subsequent request. -
search_after
(Optional, string)
-
termvectorsedit
Returns information and statistics about terms in the fields of a particular document.
client.termvectors({ index })
Argumentsedit
-
Request (object):
-
index
(string): The index in which the document resides. -
id
(Optional, string): The id of the document, when not specified a doc param should be supplied. -
doc
(Optional, object): A document. -
filter
(Optional, { max_doc_freq, max_num_terms, max_term_freq, max_word_length, min_doc_freq, min_term_freq, min_word_length }) -
per_field_analyzer
(Optional, Record<string, string>) -
fields
(Optional, string | string[]): A list of fields to return. -
field_statistics
(Optional, boolean): Specifies if document count, sum of document frequencies and sum of total term frequencies should be returned. -
offsets
(Optional, boolean): Specifies if term offsets should be returned. -
payloads
(Optional, boolean): Specifies if term payloads should be returned. -
positions
(Optional, boolean): Specifies if term positions should be returned. -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random). -
realtime
(Optional, boolean): Specifies if request is real-time as opposed to near-real-time (default: true). -
routing
(Optional, string): Specific routing value. -
term_statistics
(Optional, boolean): Specifies if total term frequency and document frequency should be returned. -
version
(Optional, number): Explicit version number for concurrency control -
version_type
(Optional, Enum("internal" | "external" | "external_gte" | "force")): Specific version type
-
updateedit
Updates a document with a script or partial document.
client.update({ id, index })
Argumentsedit
-
Request (object):
-
id
(string): Document ID -
index
(string): The name of the index -
detect_noop
(Optional, boolean): Set to false to disable setting result in the response to noop if no change to the document occurred. -
doc
(Optional, object): A partial update to an existing document. -
doc_as_upsert
(Optional, boolean): Set to true to use the contents of doc as the value of upsert -
script
(Optional, { lang, options, source } | { id }): Script to execute to update the document. -
scripted_upsert
(Optional, boolean): Set to true to execute the script whether or not the document exists. -
_source
(Optional, boolean | { excludes, includes }): Set to false to disable source retrieval. You can also specify a comma-separated list of the fields you want to retrieve. -
upsert
(Optional, object): If the document does not already exist, the contents of upsert are inserted as a new document. If the document exists, the script is executed. -
if_primary_term
(Optional, number): Only perform the operation if the document has this primary term. -
if_seq_no
(Optional, number): Only perform the operation if the document has this sequence number. -
lang
(Optional, string): The script language. -
refresh
(Optional, Enum(true | false | "wait_for")): If true, Elasticsearch refreshes the affected shards to make this operation visible to search, if wait_for then wait for a refresh to make this operation visible to search, if false do nothing with refreshes. -
require_alias
(Optional, boolean): If true, the destination must be an index alias. -
retry_on_conflict
(Optional, number): Specify how many times should the operation be retried when a conflict occurs. -
routing
(Optional, string): Custom value used to route operations to a specific shard. -
timeout
(Optional, string | -1 | 0): Period to wait for dynamic mapping updates and active shards. This guarantees Elasticsearch waits for at least the timeout before failing. The actual wait time could be longer, particularly when multiple waits occur. -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): The number of shard copies that must be active before proceeding with the operations. Set to all or any positive integer up to the total number of shards in the index (number_of_replicas+1). Defaults to 1 meaning the primary shard. -
_source_excludes
(Optional, string | string[]): Specify the source fields you want to exclude. -
_source_includes
(Optional, string | string[]): Specify the source fields you want to retrieve.
-
update_by_queryedit
Performs an update on every document in the index without changing the source, for example to pick up a mapping change.
client.updateByQuery({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names to search; use_all
or empty string to perform the operation on all indices -
max_docs
(Optional, number) -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
script
(Optional, { lang, options, source } | { id }) -
slice
(Optional, { field, id, max }) -
conflicts
(Optional, Enum("abort" | "proceed")) -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
analyzer
(Optional, string): The analyzer to use for the query string -
analyze_wildcard
(Optional, boolean): Specify whether wildcard and prefix queries should be analyzed (default: false) -
default_operator
(Optional, Enum("and" | "or")): The default operator for query string query (AND or OR) -
df
(Optional, string): The field to use as default where no field prefix is given in the query string -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
from
(Optional, number): Starting offset (default: 0) -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
lenient
(Optional, boolean): Specify whether format-based query failures (such as providing text to a numeric field) should be ignored -
pipeline
(Optional, string): Ingest pipeline to set on index requests made by this action. (default: none) -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
refresh
(Optional, boolean): Should the affected indexes be refreshed? -
request_cache
(Optional, boolean): Specify if request cache should be used for this request or not, defaults to index level setting -
requests_per_second
(Optional, float): The throttle to set on this request in sub-requests per second. -1 means no throttle. -
routing
(Optional, string): A list of specific routing values -
scroll
(Optional, string | -1 | 0): Specify how long a consistent view of the index should be maintained for scrolled search -
scroll_size
(Optional, number): Size on the scroll request powering the update by query -
search_timeout
(Optional, string | -1 | 0): Explicit timeout for each search request. Defaults to no timeout. -
search_type
(Optional, Enum("query_then_fetch" | "dfs_query_then_fetch")): Search operation type -
slices
(Optional, number | Enum("auto")): The number of slices this task should be divided into. Defaults to 1, meaning the task isn’t sliced into subtasks. Can be set toauto
. -
sort
(Optional, string[]): A list of <field>:<direction> pairs -
stats
(Optional, string[]): Specific tag of the request for logging and statistical purposes -
terminate_after
(Optional, number): The maximum number of documents to collect for each shard, upon reaching which the query execution will terminate early. -
timeout
(Optional, string | -1 | 0): Time each individual bulk request should wait for shards that are unavailable. -
version
(Optional, boolean): Specify whether to return document version as part of a hit -
version_type
(Optional, boolean): Should the document increment the version number (internal) on hit or not (reindex) -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of shard copies that must be active before proceeding with the update by query operation. Defaults to 1, meaning the primary shard only. Set toall
for all shard copies, otherwise set to any non-negative value less than or equal to the total number of copies for the shard (number of replicas + 1) -
wait_for_completion
(Optional, boolean): Should the request should block until the update by query operation is complete.
-
update_by_query_rethrottleedit
Changes the number of requests per second for a particular Update By Query operation.
client.updateByQueryRethrottle({ task_id })
Argumentsedit
-
Request (object):
-
task_id
(string): The task id to rethrottle -
requests_per_second
(Optional, float): The throttle to set on this request in floating sub-requests per second. -1 means set no throttle.
-
async_searchedit
deleteedit
Deletes an async search by ID. If the search is still running, the search request will be cancelled. Otherwise, the saved search results are deleted.
client.asyncSearch.delete({ id })
Argumentsedit
-
Request (object):
-
id
(string): A unique identifier for the async search.
-
getedit
Retrieves the results of a previously submitted async search request given its ID.
client.asyncSearch.get({ id })
Argumentsedit
-
Request (object):
-
id
(string): A unique identifier for the async search. -
keep_alive
(Optional, string | -1 | 0): Specifies how long the async search should be available in the cluster. When not specified, thekeep_alive
set with the corresponding submit async request will be used. Otherwise, it is possible to override the value and extend the validity of the request. When this period expires, the search, if still running, is cancelled. If the search is completed, its saved results are deleted. -
typed_keys
(Optional, boolean): Specify whether aggregation and suggester names should be prefixed by their respective types in the response -
wait_for_completion_timeout
(Optional, string | -1 | 0): Specifies to wait for the search to be completed up until the provided timeout. Final results will be returned if available before the timeout expires, otherwise the currently available results will be returned once the timeout expires. By default no timeout is set meaning that the currently available results will be returned without any additional wait.
-
statusedit
Retrieves the status of a previously submitted async search request given its ID.
client.asyncSearch.status({ id })
Argumentsedit
-
Request (object):
-
id
(string): A unique identifier for the async search.
-
submitedit
Executes a search request asynchronously.
client.asyncSearch.submit({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names to search; use_all
or empty string to perform the operation on all indices -
aggregations
(Optional, Record<string, { aggregations, meta, adjacency_matrix, auto_date_histogram, avg, avg_bucket, boxplot, bucket_script, bucket_selector, bucket_sort, bucket_count_ks_test, bucket_correlation, cardinality, categorize_text, children, composite, cumulative_cardinality, cumulative_sum, date_histogram, date_range, derivative, diversified_sampler, extended_stats, extended_stats_bucket, frequent_item_sets, filter, filters, geo_bounds, geo_centroid, geo_distance, geohash_grid, geo_line, geotile_grid, geohex_grid, global, histogram, ip_range, ip_prefix, inference, line, matrix_stats, max, max_bucket, median_absolute_deviation, min, min_bucket, missing, moving_avg, moving_percentiles, moving_fn, multi_terms, nested, normalize, parent, percentile_ranks, percentiles, percentiles_bucket, range, rare_terms, rate, reverse_nested, sampler, scripted_metric, serial_diff, significant_terms, significant_text, stats, stats_bucket, string_stats, sum, sum_bucket, terms, top_hits, t_test, top_metrics, value_count, weighted_avg, variable_width_histogram }>) -
collapse
(Optional, { field, inner_hits, max_concurrent_group_searches, collapse }) -
explain
(Optional, boolean): If true, returns detailed information about score computation as part of a hit. -
ext
(Optional, Record<string, User-defined value>): Configuration of search extensions defined by Elasticsearch plugins. -
from
(Optional, number): Starting document offset. By default, you cannot page through more than 10,000 hits using the from and size parameters. To page through more hits, use the search_after parameter. -
highlight
(Optional, { encoder, fields }) -
track_total_hits
(Optional, boolean | number): Number of hits matching the query to count accurately. If true, the exact number of hits is returned at the cost of some performance. If false, the response does not include the total number of hits matching the query. Defaults to 10,000 hits. -
indices_boost
(Optional, Record<string, number>[]): Boosts the _score of documents from specified indices. -
docvalue_fields
(Optional, { field, format, include_unmapped }[]): Array of wildcard (*) patterns. The request returns doc values for field names matching these patterns in the hits.fields property of the response. -
knn
(Optional, { field, query_vector, query_vector_builder, k, num_candidates, boost, filter, similarity } | { field, query_vector, query_vector_builder, k, num_candidates, boost, filter, similarity }[]): Defines the approximate kNN search to run. -
min_score
(Optional, number): Minimum _score for matching documents. Documents with a lower _score are not included in the search results. -
post_filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
profile
(Optional, boolean) -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): Defines the search definition using the Query DSL. -
rescore
(Optional, { query, window_size } | { query, window_size }[]) -
script_fields
(Optional, Record<string, { script, ignore_failure }>): Retrieve a script evaluation (based on different fields) for each hit. -
search_after
(Optional, number | number | string | boolean | null | User-defined value[]) -
size
(Optional, number): The number of hits to return. By default, you cannot page through more than 10,000 hits using the from and size parameters. To page through more hits, use the search_after parameter. -
slice
(Optional, { field, id, max }) -
sort
(Optional, string | { _score, _doc, _geo_distance, _script } | string | { _score, _doc, _geo_distance, _script }[]) -
_source
(Optional, boolean | { excludes, includes }): Indicates which source fields are returned for matching documents. These fields are returned in the hits._source property of the search response. -
fields
(Optional, { field, format, include_unmapped }[]): Array of wildcard (*) patterns. The request returns values for field names matching these patterns in the hits.fields property of the response. -
suggest
(Optional, { text }) -
terminate_after
(Optional, number): Maximum number of documents to collect for each shard. If a query reaches this limit, Elasticsearch terminates the query early. Elasticsearch collects documents before sorting. Defaults to 0, which does not terminate query execution early. -
timeout
(Optional, string): Specifies the period of time to wait for a response from each shard. If no response is received before the timeout expires, the request fails and returns an error. Defaults to no timeout. -
track_scores
(Optional, boolean): If true, calculate and return document scores, even if the scores are not used for sorting. -
version
(Optional, boolean): If true, returns document version as part of a hit. -
seq_no_primary_term
(Optional, boolean): If true, returns sequence number and primary term of the last modification of each hit. See Optimistic concurrency control. -
stored_fields
(Optional, string | string[]): List of stored fields to return as part of a hit. If no fields are specified, no stored fields are included in the response. If this field is specified, the _source parameter defaults to false. You can pass _source: true to return both source fields and stored fields in the search response. -
pit
(Optional, { id, keep_alive }): Limits the search to a point in time (PIT). If you provide a PIT, you cannot specify an <index> in the request path. -
runtime_mappings
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>): Defines one or more runtime fields in the search request. These fields take precedence over mapped fields with the same name. -
stats
(Optional, string[]): Stats groups to associate with the search. Each group maintains a statistics aggregation for its associated searches. You can retrieve these stats using the indices stats API. -
wait_for_completion_timeout
(Optional, string | -1 | 0): Blocks and waits until the search is completed up to a certain timeout. When the async search completes within the timeout, the response won’t include the ID as the results are not stored in the cluster. -
keep_on_completion
(Optional, boolean): Iftrue
, results are stored for later retrieval when the search completes within thewait_for_completion_timeout
. -
keep_alive
(Optional, string | -1 | 0): Specifies how long the async search needs to be available. Ongoing async searches and any saved search results are deleted after this period. -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
allow_partial_search_results
(Optional, boolean): Indicate if an error should be returned if there is a partial search failure or timeout -
analyzer
(Optional, string): The analyzer to use for the query string -
analyze_wildcard
(Optional, boolean): Specify whether wildcard and prefix queries should be analyzed (default: false) -
batched_reduce_size
(Optional, number): Affects how often partial results become available, which happens whenever shard results are reduced. A partial reduction is performed every time the coordinating node has received a certain number of new shard responses (5 by default). -
ccs_minimize_roundtrips
(Optional, boolean): The default value is the only supported value. -
default_operator
(Optional, Enum("and" | "or")): The default operator for query string query (AND or OR) -
df
(Optional, string): The field to use as default where no field prefix is given in the query string -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_throttled
(Optional, boolean): Whether specified concrete, expanded or aliased indices should be ignored when throttled -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
lenient
(Optional, boolean): Specify whether format-based query failures (such as providing text to a numeric field) should be ignored -
max_concurrent_shard_requests
(Optional, number): The number of concurrent shard requests per node this search executes concurrently. This value should be used to limit the impact of the search on the cluster in order to limit the number of concurrent shard requests -
min_compatible_shard_node
(Optional, string) -
preference
(Optional, string): Specify the node or shard the operation should be performed on (default: random) -
pre_filter_shard_size
(Optional, number): The default value cannot be changed, which enforces the execution of a pre-filter roundtrip to retrieve statistics from each shard so that the ones that surely don’t hold any document matching the query get skipped. -
request_cache
(Optional, boolean): Specify if request cache should be used for this request or not, defaults to true -
routing
(Optional, string): A list of specific routing values -
scroll
(Optional, string | -1 | 0) -
search_type
(Optional, Enum("query_then_fetch" | "dfs_query_then_fetch")): Search operation type -
suggest_field
(Optional, string): Specifies which field to use for suggestions. -
suggest_mode
(Optional, Enum("missing" | "popular" | "always")): Specify suggest mode -
suggest_size
(Optional, number): How many suggestions to return in response -
suggest_text
(Optional, string): The source text for which the suggestions should be returned. -
typed_keys
(Optional, boolean): Specify whether aggregation and suggester names should be prefixed by their respective types in the response -
rest_total_hits_as_int
(Optional, boolean) -
_source_excludes
(Optional, string | string[]): A list of fields to exclude from the returned _source field -
_source_includes
(Optional, string | string[]): A list of fields to extract and return from the _source field -
q
(Optional, string): Query in the Lucene query string syntax
-
catedit
aliasesedit
Shows information about currently configured aliases to indices including filter and routing infos.
client.cat.aliases({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string | string[]): A list of aliases to retrieve. Supports wildcards (*
). To retrieve all aliases, omit this parameter or use*
or_all
. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both.
-
allocationedit
Provides a snapshot of how many shards are allocated to each data node and how much disk space they are using.
client.cat.allocation({ ... })
Argumentsedit
-
Request (object):
-
node_id
(Optional, string | string[]): List of node identifiers or names used to limit the returned information. -
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit used to display byte values.
-
component_templatesedit
Returns information about existing component_templates templates.
client.cat.componentTemplates({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string): The name of the component template. Accepts wildcard expressions. If omitted, all component templates are returned.
-
countedit
Provides quick access to the document count of the entire cluster, or individual indices.
client.cat.count({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): List of data streams, indices, and aliases used to limit the request. Supports wildcards (*
). To target all data streams and indices, omit this parameter or use*
or_all
.
-
fielddataedit
Shows how much heap memory is currently being used by fielddata on every data node in the cluster.
client.cat.fielddata({ ... })
Argumentsedit
-
Request (object):
-
fields
(Optional, string | string[]): List of fields used to limit returned information. To retrieve all fields, omit this parameter. -
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit used to display byte values.
-
healthedit
Returns a concise representation of the cluster health.
client.cat.health({ ... })
Argumentsedit
-
Request (object):
-
time
(Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d")): The unit used to display time values. -
ts
(Optional, boolean): If true, returnsHH:MM:SS
and Unix epoch timestamps.
-
helpedit
Returns help for the Cat APIs.
client.cat.help()
indicesedit
Returns information about indices: number of primaries and replicas, document counts, disk size, …
client.cat.indices({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): List of data streams, indices, and aliases used to limit the request. Supports wildcards (*
). To target all data streams and indices, omit this parameter or use*
or_all
. -
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit used to display byte values. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): The type of index that wildcard patterns can match. -
health
(Optional, Enum("green" | "yellow" | "red")): The health status used to limit returned indices. By default, the response includes indices of any health status. -
include_unloaded_segments
(Optional, boolean): If true, the response includes information from segments that are not loaded into memory. -
pri
(Optional, boolean): If true, the response only includes information from primary shards. -
time
(Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d")): The unit used to display time values.
-
masteredit
Returns information about the master node.
client.cat.master()
ml_data_frame_analyticsedit
Gets configuration and usage information about data frame analytics jobs.
client.cat.mlDataFrameAnalytics({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): The ID of the data frame analytics to fetch -
allow_no_match
(Optional, boolean): Whether to ignore if a wildcard expression matches no configs. (This includes_all
string or when no configs have been specified) -
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit in which to display byte values -
h
(Optional, Enum("assignment_explanation" | "create_time" | "description" | "dest_index" | "failure_reason" | "id" | "model_memory_limit" | "node.address" | "node.ephemeral_id" | "node.id" | "node.name" | "progress" | "source_index" | "state" | "type" | "version") | Enum("assignment_explanation" | "create_time" | "description" | "dest_index" | "failure_reason" | "id" | "model_memory_limit" | "node.address" | "node.ephemeral_id" | "node.id" | "node.name" | "progress" | "source_index" | "state" | "type" | "version")[]): List of column names to display. -
s
(Optional, Enum("assignment_explanation" | "create_time" | "description" | "dest_index" | "failure_reason" | "id" | "model_memory_limit" | "node.address" | "node.ephemeral_id" | "node.id" | "node.name" | "progress" | "source_index" | "state" | "type" | "version") | Enum("assignment_explanation" | "create_time" | "description" | "dest_index" | "failure_reason" | "id" | "model_memory_limit" | "node.address" | "node.ephemeral_id" | "node.id" | "node.name" | "progress" | "source_index" | "state" | "type" | "version")[]): List of column names or column aliases used to sort the response. -
time
(Optional, string | -1 | 0): Unit used to display time values.
-
ml_datafeedsedit
Gets configuration and usage information about datafeeds.
client.cat.mlDatafeeds({ ... })
Argumentsedit
-
Request (object):
-
datafeed_id
(Optional, string): A numerical character string that uniquely identifies the datafeed. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:
-
- Contains wildcard expressions and there are no datafeeds that match.
-
Contains the
_all
string or no identifiers and there are no matches. - Contains wildcard expressions and there are only partial matches.
If true
, the API returns an empty datafeeds array when there are no matches and the subset of results when
there are partial matches. If false
, the API returns a 404 status code when there are no matches or only
partial matches.
h
(Optional, Enum("ae" | "bc" | "id" | "na" | "ne" | "ni" | "nn" | "sba" | "sc" | "seah" | "st" | "s") | Enum("ae" | "bc" | "id" | "na" | "ne" | "ni" | "nn" | "sba" | "sc" | "seah" | "st" | "s")[]): List of column names to display.
s
(Optional, Enum("ae" | "bc" | "id" | "na" | "ne" | "ni" | "nn" | "sba" | "sc" | "seah" | "st" | "s") | Enum("ae" | "bc" | "id" | "na" | "ne" | "ni" | "nn" | "sba" | "sc" | "seah" | "st" | "s")[]): List of column names or column aliases used to sort the response.
* *time
(Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d")): The unit used to display time values.
ml_jobsedit
Gets configuration and usage information about anomaly detection jobs.
client.cat.mlJobs({ ... })
Argumentsedit
-
Request (object):
-
job_id
(Optional, string): Identifier for the anomaly detection job. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:
-
- Contains wildcard expressions and there are no jobs that match.
-
Contains the
_all
string or no identifiers and there are no matches. - Contains wildcard expressions and there are only partial matches.
If true
, the API returns an empty jobs array when there are no matches and the subset of results when there
are partial matches. If false
, the API returns a 404 status code when there are no matches or only partial
matches.
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit used to display byte values.
h
(Optional, Enum("assignment_explanation" | "buckets.count" | "buckets.time.exp_avg" | "buckets.time.exp_avg_hour" | "buckets.time.max" | "buckets.time.min" | "buckets.time.total" | "data.buckets" | "data.earliest_record" | "data.empty_buckets" | "data.input_bytes" | "data.input_fields" | "data.input_records" | "data.invalid_dates" | "data.last" | "data.last_empty_bucket" | "data.last_sparse_bucket" | "data.latest_record" | "data.missing_fields" | "data.out_of_order_timestamps" | "data.processed_fields" | "data.processed_records" | "data.sparse_buckets" | "forecasts.memory.avg" | "forecasts.memory.max" | "forecasts.memory.min" | "forecasts.memory.total" | "forecasts.records.avg" | "forecasts.records.max" | "forecasts.records.min" | "forecasts.records.total" | "forecasts.time.avg" | "forecasts.time.max" | "forecasts.time.min" | "forecasts.time.total" | "forecasts.total" | "id" | "model.bucket_allocation_failures" | "model.by_fields" | "model.bytes" | "model.bytes_exceeded" | "model.categorization_status" | "model.categorized_doc_count" | "model.dead_category_count" | "model.failed_category_count" | "model.frequent_category_count" | "model.log_time" | "model.memory_limit" | "model.memory_status" | "model.over_fields" | "model.partition_fields" | "model.rare_category_count" | "model.timestamp" | "model.total_category_count" | "node.address" | "node.ephemeral_id" | "node.id" | "node.name" | "opened_time" | "state") | Enum("assignment_explanation" | "buckets.count" | "buckets.time.exp_avg" | "buckets.time.exp_avg_hour" | "buckets.time.max" | "buckets.time.min" | "buckets.time.total" | "data.buckets" | "data.earliest_record" | "data.empty_buckets" | "data.input_bytes" | "data.input_fields" | "data.input_records" | "data.invalid_dates" | "data.last" | "data.last_empty_bucket" | "data.last_sparse_bucket" | "data.latest_record" | "data.missing_fields" | "data.out_of_order_timestamps" | "data.processed_fields" | "data.processed_records" | "data.sparse_buckets" | "forecasts.memory.avg" | "forecasts.memory.max" | "forecasts.memory.min" | "forecasts.memory.total" | "forecasts.records.avg" | "forecasts.records.max" | "forecasts.records.min" | "forecasts.records.total" | "forecasts.time.avg" | "forecasts.time.max" | "forecasts.time.min" | "forecasts.time.total" | "forecasts.total" | "id" | "model.bucket_allocation_failures" | "model.by_fields" | "model.bytes" | "model.bytes_exceeded" | "model.categorization_status" | "model.categorized_doc_count" | "model.dead_category_count" | "model.failed_category_count" | "model.frequent_category_count" | "model.log_time" | "model.memory_limit" | "model.memory_status" | "model.over_fields" | "model.partition_fields" | "model.rare_category_count" | "model.timestamp" | "model.total_category_count" | "node.address" | "node.ephemeral_id" | "node.id" | "node.name" | "opened_time" | "state")[]): List of column names to display.
s
(Optional, Enum("assignment_explanation" | "buckets.count" | "buckets.time.exp_avg" | "buckets.time.exp_avg_hour" | "buckets.time.max" | "buckets.time.min" | "buckets.time.total" | "data.buckets" | "data.earliest_record" | "data.empty_buckets" | "data.input_bytes" | "data.input_fields" | "data.input_records" | "data.invalid_dates" | "data.last" | "data.last_empty_bucket" | "data.last_sparse_bucket" | "data.latest_record" | "data.missing_fields" | "data.out_of_order_timestamps" | "data.processed_fields" | "data.processed_records" | "data.sparse_buckets" | "forecasts.memory.avg" | "forecasts.memory.max" | "forecasts.memory.min" | "forecasts.memory.total" | "forecasts.records.avg" | "forecasts.records.max" | "forecasts.records.min" | "forecasts.records.total" | "forecasts.time.avg" | "forecasts.time.max" | "forecasts.time.min" | "forecasts.time.total" | "forecasts.total" | "id" | "model.bucket_allocation_failures" | "model.by_fields" | "model.bytes" | "model.bytes_exceeded" | "model.categorization_status" | "model.categorized_doc_count" | "model.dead_category_count" | "model.failed_category_count" | "model.frequent_category_count" | "model.log_time" | "model.memory_limit" | "model.memory_status" | "model.over_fields" | "model.partition_fields" | "model.rare_category_count" | "model.timestamp" | "model.total_category_count" | "node.address" | "node.ephemeral_id" | "node.id" | "node.name" | "opened_time" | "state") | Enum("assignment_explanation" | "buckets.count" | "buckets.time.exp_avg" | "buckets.time.exp_avg_hour" | "buckets.time.max" | "buckets.time.min" | "buckets.time.total" | "data.buckets" | "data.earliest_record" | "data.empty_buckets" | "data.input_bytes" | "data.input_fields" | "data.input_records" | "data.invalid_dates" | "data.last" | "data.last_empty_bucket" | "data.last_sparse_bucket" | "data.latest_record" | "data.missing_fields" | "data.out_of_order_timestamps" | "data.processed_fields" | "data.processed_records" | "data.sparse_buckets" | "forecasts.memory.avg" | "forecasts.memory.max" | "forecasts.memory.min" | "forecasts.memory.total" | "forecasts.records.avg" | "forecasts.records.max" | "forecasts.records.min" | "forecasts.records.total" | "forecasts.time.avg" | "forecasts.time.max" | "forecasts.time.min" | "forecasts.time.total" | "forecasts.total" | "id" | "model.bucket_allocation_failures" | "model.by_fields" | "model.bytes" | "model.bytes_exceeded" | "model.categorization_status" | "model.categorized_doc_count" | "model.dead_category_count" | "model.failed_category_count" | "model.frequent_category_count" | "model.log_time" | "model.memory_limit" | "model.memory_status" | "model.over_fields" | "model.partition_fields" | "model.rare_category_count" | "model.timestamp" | "model.total_category_count" | "node.address" | "node.ephemeral_id" | "node.id" | "node.name" | "opened_time" | "state")[]): List of column names or column aliases used to sort the response.
time
(Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d")): The unit used to display time values.
ml_trained_modelsedit
Gets configuration and usage information about inference trained models.
client.cat.mlTrainedModels({ ... })
Argumentsedit
-
Request (object):
-
model_id
(Optional, string): The ID of the trained models stats to fetch -
allow_no_match
(Optional, boolean): Whether to ignore if a wildcard expression matches no trained models. (This includes_all
string or when no trained models have been specified) -
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit in which to display byte values -
h
(Optional, Enum("create_time" | "created_by" | "data_frame_analytics_id" | "description" | "heap_size" | "id" | "ingest.count" | "ingest.current" | "ingest.failed" | "ingest.pipelines" | "ingest.time" | "license" | "operations" | "version") | Enum("create_time" | "created_by" | "data_frame_analytics_id" | "description" | "heap_size" | "id" | "ingest.count" | "ingest.current" | "ingest.failed" | "ingest.pipelines" | "ingest.time" | "license" | "operations" | "version")[]): List of column names to display -
s
(Optional, Enum("create_time" | "created_by" | "data_frame_analytics_id" | "description" | "heap_size" | "id" | "ingest.count" | "ingest.current" | "ingest.failed" | "ingest.pipelines" | "ingest.time" | "license" | "operations" | "version") | Enum("create_time" | "created_by" | "data_frame_analytics_id" | "description" | "heap_size" | "id" | "ingest.count" | "ingest.current" | "ingest.failed" | "ingest.pipelines" | "ingest.time" | "license" | "operations" | "version")[]): List of column names or column aliases to sort by -
from
(Optional, number): skips a number of trained models -
size
(Optional, number): specifies a max number of trained models to get
-
nodeattrsedit
Returns information about custom node attributes.
client.cat.nodeattrs()
nodesedit
Returns basic statistics about performance of cluster nodes.
client.cat.nodes({ ... })
Argumentsedit
-
Request (object):
-
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit used to display byte values. -
full_id
(Optional, boolean | string): Iftrue
, return the full node ID. Iffalse
, return the shortened node ID. -
include_unloaded_segments
(Optional, boolean): If true, the response includes information from segments that are not loaded into memory.
-
pending_tasksedit
Returns a concise representation of the cluster pending tasks.
client.cat.pendingTasks()
pluginsedit
Returns information about installed plugins across nodes node.
client.cat.plugins()
recoveryedit
Returns information about index shard recoveries, both on-going completed.
client.cat.recovery({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of data streams, indices, and aliases used to limit the request. Supports wildcards (*
). To target all data streams and indices, omit this parameter or use*
or_all
. -
active_only
(Optional, boolean): Iftrue
, the response only includes ongoing shard recoveries. -
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit used to display byte values. -
detailed
(Optional, boolean): Iftrue
, the response includes detailed information about shard recoveries.
-
repositoriesedit
Returns information about snapshot repositories registered in the cluster.
client.cat.repositories()
segmentsedit
Provides low-level information about the segments in the shards of an index.
client.cat.segments({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names to limit the returned information -
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit in which to display byte values
-
shardsedit
Provides a detailed view of shard allocation on nodes.
client.cat.shards({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names to limit the returned information -
bytes
(Optional, Enum("b" | "kb" | "mb" | "gb" | "tb" | "pb")): The unit in which to display byte values
-
snapshotsedit
Returns all snapshots in a specific repository.
client.cat.snapshots({ ... })
Argumentsedit
-
Request (object):
-
repository
(Optional, string | string[]): Name of repository from which to fetch the snapshot information -
ignore_unavailable
(Optional, boolean): Set to true to ignore unavailable snapshots
-
tasksedit
Returns information about the tasks currently executing on one or more nodes in the cluster.
client.cat.tasks({ ... })
Argumentsedit
-
Request (object):
-
actions
(Optional, string[]): A list of actions that should be returned. Leave empty to return all. -
detailed
(Optional, boolean): Return detailed task information (default: false) -
node_id
(Optional, string[]) -
parent_task
(Optional, number)
-
templatesedit
Returns information about existing templates.
client.cat.templates({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string): A pattern that returned template names must match
-
thread_pooledit
Returns cluster-wide thread pool statistics per node. By default the active, queue and rejected statistics are returned for all thread pools.
client.cat.threadPool({ ... })
Argumentsedit
-
Request (object):
-
thread_pool_patterns
(Optional, string | string[]): List of thread pool names used to limit the request. Accepts wildcard expressions. -
time
(Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d")): Unit used to display time values.
-
transformsedit
Gets configuration and usage information about transforms.
client.cat.transforms({ ... })
Argumentsedit
-
Request (object):
-
transform_id
(Optional, string): The id of the transform for which to get stats. _all or * implies all transforms -
allow_no_match
(Optional, boolean): Whether to ignore if a wildcard expression matches no transforms. (This includes_all
string or when no transforms have been specified) -
from
(Optional, number): skips a number of transform configs, defaults to 0 -
h
(Optional, Enum("changes_last_detection_time" | "checkpoint" | "checkpoint_duration_time_exp_avg" | "checkpoint_progress" | "create_time" | "delete_time" | "description" | "dest_index" | "documents_deleted" | "documents_indexed" | "docs_per_second" | "documents_processed" | "frequency" | "id" | "index_failure" | "index_time" | "index_total" | "indexed_documents_exp_avg" | "last_search_time" | "max_page_search_size" | "pages_processed" | "pipeline" | "processed_documents_exp_avg" | "processing_time" | "reason" | "search_failure" | "search_time" | "search_total" | "source_index" | "state" | "transform_type" | "trigger_count" | "version") | Enum("changes_last_detection_time" | "checkpoint" | "checkpoint_duration_time_exp_avg" | "checkpoint_progress" | "create_time" | "delete_time" | "description" | "dest_index" | "documents_deleted" | "documents_indexed" | "docs_per_second" | "documents_processed" | "frequency" | "id" | "index_failure" | "index_time" | "index_total" | "indexed_documents_exp_avg" | "last_search_time" | "max_page_search_size" | "pages_processed" | "pipeline" | "processed_documents_exp_avg" | "processing_time" | "reason" | "search_failure" | "search_time" | "search_total" | "source_index" | "state" | "transform_type" | "trigger_count" | "version")[]): List of column names to display. -
s
(Optional, Enum("changes_last_detection_time" | "checkpoint" | "checkpoint_duration_time_exp_avg" | "checkpoint_progress" | "create_time" | "delete_time" | "description" | "dest_index" | "documents_deleted" | "documents_indexed" | "docs_per_second" | "documents_processed" | "frequency" | "id" | "index_failure" | "index_time" | "index_total" | "indexed_documents_exp_avg" | "last_search_time" | "max_page_search_size" | "pages_processed" | "pipeline" | "processed_documents_exp_avg" | "processing_time" | "reason" | "search_failure" | "search_time" | "search_total" | "source_index" | "state" | "transform_type" | "trigger_count" | "version") | Enum("changes_last_detection_time" | "checkpoint" | "checkpoint_duration_time_exp_avg" | "checkpoint_progress" | "create_time" | "delete_time" | "description" | "dest_index" | "documents_deleted" | "documents_indexed" | "docs_per_second" | "documents_processed" | "frequency" | "id" | "index_failure" | "index_time" | "index_total" | "indexed_documents_exp_avg" | "last_search_time" | "max_page_search_size" | "pages_processed" | "pipeline" | "processed_documents_exp_avg" | "processing_time" | "reason" | "search_failure" | "search_time" | "search_total" | "source_index" | "state" | "transform_type" | "trigger_count" | "version")[]): List of column names or column aliases used to sort the response. -
time
(Optional, Enum("nanos" | "micros" | "ms" | "s" | "m" | "h" | "d")): Unit used to display time values. -
size
(Optional, number): specifies a max number of transforms to get, defaults to 100
-
ccredit
delete_auto_follow_patternedit
Deletes auto-follow patterns.
client.ccr.deleteAutoFollowPattern({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the auto follow pattern.
-
followedit
Creates a new follower index configured to follow the referenced leader index.
client.ccr.follow({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the follower index -
leader_index
(Optional, string) -
max_outstanding_read_requests
(Optional, number) -
max_outstanding_write_requests
(Optional, number) -
max_read_request_operation_count
(Optional, number) -
max_read_request_size
(Optional, string) -
max_retry_delay
(Optional, string | -1 | 0) -
max_write_buffer_count
(Optional, number) -
max_write_buffer_size
(Optional, string) -
max_write_request_operation_count
(Optional, number) -
max_write_request_size
(Optional, string) -
read_poll_timeout
(Optional, string | -1 | 0) -
remote_cluster
(Optional, string) -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of shard copies that must be active before returning. Defaults to 0. Set toall
for all shard copies, otherwise set to any non-negative value less than or equal to the total number of copies for the shard (number of replicas + 1)
-
follow_infoedit
Retrieves information about all follower indices, including parameters and status for each follower index
client.ccr.followInfo({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index patterns; use_all
to perform the operation on all indices
-
follow_statsedit
Retrieves follower stats. return shard-level stats about the following tasks associated with each shard for the specified indices.
client.ccr.followStats({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index patterns; use_all
to perform the operation on all indices
-
forget_followeredit
Removes the follower retention leases from the leader.
client.ccr.forgetFollower({ index })
Argumentsedit
-
Request (object):
-
index
(string): the name of the leader index for which specified follower retention leases should be removed -
follower_cluster
(Optional, string) -
follower_index
(Optional, string) -
follower_index_uuid
(Optional, string) -
leader_remote_cluster
(Optional, string)
-
get_auto_follow_patternedit
Gets configured auto-follow patterns. Returns the specified auto-follow pattern collection.
client.ccr.getAutoFollowPattern({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string): Specifies the auto-follow pattern collection that you want to retrieve. If you do not specify a name, the API returns information for all collections.
-
pause_auto_follow_patternedit
Pauses an auto-follow pattern
client.ccr.pauseAutoFollowPattern({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the auto follow pattern that should pause discovering new indices to follow.
-
pause_followedit
Pauses a follower index. The follower index will not fetch any additional operations from the leader index.
client.ccr.pauseFollow({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the follower index that should pause following its leader index.
-
put_auto_follow_patternedit
Creates a new named collection of auto-follow patterns against a specified remote cluster. Newly created indices on the remote cluster matching any of the specified patterns will be automatically configured as follower indices.
client.ccr.putAutoFollowPattern({ name, remote_cluster })
Argumentsedit
-
Request (object):
-
name
(string): The name of the collection of auto-follow patterns. -
remote_cluster
(string): The remote cluster containing the leader indices to match against. -
follow_index_pattern
(Optional, string): The name of follower index. The template {{leader_index}} can be used to derive the name of the follower index from the name of the leader index. When following a data stream, use {{leader_index}}; CCR does not support changes to the names of a follower data stream’s backing indices. -
leader_index_patterns
(Optional, string[]): An array of simple index patterns to match against indices in the remote cluster specified by the remote_cluster field. -
leader_index_exclusion_patterns
(Optional, string[]): An array of simple index patterns that can be used to exclude indices from being auto-followed. Indices in the remote cluster whose names are matching one or more leader_index_patterns and one or more leader_index_exclusion_patterns won’t be followed. -
max_outstanding_read_requests
(Optional, number): The maximum number of outstanding reads requests from the remote cluster. -
settings
(Optional, Record<string, User-defined value>): Settings to override from the leader index. Note that certain settings can not be overrode (e.g., index.number_of_shards). -
max_outstanding_write_requests
(Optional, number): The maximum number of outstanding reads requests from the remote cluster. -
read_poll_timeout
(Optional, string | -1 | 0): The maximum time to wait for new operations on the remote cluster when the follower index is synchronized with the leader index. When the timeout has elapsed, the poll for operations will return to the follower so that it can update some statistics. Then the follower will immediately attempt to read from the leader again. -
max_read_request_operation_count
(Optional, number): The maximum number of operations to pull per read from the remote cluster. -
max_read_request_size
(Optional, number | string): The maximum size in bytes of per read of a batch of operations pulled from the remote cluster. -
max_retry_delay
(Optional, string | -1 | 0): The maximum time to wait before retrying an operation that failed exceptionally. An exponential backoff strategy is employed when retrying. -
max_write_buffer_count
(Optional, number): The maximum number of operations that can be queued for writing. When this limit is reached, reads from the remote cluster will be deferred until the number of queued operations goes below the limit. -
max_write_buffer_size
(Optional, number | string): The maximum total bytes of operations that can be queued for writing. When this limit is reached, reads from the remote cluster will be deferred until the total bytes of queued operations goes below the limit. -
max_write_request_operation_count
(Optional, number): The maximum number of operations per bulk write request executed on the follower. -
max_write_request_size
(Optional, number | string): The maximum total bytes of operations per bulk write request executed on the follower.
-
resume_auto_follow_patternedit
Resumes an auto-follow pattern that has been paused
client.ccr.resumeAutoFollowPattern({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the auto follow pattern to resume discovering new indices to follow.
-
resume_followedit
Resumes a follower index that has been paused
client.ccr.resumeFollow({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the follow index to resume following. -
max_outstanding_read_requests
(Optional, number) -
max_outstanding_write_requests
(Optional, number) -
max_read_request_operation_count
(Optional, number) -
max_read_request_size
(Optional, string) -
max_retry_delay
(Optional, string | -1 | 0) -
max_write_buffer_count
(Optional, number) -
max_write_buffer_size
(Optional, string) -
max_write_request_operation_count
(Optional, number) -
max_write_request_size
(Optional, string) -
read_poll_timeout
(Optional, string | -1 | 0)
-
statsedit
Gets all stats related to cross-cluster replication.
client.ccr.stats()
unfollowedit
Stops the following task associated with a follower index and removes index metadata and settings associated with cross-cluster replication.
client.ccr.unfollow({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the follower index that should be turned into a regular index.
-
clusteredit
allocation_explainedit
Provides explanations for shard allocations in the cluster.
client.cluster.allocationExplain({ ... })
Argumentsedit
-
Request (object):
-
current_node
(Optional, string): Specifies the node ID or the name of the node to only explain a shard that is currently located on the specified node. -
index
(Optional, string): Specifies the name of the index that you would like an explanation for. -
primary
(Optional, boolean): If true, returns explanation for the primary shard for the given shard ID. -
shard
(Optional, number): Specifies the ID of the shard that you would like an explanation for. -
include_disk_info
(Optional, boolean): If true, returns information about disk usage and shard sizes. -
include_yes_decisions
(Optional, boolean): If true, returns YES decisions in explanation.
-
delete_component_templateedit
Deletes a component template
client.cluster.deleteComponentTemplate({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): List or wildcard expression of component template names used to limit the request. -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
delete_voting_config_exclusionsedit
Clears cluster voting config exclusions.
client.cluster.deleteVotingConfigExclusions({ ... })
Argumentsedit
-
Request (object):
-
wait_for_removal
(Optional, boolean): Specifies whether to wait for all excluded nodes to be removed from the cluster before clearing the voting configuration exclusions list. Defaults to true, meaning that all excluded nodes must be removed from the cluster before this API takes any action. If set to false then the voting configuration exclusions list is cleared even if some excluded nodes are still in the cluster.
-
exists_component_templateedit
Returns information about whether a particular component template exist
client.cluster.existsComponentTemplate({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): List of component template names used to limit the request. Wildcard (*) expressions are supported. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
local
(Optional, boolean): If true, the request retrieves information from the local node only. Defaults to false, which means information is retrieved from the master node.
-
get_component_templateedit
Returns one or more component templates
client.cluster.getComponentTemplate({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string): The comma separated names of the component templates -
flat_settings
(Optional, boolean) -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false) -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
include_defaults
(Optional, boolean): Return all default configurations for the component template (default: false)
-
get_settingsedit
Returns cluster settings.
client.cluster.getSettings({ ... })
Argumentsedit
-
Request (object):
-
flat_settings
(Optional, boolean): Return settings in flat format (default: false) -
include_defaults
(Optional, boolean): Whether to return all default clusters setting. -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
healthedit
Returns basic information about the health of the cluster.
client.cluster.health({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): List of data streams, indices, and index aliases used to limit the request. Wildcard expressions (*) are supported. To target all data streams and indices in a cluster, omit this parameter or use _all or *. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
level
(Optional, Enum("cluster" | "indices" | "shards")): Can be one of cluster, indices or shards. Controls the details level of the health information returned. -
local
(Optional, boolean): If true, the request retrieves information from the local node only. Defaults to false, which means information is retrieved from the master node. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error. -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): A number controlling to how many active shards to wait for, all to wait for all shards in the cluster to be active, or 0 to not wait. -
wait_for_events
(Optional, Enum("immediate" | "urgent" | "high" | "normal" | "low" | "languid")): Can be one of immediate, urgent, high, normal, low, languid. Wait until all currently queued events with the given priority are processed. -
wait_for_nodes
(Optional, string | number): The request waits until the specified number N of nodes is available. It also accepts >=N, ⇐N, >N and <N. Alternatively, it is possible to use ge(N), le(N), gt(N) and lt(N) notation. -
wait_for_no_initializing_shards
(Optional, boolean): A boolean value which controls whether to wait (until the timeout provided) for the cluster to have no shard initializations. Defaults to false, which means it will not wait for initializing shards. -
wait_for_no_relocating_shards
(Optional, boolean): A boolean value which controls whether to wait (until the timeout provided) for the cluster to have no shard relocations. Defaults to false, which means it will not wait for relocating shards. -
wait_for_status
(Optional, Enum("green" | "yellow" | "red")): One of green, yellow or red. Will wait (until the timeout provided) until the status of the cluster changes to the one provided or better, i.e. green > yellow > red. By default, will not wait for any status.
-
infoedit
Returns different information about the cluster.
client.cluster.info({ target })
Argumentsedit
-
Request (object):
-
target
(Enum("_all" | "http" | "ingest" | "thread_pool" | "script") | Enum("_all" | "http" | "ingest" | "thread_pool" | "script")[]): Limits the information returned to the specific target. Supports a list, such as http,ingest.
-
pending_tasksedit
Returns a list of any cluster-level changes (e.g. create index, update mapping, allocate or fail shard) which have not yet been executed.
client.cluster.pendingTasks({ ... })
Argumentsedit
-
Request (object):
-
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master
-
post_voting_config_exclusionsedit
Updates the cluster voting config exclusions by node ids or node names.
client.cluster.postVotingConfigExclusions({ ... })
Argumentsedit
-
Request (object):
-
node_names
(Optional, string | string[]): A list of the names of the nodes to exclude from the voting configuration. If specified, you may not also specify node_ids. -
node_ids
(Optional, string | string[]): A list of the persistent ids of the nodes to exclude from the voting configuration. If specified, you may not also specify node_names. -
timeout
(Optional, string | -1 | 0): When adding a voting configuration exclusion, the API waits for the specified nodes to be excluded from the voting configuration before returning. If the timeout expires before the appropriate condition is satisfied, the request fails and returns an error.
-
put_component_templateedit
Creates or updates a component template
client.cluster.putComponentTemplate({ name, template })
Argumentsedit
-
Request (object):
-
name
(string): The name of the template -
template
({ aliases, mappings, settings, defaults, data_stream, lifecycle }): The template to be applied which includes mappings, settings, or aliases configuration. -
version
(Optional, number): Version number used to manage component templates externally. This number isn’t automatically generated or incremented by Elasticsearch. -
_meta
(Optional, Record<string, User-defined value>): Optional user metadata about the component template. May have any contents. This map is not automatically generated by Elasticsearch. -
allow_auto_create
(Optional, boolean): This setting overrides the value of theaction.auto_create_index
cluster setting. If set totrue
in a template, then indices can be automatically created using that template even if auto-creation of indices is disabled viaactions.auto_create_index
. If set tofalse
then data streams matching the template must always be explicitly created. -
create
(Optional, boolean): Whether the index template should only be added if new or can also replace an existing one -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master
-
put_settingsedit
Updates the cluster settings.
client.cluster.putSettings({ ... })
Argumentsedit
-
Request (object):
-
persistent
(Optional, Record<string, User-defined value>) -
transient
(Optional, Record<string, User-defined value>) -
flat_settings
(Optional, boolean): Return settings in flat format (default: false) -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
remote_infoedit
Returns the information about configured remote clusters.
client.cluster.remoteInfo()
rerouteedit
Allows to manually change the allocation of individual shards in the cluster.
client.cluster.reroute({ ... })
Argumentsedit
-
Request (object):
-
commands
(Optional, { cancel, move, allocate_replica, allocate_stale_primary, allocate_empty_primary }[]): Defines the commands to perform. -
dry_run
(Optional, boolean): If true, then the request simulates the operation only and returns the resulting state. -
explain
(Optional, boolean): If true, then the response contains an explanation of why the commands can or cannot be executed. -
metric
(Optional, string | string[]): Limits the information returned to the specified metrics. -
retry_failed
(Optional, boolean): If true, then retries allocation of shards that are blocked due to too many subsequent allocation failures. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
stateedit
Returns a comprehensive information about the state of the cluster.
client.cluster.state({ ... })
Argumentsedit
-
Request (object):
-
metric
(Optional, string | string[]): Limit the information returned to the specified metrics -
index
(Optional, string | string[]): A list of index names; use_all
or empty string to perform the operation on all indices -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
flat_settings
(Optional, boolean): Return settings in flat format (default: false) -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
wait_for_metadata_version
(Optional, number): Wait for the metadata version to be equal or greater than the specified metadata version -
wait_for_timeout
(Optional, string | -1 | 0): The maximum time to wait for wait_for_metadata_version before timing out
-
statsedit
Returns high-level overview of cluster statistics.
client.cluster.stats({ ... })
Argumentsedit
-
Request (object):
-
node_id
(Optional, string | string[]): List of node filters used to limit returned information. Defaults to all nodes in the cluster. -
flat_settings
(Optional, boolean): Return settings in flat format (default: false) -
timeout
(Optional, string | -1 | 0): Period to wait for each node to respond. If a node does not respond before its timeout expires, the response does not include its stats. However, timed out nodes are included in the response’s _nodes.failed property. Defaults to no timeout.
-
dangling_indicesedit
delete_dangling_indexedit
Deletes the specified dangling index
client.danglingIndices.deleteDanglingIndex({ index_uuid, accept_data_loss })
Argumentsedit
-
Request (object):
-
index_uuid
(string): The UUID of the dangling index -
accept_data_loss
(boolean): Must be set to true in order to delete the dangling index -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
import_dangling_indexedit
Imports the specified dangling index
client.danglingIndices.importDanglingIndex({ index_uuid, accept_data_loss })
Argumentsedit
-
Request (object):
-
index_uuid
(string): The UUID of the dangling index -
accept_data_loss
(boolean): Must be set to true in order to import the dangling index -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
list_dangling_indicesedit
Returns all dangling indices.
client.danglingIndices.listDanglingIndices()
enrichedit
delete_policyedit
Deletes an existing enrich policy and its enrich index.
client.enrich.deletePolicy({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the enrich policy
-
execute_policyedit
Creates the enrich index for an existing enrich policy.
client.enrich.executePolicy({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the enrich policy -
wait_for_completion
(Optional, boolean): Should the request should block until the execution is complete.
-
get_policyedit
Gets information about an enrich policy.
client.enrich.getPolicy({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string | string[]): A list of enrich policy names
-
put_policyedit
Creates a new enrich policy.
client.enrich.putPolicy({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the enrich policy -
geo_match
(Optional, { enrich_fields, indices, match_field, query, name, elasticsearch_version }) -
match
(Optional, { enrich_fields, indices, match_field, query, name, elasticsearch_version }) -
range
(Optional, { enrich_fields, indices, match_field, query, name, elasticsearch_version })
-
statsedit
Gets enrich coordinator statistics and information about enrich policies that are currently executing.
client.enrich.stats()
eqledit
deleteedit
Deletes an async EQL search by ID. If the search is still running, the search request will be cancelled. Otherwise, the saved search results are deleted.
client.eql.delete({ id })
Argumentsedit
-
Request (object):
-
id
(string): Identifier for the search to delete.
-
getedit
Returns async results from previously executed Event Query Language (EQL) search
client.eql.get({ id })
Argumentsedit
-
Request (object):
-
id
(string): Identifier for the search. -
keep_alive
(Optional, string | -1 | 0): Period for which the search and its results are stored on the cluster. Defaults to the keep_alive value set by the search’s EQL search API request. -
wait_for_completion_timeout
(Optional, string | -1 | 0): Timeout duration to wait for the request to finish. Defaults to no timeout, meaning the request waits for complete search results.
-
get_statusedit
Returns the status of a previously submitted async or stored Event Query Language (EQL) search
client.eql.getStatus({ id })
Argumentsedit
-
Request (object):
-
id
(string): Identifier for the search.
-
searchedit
Returns results matching a query expressed in Event Query Language (EQL)
client.eql.search({ index, query })
Argumentsedit
-
Request (object):
-
index
(string | string[]): The name of the index to scope the operation -
query
(string): EQL query you wish to run. -
case_sensitive
(Optional, boolean) -
event_category_field
(Optional, string): Field containing the event classification, such as process, file, or network. -
tiebreaker_field
(Optional, string): Field used to sort hits with the same timestamp in ascending order -
timestamp_field
(Optional, string): Field containing event timestamp. Default "@timestamp" -
fetch_size
(Optional, number): Maximum number of events to search at a time for sequence queries. -
filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type } | { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }[]): Query, written in Query DSL, used to filter the events on which the EQL query runs. -
keep_alive
(Optional, string | -1 | 0) -
keep_on_completion
(Optional, boolean) -
wait_for_completion_timeout
(Optional, string | -1 | 0) -
size
(Optional, number): For basic queries, the maximum number of matching events to return. Defaults to 10 -
fields
(Optional, { field, format, include_unmapped } | { field, format, include_unmapped }[]): Array of wildcard (*) patterns. The response returns values for field names matching these patterns in the fields property of each hit. -
result_position
(Optional, Enum("tail" | "head")) -
runtime_mappings
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>) -
allow_no_indices
(Optional, boolean) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]) -
ignore_unavailable
(Optional, boolean): If true, missing or closed indices are not included in the response.
-
featuresedit
get_featuresedit
Gets a list of features which can be included in snapshots using the feature_states field when creating a snapshot
client.features.getFeatures()
reset_featuresedit
Resets the internal state of features, usually by deleting system indices
client.features.resetFeatures()
fleetedit
global_checkpointsedit
Returns the current global checkpoints for an index. This API is design for internal use by the fleet server project.
client.fleet.globalCheckpoints({ index })
Argumentsedit
-
Request (object):
-
index
(string | string): A single index or index alias that resolves to a single index. -
wait_for_advance
(Optional, boolean): A boolean value which controls whether to wait (until the timeout) for the global checkpoints to advance past the providedcheckpoints
. -
wait_for_index
(Optional, boolean): A boolean value which controls whether to wait (until the timeout) for the target index to exist and all primary shards be active. Can only be true whenwait_for_advance
is true. -
checkpoints
(Optional, number[]): A comma separated list of previous global checkpoints. When used in combination withwait_for_advance
, the API will only return once the global checkpoints advances past the checkpoints. Providing an empty list will cause Elasticsearch to immediately return the current global checkpoints. -
timeout
(Optional, string | -1 | 0): Period to wait for a global checkpoints to advance pastcheckpoints
.
-
msearchedit
Multi Search API where the search will only be executed after specified checkpoints are available due to a refresh. This API is designed for internal use by the fleet server project.
client.fleet.msearch({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string): A single target to search. If the target is an index alias, it must resolve to a single index. -
searches
(Optional, { allow_no_indices, expand_wildcards, ignore_unavailable, index, preference, request_cache, routing, search_type, ccs_minimize_roundtrips, allow_partial_search_results, ignore_throttled } | { aggregations, collapse, query, explain, ext, stored_fields, docvalue_fields, knn, from, highlight, indices_boost, min_score, post_filter, profile, rescore, script_fields, search_after, size, sort, _source, fields, terminate_after, stats, timeout, track_scores, track_total_hits, version, runtime_mappings, seq_no_primary_term, pit, suggest }[]) -
allow_no_indices
(Optional, boolean): If false, the request returns an error if any wildcard expression, index alias, or _all value targets only missing or closed indices. This behavior applies even if the request targets other open indices. For example, a request targeting foo*,bar* returns an error if an index starts with foo but no index starts with bar. -
ccs_minimize_roundtrips
(Optional, boolean): If true, network roundtrips between the coordinating node and remote clusters are minimized for cross-cluster search requests. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard expressions can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. -
ignore_throttled
(Optional, boolean): If true, concrete, expanded or aliased indices are ignored when frozen. -
ignore_unavailable
(Optional, boolean): If true, missing or closed indices are not included in the response. -
max_concurrent_searches
(Optional, number): Maximum number of concurrent searches the multi search API can execute. -
max_concurrent_shard_requests
(Optional, number): Maximum number of concurrent shard requests that each sub-search request executes per node. -
pre_filter_shard_size
(Optional, number): Defines a threshold that enforces a pre-filter roundtrip to prefilter search shards based on query rewriting if the number of shards the search request expands to exceeds the threshold. This filter roundtrip can limit the number of shards significantly if for instance a shard can not match any documents based on its rewrite method i.e., if date filters are mandatory to match but the shard bounds and the query are disjoint. -
search_type
(Optional, Enum("query_then_fetch" | "dfs_query_then_fetch")): Indicates whether global term and document frequencies should be used when scoring returned documents. -
rest_total_hits_as_int
(Optional, boolean): If true, hits.total are returned as an integer in the response. Defaults to false, which returns an object. -
typed_keys
(Optional, boolean): Specifies whether aggregation and suggester names should be prefixed by their respective types in the response. -
wait_for_checkpoints
(Optional, number[]): A comma separated list of checkpoints. When configured, the search API will only be executed on a shard after the relevant checkpoint has become visible for search. Defaults to an empty list which will cause Elasticsearch to immediately execute the search. -
allow_partial_search_results
(Optional, boolean): If true, returns partial results if there are shard request timeouts or [shard failures](https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-replication.html#shard-failures). If false, returns an error with no partial results. Defaults to the configured cluster settingsearch.default_allow_partial_results
which is true by default.
-
searchedit
Search API where the search will only be executed after specified checkpoints are available due to a refresh. This API is designed for internal use by the fleet server project.
client.fleet.search({ index })
Argumentsedit
-
Request (object):
-
index
(string | string): A single target to search. If the target is an index alias, it must resolve to a single index. -
aggregations
(Optional, Record<string, { aggregations, meta, adjacency_matrix, auto_date_histogram, avg, avg_bucket, boxplot, bucket_script, bucket_selector, bucket_sort, bucket_count_ks_test, bucket_correlation, cardinality, categorize_text, children, composite, cumulative_cardinality, cumulative_sum, date_histogram, date_range, derivative, diversified_sampler, extended_stats, extended_stats_bucket, frequent_item_sets, filter, filters, geo_bounds, geo_centroid, geo_distance, geohash_grid, geo_line, geotile_grid, geohex_grid, global, histogram, ip_range, ip_prefix, inference, line, matrix_stats, max, max_bucket, median_absolute_deviation, min, min_bucket, missing, moving_avg, moving_percentiles, moving_fn, multi_terms, nested, normalize, parent, percentile_ranks, percentiles, percentiles_bucket, range, rare_terms, rate, reverse_nested, sampler, scripted_metric, serial_diff, significant_terms, significant_text, stats, stats_bucket, string_stats, sum, sum_bucket, terms, top_hits, t_test, top_metrics, value_count, weighted_avg, variable_width_histogram }>) -
collapse
(Optional, { field, inner_hits, max_concurrent_group_searches, collapse }) -
explain
(Optional, boolean): If true, returns detailed information about score computation as part of a hit. -
ext
(Optional, Record<string, User-defined value>): Configuration of search extensions defined by Elasticsearch plugins. -
from
(Optional, number): Starting document offset. By default, you cannot page through more than 10,000 hits using the from and size parameters. To page through more hits, use the search_after parameter. -
highlight
(Optional, { encoder, fields }) -
track_total_hits
(Optional, boolean | number): Number of hits matching the query to count accurately. If true, the exact number of hits is returned at the cost of some performance. If false, the response does not include the total number of hits matching the query. Defaults to 10,000 hits. -
indices_boost
(Optional, Record<string, number>[]): Boosts the _score of documents from specified indices. -
docvalue_fields
(Optional, { field, format, include_unmapped }[]): Array of wildcard (*) patterns. The request returns doc values for field names matching these patterns in the hits.fields property of the response. -
min_score
(Optional, number): Minimum _score for matching documents. Documents with a lower _score are not included in the search results. -
post_filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
profile
(Optional, boolean) -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): Defines the search definition using the Query DSL. -
rescore
(Optional, { query, window_size } | { query, window_size }[]) -
script_fields
(Optional, Record<string, { script, ignore_failure }>): Retrieve a script evaluation (based on different fields) for each hit. -
search_after
(Optional, number | number | string | boolean | null | User-defined value[]) -
size
(Optional, number): The number of hits to return. By default, you cannot page through more than 10,000 hits using the from and size parameters. To page through more hits, use the search_after parameter. -
slice
(Optional, { field, id, max }) -
sort
(Optional, string | { _score, _doc, _geo_distance, _script } | string | { _score, _doc, _geo_distance, _script }[]) -
_source
(Optional, boolean | { excludes, includes }): Indicates which source fields are returned for matching documents. These fields are returned in the hits._source property of the search response. -
fields
(Optional, { field, format, include_unmapped }[]): Array of wildcard (*) patterns. The request returns values for field names matching these patterns in the hits.fields property of the response. -
suggest
(Optional, { text }) -
terminate_after
(Optional, number): Maximum number of documents to collect for each shard. If a query reaches this limit, Elasticsearch terminates the query early. Elasticsearch collects documents before sorting. Defaults to 0, which does not terminate query execution early. -
timeout
(Optional, string): Specifies the period of time to wait for a response from each shard. If no response is received before the timeout expires, the request fails and returns an error. Defaults to no timeout. -
track_scores
(Optional, boolean): If true, calculate and return document scores, even if the scores are not used for sorting. -
version
(Optional, boolean): If true, returns document version as part of a hit. -
seq_no_primary_term
(Optional, boolean): If true, returns sequence number and primary term of the last modification of each hit. See Optimistic concurrency control. -
stored_fields
(Optional, string | string[]): List of stored fields to return as part of a hit. If no fields are specified, no stored fields are included in the response. If this field is specified, the _source parameter defaults to false. You can pass _source: true to return both source fields and stored fields in the search response. -
pit
(Optional, { id, keep_alive }): Limits the search to a point in time (PIT). If you provide a PIT, you cannot specify an <index> in the request path. -
runtime_mappings
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>): Defines one or more runtime fields in the search request. These fields take precedence over mapped fields with the same name. -
stats
(Optional, string[]): Stats groups to associate with the search. Each group maintains a statistics aggregation for its associated searches. You can retrieve these stats using the indices stats API. -
allow_no_indices
(Optional, boolean) -
analyzer
(Optional, string) -
analyze_wildcard
(Optional, boolean) -
batched_reduce_size
(Optional, number) -
ccs_minimize_roundtrips
(Optional, boolean) -
default_operator
(Optional, Enum("and" | "or")) -
df
(Optional, string) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]) -
ignore_throttled
(Optional, boolean) -
ignore_unavailable
(Optional, boolean) -
lenient
(Optional, boolean) -
max_concurrent_shard_requests
(Optional, number) -
min_compatible_shard_node
(Optional, string) -
preference
(Optional, string) -
pre_filter_shard_size
(Optional, number) -
request_cache
(Optional, boolean) -
routing
(Optional, string) -
scroll
(Optional, string | -1 | 0) -
search_type
(Optional, Enum("query_then_fetch" | "dfs_query_then_fetch")) -
suggest_field
(Optional, string): Specifies which field to use for suggestions. -
suggest_mode
(Optional, Enum("missing" | "popular" | "always")) -
suggest_size
(Optional, number) -
suggest_text
(Optional, string): The source text for which the suggestions should be returned. -
typed_keys
(Optional, boolean) -
rest_total_hits_as_int
(Optional, boolean) -
_source_excludes
(Optional, string | string[]) -
_source_includes
(Optional, string | string[]) -
q
(Optional, string) -
wait_for_checkpoints
(Optional, number[]): A comma separated list of checkpoints. When configured, the search API will only be executed on a shard after the relevant checkpoint has become visible for search. Defaults to an empty list which will cause Elasticsearch to immediately execute the search. -
allow_partial_search_results
(Optional, boolean): If true, returns partial results if there are shard request timeouts or [shard failures](https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-replication.html#shard-failures). If false, returns an error with no partial results. Defaults to the configured cluster settingsearch.default_allow_partial_results
which is true by default.
-
graphedit
exploreedit
Explore extracted and summarized information about the documents and terms in an index.
client.graph.explore({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names to search; use_all
or empty string to perform the operation on all indices -
connections
(Optional, { connections, query, vertices }) -
controls
(Optional, { sample_diversity, sample_size, timeout, use_significance }) -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
vertices
(Optional, { exclude, field, include, min_doc_count, shard_min_doc_count, size }[]) -
routing
(Optional, string): Specific routing value -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
ilmedit
delete_lifecycleedit
Deletes the specified lifecycle policy definition. A currently used policy cannot be deleted.
client.ilm.deleteLifecycle({ policy })
Argumentsedit
-
Request (object):
-
policy
(string): Identifier for the policy. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
explain_lifecycleedit
Retrieves information about the index’s current lifecycle state, such as the currently executing phase, action, and step.
client.ilm.explainLifecycle({ index })
Argumentsedit
-
Request (object):
-
index
(string): List of data streams, indices, and aliases to target. Supports wildcards (*
). To target all data streams and indices, use*
or_all
. -
only_errors
(Optional, boolean): Filters the returned indices to only indices that are managed by ILM and are in an error state, either due to an encountering an error while executing the policy, or attempting to use a policy that does not exist. -
only_managed
(Optional, boolean): Filters the returned indices to only indices that are managed by ILM. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
get_lifecycleedit
Returns the specified policy definition. Includes the policy version and last modified date.
client.ilm.getLifecycle({ ... })
Argumentsedit
-
Request (object):
-
policy
(Optional, string): Identifier for the policy. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
get_statusedit
Retrieves the current index lifecycle management (ILM) status.
client.ilm.getStatus()
migrate_to_data_tiersedit
Migrates the indices and ILM policies away from custom node attribute allocation routing to data tiers routing
client.ilm.migrateToDataTiers({ ... })
Argumentsedit
-
Request (object):
-
legacy_template_to_delete
(Optional, string) -
node_attribute
(Optional, string) -
dry_run
(Optional, boolean): If true, simulates the migration from node attributes based allocation filters to data tiers, but does not perform the migration. This provides a way to retrieve the indices and ILM policies that need to be migrated.
-
move_to_stepedit
Manually moves an index into the specified step and executes that step.
client.ilm.moveToStep({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the index whose lifecycle step is to change -
current_step
(Optional, { action, name, phase }) -
next_step
(Optional, { action, name, phase })
-
put_lifecycleedit
Creates a lifecycle policy
client.ilm.putLifecycle({ policy })
Argumentsedit
-
Request (object):
-
policy
(string): Identifier for the policy. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
remove_policyedit
Removes the assigned lifecycle policy and stops managing the specified index
client.ilm.removePolicy({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the index to remove policy on
-
retryedit
Retries executing the policy for an index that is in the ERROR step.
client.ilm.retry({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the indices (comma-separated) whose failed lifecycle step is to be retry
-
startedit
Start the index lifecycle management (ILM) plugin.
client.ilm.start({ ... })
Argumentsedit
-
Request (object):
-
master_timeout
(Optional, string | -1 | 0) -
timeout
(Optional, string | -1 | 0)
-
stopedit
Halts all lifecycle management operations and stops the index lifecycle management (ILM) plugin
client.ilm.stop({ ... })
Argumentsedit
-
Request (object):
-
master_timeout
(Optional, string | -1 | 0) -
timeout
(Optional, string | -1 | 0)
-
indicesedit
add_blockedit
Adds a block to an index.
client.indices.addBlock({ index, block })
Argumentsedit
-
Request (object):
-
index
(string): A comma separated list of indices to add a block to -
block
(Enum("metadata" | "read" | "read_only" | "write")): The block to add (one of read, write, read_only or metadata) -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
analyzeedit
Performs the analysis process on a text and return the tokens breakdown of the text.
client.indices.analyze({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string): The name of the index to scope the operation -
analyzer
(Optional, string) -
attributes
(Optional, string[]) -
char_filter
(Optional, string | { type } | { type, mappings, mappings_path } | { type, flags, pattern, replacement } | { type, mode, name } | { type, normalize_kana, normalize_kanji }[]) -
explain
(Optional, boolean) -
field
(Optional, string) -
filter
(Optional, string | { type, preserve_original } | { type, common_words, common_words_path, ignore_case, query_mode } | { type, filter, script } | { type, delimiter, encoding } | { type, max_gram, min_gram, side, preserve_original } | { type, articles, articles_path, articles_case } | { type, max_output_size, separator } | { type, dedup, dictionary, locale, longest_only } | { type } | { type, mode, types } | { type, keep_words, keep_words_case, keep_words_path } | { type, ignore_case, keywords, keywords_path, keywords_pattern } | { type } | { type, max, min } | { type, consume_all_tokens, max_token_count } | { type, language } | { type, filters, preserve_original } | { type, max_gram, min_gram, preserve_original } | { type, stoptags } | { type, patterns, preserve_original } | { type, all, flags, pattern, replacement } | { type } | { type, script } | { type } | { type } | { type, filler_token, max_shingle_size, min_shingle_size, output_unigrams, output_unigrams_if_no_shingles, token_separator } | { type, language } | { type, rules, rules_path } | { type, language } | { type, ignore_case, remove_trailing, stopwords, stopwords_path } | { type, expand, format, lenient, synonyms, synonyms_path, tokenizer, updateable } | { type, expand, format, lenient, synonyms, synonyms_path, tokenizer, updateable } | { type } | { type, length } | { type, only_on_same_position } | { type } | { type, adjust_offsets, catenate_all, catenate_numbers, catenate_words, generate_number_parts, generate_word_parts, ignore_keywords, preserve_original, protected_words, protected_words_path, split_on_case_change, split_on_numerics, stem_english_possessive, type_table, type_table_path } | { type, catenate_all, catenate_numbers, catenate_words, generate_number_parts, generate_word_parts, preserve_original, protected_words, protected_words_path, split_on_case_change, split_on_numerics, stem_english_possessive, type_table, type_table_path } | { type, minimum_length } | { type, use_romaji } | { type, stoptags } | { type, rule_files } | { type, alternate, caseFirst, caseLevel, country, decomposition, hiraganaQuaternaryMode, language, numeric, rules, strength, variableTop, variant } | { type, unicode_set_filter } | { type, name } | { type, dir, id } | { type, encoder, languageset, max_code_len, name_type, replace, rule_type } | { type }[]) -
normalizer
(Optional, string) -
text
(Optional, string | string[]) -
tokenizer
(Optional, string | { type, tokenize_on_chars, max_token_length } | { type, custom_token_chars, max_gram, min_gram, token_chars } | { type, buffer_size } | { type } | { type } | { type, custom_token_chars, max_gram, min_gram, token_chars } | { type, decompound_mode, discard_punctuation, user_dictionary, user_dictionary_rules } | { type, buffer_size, delimiter, replacement, reverse, skip } | { type, max_token_length } | { type, max_token_length } | { type, max_token_length } | { type, discard_punctuation, mode, nbest_cost, nbest_examples, user_dictionary, user_dictionary_rules, discard_compound_token } | { type, flags, group, pattern } | { type, rule_files })
-
clear_cacheedit
Clears all or specific caches for one or more indices.
client.indices.clearCache({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index name to limit the operation -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
fielddata
(Optional, boolean): Clear field data -
fields
(Optional, string | string[]): A list of fields to clear when using thefielddata
parameter (default: all) -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
query
(Optional, boolean): Clear query caches -
request
(Optional, boolean): Clear request cache
-
cloneedit
Clones an index
client.indices.clone({ index, target })
Argumentsedit
-
Request (object):
-
index
(string): The name of the source index to clone -
target
(string): The name of the target index to clone into -
aliases
(Optional, Record<string, { filter, index_routing, is_hidden, is_write_index, routing, search_routing }>) -
settings
(Optional, Record<string, User-defined value>) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Set the number of active shards to wait for on the cloned index before the operation returns.
-
closeedit
Closes an index.
client.indices.close({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A comma separated list of indices to close -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of active shards to wait for before the operation returns.
-
createedit
Creates an index with optional settings and mappings.
client.indices.create({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the index -
aliases
(Optional, Record<string, { filter, index_routing, is_hidden, is_write_index, routing, search_routing }>) -
mappings
(Optional, { all_field, date_detection, dynamic, dynamic_date_formats, dynamic_templates, _field_names, index_field, _meta, numeric_detection, properties, _routing, _size, _source, runtime, enabled, _data_stream_timestamp }): Mapping for fields in the index. If specified, this mapping can include:- Field names
- Field data types
- Mapping parameters
-
settings
(Optional, { index, mode, routing_path, soft_deletes, sort, number_of_shards, number_of_replicas, number_of_routing_shards, check_on_startup, codec, routing_partition_size, load_fixed_bitset_filters_eagerly, hidden, auto_expand_replicas, merge, search, refresh_interval, max_result_window, max_inner_result_window, max_rescore_window, max_docvalue_fields_search, max_script_fields, max_ngram_diff, max_shingle_diff, blocks, max_refresh_listeners, analyze, highlight, max_terms_count, max_regex_length, routing, gc_deletes, default_pipeline, final_pipeline, lifecycle, provided_name, creation_date, creation_date_string, uuid, version, verified_before_close, format, max_slices_per_scroll, translog, query_string, priority, top_metrics_max_size, analysis, settings, time_series, shards, queries, similarity, mapping, indexing.slowlog, indexing_pressure, store }) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Set the number of active shards to wait for before the operation returns.
-
create_data_streamedit
Creates a data stream
client.indices.createDataStream({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the data stream
-
data_streams_statsedit
Provides statistics on operations happening in a data stream.
client.indices.dataStreamsStats({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string): A list of data stream names; use_all
or empty string to perform the operation on all data streams -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[])
-
deleteedit
Deletes an index.
client.indices.delete({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of indices to delete; use_all
or*
string to delete all indices -
allow_no_indices
(Optional, boolean): Ignore if a wildcard expression resolves to no concrete indices (default: false) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether wildcard expressions should get expanded to open, closed, or hidden indices -
ignore_unavailable
(Optional, boolean): Ignore unavailable indexes (default: false) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
delete_aliasedit
Deletes an alias.
client.indices.deleteAlias({ index, name })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names (supports wildcards); use_all
for all indices -
name
(string | string[]): A list of aliases to delete (supports wildcards); use_all
to delete all aliases for the specified indices. -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit timestamp for the document
-
delete_data_lifecycleedit
Deletes the data lifecycle of the selected data streams.
client.indices.deleteDataLifecycle({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): A list of data streams of which the data lifecycle will be deleted; use*
to get all data streams -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether wildcard expressions should get expanded to open or closed indices (default: open) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit timestamp for the document
-
delete_data_streamedit
Deletes a data stream.
client.indices.deleteDataStream({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): A list of data streams to delete; use*
to delete all data streams -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether wildcard expressions should get expanded to open or closed indices (default: open)
-
delete_index_templateedit
Deletes an index template.
client.indices.deleteIndexTemplate({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): List of index template names used to limit the request. Wildcard (*) expressions are supported. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
delete_templateedit
Deletes an index template.
client.indices.deleteTemplate({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the template -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
disk_usageedit
Analyzes the disk usage of each field of an index or data stream
client.indices.diskUsage({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): List of data streams, indices, and aliases used to limit the request. It’s recommended to execute this API with a single index (or the latest backing index of a data stream) as the API consumes resources significantly. -
allow_no_indices
(Optional, boolean): If false, the request returns an error if any wildcard expression, index alias, or _all value targets only missing or closed indices. This behavior applies even if the request targets other open indices. For example, a request targeting foo*,bar* returns an error if an index starts with foo but no index starts with bar. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard patterns can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. Supports a list of values, such as open,hidden. -
flush
(Optional, boolean): If true, the API performs a flush before analysis. If false, the response may not include uncommitted data. -
ignore_unavailable
(Optional, boolean): If true, missing or closed indices are not included in the response. -
run_expensive_tasks
(Optional, boolean): Analyzing field disk usage is resource-intensive. To use the API, this parameter must be set to true.
-
downsampleedit
Downsample an index
client.indices.downsample({ index, target_index })
Argumentsedit
-
Request (object):
-
index
(string): The index to downsample -
target_index
(string): The name of the target index to store downsampled data -
config
(Optional, { fixed_interval })
-
existsedit
Returns information about whether a particular index exists.
client.indices.exists({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names -
allow_no_indices
(Optional, boolean): Ignore if a wildcard expression resolves to no concrete indices (default: false) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether wildcard expressions should get expanded to open or closed indices (default: open) -
flat_settings
(Optional, boolean): Return settings in flat format (default: false) -
ignore_unavailable
(Optional, boolean): Ignore unavailable indexes (default: false) -
include_defaults
(Optional, boolean): Whether to return all default setting for each of the indices. -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false)
-
exists_aliasedit
Returns information about whether a particular alias exists.
client.indices.existsAlias({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): A list of alias names to return -
index
(Optional, string | string[]): A list of index names to filter aliases -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false)
-
exists_index_templateedit
Returns information about whether a particular index template exists.
client.indices.existsIndexTemplate({ name })
Argumentsedit
-
Request (object):
-
name
(string): List of index template names used to limit the request. Wildcard (*) expressions are supported. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error.
-
exists_templateedit
Returns information about whether a particular index template exists.
client.indices.existsTemplate({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): The comma separated names of the index templates -
flat_settings
(Optional, boolean): Return settings in flat format (default: false) -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false) -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node
-
explain_data_lifecycleedit
Retrieves information about the index’s current DLM lifecycle, such as any potential encountered error, time since creation etc.
client.indices.explainDataLifecycle({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): The name of the index to explain -
include_defaults
(Optional, boolean): indicates if the API should return the default values the system uses for the index’s lifecycle -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master
-
field_usage_statsedit
Returns the field usage stats for each field of an index
client.indices.fieldUsageStats({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): List or wildcard expression of index names used to limit the request. -
allow_no_indices
(Optional, boolean): If false, the request returns an error if any wildcard expression, index alias, or _all value targets only missing or closed indices. This behavior applies even if the request targets other open indices. For example, a request targetingfoo*,bar*
returns an error if an index starts withfoo
but no index starts withbar
. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard patterns can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. Supports a list of values, such asopen,hidden
. -
ignore_unavailable
(Optional, boolean): If true, missing or closed indices are not included in the response. -
fields
(Optional, string | string[]): List or wildcard expressions of fields to include in the statistics. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error. -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): The number of shard copies that must be active before proceeding with the operation. Set to all or any positive integer up to the total number of shards in the index (number_of_replicas+1
).
-
flushedit
Performs the flush operation on one or more indices.
client.indices.flush({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names; use_all
or empty string for all indices -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
force
(Optional, boolean): Whether a flush should be forced even if it is not necessarily needed ie. if no changes will be committed to the index. This is useful if transaction log IDs should be incremented even if no uncommitted changes are present. (This setting can be considered as internal) -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
wait_if_ongoing
(Optional, boolean): If set to true the flush operation will block until the flush can be executed if another flush operation is already executing. The default is true. If set to false the flush will be skipped iff if another flush operation is already running.
-
forcemergeedit
Performs the force merge operation on one or more indices.
client.indices.forcemerge({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names; use_all
or empty string to perform the operation on all indices -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
flush
(Optional, boolean): Specify whether the index should be flushed after performing the operation (default: true) -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
max_num_segments
(Optional, number): The number of segments the index should be merged into (default: dynamic) -
only_expunge_deletes
(Optional, boolean): Specify whether the operation should only expunge deleted documents -
wait_for_completion
(Optional, boolean): Should the request wait until the force merge is completed.
-
getedit
Returns information about one or more indices.
client.indices.get({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): List of data streams, indices, and index aliases used to limit the request. Wildcard expressions (*) are supported. -
allow_no_indices
(Optional, boolean): If false, the request returns an error if any wildcard expression, index alias, or _all value targets only missing or closed indices. This behavior applies even if the request targets other open indices. For example, a request targeting foo*,bar* returns an error if an index starts with foo but no index starts with bar. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard expressions can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. Supports a list of values, such as open,hidden. -
flat_settings
(Optional, boolean): If true, returns settings in flat format. -
ignore_unavailable
(Optional, boolean): If false, requests that target a missing index return an error. -
include_defaults
(Optional, boolean): If true, return all default settings in the response. -
local
(Optional, boolean): If true, the request retrieves information from the local node only. Defaults to false, which means information is retrieved from the master node. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
features
(Optional, { name, description } | { name, description }[]): Return only information on specified index features
-
get_aliasedit
Returns an alias.
client.indices.getAlias({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string | string[]): A list of alias names to return -
index
(Optional, string | string[]): A list of index names to filter aliases -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false)
-
get_data_lifecycleedit
Returns the data lifecycle of the selected data streams.
client.indices.getDataLifecycle({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): A list of data streams to get; use*
to get all data streams -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether wildcard expressions should get expanded to open or closed indices (default: open) -
include_defaults
(Optional, boolean): Return all relevant default configurations for the data stream (default: false)
-
get_data_streamedit
Returns data streams.
client.indices.getDataStream({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string | string[]): A list of data streams to get; use*
to get all data streams -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether wildcard expressions should get expanded to open or closed indices (default: open) -
include_defaults
(Optional, boolean): If true, returns all relevant default configurations for the index template.
-
get_field_mappingedit
Returns mapping for one or more fields.
client.indices.getFieldMapping({ fields })
Argumentsedit
-
Request (object):
-
fields
(string | string[]): A list of fields -
index
(Optional, string | string[]): A list of index names -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
include_defaults
(Optional, boolean): Whether the default mapping values should be returned as well -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false)
-
get_index_templateedit
Returns an index template.
client.indices.getIndexTemplate({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string): List of index template names used to limit the request. Wildcard (*) expressions are supported. -
local
(Optional, boolean): If true, the request retrieves information from the local node only. Defaults to false, which means information is retrieved from the master node. -
flat_settings
(Optional, boolean): If true, returns settings in flat format. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
include_defaults
(Optional, boolean): If true, returns all relevant default configurations for the index template.
-
get_mappingedit
Returns mappings for one or more indices.
client.indices.getMapping({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master
-
get_settingsedit
Returns settings for one or more indices.
client.indices.getSettings({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names; use_all
or empty string to perform the operation on all indices -
name
(Optional, string | string[]): The name of the settings that should be included -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
flat_settings
(Optional, boolean): Return settings in flat format (default: false) -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
include_defaults
(Optional, boolean): Whether to return all default setting for each of the indices. -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master
-
get_templateedit
Returns an index template.
client.indices.getTemplate({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string | string[]): The comma separated names of the index templates -
flat_settings
(Optional, boolean): Return settings in flat format (default: false) -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false) -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node
-
migrate_to_data_streamedit
Migrates an alias to a data stream
client.indices.migrateToDataStream({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the alias to migrate
-
modify_data_streamedit
Modifies a data stream
client.indices.modifyDataStream({ actions })
Argumentsedit
-
Request (object):
-
actions
({ add_backing_index, remove_backing_index }[]): Actions to perform.
-
openedit
Opens an index.
client.indices.open({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A comma separated list of indices to open -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Sets the number of active shards to wait for before the operation returns.
-
promote_data_streamedit
Promotes a data stream from a replicated data stream managed by CCR to a regular data stream
client.indices.promoteDataStream({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the data stream
-
put_aliasedit
Creates or updates an alias.
client.indices.putAlias({ index, name })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names the alias should point to (supports wildcards); use_all
to perform the operation on all indices. -
name
(string): The name of the alias to be created or updated -
filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
index_routing
(Optional, string) -
is_write_index
(Optional, boolean) -
routing
(Optional, string) -
search_routing
(Optional, string) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit timestamp for the document
-
put_data_lifecycleedit
Updates the data lifecycle of the selected data streams.
client.indices.putDataLifecycle({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): A list of data streams whose lifecycle will be updated; use*
to set the lifecycle to all data streams -
data_retention
(Optional, string | -1 | 0) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether wildcard expressions should get expanded to open or closed indices (default: open) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit timestamp for the document
-
put_index_templateedit
Creates or updates an index template.
client.indices.putIndexTemplate({ name })
Argumentsedit
-
Request (object):
-
name
(string): Index or template name -
index_patterns
(Optional, string | string[]) -
composed_of
(Optional, string[]) -
template
(Optional, { aliases, mappings, settings, lifecycle }) -
data_stream
(Optional, { hidden }) -
priority
(Optional, number) -
version
(Optional, number) -
_meta
(Optional, Record<string, User-defined value>) -
create
(Optional, boolean): Whether the index template should only be added if new or can also replace an existing one
-
put_mappingedit
Updates the index mappings.
client.indices.putMapping({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names the mapping should be added to (supports wildcards); use_all
or omit to add the mapping on all indices. -
date_detection
(Optional, boolean): Controls whether dynamic date detection is enabled. -
dynamic
(Optional, Enum("strict" | "runtime" | true | false)): Controls whether new fields are added dynamically. -
dynamic_date_formats
(Optional, string[]): If date detection is enabled then new string fields are checked against dynamic_date_formats and if the value matches then a new date field is added instead of string. -
dynamic_templates
(Optional, Record<string, { mapping, match, match_mapping_type, match_pattern, path_match, path_unmatch, unmatch }> | Record<string, { mapping, match, match_mapping_type, match_pattern, path_match, path_unmatch, unmatch }>[]): Specify dynamic templates for the mapping. -
_field_names
(Optional, { enabled }): Control whether field names are enabled for the index. -
_meta
(Optional, Record<string, User-defined value>): A mapping type can have custom meta data associated with it. These are not used at all by Elasticsearch, but can be used to store application-specific metadata. -
numeric_detection
(Optional, boolean): Automatically map strings into numeric data types for all fields. -
properties
(Optional, Record<string, { type } | { boost, fielddata, index, null_value, type } | { type, enabled, null_value, boost, coerce, script, on_script_error, ignore_malformed, time_series_metric, analyzer, eager_global_ordinals, index, index_options, index_phrases, index_prefixes, norms, position_increment_gap, search_analyzer, search_quote_analyzer, term_vector, format, precision_step, locale } | { relations, eager_global_ordinals, type } | { boost, eager_global_ordinals, index, index_options, normalizer, norms, null_value, split_queries_on_whitespace, time_series_dimension, type } | { type, fields, meta, copy_to } | { type } | { positive_score_impact, type } | { type } | { analyzer, index, index_options, max_shingle_size, norms, search_analyzer, search_quote_analyzer, term_vector, type } | { analyzer, boost, eager_global_ordinals, fielddata, fielddata_frequency_filter, index, index_options, index_phrases, index_prefixes, norms, position_increment_gap, search_analyzer, search_quote_analyzer, term_vector, type } | { type } | { type, null_value } | { boost, format, ignore_malformed, index, null_value, precision_step, type } | { boost, fielddata, format, ignore_malformed, index, null_value, precision_step, locale, type } | { type, default_metric, metrics, time_series_metric } | { type, dims, similarity, index, index_options } | { boost, depth_limit, doc_values, eager_global_ordinals, index, index_options, null_value, similarity, split_queries_on_whitespace, type } | { enabled, include_in_parent, include_in_root, type } | { enabled, type } | { analyzer, contexts, max_input_length, preserve_position_increments, preserve_separators, search_analyzer, type } | { value, type } | { path, type } | { ignore_malformed, type } | { boost, index, ignore_malformed, null_value, on_script_error, script, time_series_dimension, type } | { type } | { analyzer, boost, index, null_value, enable_position_increments, type } | { ignore_malformed, ignore_z_value, null_value, type } | { coerce, ignore_malformed, ignore_z_value, orientation, strategy, type } | { ignore_malformed, ignore_z_value, null_value, type } | { coerce, ignore_malformed, ignore_z_value, orientation, type } | { type, null_value } | { type, null_value } | { type, null_value } | { type, null_value } | { type, null_value } | { type, null_value } | { type, null_value, scaling_factor } | { type, null_value } | { type, null_value } | { format, type } | { type } | { type } | { type } | { type } | { type }>): Mapping for a field. For new fields, this mapping can include:- Field name
- Field data type
- Mapping parameters
-
_routing
(Optional, { required }): Enable making a routing value required on indexed documents. -
_source
(Optional, { compress, compress_threshold, enabled, excludes, includes, mode }): Control whether the _source field is enabled on the index. -
runtime
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>): Mapping of runtime fields for the index. -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
write_index_only
(Optional, boolean): When true, applies mappings only to the write index of an alias or data stream
-
put_settingsedit
Updates the index settings.
client.indices.putSettings({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names; use_all
or empty string to perform the operation on all indices -
settings
(Optional, { index, mode, routing_path, soft_deletes, sort, number_of_shards, number_of_replicas, number_of_routing_shards, check_on_startup, codec, routing_partition_size, load_fixed_bitset_filters_eagerly, hidden, auto_expand_replicas, merge, search, refresh_interval, max_result_window, max_inner_result_window, max_rescore_window, max_docvalue_fields_search, max_script_fields, max_ngram_diff, max_shingle_diff, blocks, max_refresh_listeners, analyze, highlight, max_terms_count, max_regex_length, routing, gc_deletes, default_pipeline, final_pipeline, lifecycle, provided_name, creation_date, creation_date_string, uuid, version, verified_before_close, format, max_slices_per_scroll, translog, query_string, priority, top_metrics_max_size, analysis, settings, time_series, shards, queries, similarity, mapping, indexing.slowlog, indexing_pressure, store }) -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
flat_settings
(Optional, boolean): Return settings in flat format (default: false) -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
preserve_existing
(Optional, boolean): Whether to update existing settings. If set totrue
existing settings on an index remain unchanged, the default isfalse
-
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
put_templateedit
Creates or updates an index template.
client.indices.putTemplate({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the template -
aliases
(Optional, Record<string, { filter, index_routing, is_hidden, is_write_index, routing, search_routing }>): Aliases for the index. -
index_patterns
(Optional, string | string[]): Array of wildcard expressions used to match the names of indices during creation. -
mappings
(Optional, { all_field, date_detection, dynamic, dynamic_date_formats, dynamic_templates, _field_names, index_field, _meta, numeric_detection, properties, _routing, _size, _source, runtime, enabled, _data_stream_timestamp }): Mapping for fields in the index. -
order
(Optional, number): Order in which Elasticsearch applies this template if index matches multiple templates.
-
Templates with lower order values are merged first. Templates with higher
order values are merged later, overriding templates with lower values.
settings
(Optional, Record<string, User-defined value>): Configuration options for the index.
version
(Optional, number): Version number used to manage index templates externally. This number
is not automatically generated by Elasticsearch.
create
(Optional, boolean): If true, this request cannot replace or update existing index templates.
flat_settings
(Optional, boolean)
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is
received before the timeout expires, the request fails and returns an error.
timeout
(Optional, string | -1 | 0)
recoveryedit
Returns information about ongoing index shard recoveries.
client.indices.recovery({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names; use_all
or empty string to perform the operation on all indices -
active_only
(Optional, boolean): Display only those recoveries that are currently on-going -
detailed
(Optional, boolean): Whether to display detailed information about shard recovery
-
refreshedit
Performs the refresh operation in one or more indices.
client.indices.refresh({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names; use_all
or empty string to perform the operation on all indices -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed)
-
reload_search_analyzersedit
Reloads an index’s search analyzers and their resources.
client.indices.reloadSearchAnalyzers({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): A list of index names to reload analyzers for -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed)
-
resolve_indexedit
Returns information about any matching indices, aliases, and data streams
client.indices.resolveIndex({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): A list of names or wildcard expressions -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether wildcard expressions should get expanded to open or closed indices (default: open)
-
rolloveredit
Updates an alias to point to a new index when the existing index is considered to be too large or too old.
client.indices.rollover({ alias })
Argumentsedit
-
Request (object):
-
alias
(string): The name of the alias to rollover -
new_index
(Optional, string): The name of the rollover index -
aliases
(Optional, Record<string, { filter, index_routing, is_hidden, is_write_index, routing, search_routing }>) -
conditions
(Optional, { min_age, max_age, max_age_millis, min_docs, max_docs, max_size, max_size_bytes, min_size, min_size_bytes, max_primary_shard_size, max_primary_shard_size_bytes, min_primary_shard_size, min_primary_shard_size_bytes, max_primary_shard_docs, min_primary_shard_docs }) -
mappings
(Optional, { all_field, date_detection, dynamic, dynamic_date_formats, dynamic_templates, _field_names, index_field, _meta, numeric_detection, properties, _routing, _size, _source, runtime, enabled, _data_stream_timestamp }) -
settings
(Optional, Record<string, User-defined value>) -
dry_run
(Optional, boolean): If set to true the rollover action will only be validated but not actually performed even if a condition matches. The default is false -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Set the number of active shards to wait for on the newly created rollover index before the operation returns.
-
segmentsedit
Provides low-level information about segments in a Lucene index.
client.indices.segments({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names; use_all
or empty string to perform the operation on all indices -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
verbose
(Optional, boolean): Includes detailed memory usage by Lucene.
-
shard_storesedit
Provides store information for shard copies of indices.
client.indices.shardStores({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): List of data streams, indices, and aliases used to limit the request. -
allow_no_indices
(Optional, boolean): If false, the request returns an error if any wildcard expression, index alias, or _all value targets only missing or closed indices. This behavior applies even if the request targets other open indices. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard patterns can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. -
ignore_unavailable
(Optional, boolean): If true, missing or closed indices are not included in the response. -
status
(Optional, Enum("green" | "yellow" | "red" | "all") | Enum("green" | "yellow" | "red" | "all")[]): List of shard health statuses used to limit the request.
-
shrinkedit
Allow to shrink an existing index into a new index with fewer primary shards.
client.indices.shrink({ index, target })
Argumentsedit
-
Request (object):
-
index
(string): The name of the source index to shrink -
target
(string): The name of the target index to shrink into -
aliases
(Optional, Record<string, { filter, index_routing, is_hidden, is_write_index, routing, search_routing }>) -
settings
(Optional, Record<string, User-defined value>) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Set the number of active shards to wait for on the shrunken index before the operation returns.
-
simulate_index_templateedit
Simulate matching the given index name against the index templates in the system
client.indices.simulateIndexTemplate({ name })
Argumentsedit
-
Request (object):
-
name
(string): Index or template name to simulate -
allow_auto_create
(Optional, boolean) -
index_patterns
(Optional, string | string[]) -
composed_of
(Optional, string[]) -
template
(Optional, { aliases, mappings, settings, lifecycle }) -
data_stream
(Optional, { hidden }) -
priority
(Optional, number) -
version
(Optional, number) -
_meta
(Optional, Record<string, User-defined value>) -
create
(Optional, boolean): Iftrue
, the template passed in the body is only used if no existing templates match the same index patterns. Iffalse
, the simulation uses the template with the highest priority. Note that the template is not permanently added or updated in either case; it is only used for the simulation. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
include_defaults
(Optional, boolean): If true, returns all relevant default configurations for the index template.
-
simulate_templateedit
Simulate resolving the given template name or body
client.indices.simulateTemplate({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string): Name of the index template to simulate. To test a template configuration before you add it to the cluster, omit this parameter and specify the template configuration in the request body. -
template
(Optional, { index_patterns, composed_of, template, version, priority, _meta, allow_auto_create, data_stream }) -
create
(Optional, boolean): If true, the template passed in the body is only used if no existing templates match the same index patterns. If false, the simulation uses the template with the highest priority. Note that the template is not permanently added or updated in either case; it is only used for the simulation. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
include_defaults
(Optional, boolean): If true, returns all relevant default configurations for the index template.
-
splitedit
Allows you to split an existing index into a new index with more primary shards.
client.indices.split({ index, target })
Argumentsedit
-
Request (object):
-
index
(string): The name of the source index to split -
target
(string): The name of the target index to split into -
aliases
(Optional, Record<string, { filter, index_routing, is_hidden, is_write_index, routing, search_routing }>) -
settings
(Optional, Record<string, User-defined value>) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_active_shards
(Optional, number | Enum("all" | "index-setting")): Set the number of active shards to wait for on the shrunken index before the operation returns.
-
statsedit
Provides statistics on operations happening in an index.
client.indices.stats({ ... })
Argumentsedit
-
Request (object):
-
metric
(Optional, string | string[]): Limit the information returned the specific metrics. -
index
(Optional, string | string[]): A list of index names; use_all
or empty string to perform the operation on all indices -
completion_fields
(Optional, string | string[]): List or wildcard expressions of fields to include in fielddata and suggest statistics. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard patterns can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. Supports a list of values, such asopen,hidden
. -
fielddata_fields
(Optional, string | string[]): List or wildcard expressions of fields to include in fielddata statistics. -
fields
(Optional, string | string[]): List or wildcard expressions of fields to include in the statistics. -
forbid_closed_indices
(Optional, boolean): If true, statistics are not collected from closed indices. -
groups
(Optional, string | string[]): List of search groups to include in the search statistics. -
include_segment_file_sizes
(Optional, boolean): If true, the call reports the aggregated disk usage of each one of the Lucene index files (only applies if segment stats are requested). -
include_unloaded_segments
(Optional, boolean): If true, the response includes information from segments that are not loaded into memory. -
level
(Optional, Enum("cluster" | "indices" | "shards")): Indicates whether statistics are aggregated at the cluster, index, or shard level.
-
unfreezeedit
Unfreezes an index. When a frozen index is unfrozen, the index goes through the normal recovery process and becomes writeable again.
client.indices.unfreeze({ index })
Argumentsedit
-
Request (object):
-
index
(string): The name of the index to unfreeze -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_active_shards
(Optional, string): Sets the number of active shards to wait for before the operation returns.
-
update_aliasesedit
Updates index aliases.
client.indices.updateAliases({ ... })
Argumentsedit
-
Request (object):
-
actions
(Optional, { add_backing_index, remove_backing_index }[]) -
master_timeout
(Optional, string | -1 | 0): Specify timeout for connection to master -
timeout
(Optional, string | -1 | 0): Request timeout
-
validate_queryedit
Allows a user to validate a potentially expensive query without executing it.
client.indices.validateQuery({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names to restrict the operation; use_all
or empty string to perform the operation on all indices -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
all_shards
(Optional, boolean): Execute validation on all shards instead of one random shard per index -
analyzer
(Optional, string): The analyzer to use for the query string -
analyze_wildcard
(Optional, boolean): Specify whether wildcard and prefix queries should be analyzed (default: false) -
default_operator
(Optional, Enum("and" | "or")): The default operator for query string query (AND or OR) -
df
(Optional, string): The field to use as default where no field prefix is given in the query string -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
explain
(Optional, boolean): Return detailed information about the error -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
lenient
(Optional, boolean): Specify whether format-based query failures (such as providing text to a numeric field) should be ignored -
rewrite
(Optional, boolean): Provide a more detailed explanation showing the actual Lucene query that will be executed. -
q
(Optional, string): Query in the Lucene query string syntax
-
ingestedit
delete_pipelineedit
Deletes a pipeline.
client.ingest.deletePipeline({ id })
Argumentsedit
-
Request (object):
-
id
(string): Pipeline ID -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
geo_ip_statsedit
Returns statistical information about geoip databases
client.ingest.geoIpStats()
get_pipelineedit
Returns a pipeline.
client.ingest.getPipeline({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): Comma separated list of pipeline ids. Wildcards supported -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
summary
(Optional, boolean): Return pipelines without their definitions (default: false)
-
processor_grokedit
Returns a list of the built-in patterns.
client.ingest.processorGrok()
put_pipelineedit
Creates or updates a pipeline.
client.ingest.putPipeline({ id })
Argumentsedit
-
Request (object):
-
id
(string): ID of the ingest pipeline to create or update. -
_meta
(Optional, Record<string, User-defined value>): Optional metadata about the ingest pipeline. May have any contents. This map is not automatically generated by Elasticsearch. -
description
(Optional, string): Description of the ingest pipeline. -
on_failure
(Optional, { attachment, append, csv, convert, date, date_index_name, dot_expander, enrich, fail, foreach, json, user_agent, kv, geoip, grok, gsub, join, lowercase, remove, rename, script, set, sort, split, trim, uppercase, urldecode, bytes, dissect, set_security_user, pipeline, drop, circle, inference }[]): Processors to run immediately after a processor failure. Each processor supports a processor-levelon_failure
value. If a processor without anon_failure
value fails, Elasticsearch uses this pipeline-level parameter as a fallback. The processors in this parameter run sequentially in the order specified. Elasticsearch will not attempt to run the pipeline’s remaining processors. -
processors
(Optional, { attachment, append, csv, convert, date, date_index_name, dot_expander, enrich, fail, foreach, json, user_agent, kv, geoip, grok, gsub, join, lowercase, remove, rename, script, set, sort, split, trim, uppercase, urldecode, bytes, dissect, set_security_user, pipeline, drop, circle, inference }[]): Processors used to perform transformations on documents before indexing. Processors run sequentially in the order specified. -
version
(Optional, number): Version number used by external systems to track ingest pipelines. This parameter is intended for external systems only. Elasticsearch does not use or validate pipeline version numbers. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error. -
if_version
(Optional, number): Required version for optimistic concurrency control for pipeline updates
-
simulateedit
Allows to simulate a pipeline with example documents.
client.ingest.simulate({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): Pipeline ID -
docs
(Optional, { _id, _index, _source }[]) -
pipeline
(Optional, { description, on_failure, processors, version }) -
verbose
(Optional, boolean): Verbose mode. Display data output for each processor in executed pipeline
-
licenseedit
deleteedit
Deletes licensing information for the cluster
client.license.delete()
getedit
Retrieves licensing information for the cluster
client.license.get({ ... })
Argumentsedit
-
Request (object):
-
accept_enterprise
(Optional, boolean): Iftrue
, this parameter returns enterprise for Enterprise license types. Iffalse
, this parameter returns platinum for both platinum and enterprise license types. This behavior is maintained for backwards compatibility. This parameter is deprecated and will always be set to true in 8.x. -
local
(Optional, boolean): Specifies whether to retrieve local information. The default value isfalse
, which means the information is retrieved from the master node.
-
get_basic_statusedit
Retrieves information about the status of the basic license.
client.license.getBasicStatus()
get_trial_statusedit
Retrieves information about the status of the trial license.
client.license.getTrialStatus()
postedit
Updates the license for the cluster.
client.license.post({ ... })
Argumentsedit
-
Request (object):
-
license
(Optional, { expiry_date_in_millis, issue_date_in_millis, start_date_in_millis, issued_to, issuer, max_nodes, max_resource_units, signature, type, uid }) -
licenses
(Optional, { expiry_date_in_millis, issue_date_in_millis, start_date_in_millis, issued_to, issuer, max_nodes, max_resource_units, signature, type, uid }[]): A sequence of one or more JSON documents containing the license information. -
acknowledge
(Optional, boolean): Specifies whether you acknowledge the license changes.
-
post_start_basicedit
Starts an indefinite basic license.
client.license.postStartBasic({ ... })
Argumentsedit
-
Request (object):
-
acknowledge
(Optional, boolean): whether the user has acknowledged acknowledge messages (default: false)
-
post_start_trialedit
starts a limited time trial license.
client.license.postStartTrial({ ... })
Argumentsedit
-
Request (object):
-
acknowledge
(Optional, boolean): whether the user has acknowledged acknowledge messages (default: false) -
type_query_string
(Optional, string)
-
logstashedit
delete_pipelineedit
Deletes Logstash Pipelines used by Central Management
client.logstash.deletePipeline({ id })
Argumentsedit
-
Request (object):
-
id
(string): The ID of the Pipeline
-
get_pipelineedit
Retrieves Logstash Pipelines used by Central Management
client.logstash.getPipeline({ id })
Argumentsedit
-
Request (object):
-
id
(string | string[]): A list of Pipeline IDs
-
put_pipelineedit
Adds and updates Logstash Pipelines used for Central Management
client.logstash.putPipeline({ id })
Argumentsedit
-
Request (object):
-
id
(string): The ID of the Pipeline -
pipeline
(Optional, { description, on_failure, processors, version })
-
migrationedit
deprecationsedit
Retrieves information about different cluster, node, and index level settings that use deprecated features that will be removed or changed in the next major version.
client.migration.deprecations({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string): Comma-separate list of data streams or indices to check. Wildcard (*) expressions are supported.
-
get_feature_upgrade_statusedit
Find out whether system features need to be upgraded or not
client.migration.getFeatureUpgradeStatus()
post_feature_upgradeedit
Begin upgrades for system features
client.migration.postFeatureUpgrade()
mledit
clear_trained_model_deployment_cacheedit
Clear the cached results from a trained model deployment
client.ml.clearTrainedModelDeploymentCache({ model_id })
Argumentsedit
-
Request (object):
-
model_id
(string): The unique identifier of the trained model.
-
close_jobedit
Closes one or more anomaly detection jobs. A job can be opened and closed multiple times throughout its lifecycle.
client.ml.closeJob({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. It can be a job identifier, a group name, or a wildcard expression. You can close multiple anomaly detection jobs in a single API request by using a group name, a list of jobs, or a wildcard expression. You can close all jobs by using_all
or by specifying*
as the job identifier. -
allow_no_match
(Optional, boolean): Refer to the description for theallow_no_match
query parameter. -
force
(Optional, boolean): Refer to the descriptiion for theforce
query parameter. -
timeout
(Optional, string | -1 | 0): Refer to the description for thetimeout
query parameter.
-
delete_calendaredit
Deletes a calendar.
client.ml.deleteCalendar({ calendar_id })
Argumentsedit
-
Request (object):
-
calendar_id
(string): A string that uniquely identifies a calendar.
-
delete_calendar_eventedit
Deletes scheduled events from a calendar.
client.ml.deleteCalendarEvent({ calendar_id, event_id })
Argumentsedit
-
Request (object):
-
calendar_id
(string): The ID of the calendar to modify -
event_id
(string): The ID of the event to remove from the calendar
-
delete_calendar_jobedit
Deletes anomaly detection jobs from a calendar.
client.ml.deleteCalendarJob({ calendar_id, job_id })
Argumentsedit
-
Request (object):
-
calendar_id
(string): A string that uniquely identifies a calendar. -
job_id
(string | string[]): An identifier for the anomaly detection jobs. It can be a job identifier, a group name, or a list of jobs or groups.
-
delete_data_frame_analyticsedit
Deletes an existing data frame analytics job.
client.ml.deleteDataFrameAnalytics({ id })
Argumentsedit
-
Request (object):
-
id
(string): Identifier for the data frame analytics job. -
force
(Optional, boolean): Iftrue
, it deletes a job that is not stopped; this method is quicker than stopping and deleting the job. -
timeout
(Optional, string | -1 | 0): The time to wait for the job to be deleted.
-
delete_datafeededit
Deletes an existing datafeed.
client.ml.deleteDatafeed({ datafeed_id })
Argumentsedit
-
Request (object):
-
datafeed_id
(string): A numerical character string that uniquely identifies the datafeed. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
force
(Optional, boolean): Use to forcefully delete a started datafeed; this method is quicker than stopping and deleting the datafeed.
-
delete_expired_dataedit
Deletes expired and unused machine learning data.
client.ml.deleteExpiredData({ ... })
Argumentsedit
-
Request (object):
-
job_id
(Optional, string): Identifier for an anomaly detection job. It can be a job identifier, a group name, or a wildcard expression. -
requests_per_second
(Optional, float): The desired requests per second for the deletion processes. The default behavior is no throttling. -
timeout
(Optional, string | -1 | 0): How long can the underlying delete processes run until they are canceled.
-
delete_filteredit
Deletes a filter.
client.ml.deleteFilter({ filter_id })
Argumentsedit
-
Request (object):
-
filter_id
(string): A string that uniquely identifies a filter.
-
delete_forecastedit
Deletes forecasts from a machine learning job.
client.ml.deleteForecast({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
forecast_id
(Optional, string): A list of forecast identifiers. If you do not specify this optional parameter or if you specify_all
or*
the API deletes all forecasts from the job. -
allow_no_forecasts
(Optional, boolean): Specifies whether an error occurs when there are no forecasts. In particular, if this parameter is set tofalse
and there are no forecasts associated with the job, attempts to delete all forecasts return an error. -
timeout
(Optional, string | -1 | 0): Specifies the period of time to wait for the completion of the delete operation. When this period of time elapses, the API fails and returns an error.
-
delete_jobedit
Deletes an existing anomaly detection job.
client.ml.deleteJob({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
force
(Optional, boolean): Use to forcefully delete an opened job; this method is quicker than closing and deleting the job. -
delete_user_annotations
(Optional, boolean): Specifies whether annotations that have been added by the user should be deleted along with any auto-generated annotations when the job is reset. -
wait_for_completion
(Optional, boolean): Specifies whether the request should return immediately or wait until the job deletion completes.
-
delete_model_snapshotedit
Deletes an existing model snapshot.
client.ml.deleteModelSnapshot({ job_id, snapshot_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
snapshot_id
(string): Identifier for the model snapshot.
-
delete_trained_modeledit
Deletes an existing trained inference model that is currently not referenced by an ingest pipeline.
client.ml.deleteTrainedModel({ model_id })
Argumentsedit
-
Request (object):
-
model_id
(string): The unique identifier of the trained model. -
force
(Optional, boolean): Forcefully deletes a trained model that is referenced by ingest pipelines or has a started deployment.
-
delete_trained_model_aliasedit
Deletes a model alias that refers to the trained model
client.ml.deleteTrainedModelAlias({ model_alias, model_id })
Argumentsedit
-
Request (object):
-
model_alias
(string): The model alias to delete. -
model_id
(string): The trained model ID to which the model alias refers.
-
estimate_model_memoryedit
Estimates the model memory
client.ml.estimateModelMemory({ ... })
Argumentsedit
-
Request (object):
-
analysis_config
(Optional, { bucket_span, categorization_analyzer, categorization_field_name, categorization_filters, detectors, influencers, latency, model_prune_window, multivariate_by_fields, per_partition_categorization, summary_count_field_name }): For a list of the properties that you can specify in theanalysis_config
component of the body of this API. -
max_bucket_cardinality
(Optional, Record<string, number>): Estimates of the highest cardinality in a single bucket that is observed for influencer fields over the time period that the job analyzes data. To produce a good answer, values must be provided for all influencer fields. Providing values for fields that are not listed asinfluencers
has no effect on the estimation. -
overall_cardinality
(Optional, Record<string, number>): Estimates of the cardinality that is observed for fields over the whole time period that the job analyzes data. To produce a good answer, values must be provided for fields referenced in theby_field_name
,over_field_name
andpartition_field_name
of any detectors. Providing values for other fields has no effect on the estimation. It can be omitted from the request if no detectors have aby_field_name
,over_field_name
orpartition_field_name
.
-
evaluate_data_frameedit
Evaluates the data frame analytics for an annotated index.
client.ml.evaluateDataFrame({ evaluation, index })
Argumentsedit
-
Request (object):
-
evaluation
({ classification, outlier_detection, regression }): Defines the type of evaluation you want to perform. -
index
(string): Defines theindex
in which the evaluation will be performed. -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): A query clause that retrieves a subset of data from the source index.
-
explain_data_frame_analyticsedit
Explains a data frame analytics config.
client.ml.explainDataFrameAnalytics({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): Identifier for the data frame analytics job. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
source
(Optional, { index, query, runtime_mappings, _source }): The configuration of how to source the analysis data. It requires an index. Optionally, query and _source may be specified. -
dest
(Optional, { index, results_field }): The destination configuration, consisting of index and optionally results_field (ml by default). -
analysis
(Optional, { classification, outlier_detection, regression }): The analysis configuration, which contains the information necessary to perform one of the following types of analysis: classification, outlier detection, or regression. -
description
(Optional, string): A description of the job. -
model_memory_limit
(Optional, string): The approximate maximum amount of memory resources that are permitted for analytical processing. If yourelasticsearch.yml
file contains anxpack.ml.max_model_memory_limit
setting, an error occurs when you try to create data frame analytics jobs that havemodel_memory_limit
values greater than that setting. -
max_num_threads
(Optional, number): The maximum number of threads to be used by the analysis. Using more threads may decrease the time necessary to complete the analysis at the cost of using more CPU. Note that the process may use additional threads for operational functionality other than the analysis itself. -
analyzed_fields
(Optional, { includes, excludes }): Specify includes and/or excludes patterns to select which fields will be included in the analysis. The patterns specified in excludes are applied last, therefore excludes takes precedence. In other words, if the same field is specified in both includes and excludes, then the field will not be included in the analysis. -
allow_lazy_start
(Optional, boolean): Specifies whether this job can start when there is insufficient machine learning node capacity for it to be immediately assigned to a node.
-
flush_jobedit
Forces any buffered data to be processed by the job.
client.ml.flushJob({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
advance_time
(Optional, string | Unit): Refer to the description for theadvance_time
query parameter. -
calc_interim
(Optional, boolean): Refer to the description for thecalc_interim
query parameter. -
end
(Optional, string | Unit): Refer to the description for theend
query parameter. -
skip_time
(Optional, string | Unit): Refer to the description for theskip_time
query parameter. -
start
(Optional, string | Unit): Refer to the description for thestart
query parameter.
-
forecastedit
Predicts the future behavior of a time series by using its historical behavior.
client.ml.forecast({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. The job must be open when you create a forecast; otherwise, an error occurs. -
duration
(Optional, string | -1 | 0): Refer to the description for theduration
query parameter. -
expires_in
(Optional, string | -1 | 0): Refer to the description for theexpires_in
query parameter. -
max_model_memory
(Optional, string): Refer to the description for themax_model_memory
query parameter.
-
get_bucketsedit
Retrieves anomaly detection job results for one or more buckets.
client.ml.getBuckets({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
timestamp
(Optional, string | Unit): The timestamp of a single bucket result. If you do not specify this parameter, the API returns information about all buckets. -
anomaly_score
(Optional, number): Refer to the description for theanomaly_score
query parameter. -
desc
(Optional, boolean): Refer to the description for thedesc
query parameter. -
end
(Optional, string | Unit): Refer to the description for theend
query parameter. -
exclude_interim
(Optional, boolean): Refer to the description for theexclude_interim
query parameter. -
expand
(Optional, boolean): Refer to the description for theexpand
query parameter. -
page
(Optional, { from, size }) -
sort
(Optional, string): Refer to the desription for thesort
query parameter. -
start
(Optional, string | Unit): Refer to the description for thestart
query parameter. -
from
(Optional, number): Skips the specified number of buckets. -
size
(Optional, number): Specifies the maximum number of buckets to obtain.
-
get_calendar_eventsedit
Retrieves information about the scheduled events in calendars.
client.ml.getCalendarEvents({ calendar_id })
Argumentsedit
-
Request (object):
-
calendar_id
(string): A string that uniquely identifies a calendar. You can get information for multiple calendars by using a list of ids or a wildcard expression. You can get information for all calendars by using_all
or*
or by omitting the calendar identifier. -
end
(Optional, string | Unit): Specifies to get events with timestamps earlier than this time. -
from
(Optional, number): Skips the specified number of events. -
job_id
(Optional, string): Specifies to get events for a specific anomaly detection job identifier or job group. It must be used with a calendar identifier of_all
or*
. -
size
(Optional, number): Specifies the maximum number of events to obtain. -
start
(Optional, string | Unit): Specifies to get events with timestamps after this time.
-
get_calendarsedit
Retrieves configuration information for calendars.
client.ml.getCalendars({ ... })
Argumentsedit
-
Request (object):
-
calendar_id
(Optional, string): A string that uniquely identifies a calendar. You can get information for multiple calendars by using a list of ids or a wildcard expression. You can get information for all calendars by using_all
or*
or by omitting the calendar identifier. -
page
(Optional, { from, size }): This object is supported only when you omit the calendar identifier. -
from
(Optional, number): Skips the specified number of calendars. This parameter is supported only when you omit the calendar identifier. -
size
(Optional, number): Specifies the maximum number of calendars to obtain. This parameter is supported only when you omit the calendar identifier.
-
get_categoriesedit
Retrieves anomaly detection job results for one or more categories.
client.ml.getCategories({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
category_id
(Optional, string): Identifier for the category, which is unique in the job. If you specify neither the category ID nor the partition_field_value, the API returns information about all categories. If you specify only the partition_field_value, it returns information about all categories for the specified partition. -
page
(Optional, { from, size }) -
from
(Optional, number): Skips the specified number of categories. -
partition_field_value
(Optional, string): Only return categories for the specified partition. -
size
(Optional, number): Specifies the maximum number of categories to obtain.
-
get_data_frame_analyticsedit
Retrieves configuration information for data frame analytics jobs.
client.ml.getDataFrameAnalytics({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): Identifier for the data frame analytics job. If you do not specify this option, the API returns information for the first hundred data frame analytics jobs. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no data frame analytics jobs that match.
-
Contains the
_all
string or no identifiers and there are no matches. - Contains wildcard expressions and there are only partial matches.
-
The default value returns an empty data_frame_analytics array when there
are no matches and the subset of results when there are partial matches.
If this parameter is false
, the request returns a 404 status code when
there are no matches or only partial matches.
from
(Optional, number): Skips the specified number of data frame analytics jobs.
size
(Optional, number): Specifies the maximum number of data frame analytics jobs to obtain.
* *exclude_generated
(Optional, boolean): Indicates if certain fields should be removed from the configuration on
retrieval. This allows the configuration to be in an acceptable format to
be retrieved and then added to another cluster.
get_data_frame_analytics_statsedit
Retrieves usage information for data frame analytics jobs.
client.ml.getDataFrameAnalyticsStats({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): Identifier for the data frame analytics job. If you do not specify this option, the API returns information for the first hundred data frame analytics jobs. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no data frame analytics jobs that match.
-
Contains the
_all
string or no identifiers and there are no matches. - Contains wildcard expressions and there are only partial matches.
-
The default value returns an empty data_frame_analytics array when there
are no matches and the subset of results when there are partial matches.
If this parameter is false
, the request returns a 404 status code when
there are no matches or only partial matches.
from
(Optional, number): Skips the specified number of data frame analytics jobs.
size
(Optional, number): Specifies the maximum number of data frame analytics jobs to obtain.
* *verbose
(Optional, boolean): Defines whether the stats response should be verbose.
get_datafeed_statsedit
Retrieves usage information for datafeeds.
client.ml.getDatafeedStats({ ... })
Argumentsedit
-
Request (object):
-
datafeed_id
(Optional, string | string[]): Identifier for the datafeed. It can be a datafeed identifier or a wildcard expression. If you do not specify one of these options, the API returns information about all datafeeds. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no datafeeds that match.
-
Contains the
_all
string or no identifiers and there are no matches. - Contains wildcard expressions and there are only partial matches.
-
The default value is true
, which returns an empty datafeeds
array
when there are no matches and the subset of results when there are
partial matches. If this parameter is false
, the request returns a
404
status code when there are no matches or only partial matches.
get_datafeedsedit
Retrieves configuration information for datafeeds.
client.ml.getDatafeeds({ ... })
Argumentsedit
-
Request (object):
-
datafeed_id
(Optional, string | string[]): Identifier for the datafeed. It can be a datafeed identifier or a wildcard expression. If you do not specify one of these options, the API returns information about all datafeeds. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no datafeeds that match.
-
Contains the
_all
string or no identifiers and there are no matches. - Contains wildcard expressions and there are only partial matches.
-
The default value is true
, which returns an empty datafeeds
array
when there are no matches and the subset of results when there are
partial matches. If this parameter is false
, the request returns a
404
status code when there are no matches or only partial matches.
* *exclude_generated
(Optional, boolean): Indicates if certain fields should be removed from the configuration on
retrieval. This allows the configuration to be in an acceptable format to
be retrieved and then added to another cluster.
get_filtersedit
Retrieves filters.
client.ml.getFilters({ ... })
Argumentsedit
-
Request (object):
-
filter_id
(Optional, string | string[]): A string that uniquely identifies a filter. -
from
(Optional, number): Skips the specified number of filters. -
size
(Optional, number): Specifies the maximum number of filters to obtain.
-
get_influencersedit
Retrieves anomaly detection job results for one or more influencers.
client.ml.getInfluencers({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
page
(Optional, { from, size }) -
desc
(Optional, boolean): If true, the results are sorted in descending order. -
end
(Optional, string | Unit): Returns influencers with timestamps earlier than this time. The default value means it is unset and results are not limited to specific timestamps. -
exclude_interim
(Optional, boolean): If true, the output excludes interim results. By default, interim results are included. -
influencer_score
(Optional, number): Returns influencers with anomaly scores greater than or equal to this value. -
from
(Optional, number): Skips the specified number of influencers. -
size
(Optional, number): Specifies the maximum number of influencers to obtain. -
sort
(Optional, string): Specifies the sort field for the requested influencers. By default, the influencers are sorted by theinfluencer_score
value. -
start
(Optional, string | Unit): Returns influencers with timestamps after this time. The default value means it is unset and results are not limited to specific timestamps.
-
get_job_statsedit
Retrieves usage information for anomaly detection jobs.
client.ml.getJobStats({ ... })
Argumentsedit
-
Request (object):
-
job_id
(Optional, string): Identifier for the anomaly detection job. It can be a job identifier, a group name, a list of jobs, or a wildcard expression. If you do not specify one of these options, the API returns information for all anomaly detection jobs. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no jobs that match.
- Contains the _all string or no identifiers and there are no matches.
- Contains wildcard expressions and there are only partial matches.
-
If true
, the API returns an empty jobs
array when
there are no matches and the subset of results when there are partial
matches. If false
, the API returns a 404
status
code when there are no matches or only partial matches.
get_jobsedit
Retrieves configuration information for anomaly detection jobs.
client.ml.getJobs({ ... })
Argumentsedit
-
Request (object):
-
job_id
(Optional, string | string[]): Identifier for the anomaly detection job. It can be a job identifier, a group name, or a wildcard expression. If you do not specify one of these options, the API returns information for all anomaly detection jobs. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no jobs that match.
- Contains the _all string or no identifiers and there are no matches.
- Contains wildcard expressions and there are only partial matches.
-
The default value is true
, which returns an empty jobs
array when
there are no matches and the subset of results when there are partial
matches. If this parameter is false
, the request returns a 404
status
code when there are no matches or only partial matches.
* *exclude_generated
(Optional, boolean): Indicates if certain fields should be removed from the configuration on
retrieval. This allows the configuration to be in an acceptable format to
be retrieved and then added to another cluster.
get_memory_statsedit
Returns information on how ML is using memory.
client.ml.getMemoryStats({ ... })
Argumentsedit
-
Request (object):
-
node_id
(Optional, string): The names of particular nodes in the cluster to target. For example,nodeId1,nodeId2
orml:true
-
human
(Optional, boolean): Specify this query parameter to include the fields with units in the response. Otherwise only the_in_bytes
sizes are returned in the response. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
get_model_snapshot_upgrade_statsedit
Gets stats for anomaly detection job model snapshot upgrades that are in progress.
client.ml.getModelSnapshotUpgradeStats({ job_id, snapshot_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
snapshot_id
(string): A numerical character string that uniquely identifies the model snapshot. You can get information for multiple snapshots by using a list or a wildcard expression. You can get all snapshots by using_all
, by specifying*
as the snapshot ID, or by omitting the snapshot ID. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no jobs that match.
- Contains the _all string or no identifiers and there are no matches.
- Contains wildcard expressions and there are only partial matches.
-
The default value is true, which returns an empty jobs array when there are no matches and the subset of results when there are partial matches. If this parameter is false, the request returns a 404 status code when there are no matches or only partial matches.
get_model_snapshotsedit
Retrieves information about model snapshots.
client.ml.getModelSnapshots({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
snapshot_id
(Optional, string): A numerical character string that uniquely identifies the model snapshot. You can get information for multiple snapshots by using a list or a wildcard expression. You can get all snapshots by using_all
, by specifying*
as the snapshot ID, or by omitting the snapshot ID. -
desc
(Optional, boolean): Refer to the description for thedesc
query parameter. -
end
(Optional, string | Unit): Refer to the description for theend
query parameter. -
page
(Optional, { from, size }) -
sort
(Optional, string): Refer to the description for thesort
query parameter. -
start
(Optional, string | Unit): Refer to the description for thestart
query parameter. -
from
(Optional, number): Skips the specified number of snapshots. -
size
(Optional, number): Specifies the maximum number of snapshots to obtain.
-
get_overall_bucketsedit
Retrieves overall bucket results that summarize the bucket results of multiple anomaly detection jobs.
client.ml.getOverallBuckets({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. It can be a job identifier, a group name, a list of jobs or groups, or a wildcard expression.
-
You can summarize the bucket results for all anomaly detection jobs by
using _all
or by specifying *
as the <job_id>
.
allow_no_match
(Optional, boolean): Refer to the description for the allow_no_match
query parameter.
bucket_span
(Optional, string | -1 | 0): Refer to the description for the bucket_span
query parameter.
end
(Optional, string | Unit): Refer to the description for the end
query parameter.
exclude_interim
(Optional, boolean): Refer to the description for the exclude_interim
query parameter.
overall_score
(Optional, number | string): Refer to the description for the overall_score
query parameter.
start
(Optional, string | Unit): Refer to the description for the start
query parameter.
* *top_n
(Optional, number): Refer to the description for the top_n
query parameter.
get_recordsedit
Retrieves anomaly records for an anomaly detection job.
client.ml.getRecords({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
desc
(Optional, boolean): Refer to the description for thedesc
query parameter. -
end
(Optional, string | Unit): Refer to the description for theend
query parameter. -
exclude_interim
(Optional, boolean): Refer to the description for theexclude_interim
query parameter. -
page
(Optional, { from, size }) -
record_score
(Optional, number): Refer to the description for therecord_score
query parameter. -
sort
(Optional, string): Refer to the description for thesort
query parameter. -
start
(Optional, string | Unit): Refer to the description for thestart
query parameter. -
from
(Optional, number): Skips the specified number of records. -
size
(Optional, number): Specifies the maximum number of records to obtain.
-
get_trained_modelsedit
Retrieves configuration information for a trained inference model.
client.ml.getTrainedModels({ ... })
Argumentsedit
-
Request (object):
-
model_id
(Optional, string): The unique identifier of the trained model. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no models that match.
- Contains the _all string or no identifiers and there are no matches.
- Contains wildcard expressions and there are only partial matches.
-
If true, it returns an empty array when there are no matches and the
subset of results when there are partial matches.
decompress_definition
(Optional, boolean): Specifies whether the included model definition should be returned as a
JSON map (true) or in a custom compressed format (false).
exclude_generated
(Optional, boolean): Indicates if certain fields should be removed from the configuration on
retrieval. This allows the configuration to be in an acceptable format to
be retrieved and then added to another cluster.
from
(Optional, number): Skips the specified number of models.
include
(Optional, Enum("definition" | "feature_importance_baseline" | "hyperparameters" | "total_feature_importance" | "definition_status")): A comma delimited string of optional fields to include in the response
body.
size
(Optional, number): Specifies the maximum number of models to obtain.
tags
(Optional, string): A comma delimited string of tags. A trained model can have many tags, or
none. When supplied, only trained models that contain all the supplied
tags are returned.
get_trained_models_statsedit
Retrieves usage information for trained inference models.
client.ml.getTrainedModelsStats({ ... })
Argumentsedit
-
Request (object):
-
model_id
(Optional, string | string[]): The unique identifier of the trained model or a model alias. It can be a list or a wildcard expression. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no models that match.
- Contains the _all string or no identifiers and there are no matches.
- Contains wildcard expressions and there are only partial matches.
-
If true, it returns an empty array when there are no matches and the
subset of results when there are partial matches.
from
(Optional, number): Skips the specified number of models.
size
(Optional, number): Specifies the maximum number of models to obtain.
infer_trained_modeledit
Evaluate a trained model.
client.ml.inferTrainedModel({ model_id, docs })
Argumentsedit
-
Request (object):
-
model_id
(string): The unique identifier of the trained model. -
docs
(Record<string, User-defined value>[]): An array of objects to pass to the model for inference. The objects should contain a fields matching your configured trained model input. Typically, for NLP models, the field name istext_field
. Currently, for NLP models, only a single value is allowed. -
inference_config
(Optional, { regression, classification, text_classification, zero_shot_classification, fill_mask, ner, pass_through, text_embedding, text_expansion, question_answering }): The inference configuration updates to apply on the API call -
timeout
(Optional, string | -1 | 0): Controls the amount of time to wait for inference results.
-
infoedit
Returns defaults and limits used by machine learning.
client.ml.info()
open_jobedit
Opens one or more anomaly detection jobs.
client.ml.openJob({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
timeout
(Optional, string | -1 | 0): Refer to the description for thetimeout
query parameter.
-
post_calendar_eventsedit
Posts scheduled events in a calendar.
client.ml.postCalendarEvents({ calendar_id, events })
Argumentsedit
-
Request (object):
-
calendar_id
(string): A string that uniquely identifies a calendar. -
events
({ calendar_id, event_id, description, end_time, start_time }[]): A list of one of more scheduled events. The event’s start and end times can be specified as integer milliseconds since the epoch or as a string in ISO 8601 format.
-
post_dataedit
Sends data to an anomaly detection job for analysis.
client.ml.postData({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. The job must have a state of open to receive and process the data. -
data
(Optional, TData[]) -
reset_end
(Optional, string | Unit): Specifies the end of the bucket resetting range. -
reset_start
(Optional, string | Unit): Specifies the start of the bucket resetting range.
-
preview_data_frame_analyticsedit
Previews that will be analyzed given a data frame analytics config.
client.ml.previewDataFrameAnalytics({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): Identifier for the data frame analytics job. -
config
(Optional, { source, analysis, model_memory_limit, max_num_threads, analyzed_fields }): A data frame analytics config as described in create data frame analytics jobs. Note thatid
anddest
don’t need to be provided in the context of this API.
-
preview_datafeededit
Previews a datafeed.
client.ml.previewDatafeed({ ... })
Argumentsedit
-
Request (object):
-
datafeed_id
(Optional, string): A numerical character string that uniquely identifies the datafeed. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. NOTE: If you use this path parameter, you cannot provide datafeed or anomaly detection job configuration details in the request body. -
datafeed_config
(Optional, { aggregations, chunking_config, datafeed_id, delayed_data_check_config, frequency, indices, indices_options, job_id, max_empty_searches, query, query_delay, runtime_mappings, script_fields, scroll_size }): The datafeed definition to preview. -
job_config
(Optional, { allow_lazy_open, analysis_config, analysis_limits, background_persist_interval, custom_settings, daily_model_snapshot_retention_after_days, data_description, datafeed_config, description, groups, job_id, job_type, model_plot_config, model_snapshot_retention_days, renormalization_window_days, results_index_name, results_retention_days }): The configuration details for the anomaly detection job that is associated with the datafeed. If thedatafeed_config
object does not include ajob_id
that references an existing anomaly detection job, you must supply thisjob_config
object. If you include both ajob_id
and ajob_config
, the latter information is used. You cannot specify ajob_config
object unless you also supply adatafeed_config
object. -
start
(Optional, string | Unit): The start time from where the datafeed preview should begin -
end
(Optional, string | Unit): The end time when the datafeed preview should stop
-
put_calendaredit
Instantiates a calendar.
client.ml.putCalendar({ calendar_id })
Argumentsedit
-
Request (object):
-
calendar_id
(string): A string that uniquely identifies a calendar. -
job_ids
(Optional, string[]): An array of anomaly detection job identifiers. -
description
(Optional, string): A description of the calendar.
-
put_calendar_jobedit
Adds an anomaly detection job to a calendar.
client.ml.putCalendarJob({ calendar_id, job_id })
Argumentsedit
-
Request (object):
-
calendar_id
(string): A string that uniquely identifies a calendar. -
job_id
(string): An identifier for the anomaly detection jobs. It can be a job identifier, a group name, or a list of jobs or groups.
-
put_data_frame_analyticsedit
Instantiates a data frame analytics job.
client.ml.putDataFrameAnalytics({ id, analysis, dest, source })
Argumentsedit
-
Request (object):
-
id
(string): Identifier for the data frame analytics job. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
analysis
({ classification, outlier_detection, regression }): The analysis configuration, which contains the information necessary to perform one of the following types of analysis: classification, outlier detection, or regression. -
dest
({ index, results_field }): The destination configuration. -
source
({ index, query, runtime_mappings, _source }): The configuration of how to source the analysis data. -
allow_lazy_start
(Optional, boolean): Specifies whether this job can start when there is insufficient machine learning node capacity for it to be immediately assigned to a node. If set tofalse
and a machine learning node with capacity to run the job cannot be immediately found, the API returns an error. If set totrue
, the API does not return an error; the job waits in thestarting
state until sufficient machine learning node capacity is available. This behavior is also affected by the cluster-widexpack.ml.max_lazy_ml_nodes
setting. -
analyzed_fields
(Optional, { includes, excludes }): Specifiesincludes
and/orexcludes
patterns to select which fields will be included in the analysis. The patterns specified inexcludes
are applied last, thereforeexcludes
takes precedence. In other words, if the same field is specified in bothincludes
andexcludes
, then the field will not be included in the analysis. Ifanalyzed_fields
is not set, only the relevant fields will be included. For example, all the numeric fields for outlier detection. The supported fields vary for each type of analysis. Outlier detection requires numeric orboolean
data to analyze. The algorithms don’t support missing values therefore fields that have data types other than numeric or boolean are ignored. Documents where included fields contain missing values, null values, or an array are also ignored. Therefore thedest
index may contain documents that don’t have an outlier score. Regression supports fields that are numeric,boolean
,text
,keyword
, andip
data types. It is also tolerant of missing values. Fields that are supported are included in the analysis, other fields are ignored. Documents where included fields contain an array with two or more values are also ignored. Documents in thedest
index that don’t contain a results field are not included in the regression analysis. Classification supports fields that are numeric,boolean
,text
,keyword
, andip
data types. It is also tolerant of missing values. Fields that are supported are included in the analysis, other fields are ignored. Documents where included fields contain an array with two or more values are also ignored. Documents in thedest
index that don’t contain a results field are not included in the classification analysis. Classification analysis can be improved by mapping ordinal variable values to a single number. For example, in case of age ranges, you can model the values as0-14 = 0
,15-24 = 1
,25-34 = 2
, and so on. -
description
(Optional, string): A description of the job. -
max_num_threads
(Optional, number): The maximum number of threads to be used by the analysis. Using more threads may decrease the time necessary to complete the analysis at the cost of using more CPU. Note that the process may use additional threads for operational functionality other than the analysis itself. -
model_memory_limit
(Optional, string): The approximate maximum amount of memory resources that are permitted for analytical processing. If yourelasticsearch.yml
file contains anxpack.ml.max_model_memory_limit
setting, an error occurs when you try to create data frame analytics jobs that havemodel_memory_limit
values greater than that setting. -
headers
(Optional, Record<string, string | string[]>) -
version
(Optional, string)
-
put_datafeededit
Instantiates a datafeed.
client.ml.putDatafeed({ datafeed_id })
Argumentsedit
-
Request (object):
-
datafeed_id
(string): A numerical character string that uniquely identifies the datafeed. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
aggregations
(Optional, Record<string, { aggregations, meta, adjacency_matrix, auto_date_histogram, avg, avg_bucket, boxplot, bucket_script, bucket_selector, bucket_sort, bucket_count_ks_test, bucket_correlation, cardinality, categorize_text, children, composite, cumulative_cardinality, cumulative_sum, date_histogram, date_range, derivative, diversified_sampler, extended_stats, extended_stats_bucket, frequent_item_sets, filter, filters, geo_bounds, geo_centroid, geo_distance, geohash_grid, geo_line, geotile_grid, geohex_grid, global, histogram, ip_range, ip_prefix, inference, line, matrix_stats, max, max_bucket, median_absolute_deviation, min, min_bucket, missing, moving_avg, moving_percentiles, moving_fn, multi_terms, nested, normalize, parent, percentile_ranks, percentiles, percentiles_bucket, range, rare_terms, rate, reverse_nested, sampler, scripted_metric, serial_diff, significant_terms, significant_text, stats, stats_bucket, string_stats, sum, sum_bucket, terms, top_hits, t_test, top_metrics, value_count, weighted_avg, variable_width_histogram }>): If set, the datafeed performs aggregation searches. Support for aggregations is limited and should be used only with low cardinality data. -
chunking_config
(Optional, { mode, time_span }): Datafeeds might be required to search over long time periods, for several months or years. This search is split into time chunks in order to ensure the load on Elasticsearch is managed. Chunking configuration controls how the size of these time chunks are calculated; it is an advanced configuration option. -
delayed_data_check_config
(Optional, { check_window, enabled }): Specifies whether the datafeed checks for missing data and the size of the window. The datafeed can optionally search over indices that have already been read in an effort to determine whether any data has subsequently been added to the index. If missing data is found, it is a good indication that thequery_delay
is set too low and the data is being indexed after the datafeed has passed that moment in time. This check runs only on real-time datafeeds. -
frequency
(Optional, string | -1 | 0): The interval at which scheduled queries are made while the datafeed runs in real time. The default value is either the bucket span for short bucket spans, or, for longer bucket spans, a sensible fraction of the bucket span. Whenfrequency
is shorter than the bucket span, interim results for the last (partial) bucket are written then eventually overwritten by the full bucket results. If the datafeed uses aggregations, this value must be divisible by the interval of the date histogram aggregation. -
indices
(Optional, string | string[]): An array of index names. Wildcards are supported. If any of the indices are in remote clusters, the machine learning nodes must have theremote_cluster_client
role. -
indices_options
(Optional, { allow_no_indices, expand_wildcards, ignore_unavailable, ignore_throttled }): Specifies index expansion options that are used during search -
job_id
(Optional, string): Identifier for the anomaly detection job. -
max_empty_searches
(Optional, number): If a real-time datafeed has never seen any data (including during any initial training period), it automatically stops and closes the associated job after this many real-time searches return no documents. In other words, it stops afterfrequency
timesmax_empty_searches
of real-time operation. If not set, a datafeed with no end time that sees no data remains started until it is explicitly stopped. By default, it is not set. -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): The Elasticsearch query domain-specific language (DSL). This value corresponds to the query object in an Elasticsearch search POST body. All the options that are supported by Elasticsearch can be used, as this object is passed verbatim to Elasticsearch. -
query_delay
(Optional, string | -1 | 0): The number of seconds behind real time that data is queried. For example, if data from 10:04 a.m. might not be searchable in Elasticsearch until 10:06 a.m., set this property to 120 seconds. The default value is randomly selected between60s
and120s
. This randomness improves the query performance when there are multiple jobs running on the same node. -
runtime_mappings
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>): Specifies runtime fields for the datafeed search. -
script_fields
(Optional, Record<string, { script, ignore_failure }>): Specifies scripts that evaluate custom expressions and returns script fields to the datafeed. The detector configuration objects in a job can contain functions that use these script fields. -
scroll_size
(Optional, number): The size parameter that is used in Elasticsearch searches when the datafeed does not use aggregations. The maximum value is the value ofindex.max_result_window
, which is 10,000 by default. -
headers
(Optional, Record<string, string | string[]>) -
allow_no_indices
(Optional, boolean): If true, wildcard indices expressions that resolve into no concrete indices are ignored. This includes the_all
string or when no indices are specified. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard patterns can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. Supports a list of values. -
ignore_throttled
(Optional, boolean): If true, concrete, expanded, or aliased indices are ignored when frozen. -
ignore_unavailable
(Optional, boolean): If true, unavailable indices (missing or closed) are ignored.
-
put_filteredit
Instantiates a filter.
client.ml.putFilter({ filter_id })
Argumentsedit
-
Request (object):
-
filter_id
(string): A string that uniquely identifies a filter. -
description
(Optional, string): A description of the filter. -
items
(Optional, string[]): The items of the filter. A wildcard*
can be used at the beginning or the end of an item. Up to 10000 items are allowed in each filter.
-
put_jobedit
Instantiates an anomaly detection job.
client.ml.putJob({ job_id, analysis_config, data_description })
Argumentsedit
-
Request (object):
-
job_id
(string): The identifier for the anomaly detection job. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
analysis_config
({ bucket_span, categorization_analyzer, categorization_field_name, categorization_filters, detectors, influencers, latency, model_prune_window, multivariate_by_fields, per_partition_categorization, summary_count_field_name }): Specifies how to analyze the data. After you create a job, you cannot change the analysis configuration; all the properties are informational. -
data_description
({ format, time_field, time_format, field_delimiter }): Defines the format of the input data when you send data to the job by using the post data API. Note that when configure a datafeed, these properties are automatically set. When data is received via the post data API, it is not stored in Elasticsearch. Only the results for anomaly detection are retained. -
allow_lazy_open
(Optional, boolean): Advanced configuration option. Specifies whether this job can open when there is insufficient machine learning node capacity for it to be immediately assigned to a node. By default, if a machine learning node with capacity to run the job cannot immediately be found, the open anomaly detection jobs API returns an error. However, this is also subject to the cluster-widexpack.ml.max_lazy_ml_nodes
setting. If this option is set to true, the open anomaly detection jobs API does not return an error and the job waits in the opening state until sufficient machine learning node capacity is available. -
analysis_limits
(Optional, { categorization_examples_limit, model_memory_limit }): Limits can be applied for the resources required to hold the mathematical models in memory. These limits are approximate and can be set per job. They do not control the memory used by other processes, for example the Elasticsearch Java processes. -
background_persist_interval
(Optional, string | -1 | 0): Advanced configuration option. The time between each periodic persistence of the model. The default value is a randomized value between 3 to 4 hours, which avoids all jobs persisting at exactly the same time. The smallest allowed value is 1 hour. For very large models (several GB), persistence could take 10-20 minutes, so do not set thebackground_persist_interval
value too low. -
custom_settings
(Optional, User-defined value): Advanced configuration option. Contains custom meta data about the job. -
daily_model_snapshot_retention_after_days
(Optional, number): Advanced configuration option, which affects the automatic removal of old model snapshots for this job. It specifies a period of time (in days) after which only the first snapshot per day is retained. This period is relative to the timestamp of the most recent snapshot for this job. Valid values range from 0 tomodel_snapshot_retention_days
. -
datafeed_config
(Optional, { aggregations, chunking_config, datafeed_id, delayed_data_check_config, frequency, indices, indices_options, job_id, max_empty_searches, query, query_delay, runtime_mappings, script_fields, scroll_size }): Defines a datafeed for the anomaly detection job. If Elasticsearch security features are enabled, your datafeed remembers which roles the user who created it had at the time of creation and runs the query using those same roles. If you provide secondary authorization headers, those credentials are used instead. -
description
(Optional, string): A description of the job. -
groups
(Optional, string[]): A list of job groups. A job can belong to no groups or many. -
model_plot_config
(Optional, { annotations_enabled, enabled, terms }): This advanced configuration option stores model information along with the results. It provides a more detailed view into anomaly detection. If you enable model plot it can add considerable overhead to the performance of the system; it is not feasible for jobs with many entities. Model plot provides a simplified and indicative view of the model and its bounds. It does not display complex features such as multivariate correlations or multimodal data. As such, anomalies may occasionally be reported which cannot be seen in the model plot. Model plot config can be configured when the job is created or updated later. It must be disabled if performance issues are experienced. -
model_snapshot_retention_days
(Optional, number): Advanced configuration option, which affects the automatic removal of old model snapshots for this job. It specifies the maximum period of time (in days) that snapshots are retained. This period is relative to the timestamp of the most recent snapshot for this job. By default, snapshots ten days older than the newest snapshot are deleted. -
renormalization_window_days
(Optional, number): Advanced configuration option. The period over which adjustments to the score are applied, as new data is seen. The default value is the longer of 30 days or 100 bucket spans. -
results_index_name
(Optional, string): A text string that affects the name of the machine learning results index. By default, the job generates an index named.ml-anomalies-shared
. -
results_retention_days
(Optional, number): Advanced configuration option. The period of time (in days) that results are retained. Age is calculated relative to the timestamp of the latest bucket result. If this property has a non-null value, once per day at 00:30 (server time), results that are the specified number of days older than the latest bucket result are deleted from Elasticsearch. The default value is null, which means all results are retained. Annotations generated by the system also count as results for retention purposes; they are deleted after the same number of days as results. Annotations added by users are retained forever.
-
put_trained_modeledit
Creates an inference trained model.
client.ml.putTrainedModel({ model_id })
Argumentsedit
-
Request (object):
-
model_id
(string): The unique identifier of the trained model. -
compressed_definition
(Optional, string): The compressed (GZipped and Base64 encoded) inference definition of the model. If compressed_definition is specified, then definition cannot be specified. -
definition
(Optional, { preprocessors, trained_model }): The inference definition for the model. If definition is specified, then compressed_definition cannot be specified. -
description
(Optional, string): A human-readable description of the inference trained model. -
inference_config
(Optional, { regression, classification, text_classification, zero_shot_classification, fill_mask, ner, pass_through, text_embedding, text_expansion, question_answering }): The default configuration for inference. This can be either a regression or classification configuration. It must match the underlying definition.trained_model’s target_type. For pre-packaged models such as ELSER the config is not required. -
input
(Optional, { field_names }): The input field names for the model definition. -
metadata
(Optional, User-defined value): An object map that contains metadata about the model. -
model_type
(Optional, Enum("tree_ensemble" | "lang_ident" | "pytorch")): The model type. -
model_size_bytes
(Optional, number): The estimated memory usage in bytes to keep the trained model in memory. This property is supported only if defer_definition_decompression is true or the model definition is not supplied. -
tags
(Optional, string[]): An array of tags to organize the model. -
defer_definition_decompression
(Optional, boolean): If set totrue
and acompressed_definition
is provided, the request defers definition decompression and skips relevant validations.
-
put_trained_model_aliasedit
Creates a new model alias (or reassigns an existing one) to refer to the trained model
client.ml.putTrainedModelAlias({ model_alias, model_id })
Argumentsedit
-
Request (object):
-
model_alias
(string): The alias to create or update. This value cannot end in numbers. -
model_id
(string): The identifier for the trained model that the alias refers to. -
reassign
(Optional, boolean): Specifies whether the alias gets reassigned to the specified trained model if it is already assigned to a different model. If the alias is already assigned and this parameter is false, the API returns an error.
-
put_trained_model_definition_partedit
Creates part of a trained model definition
client.ml.putTrainedModelDefinitionPart({ model_id, part, definition, total_definition_length, total_parts })
Argumentsedit
-
Request (object):
-
model_id
(string): The unique identifier of the trained model. -
part
(number): The definition part number. When the definition is loaded for inference the definition parts are streamed in the order of their part number. The first part must be0
and the final part must betotal_parts - 1
. -
definition
(string): The definition part for the model. Must be a base64 encoded string. -
total_definition_length
(number): The total uncompressed definition length in bytes. Not base64 encoded. -
total_parts
(number): The total number of parts that will be uploaded. Must be greater than 0.
-
put_trained_model_vocabularyedit
Creates a trained model vocabulary
client.ml.putTrainedModelVocabulary({ model_id, vocabulary })
Argumentsedit
-
Request (object):
-
model_id
(string): The unique identifier of the trained model. -
vocabulary
(string[]): The model vocabulary, which must not be empty. -
merges
(Optional, string[]): The optional model merges if required by the tokenizer.
-
reset_jobedit
Resets an existing anomaly detection job.
client.ml.resetJob({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): The ID of the job to reset. -
wait_for_completion
(Optional, boolean): Should this request wait until the operation has completed before returning. -
delete_user_annotations
(Optional, boolean): Specifies whether annotations that have been added by the user should be deleted along with any auto-generated annotations when the job is reset.
-
revert_model_snapshotedit
Reverts to a specific snapshot.
client.ml.revertModelSnapshot({ job_id, snapshot_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
snapshot_id
(string): You can specifyempty
as the <snapshot_id>. Reverting to the empty snapshot means the anomaly detection job starts learning a new model from scratch when it is started. -
delete_intervening_results
(Optional, boolean): Refer to the description for thedelete_intervening_results
query parameter.
-
set_upgrade_modeedit
Sets a cluster wide upgrade_mode setting that prepares machine learning indices for an upgrade.
client.ml.setUpgradeMode({ ... })
Argumentsedit
-
Request (object):
-
enabled
(Optional, boolean): Whentrue
, it enablesupgrade_mode
which temporarily halts all job and datafeed tasks and prohibits new job and datafeed tasks from starting. -
timeout
(Optional, string | -1 | 0): The time to wait for the request to be completed.
-
start_data_frame_analyticsedit
Starts a data frame analytics job.
client.ml.startDataFrameAnalytics({ id })
Argumentsedit
-
Request (object):
-
id
(string): Identifier for the data frame analytics job. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
timeout
(Optional, string | -1 | 0): Controls the amount of time to wait until the data frame analytics job starts.
-
start_datafeededit
Starts one or more datafeeds.
client.ml.startDatafeed({ datafeed_id })
Argumentsedit
-
Request (object):
-
datafeed_id
(string): A numerical character string that uniquely identifies the datafeed. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
end
(Optional, string | Unit): Refer to the description for theend
query parameter. -
start
(Optional, string | Unit): Refer to the description for thestart
query parameter. -
timeout
(Optional, string | -1 | 0): Refer to the description for thetimeout
query parameter.
-
start_trained_model_deploymentedit
Start a trained model deployment.
client.ml.startTrainedModelDeployment({ model_id })
Argumentsedit
-
Request (object):
-
model_id
(string): The unique identifier of the trained model. Currently, only PyTorch models are supported. -
cache_size
(Optional, number | string): The inference cache size (in memory outside the JVM heap) per node for the model. The default value is the same size as themodel_size_bytes
. To disable the cache,0b
can be provided. -
number_of_allocations
(Optional, number): The number of model allocations on each node where the model is deployed. All allocations on a node share the same copy of the model in memory but use a separate set of threads to evaluate the model. Increasing this value generally increases the throughput. If this setting is greater than the number of hardware threads it will automatically be changed to a value less than the number of hardware threads. -
priority
(Optional, Enum("normal" | "low")): The deployment priority. -
queue_capacity
(Optional, number): Specifies the number of inference requests that are allowed in the queue. After the number of requests exceeds this value, new requests are rejected with a 429 error. -
threads_per_allocation
(Optional, number): Sets the number of threads used by each model allocation during inference. This generally increases the inference speed. The inference process is a compute-bound process; any number greater than the number of available hardware threads on the machine does not increase the inference speed. If this setting is greater than the number of hardware threads it will automatically be changed to a value less than the number of hardware threads. -
timeout
(Optional, string | -1 | 0): Specifies the amount of time to wait for the model to deploy. -
wait_for
(Optional, Enum("started" | "starting" | "fully_allocated")): Specifies the allocation status to wait for before returning.
-
stop_data_frame_analyticsedit
Stops one or more data frame analytics jobs.
client.ml.stopDataFrameAnalytics({ id })
Argumentsedit
-
Request (object):
-
id
(string): Identifier for the data frame analytics job. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no data frame analytics jobs that match.
- Contains the _all string or no identifiers and there are no matches.
- Contains wildcard expressions and there are only partial matches.
-
The default value is true, which returns an empty data_frame_analytics
array when there are no matches and the subset of results when there are
partial matches. If this parameter is false, the request returns a 404
status code when there are no matches or only partial matches.
force
(Optional, boolean): If true, the data frame analytics job is stopped forcefully.
timeout
(Optional, string | -1 | 0): Controls the amount of time to wait until the data frame analytics job
stops. Defaults to 20 seconds.
stop_datafeededit
Stops one or more datafeeds.
client.ml.stopDatafeed({ datafeed_id })
Argumentsedit
-
Request (object):
-
datafeed_id
(string): Identifier for the datafeed. You can stop multiple datafeeds in a single API request by using a comma-separated list of datafeeds or a wildcard expression. You can close all datafeeds by using_all
or by specifying*
as the identifier. -
allow_no_match
(Optional, boolean): Refer to the description for theallow_no_match
query parameter. -
force
(Optional, boolean): Refer to the description for theforce
query parameter. -
timeout
(Optional, string | -1 | 0): Refer to the description for thetimeout
query parameter.
-
stop_trained_model_deploymentedit
Stop a trained model deployment.
client.ml.stopTrainedModelDeployment({ model_id })
Argumentsedit
-
Request (object):
-
model_id
(string): The unique identifier of the trained model. -
allow_no_match
(Optional, boolean): Specifies what to do when the request: contains wildcard expressions and there are no deployments that match; contains the_all
string or no identifiers and there are no matches; or contains wildcard expressions and there are only partial matches. By default, it returns an empty array when there are no matches and the subset of results when there are partial matches. Iffalse
, the request returns a 404 status code when there are no matches or only partial matches. -
force
(Optional, boolean): Forcefully stops the deployment, even if it is used by ingest pipelines. You can’t use these pipelines until you restart the model deployment.
-
update_data_frame_analyticsedit
Updates certain properties of a data frame analytics job.
client.ml.updateDataFrameAnalytics({ id })
Argumentsedit
-
Request (object):
-
id
(string): Identifier for the data frame analytics job. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
description
(Optional, string): A description of the job. -
model_memory_limit
(Optional, string): The approximate maximum amount of memory resources that are permitted for analytical processing. If yourelasticsearch.yml
file contains anxpack.ml.max_model_memory_limit
setting, an error occurs when you try to create data frame analytics jobs that havemodel_memory_limit
values greater than that setting. -
max_num_threads
(Optional, number): The maximum number of threads to be used by the analysis. Using more threads may decrease the time necessary to complete the analysis at the cost of using more CPU. Note that the process may use additional threads for operational functionality other than the analysis itself. -
allow_lazy_start
(Optional, boolean): Specifies whether this job can start when there is insufficient machine learning node capacity for it to be immediately assigned to a node.
-
update_datafeededit
Updates certain properties of a datafeed.
client.ml.updateDatafeed({ datafeed_id })
Argumentsedit
-
Request (object):
-
datafeed_id
(string): A numerical character string that uniquely identifies the datafeed. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It must start and end with alphanumeric characters. -
aggregations
(Optional, Record<string, { aggregations, meta, adjacency_matrix, auto_date_histogram, avg, avg_bucket, boxplot, bucket_script, bucket_selector, bucket_sort, bucket_count_ks_test, bucket_correlation, cardinality, categorize_text, children, composite, cumulative_cardinality, cumulative_sum, date_histogram, date_range, derivative, diversified_sampler, extended_stats, extended_stats_bucket, frequent_item_sets, filter, filters, geo_bounds, geo_centroid, geo_distance, geohash_grid, geo_line, geotile_grid, geohex_grid, global, histogram, ip_range, ip_prefix, inference, line, matrix_stats, max, max_bucket, median_absolute_deviation, min, min_bucket, missing, moving_avg, moving_percentiles, moving_fn, multi_terms, nested, normalize, parent, percentile_ranks, percentiles, percentiles_bucket, range, rare_terms, rate, reverse_nested, sampler, scripted_metric, serial_diff, significant_terms, significant_text, stats, stats_bucket, string_stats, sum, sum_bucket, terms, top_hits, t_test, top_metrics, value_count, weighted_avg, variable_width_histogram }>): If set, the datafeed performs aggregation searches. Support for aggregations is limited and should be used only with low cardinality data. -
chunking_config
(Optional, { mode, time_span }): Datafeeds might search over long time periods, for several months or years. This search is split into time chunks in order to ensure the load on Elasticsearch is managed. Chunking configuration controls how the size of these time chunks are calculated; it is an advanced configuration option. -
delayed_data_check_config
(Optional, { check_window, enabled }): Specifies whether the datafeed checks for missing data and the size of the window. The datafeed can optionally search over indices that have already been read in an effort to determine whether any data has subsequently been added to the index. If missing data is found, it is a good indication that thequery_delay
is set too low and the data is being indexed after the datafeed has passed that moment in time. This check runs only on real-time datafeeds. -
frequency
(Optional, string | -1 | 0): The interval at which scheduled queries are made while the datafeed runs in real time. The default value is either the bucket span for short bucket spans, or, for longer bucket spans, a sensible fraction of the bucket span. Whenfrequency
is shorter than the bucket span, interim results for the last (partial) bucket are written then eventually overwritten by the full bucket results. If the datafeed uses aggregations, this value must be divisible by the interval of the date histogram aggregation. -
indices
(Optional, string[]): An array of index names. Wildcards are supported. If any of the indices are in remote clusters, the machine learning nodes must have theremote_cluster_client
role. -
indices_options
(Optional, { allow_no_indices, expand_wildcards, ignore_unavailable, ignore_throttled }): Specifies index expansion options that are used during search. -
job_id
(Optional, string) -
max_empty_searches
(Optional, number): If a real-time datafeed has never seen any data (including during any initial training period), it automatically stops and closes the associated job after this many real-time searches return no documents. In other words, it stops afterfrequency
timesmax_empty_searches
of real-time operation. If not set, a datafeed with no end time that sees no data remains started until it is explicitly stopped. By default, it is not set. -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): The Elasticsearch query domain-specific language (DSL). This value corresponds to the query object in an Elasticsearch search POST body. All the options that are supported by Elasticsearch can be used, as this object is passed verbatim to Elasticsearch. Note that if you change the query, the analyzed data is also changed. Therefore, the time required to learn might be long and the understandability of the results is unpredictable. If you want to make significant changes to the source data, it is recommended that you clone the job and datafeed and make the amendments in the clone. Let both run in parallel and close one when you are satisfied with the results of the job. -
query_delay
(Optional, string | -1 | 0): The number of seconds behind real time that data is queried. For example, if data from 10:04 a.m. might not be searchable in Elasticsearch until 10:06 a.m., set this property to 120 seconds. The default value is randomly selected between60s
and120s
. This randomness improves the query performance when there are multiple jobs running on the same node. -
runtime_mappings
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>): Specifies runtime fields for the datafeed search. -
script_fields
(Optional, Record<string, { script, ignore_failure }>): Specifies scripts that evaluate custom expressions and returns script fields to the datafeed. The detector configuration objects in a job can contain functions that use these script fields. -
scroll_size
(Optional, number): The size parameter that is used in Elasticsearch searches when the datafeed does not use aggregations. The maximum value is the value ofindex.max_result_window
. -
allow_no_indices
(Optional, boolean): Iftrue
, wildcard indices expressions that resolve into no concrete indices are ignored. This includes the_all
string or when no indices are specified. -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Type of index that wildcard patterns can match. If the request can target data streams, this argument determines whether wildcard expressions match hidden data streams. Supports a list of values. Valid values are:
-
-
all
: Match any data stream or index, including hidden ones. -
closed
: Match closed, non-hidden indices. Also matches any non-hidden data stream. Data streams cannot be closed. -
hidden
: Match hidden data streams and hidden indices. Must be combined withopen
,closed
, or both. -
none
: Wildcard patterns are not accepted. -
open
: Match open, non-hidden indices. Also matches any non-hidden data stream.-
ignore_throttled
(Optional, boolean): Iftrue
, concrete, expanded or aliased indices are ignored when frozen. -
ignore_unavailable
(Optional, boolean): Iftrue
, unavailable indices (missing or closed) are ignored.
-
update_filteredit
Updates the description of a filter, adds items, or removes items.
client.ml.updateFilter({ filter_id })
Argumentsedit
-
Request (object):
-
filter_id
(string): A string that uniquely identifies a filter. -
add_items
(Optional, string[]): The items to add to the filter. -
description
(Optional, string): A description for the filter. -
remove_items
(Optional, string[]): The items to remove from the filter.
-
update_jobedit
Updates certain properties of an anomaly detection job.
client.ml.updateJob({ job_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the job. -
allow_lazy_open
(Optional, boolean): Advanced configuration option. Specifies whether this job can open when there is insufficient machine learning node capacity for it to be immediately assigned to a node. Iffalse
and a machine learning node with capacity to run the job cannot immediately be found, the open anomaly detection jobs API returns an error. However, this is also subject to the cluster-widexpack.ml.max_lazy_ml_nodes
setting. If this option is set totrue
, the open anomaly detection jobs API does not return an error and the job waits in the opening state until sufficient machine learning node capacity is available. -
analysis_limits
(Optional, { model_memory_limit }) -
background_persist_interval
(Optional, string | -1 | 0): Advanced configuration option. The time between each periodic persistence of the model. The default value is a randomized value between 3 to 4 hours, which avoids all jobs persisting at exactly the same time. The smallest allowed value is 1 hour. For very large models (several GB), persistence could take 10-20 minutes, so do not set the value too low. If the job is open when you make the update, you must stop the datafeed, close the job, then reopen the job and restart the datafeed for the changes to take effect. -
custom_settings
(Optional, Record<string, User-defined value>): Advanced configuration option. Contains custom meta data about the job. For example, it can contain custom URL information as shown in Adding custom URLs to machine learning results. -
categorization_filters
(Optional, string[]) -
description
(Optional, string): A description of the job. -
model_plot_config
(Optional, { annotations_enabled, enabled, terms }) -
model_prune_window
(Optional, string | -1 | 0) -
daily_model_snapshot_retention_after_days
(Optional, number): Advanced configuration option, which affects the automatic removal of old model snapshots for this job. It specifies a period of time (in days) after which only the first snapshot per day is retained. This period is relative to the timestamp of the most recent snapshot for this job. Valid values range from 0 tomodel_snapshot_retention_days
. For jobs created before version 7.8.0, the default value matchesmodel_snapshot_retention_days
. -
model_snapshot_retention_days
(Optional, number): Advanced configuration option, which affects the automatic removal of old model snapshots for this job. It specifies the maximum period of time (in days) that snapshots are retained. This period is relative to the timestamp of the most recent snapshot for this job. -
renormalization_window_days
(Optional, number): Advanced configuration option. The period over which adjustments to the score are applied, as new data is seen. -
results_retention_days
(Optional, number): Advanced configuration option. The period of time (in days) that results are retained. Age is calculated relative to the timestamp of the latest bucket result. If this property has a non-null value, once per day at 00:30 (server time), results that are the specified number of days older than the latest bucket result are deleted from Elasticsearch. The default value is null, which means all results are retained. -
groups
(Optional, string[]): A list of job groups. A job can belong to no groups or many. -
detectors
(Optional, { by_field_name, custom_rules, detector_description, detector_index, exclude_frequent, field_name, function, over_field_name, partition_field_name, use_null }[]): An array of detector update objects. -
per_partition_categorization
(Optional, { enabled, stop_on_warn }): Settings related to how categorization interacts with partition fields.
-
update_model_snapshotedit
Updates certain properties of a snapshot.
client.ml.updateModelSnapshot({ job_id, snapshot_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
snapshot_id
(string): Identifier for the model snapshot. -
description
(Optional, string): A description of the model snapshot. -
retain
(Optional, boolean): Iftrue
, this snapshot will not be deleted during automatic cleanup of snapshots older thanmodel_snapshot_retention_days
. However, this snapshot will be deleted when the job is deleted.
-
update_trained_model_deploymentedit
Updates certain properties of trained model deployment.
client.ml.updateTrainedModelDeployment()
upgrade_job_snapshotedit
Upgrades a given job snapshot to the current major version.
client.ml.upgradeJobSnapshot({ job_id, snapshot_id })
Argumentsedit
-
Request (object):
-
job_id
(string): Identifier for the anomaly detection job. -
snapshot_id
(string): A numerical character string that uniquely identifies the model snapshot. -
wait_for_completion
(Optional, boolean): When true, the API won’t respond until the upgrade is complete. Otherwise, it responds as soon as the upgrade task is assigned to a node. -
timeout
(Optional, string | -1 | 0): Controls the time to wait for the request to complete.
-
nodesedit
clear_repositories_metering_archiveedit
Removes the archived repositories metering information present in the cluster.
client.nodes.clearRepositoriesMeteringArchive({ node_id, max_archive_version })
Argumentsedit
-
Request (object):
-
node_id
(string | string[]): List of node IDs or names used to limit returned information. All the nodes selective options are explained [here](https://www.elastic.co/guide/en/elasticsearch/reference/current/cluster.html#cluster-nodes). -
max_archive_version
(number): Specifies the maximum [archive_version](https://www.elastic.co/guide/en/elasticsearch/reference/current/get-repositories-metering-api.html#get-repositories-metering-api-response-body) to be cleared from the archive.
-
get_repositories_metering_infoedit
Returns cluster repositories metering information.
client.nodes.getRepositoriesMeteringInfo({ node_id })
Argumentsedit
-
Request (object):
-
node_id
(string | string[]): List of node IDs or names used to limit returned information. All the nodes selective options are explained [here](https://www.elastic.co/guide/en/elasticsearch/reference/current/cluster.html#cluster-nodes).
-
hot_threadsedit
Returns information about hot threads on each node in the cluster.
client.nodes.hotThreads({ ... })
Argumentsedit
-
Request (object):
-
node_id
(Optional, string | string[]): List of node IDs or names used to limit returned information. -
ignore_idle_threads
(Optional, boolean): If true, known idle threads (e.g. waiting in a socket select, or to get a task from an empty queue) are filtered out. -
interval
(Optional, string | -1 | 0): The interval to do the second sampling of threads. -
snapshots
(Optional, number): Number of samples of thread stacktrace. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
threads
(Optional, number): Specifies the number of hot threads to provide information for. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error. -
type
(Optional, Enum("cpu" | "wait" | "block" | "gpu" | "mem")): The type to sample. -
sort
(Optional, Enum("cpu" | "wait" | "block" | "gpu" | "mem")): The sort order for cpu type (default: total)
-
infoedit
Returns information about nodes in the cluster.
client.nodes.info({ ... })
Argumentsedit
-
Request (object):
-
node_id
(Optional, string | string[]): List of node IDs or names used to limit returned information. -
metric
(Optional, string | string[]): Limits the information returned to the specific metrics. Supports a list, such as http,ingest. -
flat_settings
(Optional, boolean): If true, returns settings in flat format. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
reload_secure_settingsedit
Reloads secure settings.
client.nodes.reloadSecureSettings({ ... })
Argumentsedit
-
Request (object):
-
node_id
(Optional, string | string[]): A list of node IDs to span the reload/reinit call. Should stay empty because reloading usually involves all cluster nodes. -
secure_settings_password
(Optional, string) -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
statsedit
Returns statistical information about nodes in the cluster.
client.nodes.stats({ ... })
Argumentsedit
-
Request (object):
-
node_id
(Optional, string | string[]): List of node IDs or names used to limit returned information. -
metric
(Optional, string | string[]): Limit the information returned to the specified metrics -
index_metric
(Optional, string | string[]): Limit the information returned for indices metric to the specific index metrics. It can be used only if indices (or all) metric is specified. -
completion_fields
(Optional, string | string[]): List or wildcard expressions of fields to include in fielddata and suggest statistics. -
fielddata_fields
(Optional, string | string[]): List or wildcard expressions of fields to include in fielddata statistics. -
fields
(Optional, string | string[]): List or wildcard expressions of fields to include in the statistics. -
groups
(Optional, boolean): List of search groups to include in the search statistics. -
include_segment_file_sizes
(Optional, boolean): If true, the call reports the aggregated disk usage of each one of the Lucene index files (only applies if segment stats are requested). -
level
(Optional, Enum("cluster" | "indices" | "shards")): Indicates whether statistics are aggregated at the cluster, index, or shard level. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error. -
types
(Optional, string[]): A list of document types for the indexing index metric. -
include_unloaded_segments
(Optional, boolean): If set to true segment stats will include stats for segments that are not currently loaded into memory
-
usageedit
Returns low-level information about REST actions usage on nodes.
client.nodes.usage({ ... })
Argumentsedit
-
Request (object):
-
node_id
(Optional, string | string[]): A list of node IDs or names to limit the returned information; use_local
to return information from the node you’re connecting to, leave empty to get information from all nodes -
metric
(Optional, string | string[]): Limit the information returned to the specified metrics -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
rollupedit
delete_jobedit
Deletes an existing rollup job.
client.rollup.deleteJob({ id })
Argumentsedit
-
Request (object):
-
id
(string): The ID of the job to delete
-
get_jobsedit
Retrieves the configuration, stats, and status of rollup jobs.
client.rollup.getJobs({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): The ID of the job(s) to fetch. Accepts glob patterns, or left blank for all jobs
-
get_rollup_capsedit
Returns the capabilities of any rollup jobs that have been configured for a specific index or index pattern.
client.rollup.getRollupCaps({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): The ID of the index to check rollup capabilities on, or left blank for all jobs
-
get_rollup_index_capsedit
Returns the rollup capabilities of all jobs inside of a rollup index (e.g. the index where rollup data is stored).
client.rollup.getRollupIndexCaps({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): The rollup index or index pattern to obtain rollup capabilities from.
-
put_jobedit
Creates a rollup job.
client.rollup.putJob({ id, cron, groups, index_pattern, page_size, rollup_index })
Argumentsedit
-
Request (object):
-
id
(string): Identifier for the rollup job. This can be any alphanumeric string and uniquely identifies the data that is associated with the rollup job. The ID is persistent; it is stored with the rolled up data. If you create a job, let it run for a while, then delete the job, the data that the job rolled up is still be associated with this job ID. You cannot create a new job with the same ID since that could lead to problems with mismatched job configurations. -
cron
(string): A cron string which defines the intervals when the rollup job should be executed. When the interval triggers, the indexer attempts to rollup the data in the index pattern. The cron pattern is unrelated to the time interval of the data being rolled up. For example, you may wish to create hourly rollups of your document but to only run the indexer on a daily basis at midnight, as defined by the cron. The cron pattern is defined just like a Watcher cron schedule. -
groups
({ date_histogram, histogram, terms }): Defines the grouping fields and aggregations that are defined for this rollup job. These fields will then be available later for aggregating into buckets. These aggs and fields can be used in any combination. Think of the groups configuration as defining a set of tools that can later be used in aggregations to partition the data. Unlike raw data, we have to think ahead to which fields and aggregations might be used. Rollups provide enough flexibility that you simply need to determine which fields are needed, not in what order they are needed. -
index_pattern
(string): The index or index pattern to roll up. Supports wildcard-style patterns (logstash-*
). The job attempts to rollup the entire index or index-pattern. -
page_size
(number): The number of bucket results that are processed on each iteration of the rollup indexer. A larger value tends to execute faster, but requires more memory during processing. This value has no effect on how the data is rolled up; it is merely used for tweaking the speed or memory cost of the indexer. -
rollup_index
(string): The index that contains the rollup results. The index can be shared with other rollup jobs. The data is stored so that it doesn’t interfere with unrelated jobs. -
metrics
(Optional, { field, metrics }[]): Defines the metrics to collect for each grouping tuple. By default, only the doc_counts are collected for each group. To make rollup useful, you will often add metrics like averages, mins, maxes, etc. Metrics are defined on a per-field basis and for each field you configure which metric should be collected. -
timeout
(Optional, string | -1 | 0): Time to wait for the request to complete. -
headers
(Optional, Record<string, string | string[]>)
-
rollup_searchedit
Enables searching rolled-up data using the standard query DSL.
client.rollup.rollupSearch({ index })
Argumentsedit
-
Request (object):
-
index
(string | string[]): The indices or index-pattern(s) (containing rollup or regular data) that should be searched -
aggregations
(Optional, Record<string, { aggregations, meta, adjacency_matrix, auto_date_histogram, avg, avg_bucket, boxplot, bucket_script, bucket_selector, bucket_sort, bucket_count_ks_test, bucket_correlation, cardinality, categorize_text, children, composite, cumulative_cardinality, cumulative_sum, date_histogram, date_range, derivative, diversified_sampler, extended_stats, extended_stats_bucket, frequent_item_sets, filter, filters, geo_bounds, geo_centroid, geo_distance, geohash_grid, geo_line, geotile_grid, geohex_grid, global, histogram, ip_range, ip_prefix, inference, line, matrix_stats, max, max_bucket, median_absolute_deviation, min, min_bucket, missing, moving_avg, moving_percentiles, moving_fn, multi_terms, nested, normalize, parent, percentile_ranks, percentiles, percentiles_bucket, range, rare_terms, rate, reverse_nested, sampler, scripted_metric, serial_diff, significant_terms, significant_text, stats, stats_bucket, string_stats, sum, sum_bucket, terms, top_hits, t_test, top_metrics, value_count, weighted_avg, variable_width_histogram }>) -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
size
(Optional, number): Must be zero if set, as rollups work on pre-aggregated data -
rest_total_hits_as_int
(Optional, boolean): Indicates whether hits.total should be rendered as an integer or an object in the rest search response -
typed_keys
(Optional, boolean): Specify whether aggregation and suggester names should be prefixed by their respective types in the response
-
start_jobedit
Starts an existing, stopped rollup job.
client.rollup.startJob({ id })
Argumentsedit
-
Request (object):
-
id
(string): The ID of the job to start
-
stop_jobedit
Stops an existing, started rollup job.
client.rollup.stopJob({ id })
Argumentsedit
-
Request (object):
-
id
(string): The ID of the job to stop -
timeout
(Optional, string | -1 | 0): Block for (at maximum) the specified duration while waiting for the job to stop. Defaults to 30s. -
wait_for_completion
(Optional, boolean): True if the API should block until the job has fully stopped, false if should be executed async. Defaults to false.
-
search_applicationedit
deleteedit
Deletes a search application.
client.searchApplication.delete({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the search application to delete
-
delete_behavioral_analyticsedit
Delete a behavioral analytics collection.
client.searchApplication.deleteBehavioralAnalytics({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the analytics collection to be deleted
-
getedit
Returns the details about a search application.
client.searchApplication.get({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the search application
-
get_behavioral_analyticsedit
Returns the existing behavioral analytics collections.
client.searchApplication.getBehavioralAnalytics({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string[]): A list of analytics collections to limit the returned information
-
listedit
Returns the existing search applications.
client.searchApplication.list({ ... })
Argumentsedit
-
Request (object):
-
q
(Optional, string): Query in the Lucene query string syntax" -
from
(Optional, number): Starting offset (default: 0) -
size
(Optional, number): specifies a max number of results to get
-
post_behavioral_analytics_eventedit
Creates a behavioral analytics event for existing collection.
client.searchApplication.postBehavioralAnalyticsEvent()
putedit
Creates or updates a search application.
client.searchApplication.put({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the search application to be created or updated -
search_application
(Optional, { name, indices, updated_at_millis, analytics_collection_name, template }) -
create
(Optional, boolean): If true, requires that a search application with the specified resource_id does not already exist. (default: false)
-
put_behavioral_analyticsedit
Creates a behavioral analytics collection.
client.searchApplication.putBehavioralAnalytics({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the analytics collection to be created or updated
-
render_queryedit
Renders a query for given search application search parameters
client.searchApplication.renderQuery()
searchedit
Perform a search against a search application
client.searchApplication.search({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the search application to be searched -
params
(Optional, Record<string, User-defined value>)
-
searchable_snapshotsedit
cache_statsedit
Retrieve node-level cache statistics about searchable snapshots.
client.searchableSnapshots.cacheStats({ ... })
Argumentsedit
-
Request (object):
-
node_id
(Optional, string | string[]): A list of node IDs or names to limit the returned information; use_local
to return information from the node you’re connecting to, leave empty to get information from all nodes -
master_timeout
(Optional, string | -1 | 0)
-
clear_cacheedit
Clear the cache of searchable snapshots.
client.searchableSnapshots.clearCache({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names -
expand_wildcards
(Optional, Enum("all" | "open" | "closed" | "hidden" | "none") | Enum("all" | "open" | "closed" | "hidden" | "none")[]): Whether to expand wildcard expression to concrete indices that are open, closed or both. -
allow_no_indices
(Optional, boolean): Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes_all
string or when no indices have been specified) -
ignore_unavailable
(Optional, boolean): Whether specified concrete indices should be ignored when unavailable (missing or closed) -
pretty
(Optional, boolean) -
human
(Optional, boolean)
-
mountedit
Mount a snapshot as a searchable index.
client.searchableSnapshots.mount({ repository, snapshot, index })
Argumentsedit
-
Request (object):
-
repository
(string): The name of the repository containing the snapshot of the index to mount -
snapshot
(string): The name of the snapshot of the index to mount -
index
(string) -
renamed_index
(Optional, string) -
index_settings
(Optional, Record<string, User-defined value>) -
ignore_index_settings
(Optional, string[]) -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
wait_for_completion
(Optional, boolean): Should this request wait until the operation has completed before returning -
storage
(Optional, string): Selects the kind of local storage used to accelerate searches. Experimental, and defaults tofull_copy
-
statsedit
Retrieve shard-level statistics about searchable snapshots.
client.searchableSnapshots.stats({ ... })
Argumentsedit
-
Request (object):
-
index
(Optional, string | string[]): A list of index names -
level
(Optional, Enum("cluster" | "indices" | "shards")): Return stats aggregated at cluster, index or shard level
-
securityedit
authenticateedit
Enables authentication as a user and retrieve information about the authenticated user.
client.security.authenticate()
bulk_update_api_keysedit
Updates the attributes of multiple existing API keys.
client.security.bulkUpdateApiKeys()
change_passwordedit
Changes the passwords of users in the native realm and built-in users.
client.security.changePassword({ ... })
Argumentsedit
-
Request (object):
-
username
(Optional, string): The user whose password you want to change. If you do not specify this parameter, the password is changed for the current user. -
password
(Optional, string): The new password value. Passwords must be at least 6 characters long. -
password_hash
(Optional, string): A hash of the new password value. This must be produced using the same hashing algorithm as has been configured for password storage. For more details, see the explanation of thexpack.security.authc.password_hashing.algorithm
setting. -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
clear_api_key_cacheedit
Clear a subset or all entries from the API key cache.
client.security.clearApiKeyCache({ ids })
Argumentsedit
-
Request (object):
-
ids
(string | string[]): A list of IDs of API keys to clear from the cache
-
clear_cached_privilegesedit
Evicts application privileges from the native application privileges cache.
client.security.clearCachedPrivileges({ application })
Argumentsedit
-
Request (object):
-
application
(string): A list of application names
-
clear_cached_realmsedit
Evicts users from the user cache. Can completely clear the cache or evict specific users.
client.security.clearCachedRealms({ realms })
Argumentsedit
-
Request (object):
-
realms
(string | string[]): List of realms to clear -
usernames
(Optional, string[]): List of usernames to clear from the cache
-
clear_cached_rolesedit
Evicts roles from the native role cache.
client.security.clearCachedRoles({ name })
Argumentsedit
-
Request (object):
-
name
(string | string[]): Role name
-
clear_cached_service_tokensedit
Evicts tokens from the service account token caches.
client.security.clearCachedServiceTokens({ namespace, service, name })
Argumentsedit
-
Request (object):
-
namespace
(string): An identifier for the namespace -
service
(string): An identifier for the service name -
name
(string | string[]): A list of service token names
-
create_api_keyedit
Creates an API key for access without requiring basic authentication.
client.security.createApiKey({ ... })
Argumentsedit
-
Request (object):
-
expiration
(Optional, string | -1 | 0): Expiration time for the API key. By default, API keys never expire. -
name
(Optional, string): Specifies the name for this API key. -
role_descriptors
(Optional, Record<string, { cluster, indices, global, applications, metadata, run_as, transient_metadata }>): An array of role descriptors for this API key. This parameter is optional. When it is not specified or is an empty array, then the API key will have a point in time snapshot of permissions of the authenticated user. If you supply role descriptors then the resultant permissions would be an intersection of API keys permissions and authenticated user’s permissions thereby limiting the access scope for API keys. The structure of role descriptor is the same as the request for create role API. For more details, see create or update roles API. -
metadata
(Optional, Record<string, User-defined value>): Arbitrary metadata that you want to associate with the API key. It supports nested data structure. Within the metadata object, keys beginning with _ are reserved for system usage. -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
create_service_tokenedit
Creates a service account token for access without requiring basic authentication.
client.security.createServiceToken({ namespace, service })
Argumentsedit
-
Request (object):
-
namespace
(string): An identifier for the namespace -
service
(string): An identifier for the service name -
name
(Optional, string): An identifier for the token name -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
then refresh the affected shards to make this operation visible to search, ifwait_for
(the default) then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
delete_privilegesedit
Removes application privileges.
client.security.deletePrivileges({ application, name })
Argumentsedit
-
Request (object):
-
application
(string): Application name -
name
(string | string[]): Privilege name -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
delete_roleedit
Removes roles in the native realm.
client.security.deleteRole({ name })
Argumentsedit
-
Request (object):
-
name
(string): Role name -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
delete_role_mappingedit
Removes role mappings.
client.security.deleteRoleMapping({ name })
Argumentsedit
-
Request (object):
-
name
(string): Role-mapping name -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
delete_service_tokenedit
Deletes a service account token.
client.security.deleteServiceToken({ namespace, service, name })
Argumentsedit
-
Request (object):
-
namespace
(string): An identifier for the namespace -
service
(string): An identifier for the service name -
name
(string): An identifier for the token name -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
then refresh the affected shards to make this operation visible to search, ifwait_for
(the default) then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
delete_useredit
Deletes users from the native realm.
client.security.deleteUser({ username })
Argumentsedit
-
Request (object):
-
username
(string): username -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
disable_useredit
Disables users in the native realm.
client.security.disableUser({ username })
Argumentsedit
-
Request (object):
-
username
(string): The username of the user to disable -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
enable_useredit
Enables users in the native realm.
client.security.enableUser({ username })
Argumentsedit
-
Request (object):
-
username
(string): The username of the user to enable -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
enroll_kibanaedit
Allows a kibana instance to configure itself to communicate with a secured elasticsearch cluster.
client.security.enrollKibana()
enroll_nodeedit
Allows a new node to enroll to an existing cluster with security enabled.
client.security.enrollNode()
get_api_keyedit
Retrieves information for one or more API keys.
client.security.getApiKey({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): API key id of the API key to be retrieved -
name
(Optional, string): API key name of the API key to be retrieved -
owner
(Optional, boolean): flag to query API keys owned by the currently authenticated user -
realm_name
(Optional, string): realm name of the user who created this API key to be retrieved -
username
(Optional, string): user name of the user who created this API key to be retrieved -
with_limited_by
(Optional, boolean): Return the snapshot of the owner user’s role descriptors associated with the API key. An API key’s actual permission is the intersection of its assigned role descriptors and the owner user’s role descriptors.
-
get_builtin_privilegesedit
Retrieves the list of cluster privileges and index privileges that are available in this version of Elasticsearch.
client.security.getBuiltinPrivileges()
get_privilegesedit
Retrieves application privileges.
client.security.getPrivileges({ ... })
Argumentsedit
-
Request (object):
-
application
(Optional, string): Application name -
name
(Optional, string | string[]): Privilege name
-
get_roleedit
Retrieves roles in the native realm.
client.security.getRole({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string | string[]): The name of the role. You can specify multiple roles as a list. If you do not specify this parameter, the API returns information about all roles.
-
get_role_mappingedit
Retrieves role mappings.
client.security.getRoleMapping({ ... })
Argumentsedit
-
Request (object):
-
name
(Optional, string | string[]): The distinct name that identifies the role mapping. The name is used solely as an identifier to facilitate interaction via the API; it does not affect the behavior of the mapping in any way. You can specify multiple mapping names as a list. If you do not specify this parameter, the API returns information about all role mappings.
-
get_service_accountsedit
Retrieves information about service accounts.
client.security.getServiceAccounts({ ... })
Argumentsedit
-
Request (object):
-
namespace
(Optional, string): Name of the namespace. Omit this parameter to retrieve information about all service accounts. If you omit this parameter, you must also omit theservice
parameter. -
service
(Optional, string): Name of the service name. Omit this parameter to retrieve information about all service accounts that belong to the specifiednamespace
.
-
get_service_credentialsedit
Retrieves information of all service credentials for a service account.
client.security.getServiceCredentials({ namespace, service })
Argumentsedit
-
Request (object):
-
namespace
(string): Name of the namespace. -
service
(string): Name of the service name.
-
get_tokenedit
Creates a bearer token for access without requiring basic authentication.
client.security.getToken({ ... })
Argumentsedit
-
Request (object):
-
grant_type
(Optional, Enum("password" | "client_credentials" | "_kerberos" | "refresh_token")) -
scope
(Optional, string) -
password
(Optional, string) -
kerberos_ticket
(Optional, string) -
refresh_token
(Optional, string) -
username
(Optional, string)
-
get_useredit
Retrieves information about users in the native realm and built-in users.
client.security.getUser({ ... })
Argumentsedit
-
Request (object):
-
username
(Optional, string | string[]): An identifier for the user. You can specify multiple usernames as a list. If you omit this parameter, the API retrieves information about all users. -
with_profile_uid
(Optional, boolean): If true will return the User Profile ID for a user, if any.
-
get_user_privilegesedit
Retrieves security privileges for the logged in user.
client.security.getUserPrivileges({ ... })
Argumentsedit
-
Request (object):
-
application
(Optional, string): The name of the application. Application privileges are always associated with exactly one application. If you do not specify this parameter, the API returns information about all privileges for all applications. -
priviledge
(Optional, string): The name of the privilege. If you do not specify this parameter, the API returns information about all privileges for the requested application. -
username
(Optional, string | null)
-
grant_api_keyedit
Creates an API key on behalf of another user.
client.security.grantApiKey({ api_key, grant_type })
Argumentsedit
-
Request (object):
-
api_key
({ name, expiration, role_descriptors, metadata }) -
grant_type
(Enum("access_token" | "password")) -
access_token
(Optional, string) -
username
(Optional, string) -
password
(Optional, string) -
run_as
(Optional, string)
-
has_privilegesedit
Determines whether the specified user has a specified list of privileges.
client.security.hasPrivileges({ ... })
Argumentsedit
-
Request (object):
-
user
(Optional, string): Username -
application
(Optional, { application, privileges, resources }[]) -
cluster
(Optional, Enum("all" | "cancel_task" | "create_snapshot" | "grant_api_key" | "manage" | "manage_api_key" | "manage_ccr" | "manage_enrich" | "manage_ilm" | "manage_index_templates" | "manage_ingest_pipelines" | "manage_logstash_pipelines" | "manage_ml" | "manage_oidc" | "manage_own_api_key" | "manage_pipeline" | "manage_rollup" | "manage_saml" | "manage_security" | "manage_service_account" | "manage_slm" | "manage_token" | "manage_transform" | "manage_user_profile" | "manage_watcher" | "monitor" | "monitor_ml" | "monitor_rollup" | "monitor_snapshot" | "monitor_text_structure" | "monitor_transform" | "monitor_watcher" | "read_ccr" | "read_ilm" | "read_pipeline" | "read_slm" | "transport_client")[]): A list of the cluster privileges that you want to check. -
index
(Optional, { names, privileges, allow_restricted_indices }[])
-
invalidate_api_keyedit
Invalidates one or more API keys.
client.security.invalidateApiKey({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string) -
ids
(Optional, string[]) -
name
(Optional, string) -
owner
(Optional, boolean) -
realm_name
(Optional, string) -
username
(Optional, string)
-
invalidate_tokenedit
Invalidates one or more access tokens or refresh tokens.
client.security.invalidateToken({ ... })
Argumentsedit
-
Request (object):
-
token
(Optional, string) -
refresh_token
(Optional, string) -
realm_name
(Optional, string) -
username
(Optional, string)
-
oidc_authenticateedit
Exchanges an OpenID Connection authentication response message for an Elasticsearch access token and refresh token pair
client.security.oidcAuthenticate()
oidc_logoutedit
Invalidates a refresh token and access token that was generated from the OpenID Connect Authenticate API
client.security.oidcLogout()
oidc_prepare_authenticationedit
Creates an OAuth 2.0 authentication request as a URL string
client.security.oidcPrepareAuthentication()
put_privilegesedit
Adds or updates application privileges.
client.security.putPrivileges({ ... })
Argumentsedit
-
Request (object):
-
privileges
(Optional, Record<string, Record<string, User-defined value>>) -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
put_roleedit
Adds and updates roles in the native realm.
client.security.putRole({ name })
Argumentsedit
-
Request (object):
-
name
(string): The name of the role. -
applications
(Optional, { application, privileges, resources }[]): A list of application privilege entries. -
cluster
(Optional, Enum("all" | "cancel_task" | "create_snapshot" | "grant_api_key" | "manage" | "manage_api_key" | "manage_ccr" | "manage_enrich" | "manage_ilm" | "manage_index_templates" | "manage_ingest_pipelines" | "manage_logstash_pipelines" | "manage_ml" | "manage_oidc" | "manage_own_api_key" | "manage_pipeline" | "manage_rollup" | "manage_saml" | "manage_security" | "manage_service_account" | "manage_slm" | "manage_token" | "manage_transform" | "manage_user_profile" | "manage_watcher" | "monitor" | "monitor_ml" | "monitor_rollup" | "monitor_snapshot" | "monitor_text_structure" | "monitor_transform" | "monitor_watcher" | "read_ccr" | "read_ilm" | "read_pipeline" | "read_slm" | "transport_client")[]): A list of cluster privileges. These privileges define the cluster-level actions for users with this role. -
global
(Optional, Record<string, User-defined value>): An object defining global privileges. A global privilege is a form of cluster privilege that is request-aware. Support for global privileges is currently limited to the management of application privileges. -
indices
(Optional, { field_security, names, privileges, query, allow_restricted_indices }[]): A list of indices permissions entries. -
metadata
(Optional, Record<string, User-defined value>): Optional metadata. Within the metadata object, keys that begin with an underscore (_
) are reserved for system use. -
run_as
(Optional, string[]): A list of users that the owners of this role can impersonate. -
transient_metadata
(Optional, { enabled }): Indicates roles that might be incompatible with the current cluster license, specifically roles with document and field level security. When the cluster license doesn’t allow certain features for a given role, this parameter is updated dynamically to list the incompatible features. Ifenabled
isfalse
, the role is ignored, but is still listed in the response from the authenticate API. -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
put_role_mappingedit
Creates and updates role mappings.
client.security.putRoleMapping({ name })
Argumentsedit
-
Request (object):
-
name
(string): Role-mapping name -
enabled
(Optional, boolean) -
metadata
(Optional, Record<string, User-defined value>) -
roles
(Optional, string[]) -
rules
(Optional, { any, all, field, except }) -
run_as
(Optional, string[]) -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
put_useredit
Adds and updates users in the native realm. These users are commonly referred to as native users.
client.security.putUser({ username })
Argumentsedit
-
Request (object):
-
username
(string): The username of the User -
email
(Optional, string | null) -
full_name
(Optional, string | null) -
metadata
(Optional, Record<string, User-defined value>) -
password
(Optional, string) -
password_hash
(Optional, string) -
roles
(Optional, string[]) -
enabled
(Optional, boolean) -
refresh
(Optional, Enum(true | false | "wait_for")): Iftrue
(the default) then refresh the affected shards to make this operation visible to search, ifwait_for
then wait for a refresh to make this operation visible to search, iffalse
then do nothing with refreshes.
-
query_api_keysedit
Retrieves information for API keys using a subset of query DSL
client.security.queryApiKeys({ ... })
Argumentsedit
-
Request (object):
-
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): A query to filter which API keys to return. The query supports a subset of query types, including match_all, bool, term, terms, ids, prefix, wildcard, and range. You can query all public information associated with an API key -
from
(Optional, number): Starting document offset. By default, you cannot page through more than 10,000 hits using the from and size parameters. To page through more hits, use the search_after parameter. -
sort
(Optional, string | { _score, _doc, _geo_distance, _script } | string | { _score, _doc, _geo_distance, _script }[]) -
size
(Optional, number): The number of hits to return. By default, you cannot page through more than 10,000 hits using the from and size parameters. To page through more hits, use the search_after parameter. -
search_after
(Optional, number | number | string | boolean | null | User-defined value[]) -
with_limited_by
(Optional, boolean): Return the snapshot of the owner user’s role descriptors associated with the API key. An API key’s actual permission is the intersection of its assigned role descriptors and the owner user’s role descriptors.
-
saml_authenticateedit
Exchanges a SAML Response message for an Elasticsearch access token and refresh token pair
client.security.samlAuthenticate({ content, ids })
Argumentsedit
-
Request (object):
-
content
(string): The SAML response as it was sent by the user’s browser, usually a Base64 encoded XML document. -
ids
(string | string[]): A json array with all the valid SAML Request Ids that the caller of the API has for the current user. -
realm
(Optional, string): The name of the realm that should authenticate the SAML response. Useful in cases where many SAML realms are defined.
-
saml_complete_logoutedit
Verifies the logout response sent from the SAML IdP
client.security.samlCompleteLogout({ realm, ids })
Argumentsedit
-
Request (object):
-
realm
(string): The name of the SAML realm in Elasticsearch for which the configuration is used to verify the logout response. -
ids
(string | string[]): A json array with all the valid SAML Request Ids that the caller of the API has for the current user. -
query_string
(Optional, string): If the SAML IdP sends the logout response with the HTTP-Redirect binding, this field must be set to the query string of the redirect URI. -
content
(Optional, string): If the SAML IdP sends the logout response with the HTTP-Post binding, this field must be set to the value of the SAMLResponse form parameter from the logout response.
-
saml_invalidateedit
Consumes a SAML LogoutRequest
client.security.samlInvalidate({ query_string })
Argumentsedit
-
Request (object):
-
query_string
(string): The query part of the URL that the user was redirected to by the SAML IdP to initiate the Single Logout. This query should include a single parameter named SAMLRequest that contains a SAML logout request that is deflated and Base64 encoded. If the SAML IdP has signed the logout request, the URL should include two extra parameters named SigAlg and Signature that contain the algorithm used for the signature and the signature value itself. In order for Elasticsearch to be able to verify the IdP’s signature, the value of the query_string field must be an exact match to the string provided by the browser. The client application must not attempt to parse or process the string in any way. -
acs
(Optional, string): The Assertion Consumer Service URL that matches the one of the SAML realm in Elasticsearch that should be used. You must specify either this parameter or the realm parameter. -
realm
(Optional, string): The name of the SAML realm in Elasticsearch the configuration. You must specify either this parameter or the acs parameter.
-
saml_logoutedit
Invalidates an access token and a refresh token that were generated via the SAML Authenticate API
client.security.samlLogout({ token })
Argumentsedit
-
Request (object):
-
token
(string): The access token that was returned as a response to calling the SAML authenticate API. Alternatively, the most recent token that was received after refreshing the original one by using a refresh_token. -
refresh_token
(Optional, string): The refresh token that was returned as a response to calling the SAML authenticate API. Alternatively, the most recent refresh token that was received after refreshing the original access token.
-
saml_prepare_authenticationedit
Creates a SAML authentication request
client.security.samlPrepareAuthentication({ ... })
Argumentsedit
-
Request (object):
-
acs
(Optional, string): The Assertion Consumer Service URL that matches the one of the SAML realms in Elasticsearch. The realm is used to generate the authentication request. You must specify either this parameter or the realm parameter. -
realm
(Optional, string): The name of the SAML realm in Elasticsearch for which the configuration is used to generate the authentication request. You must specify either this parameter or the acs parameter. -
relay_state
(Optional, string): A string that will be included in the redirect URL that this API returns as the RelayState query parameter. If the Authentication Request is signed, this value is used as part of the signature computation.
-
saml_service_provider_metadataedit
Generates SAML metadata for the Elastic stack SAML 2.0 Service Provider
client.security.samlServiceProviderMetadata({ realm_name })
Argumentsedit
-
Request (object):
-
realm_name
(string): The name of the SAML realm in Elasticsearch.
-
update_api_keyedit
Updates attributes of an existing API key.
client.security.updateApiKey({ id })
Argumentsedit
-
Request (object):
-
id
(string): The ID of the API key to update. -
role_descriptors
(Optional, Record<string, { cluster, indices, global, applications, metadata, run_as, transient_metadata }>): An array of role descriptors for this API key. This parameter is optional. When it is not specified or is an empty array, then the API key will have a point in time snapshot of permissions of the authenticated user. If you supply role descriptors then the resultant permissions would be an intersection of API keys permissions and authenticated user’s permissions thereby limiting the access scope for API keys. The structure of role descriptor is the same as the request for create role API. For more details, see create or update roles API. -
metadata
(Optional, Record<string, User-defined value>): Arbitrary metadata that you want to associate with the API key. It supports nested data structure. Within the metadata object, keys beginning with _ are reserved for system usage.
-
slmedit
delete_lifecycleedit
Deletes an existing snapshot lifecycle policy.
client.slm.deleteLifecycle({ policy_id })
Argumentsedit
-
Request (object):
-
policy_id
(string): The id of the snapshot lifecycle policy to remove
-
execute_lifecycleedit
Immediately creates a snapshot according to the lifecycle policy, without waiting for the scheduled time.
client.slm.executeLifecycle({ policy_id })
Argumentsedit
-
Request (object):
-
policy_id
(string): The id of the snapshot lifecycle policy to be executed
-
execute_retentionedit
Deletes any snapshots that are expired according to the policy’s retention rules.
client.slm.executeRetention()
get_lifecycleedit
Retrieves one or more snapshot lifecycle policy definitions and information about the latest snapshot attempts.
client.slm.getLifecycle({ ... })
Argumentsedit
-
Request (object):
-
policy_id
(Optional, string | string[]): List of snapshot lifecycle policies to retrieve
-
get_statsedit
Returns global and policy-level statistics about actions taken by snapshot lifecycle management.
client.slm.getStats()
get_statusedit
Retrieves the status of snapshot lifecycle management (SLM).
client.slm.getStatus()
put_lifecycleedit
Creates or updates a snapshot lifecycle policy.
client.slm.putLifecycle({ policy_id })
Argumentsedit
-
Request (object):
-
policy_id
(string): ID for the snapshot lifecycle policy you want to create or update. -
config
(Optional, { ignore_unavailable, indices, include_global_state, feature_states, metadata, partial }): Configuration for each snapshot created by the policy. -
name
(Optional, string): Name automatically assigned to each snapshot created by the policy. Date math is supported. To prevent conflicting snapshot names, a UUID is automatically appended to each snapshot name. -
repository
(Optional, string): Repository used to store snapshots created by this policy. This repository must exist prior to the policy’s creation. You can create a repository using the snapshot repository API. -
retention
(Optional, { expire_after, max_count, min_count }): Retention rules used to retain and delete snapshots created by the policy. -
schedule
(Optional, string): Periodic or absolute schedule at which the policy creates snapshots. SLM applies schedule changes immediately. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
startedit
Turns on snapshot lifecycle management (SLM).
client.slm.start()
stopedit
Turns off snapshot lifecycle management (SLM).
client.slm.stop()
snapshotedit
cleanup_repositoryedit
Removes stale data from repository.
client.snapshot.cleanupRepository({ repository })
Argumentsedit
-
Request (object):
-
repository
(string): Snapshot repository to clean up. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. -
timeout
(Optional, string | -1 | 0): Period to wait for a response.
-
cloneedit
Clones indices from one snapshot into another snapshot in the same repository.
client.snapshot.clone({ repository, snapshot, target_snapshot, indices })
Argumentsedit
-
Request (object):
-
repository
(string): A repository name -
snapshot
(string): The name of the snapshot to clone from -
target_snapshot
(string): The name of the cloned snapshot to create -
indices
(string) -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
timeout
(Optional, string | -1 | 0)
-
createedit
Creates a snapshot in a repository.
client.snapshot.create({ repository, snapshot })
Argumentsedit
-
Request (object):
-
repository
(string): Repository for the snapshot. -
snapshot
(string): Name of the snapshot. Must be unique in the repository. -
ignore_unavailable
(Optional, boolean): Iftrue
, the request ignores data streams and indices inindices
that are missing or closed. Iffalse
, the request returns an error for any data stream or index that is missing or closed. -
include_global_state
(Optional, boolean): Iftrue
, the current cluster state is included in the snapshot. The cluster state includes persistent cluster settings, composable index templates, legacy index templates, ingest pipelines, and ILM policies. It also includes data stored in system indices, such as Watches and task records (configurable viafeature_states
). -
indices
(Optional, string | string[]): Data streams and indices to include in the snapshot. Supports multi-target syntax. Includes all data streams and indices by default. -
feature_states
(Optional, string[]): Feature states to include in the snapshot. Each feature state includes one or more system indices containing related data. You can view a list of eligible features using the get features API. Ifinclude_global_state
istrue
, all current feature states are included by default. Ifinclude_global_state
isfalse
, no feature states are included by default. -
metadata
(Optional, Record<string, User-defined value>): Optional metadata for the snapshot. May have any contents. Must be less than 1024 bytes. This map is not automatically generated by Elasticsearch. -
partial
(Optional, boolean): Iftrue
, allows restoring a partial snapshot of indices with unavailable shards. Only shards that were successfully included in the snapshot will be restored. All missing shards will be recreated as empty. Iffalse
, the entire restore operation will fail if one or more indices included in the snapshot do not have all primary shards available. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
wait_for_completion
(Optional, boolean): Iftrue
, the request returns a response when the snapshot is complete. Iffalse
, the request returns a response when the snapshot initializes.
-
create_repositoryedit
Creates a repository.
client.snapshot.createRepository({ repository, type, settings })
Argumentsedit
-
Request (object):
-
repository
(string): A repository name -
type
(string) -
settings
({ chunk_size, compress, concurrent_streams, location, read_only }) -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
verify
(Optional, boolean): Whether to verify the repository after creation
-
deleteedit
Deletes one or more snapshots.
client.snapshot.delete({ repository, snapshot })
Argumentsedit
-
Request (object):
-
repository
(string): A repository name -
snapshot
(string): A list of snapshot names -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node
-
delete_repositoryedit
Deletes a repository.
client.snapshot.deleteRepository({ repository })
Argumentsedit
-
Request (object):
-
repository
(string | string[]): Name of the snapshot repository to unregister. Wildcard (*
) patterns are supported. -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
getedit
Returns information about a snapshot.
client.snapshot.get({ repository, snapshot })
Argumentsedit
-
Request (object):
-
repository
(string): List of snapshot repository names used to limit the request. Wildcard (*) expressions are supported. -
snapshot
(string | string[]): List of snapshot names to retrieve. Also accepts wildcards (*).- To get information about all snapshots in a registered repository, use a wildcard (*) or _all.
- To get information about any snapshots that are currently running, use _current.
-
ignore_unavailable
(Optional, boolean): If false, the request returns an error for any snapshots that are unavailable. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
verbose
(Optional, boolean): If true, returns additional information about each snapshot such as the version of Elasticsearch which took the snapshot, the start and end times of the snapshot, and the number of shards snapshotted. -
index_details
(Optional, boolean): If true, returns additional information about each index in the snapshot comprising the number of shards in the index, the total size of the index in bytes, and the maximum number of segments per shard in the index. Defaults to false, meaning that this information is omitted. -
index_names
(Optional, boolean): If true, returns the name of each index in each snapshot. -
include_repository
(Optional, boolean): If true, returns the repository name in each snapshot. -
sort
(Optional, Enum("start_time" | "duration" | "name" | "index_count" | "repository" | "shard_count" | "failed_shard_count")): Allows setting a sort order for the result. Defaults to start_time, i.e. sorting by snapshot start time stamp. -
size
(Optional, number): Maximum number of snapshots to return. Defaults to 0 which means return all that match the request without limit. -
order
(Optional, Enum("asc" | "desc")): Sort order. Valid values are asc for ascending and desc for descending order. Defaults to asc, meaning ascending order. -
after
(Optional, string): Offset identifier to start pagination from as returned by the next field in the response body. -
offset
(Optional, number): Numeric offset to start pagination from based on the snapshots matching this request. Using a non-zero value for this parameter is mutually exclusive with using the after parameter. Defaults to 0. -
from_sort_value
(Optional, string): Value of the current sort column at which to start retrieval. Can either be a string snapshot- or repository name when sorting by snapshot or repository name, a millisecond time value or a number when sorting by index- or shard count. -
slm_policy_filter
(Optional, string): Filter snapshots by a list of SLM policy names that snapshots belong to. Also accepts wildcards (*) and combinations of wildcards followed by exclude patterns starting with -. To include snapshots not created by an SLM policy you can use the special pattern _none that will match all snapshots without an SLM policy.
-
get_repositoryedit
Returns information about a repository.
client.snapshot.getRepository({ ... })
Argumentsedit
-
Request (object):
-
repository
(Optional, string | string[]): A list of repository names -
local
(Optional, boolean): Return local information, do not retrieve the state from master node (default: false) -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node
-
repository_analyzeedit
Analyzes a repository for correctness and performance
client.snapshot.repositoryAnalyze()
restoreedit
Restores a snapshot.
client.snapshot.restore({ repository, snapshot })
Argumentsedit
-
Request (object):
-
repository
(string): A repository name -
snapshot
(string): A snapshot name -
feature_states
(Optional, string[]) -
ignore_index_settings
(Optional, string[]) -
ignore_unavailable
(Optional, boolean) -
include_aliases
(Optional, boolean) -
include_global_state
(Optional, boolean) -
index_settings
(Optional, { index, mode, routing_path, soft_deletes, sort, number_of_shards, number_of_replicas, number_of_routing_shards, check_on_startup, codec, routing_partition_size, load_fixed_bitset_filters_eagerly, hidden, auto_expand_replicas, merge, search, refresh_interval, max_result_window, max_inner_result_window, max_rescore_window, max_docvalue_fields_search, max_script_fields, max_ngram_diff, max_shingle_diff, blocks, max_refresh_listeners, analyze, highlight, max_terms_count, max_regex_length, routing, gc_deletes, default_pipeline, final_pipeline, lifecycle, provided_name, creation_date, creation_date_string, uuid, version, verified_before_close, format, max_slices_per_scroll, translog, query_string, priority, top_metrics_max_size, analysis, settings, time_series, shards, queries, similarity, mapping, indexing.slowlog, indexing_pressure, store }) -
indices
(Optional, string | string[]) -
partial
(Optional, boolean) -
rename_pattern
(Optional, string) -
rename_replacement
(Optional, string) -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
wait_for_completion
(Optional, boolean): Should this request wait until the operation has completed before returning
-
statusedit
Returns information about the status of a snapshot.
client.snapshot.status({ ... })
Argumentsedit
-
Request (object):
-
repository
(Optional, string): A repository name -
snapshot
(Optional, string | string[]): A list of snapshot names -
ignore_unavailable
(Optional, boolean): Whether to ignore unavailable snapshots, defaults to false which means a SnapshotMissingException is thrown -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node
-
verify_repositoryedit
Verifies a repository.
client.snapshot.verifyRepository({ repository })
Argumentsedit
-
Request (object):
-
repository
(string): A repository name -
master_timeout
(Optional, string | -1 | 0): Explicit operation timeout for connection to master node -
timeout
(Optional, string | -1 | 0): Explicit operation timeout
-
sqledit
clear_cursoredit
Clears the SQL cursor
client.sql.clearCursor({ cursor })
Argumentsedit
-
Request (object):
-
cursor
(string)
-
delete_asyncedit
Deletes an async SQL search or a stored synchronous SQL search. If the search is still running, the API cancels it.
client.sql.deleteAsync({ id })
Argumentsedit
-
Request (object):
-
id
(string): The async search ID
-
get_asyncedit
Returns the current status and available results for an async SQL search or stored synchronous SQL search
client.sql.getAsync({ id })
Argumentsedit
-
Request (object):
-
id
(string): The async search ID -
delimiter
(Optional, string): Separator for CSV results. The API only supports this parameter for CSV responses. -
format
(Optional, string): Format for the response. You must specify a format using this parameter or the Accept HTTP header. If you specify both, the API uses this parameter. -
keep_alive
(Optional, string | -1 | 0): Retention period for the search and its results. Defaults to thekeep_alive
period for the original SQL search. -
wait_for_completion_timeout
(Optional, string | -1 | 0): Period to wait for complete results. Defaults to no timeout, meaning the request waits for complete search results.
-
get_async_statusedit
Returns the current status of an async SQL search or a stored synchronous SQL search
client.sql.getAsyncStatus({ id })
Argumentsedit
-
Request (object):
-
id
(string): The async search ID
-
queryedit
Executes a SQL request
client.sql.query({ ... })
Argumentsedit
-
Request (object):
-
catalog
(Optional, string): Default catalog (cluster) for queries. If unspecified, the queries execute on the data in the local cluster only. -
columnar
(Optional, boolean): If true, the results in a columnar fashion: one row represents all the values of a certain column from the current page of results. -
cursor
(Optional, string) -
fetch_size
(Optional, number): The maximum number of rows (or entries) to return in one response -
filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): Optional Elasticsearch query DSL for additional filtering. -
query
(Optional, string): SQL query to execute -
request_timeout
(Optional, string | -1 | 0): The timeout before the request fails. -
page_timeout
(Optional, string | -1 | 0): The timeout before a pagination request fails. -
time_zone
(Optional, string): Time-zone in ISO 8601 used for executing the query on the server. More information available here. -
field_multi_value_leniency
(Optional, boolean): Throw an exception when encountering multiple values for a field (default) or be lenient and return the first value from the list (without any guarantees of what that will be - typically the first in natural ascending order). -
runtime_mappings
(Optional, Record<string, { fetch_fields, format, input_field, target_field, target_index, script, type }>): Defines one or more runtime fields in the search request. These fields take precedence over mapped fields with the same name. -
wait_for_completion_timeout
(Optional, string | -1 | 0): Period to wait for complete results. Defaults to no timeout, meaning the request waits for complete search results. If the search doesn’t finish within this period, the search becomes async. -
params
(Optional, Record<string, User-defined value>): Values for parameters in the query. -
keep_alive
(Optional, string | -1 | 0): Retention period for an async or saved synchronous search. -
keep_on_completion
(Optional, boolean): If true, Elasticsearch stores synchronous searches if you also specify the wait_for_completion_timeout parameter. If false, Elasticsearch only stores async searches that don’t finish before the wait_for_completion_timeout. -
index_using_frozen
(Optional, boolean): If true, the search can run on frozen indices. Defaults to false. -
format
(Optional, string): a short version of the Accept header, e.g. json, yaml
-
translateedit
Translates SQL into Elasticsearch queries
client.sql.translate({ query })
Argumentsedit
-
Request (object):
-
query
(string) -
fetch_size
(Optional, number) -
filter
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }) -
time_zone
(Optional, string)
-
ssledit
certificatesedit
Retrieves information about the X.509 certificates used to encrypt communications in the cluster.
client.ssl.certificates()
tasksedit
canceledit
Cancels a task, if it can be cancelled through an API.
client.tasks.cancel({ ... })
Argumentsedit
-
Request (object):
-
task_id
(Optional, string | number): Cancel the task with specified task id (node_id:task_number) -
actions
(Optional, string | string[]): A list of actions that should be cancelled. Leave empty to cancel all. -
nodes
(Optional, string[]): A list of node IDs or names to limit the returned information; use_local
to return information from the node you’re connecting to, leave empty to get information from all nodes -
parent_task_id
(Optional, string): Cancel tasks with specified parent task id (node_id:task_number). Set to -1 to cancel all. -
wait_for_completion
(Optional, boolean): Should the request block until the cancellation of the task and its descendant tasks is completed. Defaults to false
-
getedit
Returns information about a task.
client.tasks.get({ task_id })
Argumentsedit
-
Request (object):
-
task_id
(string): Return the task with specified id (node_id:task_number) -
timeout
(Optional, string | -1 | 0): Explicit operation timeout -
wait_for_completion
(Optional, boolean): Wait for the matching tasks to complete (default: false)
-
listedit
Returns a list of tasks.
client.tasks.list({ ... })
Argumentsedit
-
Request (object):
-
actions
(Optional, string | string[]): List or wildcard expression of actions used to limit the request. -
detailed
(Optional, boolean): Iftrue
, the response includes detailed information about shard recoveries. -
group_by
(Optional, Enum("nodes" | "parents" | "none")): Key used to group tasks in the response. -
node_id
(Optional, string[]): List of node IDs or names used to limit returned information. -
parent_task_id
(Optional, string): Parent task ID used to limit returned information. To return all tasks, omit this parameter or use a value of-1
. -
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error. -
wait_for_completion
(Optional, boolean): Iftrue
, the request blocks until the operation is complete.
-
text_structureedit
find_structureedit
Finds the structure of a text file. The text file must contain data that is suitable to be ingested into Elasticsearch.
client.textStructure.findStructure({ ... })
Argumentsedit
-
Request (object):
-
text_files
(Optional, TJsonDocument[]) -
charset
(Optional, string): The text’s character set. It must be a character set that is supported by the JVM that Elasticsearch uses. For example, UTF-8, UTF-16LE, windows-1252, or EUC-JP. If this parameter is not specified, the structure finder chooses an appropriate character set. -
column_names
(Optional, string): If you have set format to delimited, you can specify the column names in a list. If this parameter is not specified, the structure finder uses the column names from the header row of the text. If the text does not have a header role, columns are named "column1", "column2", "column3", etc. -
delimiter
(Optional, string): If you have set format to delimited, you can specify the character used to delimit the values in each row. Only a single character is supported; the delimiter cannot have multiple characters. By default, the API considers the following possibilities: comma, tab, semi-colon, and pipe (|). In this default scenario, all rows must have the same number of fields for the delimited format to be detected. If you specify a delimiter, up to 10% of the rows can have a different number of columns than the first row. -
explain
(Optional, boolean): If this parameter is set to true, the response includes a field named explanation, which is an array of strings that indicate how the structure finder produced its result. -
format
(Optional, string): The high level structure of the text. Valid values are ndjson, xml, delimited, and semi_structured_text. By default, the API chooses the format. In this default scenario, all rows must have the same number of fields for a delimited format to be detected. If the format is set to delimited and the delimiter is not set, however, the API tolerates up to 5% of rows that have a different number of columns than the first row. -
grok_pattern
(Optional, string): If you have set format to semi_structured_text, you can specify a Grok pattern that is used to extract fields from every message in the text. The name of the timestamp field in the Grok pattern must match what is specified in the timestamp_field parameter. If that parameter is not specified, the name of the timestamp field in the Grok pattern must match "timestamp". If grok_pattern is not specified, the structure finder creates a Grok pattern. -
has_header_row
(Optional, boolean): If you have set format to delimited, you can use this parameter to indicate whether the column names are in the first row of the text. If this parameter is not specified, the structure finder guesses based on the similarity of the first row of the text to other rows. -
line_merge_size_limit
(Optional, number): The maximum number of characters in a message when lines are merged to form messages while analyzing semi-structured text. If you have extremely long messages you may need to increase this, but be aware that this may lead to very long processing times if the way to group lines into messages is misdetected. -
lines_to_sample
(Optional, number): The number of lines to include in the structural analysis, starting from the beginning of the text. The minimum is 2; If the value of this parameter is greater than the number of lines in the text, the analysis proceeds (as long as there are at least two lines in the text) for all of the lines. -
quote
(Optional, string): If you have set format to delimited, you can specify the character used to quote the values in each row if they contain newlines or the delimiter character. Only a single character is supported. If this parameter is not specified, the default value is a double quote ("). If your delimited text format does not use quoting, a workaround is to set this argument to a character that does not appear anywhere in the sample. -
should_trim_fields
(Optional, boolean): If you have set format to delimited, you can specify whether values between delimiters should have whitespace trimmed from them. If this parameter is not specified and the delimiter is pipe (|), the default value is true. Otherwise, the default value is false. -
timeout
(Optional, string | -1 | 0): Sets the maximum amount of time that the structure analysis make take. If the analysis is still running when the timeout expires then it will be aborted. -
timestamp_field
(Optional, string): Optional parameter to specify the timestamp field in the file -
timestamp_format
(Optional, string): The Java time format of the timestamp field in the text.
-
transformedit
delete_transformedit
Deletes an existing transform.
client.transform.deleteTransform({ transform_id })
Argumentsedit
-
Request (object):
-
transform_id
(string): Identifier for the transform. -
force
(Optional, boolean): If this value is false, the transform must be stopped before it can be deleted. If true, the transform is deleted regardless of its current state. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
get_transformedit
Retrieves configuration information for transforms.
client.transform.getTransform({ ... })
Argumentsedit
-
Request (object):
-
transform_id
(Optional, string | string[]): Identifier for the transform. It can be a transform identifier or a wildcard expression. You can get information for all transforms by using_all
, by specifying*
as the<transform_id>
, or by omitting the<transform_id>
. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no transforms that match.
- Contains the _all string or no identifiers and there are no matches.
- Contains wildcard expressions and there are only partial matches.
-
If this parameter is false, the request returns a 404 status code when
there are no matches or only partial matches.
from
(Optional, number): Skips the specified number of transforms.
size
(Optional, number): Specifies the maximum number of transforms to obtain.
* *exclude_generated
(Optional, boolean): Excludes fields that were automatically added when creating the
transform. This allows the configuration to be in an acceptable format to
be retrieved and then added to another cluster.
get_transform_statsedit
Retrieves usage information for transforms.
client.transform.getTransformStats({ transform_id })
Argumentsedit
-
Request (object):
-
transform_id
(string | string[]): Identifier for the transform. It can be a transform identifier or a wildcard expression. You can get information for all transforms by using_all
, by specifying*
as the<transform_id>
, or by omitting the<transform_id>
. -
allow_no_match
(Optional, boolean): Specifies what to do when the request:- Contains wildcard expressions and there are no transforms that match.
- Contains the _all string or no identifiers and there are no matches.
- Contains wildcard expressions and there are only partial matches.
-
If this parameter is false, the request returns a 404 status code when
there are no matches or only partial matches.
from
(Optional, number): Skips the specified number of transforms.
size
(Optional, number): Specifies the maximum number of transforms to obtain.
* *timeout
(Optional, string | -1 | 0): Controls the time to wait for the stats
preview_transformedit
Previews a transform.
client.transform.previewTransform({ ... })
Argumentsedit
-
Request (object):
-
transform_id
(Optional, string): Identifier for the transform to preview. If you specify this path parameter, you cannot provide transform configuration details in the request body. -
dest
(Optional, { index, op_type, pipeline, routing, version_type }): The destination for the transform. -
description
(Optional, string): Free text description of the transform. -
frequency
(Optional, string | -1 | 0): The interval between checks for changes in the source indices when the transform is running continuously. Also determines the retry interval in the event of transient failures while the transform is searching or indexing. The minimum value is 1s and the maximum is 1h. -
pivot
(Optional, { aggregations, group_by }): The pivot method transforms the data by aggregating and grouping it. These objects define the group by fields and the aggregation to reduce the data. -
source
(Optional, { index, query, remote, size, slice, sort, _source, runtime_mappings }): The source of the data for the transform. -
settings
(Optional, { align_checkpoints, dates_as_epoch_millis, deduce_mappings, docs_per_second, max_page_search_size, unattended }): Defines optional transform settings. -
sync
(Optional, { time }): Defines the properties transforms require to run continuously. -
retention_policy
(Optional, { time }): Defines a retention policy for the transform. Data that meets the defined criteria is deleted from the destination index. -
latest
(Optional, { sort, unique_key }): The latest method transforms the data by finding the latest document for each unique key. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
put_transformedit
Instantiates a transform.
client.transform.putTransform({ transform_id, dest, source })
Argumentsedit
-
Request (object):
-
transform_id
(string): Identifier for the transform. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It has a 64 character limit and must start and end with alphanumeric characters. -
dest
({ index, op_type, pipeline, routing, version_type }): The destination for the transform. -
source
({ index, query, remote, size, slice, sort, _source, runtime_mappings }): The source of the data for the transform. -
description
(Optional, string): Free text description of the transform. -
frequency
(Optional, string | -1 | 0): The interval between checks for changes in the source indices when the transform is running continuously. Also determines the retry interval in the event of transient failures while the transform is searching or indexing. The minimum value is1s
and the maximum is1h
. -
latest
(Optional, { sort, unique_key }): The latest method transforms the data by finding the latest document for each unique key. -
_meta
(Optional, Record<string, User-defined value>): Defines optional transform metadata. -
pivot
(Optional, { aggregations, group_by }): The pivot method transforms the data by aggregating and grouping it. These objects define the group by fields and the aggregation to reduce the data. -
retention_policy
(Optional, { time }): Defines a retention policy for the transform. Data that meets the defined criteria is deleted from the destination index. -
settings
(Optional, { align_checkpoints, dates_as_epoch_millis, deduce_mappings, docs_per_second, max_page_search_size, unattended }): Defines optional transform settings. -
sync
(Optional, { time }): Defines the properties transforms require to run continuously. -
defer_validation
(Optional, boolean): When the transform is created, a series of validations occur to ensure its success. For example, there is a check for the existence of the source indices and a check that the destination index is not part of the source index pattern. You can use this parameter to skip the checks, for example when the source index does not exist until after the transform is created. The validations are always run when you start the transform, however, with the exception of privilege checks. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
reset_transformedit
Resets an existing transform.
client.transform.resetTransform({ transform_id })
Argumentsedit
-
Request (object):
-
transform_id
(string): Identifier for the transform. This identifier can contain lowercase alphanumeric characters (a-z and 0-9), hyphens, and underscores. It has a 64 character limit and must start and end with alphanumeric characters. -
force
(Optional, boolean): If this value istrue
, the transform is reset regardless of its current state. If it’sfalse
, the transform must be stopped before it can be reset.
-
schedule_now_transformedit
Schedules now a transform.
client.transform.scheduleNowTransform({ transform_id })
Argumentsedit
-
Request (object):
-
transform_id
(string): Identifier for the transform. -
timeout
(Optional, string | -1 | 0): Controls the time to wait for the scheduling to take place
-
start_transformedit
Starts one or more transforms.
client.transform.startTransform({ transform_id })
Argumentsedit
-
Request (object):
-
transform_id
(string): Identifier for the transform. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error. -
from
(Optional, string): Restricts the set of transformed entities to those changed after this time. Relative times like now-30d are supported. Only applicable for continuous transforms.
-
stop_transformedit
Stops one or more transforms.
client.transform.stopTransform({ transform_id })
Argumentsedit
-
Request (object):
-
transform_id
(string): Identifier for the transform. To stop multiple transforms, use a list or a wildcard expression. To stop all transforms, use_all
or*
as the identifier. -
allow_no_match
(Optional, boolean): Specifies what to do when the request: contains wildcard expressions and there are no transforms that match; contains the_all
string or no identifiers and there are no matches; contains wildcard expressions and there are only partial matches.
-
If it is true, the API returns a successful acknowledgement message when there are no matches. When there are only partial matches, the API stops the appropriate transforms.
If it is false, the request returns a 404 status code when there are no matches or only partial matches.
force
(Optional, boolean): If it is true, the API forcefully stops the transforms.
timeout
(Optional, string | -1 | 0): Period to wait for a response when wait_for_completion
is true
. If no response is received before the
timeout expires, the request returns a timeout exception. However, the request continues processing and
eventually moves the transform to a STOPPED state.
wait_for_checkpoint
(Optional, boolean): If it is true, the transform does not completely stop until the current checkpoint is completed. If it is false,
the transform stops as soon as possible.
wait_for_completion
(Optional, boolean): If it is true, the API blocks until the indexer state completely stops. If it is false, the API returns
immediately and the indexer is stopped asynchronously in the background.
update_transformedit
Updates certain properties of a transform.
client.transform.updateTransform({ transform_id })
Argumentsedit
-
Request (object):
-
transform_id
(string): Identifier for the transform. -
dest
(Optional, { index, op_type, pipeline, routing, version_type }): The destination for the transform. -
description
(Optional, string): Free text description of the transform. -
frequency
(Optional, string | -1 | 0): The interval between checks for changes in the source indices when the transform is running continuously. Also determines the retry interval in the event of transient failures while the transform is searching or indexing. The minimum value is 1s and the maximum is 1h. -
_meta
(Optional, Record<string, User-defined value>): Defines optional transform metadata. -
source
(Optional, { index, query, remote, size, slice, sort, _source, runtime_mappings }): The source of the data for the transform. -
settings
(Optional, { align_checkpoints, dates_as_epoch_millis, deduce_mappings, docs_per_second, max_page_search_size, unattended }): Defines optional transform settings. -
sync
(Optional, { time }): Defines the properties transforms require to run continuously. -
retention_policy
(Optional, { time } | null): Defines a retention policy for the transform. Data that meets the defined criteria is deleted from the destination index. -
defer_validation
(Optional, boolean): When true, deferrable validations are not run. This behavior may be desired if the source index does not exist until after the transform is created. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
upgrade_transformsedit
Upgrades all transforms.
client.transform.upgradeTransforms({ ... })
Argumentsedit
-
Request (object):
-
dry_run
(Optional, boolean): When true, the request checks for updates but does not run them. -
timeout
(Optional, string | -1 | 0): Period to wait for a response. If no response is received before the timeout expires, the request fails and returns an error.
-
watcheredit
ack_watchedit
Acknowledges a watch, manually throttling the execution of the watch’s actions.
client.watcher.ackWatch({ watch_id })
Argumentsedit
-
Request (object):
-
watch_id
(string): Watch ID -
action_id
(Optional, string | string[]): A list of the action ids to be acked
-
activate_watchedit
Activates a currently inactive watch.
client.watcher.activateWatch({ watch_id })
Argumentsedit
-
Request (object):
-
watch_id
(string): Watch ID
-
deactivate_watchedit
Deactivates a currently active watch.
client.watcher.deactivateWatch({ watch_id })
Argumentsedit
-
Request (object):
-
watch_id
(string): Watch ID
-
delete_watchedit
Removes a watch from Watcher.
client.watcher.deleteWatch({ id })
Argumentsedit
-
Request (object):
-
id
(string): Watch ID
-
execute_watchedit
Forces the execution of a stored watch.
client.watcher.executeWatch({ ... })
Argumentsedit
-
Request (object):
-
id
(Optional, string): Identifier for the watch. -
action_modes
(Optional, Record<string, Enum("simulate" | "force_simulate" | "execute" | "force_execute" | "skip")>): Determines how to handle the watch actions as part of the watch execution. -
alternative_input
(Optional, Record<string, User-defined value>): When present, the watch uses this object as a payload instead of executing its own input. -
ignore_condition
(Optional, boolean): When set totrue
, the watch execution uses the always condition. This can also be specified as an HTTP parameter. -
record_execution
(Optional, boolean): When set totrue
, the watch record representing the watch execution result is persisted to the.watcher-history
index for the current time. In addition, the status of the watch is updated, possibly throttling subsequent executions. This can also be specified as an HTTP parameter. -
simulated_actions
(Optional, { actions, all, use_all }) -
trigger_data
(Optional, { scheduled_time, triggered_time }): This structure is parsed as the data of the trigger event that will be used during the watch execution -
watch
(Optional, { actions, condition, input, metadata, status, throttle_period, throttle_period_in_millis, transform, trigger }): When present, this watch is used instead of the one specified in the request. This watch is not persisted to the index and record_execution cannot be set. -
debug
(Optional, boolean): Defines whether the watch runs in debug mode.
-
get_settingsedit
Retrieve settings for the watcher system index
client.watcher.getSettings()
get_watchedit
Retrieves a watch by its ID.
client.watcher.getWatch({ id })
Argumentsedit
-
Request (object):
-
id
(string): Watch ID
-
put_watchedit
Creates a new watch, or updates an existing one.
client.watcher.putWatch({ id })
Argumentsedit
-
Request (object):
-
id
(string): Watch ID -
actions
(Optional, Record<string, { add_backing_index, remove_backing_index }>) -
condition
(Optional, { always, array_compare, compare, never, script }) -
input
(Optional, { chain, http, search, simple }) -
metadata
(Optional, Record<string, User-defined value>) -
throttle_period
(Optional, string) -
transform
(Optional, { chain, script, search }) -
trigger
(Optional, { schedule }) -
active
(Optional, boolean): Specify whether the watch is in/active by default -
if_primary_term
(Optional, number): only update the watch if the last operation that has changed the watch has the specified primary term -
if_seq_no
(Optional, number): only update the watch if the last operation that has changed the watch has the specified sequence number -
version
(Optional, number): Explicit version number for concurrency control
-
query_watchesedit
Retrieves stored watches.
client.watcher.queryWatches({ ... })
Argumentsedit
-
Request (object):
-
from
(Optional, number): The offset from the first result to fetch. Needs to be non-negative. -
size
(Optional, number): The number of hits to return. Needs to be non-negative. -
query
(Optional, { bool, boosting, common, combined_fields, constant_score, dis_max, distance_feature, exists, function_score, fuzzy, geo_bounding_box, geo_distance, geo_polygon, geo_shape, has_child, has_parent, ids, intervals, match, match_all, match_bool_prefix, match_none, match_phrase, match_phrase_prefix, more_like_this, multi_match, nested, parent_id, percolate, pinned, prefix, query_string, range, rank_feature, regexp, script, script_score, shape, simple_query_string, span_containing, field_masking_span, span_first, span_multi, span_near, span_not, span_or, span_term, span_within, term, terms, terms_set, text_expansion, wildcard, wrapper, type }): Optional, query filter watches to be returned. -
sort
(Optional, string | { _score, _doc, _geo_distance, _script } | string | { _score, _doc, _geo_distance, _script }[]): Optional sort definition. -
search_after
(Optional, number | number | string | boolean | null | User-defined value[]): Optional search After to do pagination using last hit’s sort values.
-
startedit
Starts Watcher if it is not already running.
client.watcher.start()
statsedit
Retrieves the current Watcher metrics.
client.watcher.stats({ ... })
Argumentsedit
-
Request (object):
-
metric
(Optional, Enum("_all" | "queued_watches" | "current_watches" | "pending_watches") | Enum("_all" | "queued_watches" | "current_watches" | "pending_watches")[]): Defines which additional metrics are included in the response. -
emit_stacktraces
(Optional, boolean): Defines whether stack traces are generated for each watch that is running.
-
stopedit
Stops Watcher if it is running.
client.watcher.stop()
update_settingsedit
Update settings for the watcher system index
client.watcher.updateSettings()
xpackedit
infoedit
Retrieves information about the installed X-Pack features.
client.xpack.info({ ... })
Argumentsedit
-
Request (object):
-
categories
(Optional, string[]): A list of the information categories to include in the response. For example,build,license,features
. -
accept_enterprise
(Optional, boolean): If this param is used it must be set to true -
human
(Optional, boolean): Defines whether additional human-readable information is included in the response. In particular, it adds descriptions and a tag line.
-
usageedit
Retrieves usage information about the installed X-Pack features.
client.xpack.usage({ ... })
Argumentsedit
-
Request (object):
-
master_timeout
(Optional, string | -1 | 0): Period to wait for a connection to the master node. If no response is received before the timeout expires, the request fails and returns an error.
-