The index analysis module acts as a configurable registry of Analyzers
that can be used in order to break down indexed (analyzed) fields when a
document is indexed as well as to process query strings. It maps to the Lucene
Analyzers are (generally) composed of a single
Tokenizer and zero or
TokenFilters. A set of
CharFilters can be associated with an
analyzer to process the characters prior to other analysis steps. The
analysis module allows one to register
Analyzers under logical names that can then be referenced either in
mapping definitions or in certain APIs. The Analysis module
automatically registers (if not explicitly defined) built in
analyzers, token filters, and tokenizers.
See Analysis for configuration details.