IMPORTANT: No additional bug fixes or documentation updates will be released for this version. For the latest information, see the current release documentation.
Token filters accept a stream of tokens from a tokenizer and can modify tokens (eg lowercasing), delete tokens (eg remove stopwords) or add tokens (eg synonyms).
Elasticsearch has a number of built-in token filters you can use to build custom analyzers.
Intro to Kibana
ELK for Logs & Metrics