Elision Token Filteredit

A token filter which removes elisions. For example, "l’avion" (the plane) will tokenized as "avion" (plane).

Requires either an articles parameter which is a set of stop word articles, or articles_path which points to a text file containing the stop set. Also optionally accepts articles_case, which indicates whether the filter treats those articles as case sensitive.

For example:

PUT /elision_example
{
    "settings" : {
        "analysis" : {
            "analyzer" : {
                "default" : {
                    "tokenizer" : "standard",
                    "filter" : ["elision"]
                }
            },
            "filter" : {
                "elision" : {
                    "type" : "elision",
                    "articles_case": true,
                    "articles" : ["l", "m", "t", "qu", "n", "s", "j"]
                }
            }
        }
    }
}