2013-08-28 19:24:34 -04:00
|
|
|
[[analysis-standard-tokenizer]]
|
|
|
|
=== Standard Tokenizer
|
|
|
|
|
|
|
|
A tokenizer of type `standard` providing grammar based tokenizer that is
|
|
|
|
a good tokenizer for most European language documents. The tokenizer
|
|
|
|
implements the Unicode Text Segmentation algorithm, as specified in
|
|
|
|
http://unicode.org/reports/tr29/[Unicode Standard Annex #29].
|
|
|
|
|
|
|
|
The following are settings that can be set for a `standard` tokenizer
|
|
|
|
type:
|
|
|
|
|
|
|
|
[cols="<,<",options="header",]
|
|
|
|
|=======================================================================
|
|
|
|
|Setting |Description
|
|
|
|
|`max_token_length` |The maximum token length. If a token is seen that
|
2016-03-18 05:57:16 -04:00
|
|
|
exceeds this length then it is split at `max_token_length` intervals. Defaults to `255`.
|
2013-08-28 19:24:34 -04:00
|
|
|
|=======================================================================
|
|
|
|
|