OpenSearch/docs/reference/analysis/tokenizers/thai-tokenizer.asciidoc

10 lines
304 B
Plaintext

[[analysis-thai-tokenizer]]
=== Thai Tokenizer
added[1.3.0]
A tokenizer of type `thai` that segments Thai text into words. This tokenizer
uses the built-in Thai segmentation algorithm included with Java to divide
up Thai text. Text in other languages in general will be treated the same
as `standard`.