Update index.md (#6223)
Signed-off-by: kolchfa-aws <105444904+kolchfa-aws@users.noreply.github.com>
This commit is contained in:
parent
49767585c4
commit
05e44c0730
|
@ -14,7 +14,7 @@ The output of a tokenizer is a stream of tokens. Tokenizers also maintain the fo
|
||||||
|
|
||||||
- The **order** or **position** of each token: This information is used for word and phrase proximity queries.
|
- The **order** or **position** of each token: This information is used for word and phrase proximity queries.
|
||||||
- The starting and ending positions (**offsets**) of the tokens in the text: This information is used for highlighting search terms.
|
- The starting and ending positions (**offsets**) of the tokens in the text: This information is used for highlighting search terms.
|
||||||
- The token **type**: Some tokenizers (for example, `standard`) classify tokens by type, for example, <ALPHANUM> or <NUM>. Simpler tokenizers (for example, `letter`) only classify tokens as type `word`.
|
- The token **type**: Some tokenizers (for example, `standard`) classify tokens by type, for example, `<ALPHANUM>` or `<NUM>`. Simpler tokenizers (for example, `letter`) only classify tokens as type `word`.
|
||||||
|
|
||||||
You can use tokenizers to define custom analyzers.
|
You can use tokenizers to define custom analyzers.
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue