[DOCS] Correct custom analyzer callouts (#46030)
This commit is contained in:
parent
322d95f2f6
commit
3e62cf9d74
|
@ -162,26 +162,26 @@ PUT my_index
|
||||||
"settings": {
|
"settings": {
|
||||||
"analysis": {
|
"analysis": {
|
||||||
"analyzer": {
|
"analyzer": {
|
||||||
"my_custom_analyzer": {
|
"my_custom_analyzer": { <1>
|
||||||
"type": "custom",
|
"type": "custom",
|
||||||
"char_filter": [
|
"char_filter": [
|
||||||
"emoticons" <1>
|
"emoticons"
|
||||||
],
|
],
|
||||||
"tokenizer": "punctuation", <1>
|
"tokenizer": "punctuation",
|
||||||
"filter": [
|
"filter": [
|
||||||
"lowercase",
|
"lowercase",
|
||||||
"english_stop" <1>
|
"english_stop"
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"tokenizer": {
|
"tokenizer": {
|
||||||
"punctuation": { <1>
|
"punctuation": { <2>
|
||||||
"type": "pattern",
|
"type": "pattern",
|
||||||
"pattern": "[ .,!?]"
|
"pattern": "[ .,!?]"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"char_filter": {
|
"char_filter": {
|
||||||
"emoticons": { <1>
|
"emoticons": { <3>
|
||||||
"type": "mapping",
|
"type": "mapping",
|
||||||
"mappings": [
|
"mappings": [
|
||||||
":) => _happy_",
|
":) => _happy_",
|
||||||
|
@ -190,7 +190,7 @@ PUT my_index
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"filter": {
|
"filter": {
|
||||||
"english_stop": { <1>
|
"english_stop": { <4>
|
||||||
"type": "stop",
|
"type": "stop",
|
||||||
"stopwords": "_english_"
|
"stopwords": "_english_"
|
||||||
}
|
}
|
||||||
|
@ -207,9 +207,12 @@ POST my_index/_analyze
|
||||||
--------------------------------------------------
|
--------------------------------------------------
|
||||||
// CONSOLE
|
// CONSOLE
|
||||||
|
|
||||||
<1> The `emoticons` character filter, `punctuation` tokenizer and
|
<1> Assigns the index a default custom analyzer, `my_custom_analyzer`. This
|
||||||
`english_stop` token filter are custom implementations which are defined
|
analyzer uses a custom tokenizer, character filter, and token filter that
|
||||||
in the same index settings.
|
are defined later in the request.
|
||||||
|
<2> Defines the custom `punctuation` tokenizer.
|
||||||
|
<3> Defines the custom `emoticons` character filter.
|
||||||
|
<4> Defines the custom `english_stop` token filter.
|
||||||
|
|
||||||
/////////////////////
|
/////////////////////
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue