2013-08-28 19:24:34 -04:00
|
|
|
[[indices-analyze]]
|
|
|
|
== Analyze
|
|
|
|
|
|
|
|
Performs the analysis process on a text and return the tokens breakdown
|
|
|
|
of the text.
|
|
|
|
|
|
|
|
Can be used without specifying an index against one of the many built in
|
|
|
|
analyzers:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
GET _analyze
|
2015-04-02 21:51:15 -04:00
|
|
|
{
|
|
|
|
"analyzer" : "standard",
|
|
|
|
"text" : "this is a test"
|
2017-02-07 14:15:09 -05:00
|
|
|
}
|
2013-08-28 19:24:34 -04:00
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
// CONSOLE
|
2013-08-28 19:24:34 -04:00
|
|
|
|
2015-04-26 21:55:21 -04:00
|
|
|
If text parameter is provided as array of strings, it is analyzed as a multi-valued field.
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
GET _analyze
|
2015-04-26 21:55:21 -04:00
|
|
|
{
|
|
|
|
"analyzer" : "standard",
|
|
|
|
"text" : ["this is a test", "the second text"]
|
2017-02-07 14:15:09 -05:00
|
|
|
}
|
2015-04-26 21:55:21 -04:00
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
// CONSOLE
|
2015-04-26 21:55:21 -04:00
|
|
|
|
2014-02-17 23:25:12 -05:00
|
|
|
Or by building a custom transient analyzer out of tokenizers,
|
2016-04-07 12:58:10 -04:00
|
|
|
token filters and char filters. Token filters can use the shorter 'filter'
|
2014-02-17 23:25:12 -05:00
|
|
|
parameter name:
|
2013-08-28 19:24:34 -04:00
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
GET _analyze
|
2015-04-02 21:51:15 -04:00
|
|
|
{
|
|
|
|
"tokenizer" : "keyword",
|
2016-04-07 12:58:10 -04:00
|
|
|
"filter" : ["lowercase"],
|
2015-04-02 21:51:15 -04:00
|
|
|
"text" : "this is a test"
|
2017-02-07 14:15:09 -05:00
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
2014-02-17 23:25:12 -05:00
|
|
|
|
2017-02-07 14:15:09 -05:00
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
GET _analyze
|
2015-04-02 21:51:15 -04:00
|
|
|
{
|
|
|
|
"tokenizer" : "keyword",
|
2016-09-01 11:04:13 -04:00
|
|
|
"filter" : ["lowercase"],
|
2016-04-07 12:58:10 -04:00
|
|
|
"char_filter" : ["html_strip"],
|
2015-04-02 21:51:15 -04:00
|
|
|
"text" : "this is a <b>test</b>"
|
2017-02-07 14:15:09 -05:00
|
|
|
}
|
2013-08-28 19:24:34 -04:00
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
// CONSOLE
|
2013-08-28 19:24:34 -04:00
|
|
|
|
2016-09-01 11:04:13 -04:00
|
|
|
deprecated[5.0.0, Use `filter`/`char_filter` instead of `filters`/`char_filters` and `token_filters` has been removed]
|
2016-04-07 12:58:10 -04:00
|
|
|
|
2015-09-28 06:06:47 -04:00
|
|
|
Custom tokenizers, token filters, and character filters can be specified in the request body as follows:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
GET _analyze
|
2015-09-28 06:06:47 -04:00
|
|
|
{
|
|
|
|
"tokenizer" : "whitespace",
|
|
|
|
"filter" : ["lowercase", {"type": "stop", "stopwords": ["a", "is", "this"]}],
|
|
|
|
"text" : "this is a test"
|
2017-02-07 14:15:09 -05:00
|
|
|
}
|
2015-09-28 06:06:47 -04:00
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
// CONSOLE
|
2015-09-28 06:06:47 -04:00
|
|
|
|
2013-08-28 19:24:34 -04:00
|
|
|
It can also run against a specific index:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2017-07-04 06:16:56 -04:00
|
|
|
GET analyze_sample/_analyze
|
2015-04-02 21:51:15 -04:00
|
|
|
{
|
|
|
|
"text" : "this is a test"
|
2017-02-07 14:15:09 -05:00
|
|
|
}
|
2013-08-28 19:24:34 -04:00
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
// CONSOLE
|
2017-07-04 06:16:56 -04:00
|
|
|
// TEST[setup:analyze_sample]
|
2013-08-28 19:24:34 -04:00
|
|
|
|
|
|
|
The above will run an analysis on the "this is a test" text, using the
|
2017-07-04 06:16:56 -04:00
|
|
|
default index analyzer associated with the `analyze_sample` index. An `analyzer`
|
2013-08-28 19:24:34 -04:00
|
|
|
can also be provided to use a different analyzer:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2017-07-04 06:16:56 -04:00
|
|
|
GET analyze_sample/_analyze
|
2015-04-02 21:51:15 -04:00
|
|
|
{
|
|
|
|
"analyzer" : "whitespace",
|
2016-07-11 09:49:39 -04:00
|
|
|
"text" : "this is a test"
|
2017-02-07 14:15:09 -05:00
|
|
|
}
|
2013-08-28 19:24:34 -04:00
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
// CONSOLE
|
2017-07-04 06:16:56 -04:00
|
|
|
// TEST[setup:analyze_sample]
|
2013-08-28 19:24:34 -04:00
|
|
|
|
|
|
|
Also, the analyzer can be derived based on a field mapping, for example:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2017-07-04 06:16:56 -04:00
|
|
|
GET analyze_sample/_analyze
|
2015-04-02 21:51:15 -04:00
|
|
|
{
|
|
|
|
"field" : "obj1.field1",
|
|
|
|
"text" : "this is a test"
|
2017-02-07 14:15:09 -05:00
|
|
|
}
|
2013-08-28 19:24:34 -04:00
|
|
|
--------------------------------------------------
|
2017-02-07 14:15:09 -05:00
|
|
|
// CONSOLE
|
2017-07-04 06:16:56 -04:00
|
|
|
// TEST[setup:analyze_sample]
|
2013-08-28 19:24:34 -04:00
|
|
|
|
2014-03-07 08:21:45 -05:00
|
|
|
Will cause the analysis to happen based on the analyzer configured in the
|
2013-08-28 19:24:34 -04:00
|
|
|
mapping for `obj1.field1` (and if not, the default index analyzer).
|
|
|
|
|
2017-07-04 06:16:56 -04:00
|
|
|
A `normalizer` can be provided for keyword field with normalizer associated with the `analyze_sample` index.
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
GET analyze_sample/_analyze
|
|
|
|
{
|
|
|
|
"normalizer" : "my_normalizer",
|
|
|
|
"text" : "BaR"
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
// TEST[setup:analyze_sample]
|
|
|
|
|
|
|
|
Or by building a custom transient normalizer out of token filters and char filters.
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
GET _analyze
|
|
|
|
{
|
|
|
|
"filter" : ["lowercase"],
|
|
|
|
"text" : "BaR"
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
|
|
|
// CONSOLE
|
|
|
|
|
2015-06-15 03:32:44 -04:00
|
|
|
=== Explain Analyze
|
|
|
|
|
|
|
|
If you want to get more advanced details, set `explain` to `true` (defaults to `false`). It will output all token attributes for each token.
|
|
|
|
You can filter token attributes you want to output by setting `attributes` option.
|
|
|
|
|
|
|
|
experimental[The format of the additional detail information is experimental and can change at any time]
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
2016-04-29 10:42:03 -04:00
|
|
|
GET _analyze
|
2015-06-15 03:32:44 -04:00
|
|
|
{
|
|
|
|
"tokenizer" : "standard",
|
2016-08-02 17:35:31 -04:00
|
|
|
"filter" : ["snowball"],
|
2015-06-15 03:32:44 -04:00
|
|
|
"text" : "detailed output",
|
|
|
|
"explain" : true,
|
|
|
|
"attributes" : ["keyword"] <1>
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
2016-05-09 09:42:23 -04:00
|
|
|
// CONSOLE
|
2015-06-15 03:32:44 -04:00
|
|
|
<1> Set "keyword" to output "keyword" attribute only
|
|
|
|
|
|
|
|
The request returns the following result:
|
|
|
|
|
|
|
|
[source,js]
|
|
|
|
--------------------------------------------------
|
|
|
|
{
|
|
|
|
"detail" : {
|
|
|
|
"custom_analyzer" : true,
|
|
|
|
"charfilters" : [ ],
|
|
|
|
"tokenizer" : {
|
|
|
|
"name" : "standard",
|
|
|
|
"tokens" : [ {
|
|
|
|
"token" : "detailed",
|
|
|
|
"start_offset" : 0,
|
|
|
|
"end_offset" : 8,
|
|
|
|
"type" : "<ALPHANUM>",
|
|
|
|
"position" : 0
|
|
|
|
}, {
|
|
|
|
"token" : "output",
|
|
|
|
"start_offset" : 9,
|
|
|
|
"end_offset" : 15,
|
|
|
|
"type" : "<ALPHANUM>",
|
|
|
|
"position" : 1
|
|
|
|
} ]
|
|
|
|
},
|
|
|
|
"tokenfilters" : [ {
|
|
|
|
"name" : "snowball",
|
|
|
|
"tokens" : [ {
|
|
|
|
"token" : "detail",
|
|
|
|
"start_offset" : 0,
|
|
|
|
"end_offset" : 8,
|
|
|
|
"type" : "<ALPHANUM>",
|
|
|
|
"position" : 0,
|
|
|
|
"keyword" : false <1>
|
|
|
|
}, {
|
|
|
|
"token" : "output",
|
|
|
|
"start_offset" : 9,
|
|
|
|
"end_offset" : 15,
|
|
|
|
"type" : "<ALPHANUM>",
|
|
|
|
"position" : 1,
|
|
|
|
"keyword" : false <1>
|
|
|
|
} ]
|
|
|
|
} ]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
--------------------------------------------------
|
2016-09-01 13:05:22 -04:00
|
|
|
// TESTRESPONSE
|
2015-06-15 03:32:44 -04:00
|
|
|
<1> Output only "keyword" attribute, since specify "attributes" in the request.
|