OpenSearch/docs/reference/analysis
Nik Everett f9dc86836d
Docs: Test examples that recreate lang analyzers (#29535)
We have a pile of documentation describing how to rebuild the built in
language analyzers and, previously, our documentation testing framework
made sure that the examples successfully built *an* analyzer but they
didn't assert that the analyzer built by the documentation matches the
built in anlayzer. Unsuprisingly, some of the examples aren't quite
right.

This adds a mechanism that tests that the analyzers built by the docs.
The mechanism is fairly simple and brutal but it seems to be working:
build a hundred random unicode sequences and send them through the
`_analyze` API with the rebuilt analyzer and then again through the
built in analyzer. Then make sure both APIs return the same results.
Each of these calls to `_anlayze` takes about 20ms on my laptop which
seems fine.
2018-05-09 09:23:10 -04:00
..
analyzers Docs: Test examples that recreate lang analyzers (#29535) 2018-05-09 09:23:10 -04:00
charfilters Allow `_doc` as a type. (#27816) 2017-12-14 17:47:53 +01:00
tokenfilters Mark synonym_graph as beta in the docs (#28496) 2018-02-02 16:33:48 +01:00
tokenizers Allow `_doc` as a type. (#27816) 2017-12-14 17:47:53 +01:00
analyzers.asciidoc First pass at improving analyzer docs (#18269) 2016-05-11 14:17:56 +02:00
anatomy.asciidoc Correction of the names of numirals (#21531) 2016-11-25 14:30:49 +01:00
charfilters.asciidoc Hindu-Arabico-Latino Numerals (#22476) 2017-01-10 15:24:56 +01:00
normalizers.asciidoc [DOCS] Add supported token filters 2018-02-13 14:10:25 -08:00
testing.asciidoc Allow `_doc` as a type. (#27816) 2017-12-14 17:47:53 +01:00
tokenfilters.asciidoc Add missing link for the WordDelimiterGraphFilter 2017-04-28 17:12:38 +02:00
tokenizers.asciidoc Rename simple pattern tokenizers (#25300) 2017-06-19 13:48:43 -07:00