diff --git a/docs/reference/analysis.asciidoc b/docs/reference/analysis.asciidoc index b5ac7883a5c..10a1cc6a89f 100644 --- a/docs/reference/analysis.asciidoc +++ b/docs/reference/analysis.asciidoc @@ -1,6 +1,8 @@ [[analysis]] = Text analysis +:lucene-analysis-docs: https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis + [partintro] -- diff --git a/docs/reference/analysis/tokenfilters/apostrophe-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/apostrophe-tokenfilter.asciidoc index 330ab16b4ef..228feeca34b 100644 --- a/docs/reference/analysis/tokenfilters/apostrophe-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/apostrophe-tokenfilter.asciidoc @@ -8,8 +8,8 @@ Strips all characters after an apostrophe, including the apostrophe itself. This filter is included in {es}'s built-in <>. It uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/tr/ApostropheFilter.html[ApostropheFilter], -which was built for the Turkish language. +{lucene-analysis-docs}/tr/ApostropheFilter.html[ApostropheFilter], which was +built for the Turkish language. [[analysis-apostrophe-tokenfilter-analyze-ex]] diff --git a/docs/reference/analysis/tokenfilters/asciifolding-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/asciifolding-tokenfilter.asciidoc index 432f4709d8c..80b1821c1e5 100644 --- a/docs/reference/analysis/tokenfilters/asciifolding-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/asciifolding-tokenfilter.asciidoc @@ -9,7 +9,7 @@ Latin Unicode block (first 127 ASCII characters) to their ASCII equivalent, if one exists. For example, the filter changes `à` to `a`. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/miscellaneous/ASCIIFoldingFilter.html[ASCIIFoldingFilter]. +{lucene-analysis-docs}/miscellaneous/ASCIIFoldingFilter.html[ASCIIFoldingFilter]. [[analysis-asciifolding-tokenfilter-analyze-ex]] ==== Example diff --git a/docs/reference/analysis/tokenfilters/cjk-bigram-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/cjk-bigram-tokenfilter.asciidoc index 712538ec278..9a6f80352f6 100644 --- a/docs/reference/analysis/tokenfilters/cjk-bigram-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/cjk-bigram-tokenfilter.asciidoc @@ -9,7 +9,7 @@ Japanese, and Korean) tokens. This filter is included in {es}'s built-in <>. It uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/cjk/CJKBigramFilter.html[CJKBigramFilter]. +{lucene-analysis-docs}/cjk/CJKBigramFilter.html[CJKBigramFilter]. [[analysis-cjk-bigram-tokenfilter-analyze-ex]] diff --git a/docs/reference/analysis/tokenfilters/cjk-width-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/cjk-width-tokenfilter.asciidoc index 83b3ba8dee7..a41e994c35b 100644 --- a/docs/reference/analysis/tokenfilters/cjk-width-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/cjk-width-tokenfilter.asciidoc @@ -14,7 +14,7 @@ characters This filter is included in {es}'s built-in <>. It uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/cjk/CJKWidthFilter.html[CJKWidthFilter]. +{lucene-analysis-docs}/cjk/CJKWidthFilter.html[CJKWidthFilter]. NOTE: This token filter can be viewed as a subset of NFKC/NFKD Unicode normalization. See the diff --git a/docs/reference/analysis/tokenfilters/classic-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/classic-tokenfilter.asciidoc index 2b281f541d1..4f25b6da3cc 100644 --- a/docs/reference/analysis/tokenfilters/classic-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/classic-tokenfilter.asciidoc @@ -9,7 +9,7 @@ Performs optional post-processing of terms generated by the This filter removes the english possessive (`'s`) from the end of words and removes dots from acronyms. It uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/standard/ClassicFilter.html[ClassicFilter]. +{lucene-analysis-docs}/standard/ClassicFilter.html[ClassicFilter]. [[analysis-classic-tokenfilter-analyze-ex]] ==== Example diff --git a/docs/reference/analysis/tokenfilters/common-grams-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/common-grams-tokenfilter.asciidoc index 1b253668570..e2b0a57d3df 100644 --- a/docs/reference/analysis/tokenfilters/common-grams-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/common-grams-tokenfilter.asciidoc @@ -16,7 +16,7 @@ You can use the `common_grams` filter in place of the completely ignore common words. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/commongrams/CommonGramsFilter.html[CommonGramsFilter]. +{lucene-analysis-docs}/commongrams/CommonGramsFilter.html[CommonGramsFilter]. [[analysis-common-grams-analyze-ex]] ==== Example diff --git a/docs/reference/analysis/tokenfilters/condition-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/condition-tokenfilter.asciidoc index 3e81d1536ae..a33a41e85a8 100644 --- a/docs/reference/analysis/tokenfilters/condition-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/condition-tokenfilter.asciidoc @@ -8,7 +8,7 @@ Applies a set of token filters to tokens that match conditions in a provided predicate script. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/miscellaneous/ConditionalTokenFilter.html[ConditionalTokenFilter]. +{lucene-analysis-docs}/miscellaneous/ConditionalTokenFilter.html[ConditionalTokenFilter]. [[analysis-condition-analyze-ex]] ==== Example diff --git a/docs/reference/analysis/tokenfilters/decimal-digit-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/decimal-digit-tokenfilter.asciidoc index 58303185c13..76d085cf4b0 100644 --- a/docs/reference/analysis/tokenfilters/decimal-digit-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/decimal-digit-tokenfilter.asciidoc @@ -8,7 +8,7 @@ Converts all digits in the Unicode `Decimal_Number` General Category to `0-9`. For example, the filter changes the Bengali numeral `৩` to `3`. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysiscore/DecimalDigitFilter.html[DecimalDigitFilter]. +{lucene-analysis-docs}/core/DecimalDigitFilter.html[DecimalDigitFilter]. [[analysis-decimal-digit-tokenfilter-analyze-ex]] ==== Example diff --git a/docs/reference/analysis/tokenfilters/delimited-payload-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/delimited-payload-tokenfilter.asciidoc index 6498a90a4df..7a08982ff0c 100644 --- a/docs/reference/analysis/tokenfilters/delimited-payload-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/delimited-payload-tokenfilter.asciidoc @@ -18,7 +18,7 @@ split `the|1 quick|2 fox|3` into the tokens `the`, `quick`, and `fox` with respective payloads of `1`, `2`, and `3`. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilter.html[DelimitedPayloadTokenFilter]. +{lucene-analysis-docs}/payloads/DelimitedPayloadTokenFilter.html[DelimitedPayloadTokenFilter]. [NOTE] .Payloads diff --git a/docs/reference/analysis/tokenfilters/dictionary-decompounder-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/dictionary-decompounder-tokenfilter.asciidoc index cc269a12829..2c399b975aa 100644 --- a/docs/reference/analysis/tokenfilters/dictionary-decompounder-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/dictionary-decompounder-tokenfilter.asciidoc @@ -17,7 +17,7 @@ Uses a specified list of words and a brute force approach to find subwords in compound words. If found, these subwords are included in the token output. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/compound/DictionaryCompoundWordTokenFilter.html[DictionaryCompoundWordTokenFilter], +{lucene-analysis-docs}/compound/DictionaryCompoundWordTokenFilter.html[DictionaryCompoundWordTokenFilter], which was built for Germanic languages. [[analysis-dict-decomp-tokenfilter-analyze-ex]] diff --git a/docs/reference/analysis/tokenfilters/edgengram-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/edgengram-tokenfilter.asciidoc index 3c7f429c97b..e10289b6881 100644 --- a/docs/reference/analysis/tokenfilters/edgengram-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/edgengram-tokenfilter.asciidoc @@ -13,7 +13,7 @@ For example, you can use the `edge_ngram` token filter to change `quick` to When not customized, the filter creates 1-character edge n-grams by default. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilter.html[EdgeNGramTokenFilter]. +{lucene-analysis-docs}/ngram/EdgeNGramTokenFilter.html[EdgeNGramTokenFilter]. [NOTE] ==== diff --git a/docs/reference/analysis/tokenfilters/elision-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/elision-tokenfilter.asciidoc index f3b6fe0cdf4..fd3e1124552 100644 --- a/docs/reference/analysis/tokenfilters/elision-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/elision-tokenfilter.asciidoc @@ -22,7 +22,7 @@ Customized versions of this filter are included in several of {es}'s built-in * <> This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/util/ElisionFilter.html[ElisionFilter]. +{lucene-analysis-docs}/util/ElisionFilter.html[ElisionFilter]. [[analysis-elision-tokenfilter-analyze-ex]] ==== Example diff --git a/docs/reference/analysis/tokenfilters/fingerprint-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/fingerprint-tokenfilter.asciidoc index 0730ddae6a2..055473ec246 100644 --- a/docs/reference/analysis/tokenfilters/fingerprint-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/fingerprint-tokenfilter.asciidoc @@ -22,7 +22,7 @@ https://github.com/OpenRefine/OpenRefine/wiki/Clustering-In-Depth#fingerprint[Op project]. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene//analysis/miscellaneous/FingerprintFilter.html[FingerprintFilter]. +{lucene-analysis-docs}/miscellaneous/FingerprintFilter.html[FingerprintFilter]. [[analysis-fingerprint-tokenfilter-analyze-ex]] ==== Example diff --git a/docs/reference/analysis/tokenfilters/hyphenation-decompounder-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/hyphenation-decompounder-tokenfilter.asciidoc index 9777937eec8..553b44379c3 100644 --- a/docs/reference/analysis/tokenfilters/hyphenation-decompounder-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/hyphenation-decompounder-tokenfilter.asciidoc @@ -9,7 +9,7 @@ words. These subwords are then checked against the specified word list. Subwords in the list are excluded from the token output. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/compound/HyphenationCompoundWordTokenFilter.html[HyphenationCompoundWordTokenFilter], +{lucene-analysis-docs}/compound/HyphenationCompoundWordTokenFilter.html[HyphenationCompoundWordTokenFilter], which was built for Germanic languages. [[analysis-hyp-decomp-tokenfilter-analyze-ex]] diff --git a/docs/reference/analysis/tokenfilters/keep-types-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/keep-types-tokenfilter.asciidoc index 838c1252835..9fca0275e80 100644 --- a/docs/reference/analysis/tokenfilters/keep-types-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/keep-types-tokenfilter.asciidoc @@ -26,7 +26,7 @@ type. ==== This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/core/TypeTokenFilter.html[TypeTokenFilter]. +{lucene-analysis-docs}/core/TypeTokenFilter.html[TypeTokenFilter]. [[analysis-keep-types-tokenfilter-analyze-include-ex]] ==== Include example diff --git a/docs/reference/analysis/tokenfilters/keep-words-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/keep-words-tokenfilter.asciidoc index a7dc784c3b7..d8346990a26 100644 --- a/docs/reference/analysis/tokenfilters/keep-words-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/keep-words-tokenfilter.asciidoc @@ -7,7 +7,7 @@ Keeps only tokens contained in a specified word list. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/miscellaneous/KeepWordFilter.html[KeepWordFilter]. +{lucene-analysis-docs}/miscellaneous/KeepWordFilter.html[KeepWordFilter]. [NOTE] ==== diff --git a/docs/reference/analysis/tokenfilters/length-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/length-tokenfilter.asciidoc index 4bbe60e52be..4eced2d39b1 100644 --- a/docs/reference/analysis/tokenfilters/length-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/length-tokenfilter.asciidoc @@ -9,7 +9,7 @@ For example, you can use the `length` filter to exclude tokens shorter than 2 characters and tokens longer than 5 characters. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/miscellaneous/LengthFilter.html[LengthFilter]. +{lucene-analysis-docs}/miscellaneous/LengthFilter.html[LengthFilter]. [TIP] ==== diff --git a/docs/reference/analysis/tokenfilters/limit-token-count-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/limit-token-count-tokenfilter.asciidoc index 61e7ec8706d..5268571f7e9 100644 --- a/docs/reference/analysis/tokenfilters/limit-token-count-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/limit-token-count-tokenfilter.asciidoc @@ -12,7 +12,7 @@ example, the filter can change the token stream `[ one, two, three ]` to `[ one ]`. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/miscellaneous/LimitTokenCountFilter.html[LimitTokenCountFilter]. +{lucene-analysis-docs}/miscellaneous/LimitTokenCountFilter.html[LimitTokenCountFilter]. [TIP] ==== diff --git a/docs/reference/analysis/tokenfilters/lowercase-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/lowercase-tokenfilter.asciidoc index 6a1ef08c0b1..9b39b8ed771 100644 --- a/docs/reference/analysis/tokenfilters/lowercase-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/lowercase-tokenfilter.asciidoc @@ -104,13 +104,17 @@ PUT lowercase_example (Optional, string) Language-specific lowercase token filter to use. Valid values include: -`greek`::: Uses Lucene's https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/el/GreekLowerCaseFilter.html[GreekLowerCaseFilter] +`greek`::: Uses Lucene's +{lucene-analysis-docs}/el/GreekLowerCaseFilter.html[GreekLowerCaseFilter] -`irish`::: Uses Lucene's http://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/ga/IrishLowerCaseFilter.html[IrishLowerCaseFilter] +`irish`::: Uses Lucene's +http://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/ga/IrishLowerCaseFilter.html[IrishLowerCaseFilter] -`turkish`::: Uses Lucene's https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/tr/TurkishLowerCaseFilter.html[TurkishLowerCaseFilter] +`turkish`::: Uses Lucene's +{lucene-analysis-docs}/tr/TurkishLowerCaseFilter.html[TurkishLowerCaseFilter] -If not specified, defaults to Lucene's https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/core/LowerCaseFilter.html[LowerCaseFilter]. +If not specified, defaults to Lucene's +{lucene-analysis-docs}/core/LowerCaseFilter.html[LowerCaseFilter]. -- [[analysis-lowercase-tokenfilter-customize]] diff --git a/docs/reference/analysis/tokenfilters/ngram-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/ngram-tokenfilter.asciidoc index 7bdc913c579..1dbe6886c53 100644 --- a/docs/reference/analysis/tokenfilters/ngram-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/ngram-tokenfilter.asciidoc @@ -11,7 +11,7 @@ For example, you can use the `ngram` token filter to change `fox` to `[ f, fo, o, ox, x ]`. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/ngram/NGramTokenFilter.html[NGramTokenFilter]. +{lucene-analysis-docs}/ngram/NGramTokenFilter.html[NGramTokenFilter]. [NOTE] ==== diff --git a/docs/reference/analysis/tokenfilters/reverse-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/reverse-tokenfilter.asciidoc index e0c9f93c97b..d66e143b4c4 100644 --- a/docs/reference/analysis/tokenfilters/reverse-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/reverse-tokenfilter.asciidoc @@ -12,7 +12,7 @@ such as finding words that end in `-ion` or searching file names by their extension. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/reverse/ReverseStringFilter.html[ReverseStringFilter]. +{lucene-analysis-docs}/reverse/ReverseStringFilter.html[ReverseStringFilter]. [[analysis-reverse-tokenfilter-analyze-ex]] ==== Example diff --git a/docs/reference/analysis/tokenfilters/truncate-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/truncate-tokenfilter.asciidoc index 1ca6f86cc50..a77387d5fd4 100644 --- a/docs/reference/analysis/tokenfilters/truncate-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/truncate-tokenfilter.asciidoc @@ -11,7 +11,7 @@ For example, you can use the `truncate` filter to shorten all tokens to `3` characters or fewer, changing `jumping fox` to `jum fox`. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/miscellaneous/TruncateTokenFilter.html[TruncateTokenFilter]. +{lucene-analysis-docs}/miscellaneous/TruncateTokenFilter.html[TruncateTokenFilter]. [[analysis-truncate-tokenfilter-analyze-ex]] ==== Example diff --git a/docs/reference/analysis/tokenfilters/uppercase-tokenfilter.asciidoc b/docs/reference/analysis/tokenfilters/uppercase-tokenfilter.asciidoc index 06ea2c3279c..780e09fa951 100644 --- a/docs/reference/analysis/tokenfilters/uppercase-tokenfilter.asciidoc +++ b/docs/reference/analysis/tokenfilters/uppercase-tokenfilter.asciidoc @@ -8,7 +8,7 @@ Changes token text to uppercase. For example, you can use the `uppercase` filter to change `the Lazy DoG` to `THE LAZY DOG`. This filter uses Lucene's -https://lucene.apache.org/core/{lucene_version_path}/analyzers-common/org/apache/lucene/analysis/core/UpperCaseFilter.html[UpperCaseFilter]. +{lucene-analysis-docs}/core/UpperCaseFilter.html[UpperCaseFilter]. [WARNING] ====