diff --git a/solr/CHANGES.txt b/solr/CHANGES.txt index 48ed8403d7f..9157bb3c494 100644 --- a/solr/CHANGES.txt +++ b/solr/CHANGES.txt @@ -339,6 +339,9 @@ Other Changes * SOLR-12690: Regularize LoggerFactory declarations (Erick Erickson) +* SOLR-12590: Improve Solr resource loader coverage in the ref guide. + (Steve Rowe, Cassandra Targett, Christine Poerschke) + ================== 7.4.0 ================== Consult the LUCENE_CHANGES.txt file for additional, low level, changes in this release. diff --git a/solr/solr-ref-guide/src/analytics.adoc b/solr/solr-ref-guide/src/analytics.adoc index fe9b1105ce2..d7407d17ea7 100644 --- a/solr/solr-ref-guide/src/analytics.adoc +++ b/solr/solr-ref-guide/src/analytics.adoc @@ -33,7 +33,7 @@ Since the Analytics framework is a _search component_, it must be declared as su For distributed analytics requests over cloud collections, the component uses the `AnalyticsHandler` strictly for inter-shard communication. The Analytics Handler should not be used by users to submit analytics requests. -To configure Solr to use the Analytics Component, the first step is to add a `lib` directive so Solr loads the Analytic Component classes (for more about the `lib` directive, see <>). In the section of `solrconfig.xml` where the default `lib` directive are, add a line: +To configure Solr to use the Analytics Component, the first step is to add a `` directive so Solr loads the Analytic Component classes (for more about the `` directive, see <>). In the section of `solrconfig.xml` where the default `` directives are, add a line: [source,xml] diff --git a/solr/solr-ref-guide/src/configuring-solrconfig-xml.adoc b/solr/solr-ref-guide/src/configuring-solrconfig-xml.adoc index 83febafd0a1..d2570faedd9 100644 --- a/solr/solr-ref-guide/src/configuring-solrconfig-xml.adoc +++ b/solr/solr-ref-guide/src/configuring-solrconfig-xml.adoc @@ -1,5 +1,5 @@ = Configuring solrconfig.xml -:page-children: datadir-and-directoryfactory-in-solrconfig, lib-directives-in-solrconfig, schema-factory-definition-in-solrconfig, indexconfig-in-solrconfig, requesthandlers-and-searchcomponents-in-solrconfig, initparams-in-solrconfig, updatehandlers-in-solrconfig, query-settings-in-solrconfig, requestdispatcher-in-solrconfig, update-request-processors, codec-factory +:page-children: datadir-and-directoryfactory-in-solrconfig, resource-and-plugin-loading, schema-factory-definition-in-solrconfig, indexconfig-in-solrconfig, requesthandlers-and-searchcomponents-in-solrconfig, initparams-in-solrconfig, updatehandlers-in-solrconfig, query-settings-in-solrconfig, requestdispatcher-in-solrconfig, update-request-processors, codec-factory // Licensed to the Apache Software Foundation (ASF) under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information @@ -38,7 +38,7 @@ The `solrconfig.xml` file is located in the `conf/` directory for each collectio We've covered the options in the following sections: * <> -* <> +* <> * <> * <> * <> diff --git a/solr/solr-ref-guide/src/detecting-languages-during-indexing.adoc b/solr/solr-ref-guide/src/detecting-languages-during-indexing.adoc index 8b0556b890e..8d446a29fe7 100644 --- a/solr/solr-ref-guide/src/detecting-languages-during-indexing.adoc +++ b/solr/solr-ref-guide/src/detecting-languages-during-indexing.adoc @@ -80,7 +80,7 @@ Here is an example of a minimal OpenNLP `langid` configuration in `solrconfig.xm ==== OpenNLP-specific Parameters `langid.model`:: -An OpenNLP language detection model. The OpenNLP project provides a pre-trained 103 language model on the http://opennlp.apache.org/models.html[OpenNLP site's model dowload page]. Model training instructions are provided on the http://opennlp.apache.org/docs/{ivy-opennlp-version}/manual/opennlp.html#tools.langdetect[OpenNLP website]. This parameter is required. +An OpenNLP language detection model. The OpenNLP project provides a pre-trained 103 language model on the http://opennlp.apache.org/models.html[OpenNLP site's model dowload page]. Model training instructions are provided on the http://opennlp.apache.org/docs/{ivy-opennlp-version}/manual/opennlp.html#tools.langdetect[OpenNLP website]. This parameter is required. See <> for information on where to put the model. ==== OpenNLP Language Codes diff --git a/solr/solr-ref-guide/src/filter-descriptions.adoc b/solr/solr-ref-guide/src/filter-descriptions.adoc index f5179012bc6..7fabb750cc5 100644 --- a/solr/solr-ref-guide/src/filter-descriptions.adoc +++ b/solr/solr-ref-guide/src/filter-descriptions.adoc @@ -471,7 +471,7 @@ Note that for this filter to work properly, the upstream tokenizer must not remo This filter is a custom Unicode normalization form that applies the foldings specified in http://www.unicode.org/reports/tr30/tr30-4.html[Unicode TR #30: Character Foldings] in addition to the `NFKC_Casefold` normalization form as described in <>. This filter is a better substitute for the combined behavior of the <>, <>, and <>. -To use this filter, see `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add to your `solr_home/lib`. For more information about adding jars, see the section <>. +To use this filter, you must add additional .jars to Solr's classpath (as described in the section <>). See `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add. *Factory class:* `solr.ICUFoldingFilterFactory` @@ -543,7 +543,7 @@ This filter factory normalizes text according to one of five Unicode Normalizati For detailed information about these normalization forms, see http://unicode.org/reports/tr15/[Unicode Normalization Forms]. -To use this filter, see `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add to your `solr_home/lib`. +To use this filter, you must add additional .jars to Solr's classpath (as described in the section <>). See `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add. == ICU Transform Filter @@ -567,7 +567,7 @@ This filter applies http://userguide.icu-project.org/transforms/general[ICU Tran For detailed information about ICU Transforms, see http://userguide.icu-project.org/transforms/general. -To use this filter, see `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add to your `solr_home/lib`. +To use this filter, you must add additional .jars to Solr's classpath (as described in the section <>). See `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add. == Keep Word Filter @@ -1501,7 +1501,7 @@ NOTE: Although this filter produces correct token graphs, it cannot consume an i *Arguments:* -`synonyms`:: (required) The path of a file that contains a list of synonyms, one per line. In the (default) `solr` format - see the `format` argument below for alternatives - blank lines and lines that begin with "`#`" are ignored. This may be a comma-separated list of absolute paths, or paths relative to the Solr config directory. +`synonyms`:: (required) The path of a file that contains a list of synonyms, one per line. In the (default) `solr` format - see the `format` argument below for alternatives - blank lines and lines that begin with "`#`" are ignored. This may be a comma-separated list of paths. See <> for more information. + There are two ways to specify synonym mappings: + diff --git a/solr/solr-ref-guide/src/language-analysis.adoc b/solr/solr-ref-guide/src/language-analysis.adoc index cd893f9936e..2749ce1dc46 100644 --- a/solr/solr-ref-guide/src/language-analysis.adoc +++ b/solr/solr-ref-guide/src/language-analysis.adoc @@ -94,7 +94,7 @@ Compound words are most commonly found in Germanic languages. *Arguments:* -`dictionary`:: (required) The path of a file that contains a list of simple words, one per line. Blank lines and lines that begin with "#" are ignored. This path may be an absolute path, or path relative to the Solr config directory. +`dictionary`:: (required) The path of a file that contains a list of simple words, one per line. Blank lines and lines that begin with "#" are ignored. See <> for more information. `minWordSize`:: (integer, default 5) Any token shorter than this is not decompounded. @@ -130,7 +130,7 @@ Unicode Collation in Solr is fast, because all the work is done at index time. Rather than specifying an analyzer within ``, the `solr.CollationField` and `solr.ICUCollationField` field type classes provide this functionality. `solr.ICUCollationField`, which is backed by http://site.icu-project.org[the ICU4J library], provides more flexible configuration, has more locales, is significantly faster, and requires less memory and less index space, since its keys are smaller than those produced by the JDK implementation that backs `solr.CollationField`. -`solr.ICUCollationField` is included in the Solr `analysis-extras` contrib - see `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add to your `SOLR_HOME/lib` in order to use it. +To use `solr.ICUCollationField`, you must add additional .jars to Solr's classpath (as described in the section <>). See `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add. `solr.ICUCollationField` and `solr.CollationField` fields can be created in two ways: @@ -361,7 +361,7 @@ The `lucene/analysis/opennlp` module provides OpenNLP integration via several an NOTE: The <> must be used with all other OpenNLP analysis components, for two reasons: first, the OpenNLP Tokenizer detects and marks the sentence boundaries required by all the OpenNLP filters; and second, since the pre-trained OpenNLP models used by these filters were trained using the corresponding language-specific sentence-detection/tokenization models, the same tokenization, using the same models, must be used at runtime for optimal performance. -See `solr/contrib/analysis-extras/README.txt` for information on which jars you need to add to your `SOLR_HOME/lib`. +To use the OpenNLP components, you must add additional .jars to Solr's classpath (as described in the section <>). See `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add. === OpenNLP Tokenizer @@ -371,9 +371,9 @@ The OpenNLP Tokenizer takes two language-specific binary model files as paramete *Arguments:* -`sentenceModel`:: (required) The path of a language-specific OpenNLP sentence detection model file. This path may be an absolute path, or path relative to the Solr config directory. +`sentenceModel`:: (required) The path of a language-specific OpenNLP sentence detection model file. See <> for more information. -`tokenizerModel`:: (required) The path of a language-specific OpenNLP tokenization model file. This path may be an absolute path, or path relative to the Solr config directory. +`tokenizerModel`:: (required) The path of a language-specific OpenNLP tokenization model file. See <> for more information. *Example:* @@ -396,7 +396,7 @@ NOTE: Lucene currently does not index token types, so if you want to keep this i *Arguments:* -`posTaggerModel`:: (required) The path of a language-specific OpenNLP POS tagger model file. This path may be an absolute path, or path relative to the Solr config directory. +`posTaggerModel`:: (required) The path of a language-specific OpenNLP POS tagger model file. See <> for more information. *Examples:* @@ -469,7 +469,7 @@ NOTE: Lucene currently does not index token types, so if you want to keep this i *Arguments:* -`chunkerModel`:: (required) The path of a language-specific OpenNLP phrase chunker model file. This path may be an absolute path, or path relative to the Solr config directory. +`chunkerModel`:: (required) The path of a language-specific OpenNLP phrase chunker model file. See <> for more information. *Examples*: @@ -511,9 +511,9 @@ This filter replaces the text of each token with its lemma. Both a dictionary-ba Either `dictionary` or `lemmatizerModel` must be provided, and both may be provided - see the examples below: -`dictionary`:: (optional) The path of a lemmatization dictionary file. This path may be an absolute path, or path relative to the Solr config directory. The dictionary file must be encoded as UTF-8, with one entry per line, in the form `word[tab]lemma[tab]part-of-speech`, e.g., `wrote[tab]write[tab]VBD`. +`dictionary`:: (optional) The path of a lemmatization dictionary file. See <> for more information. The dictionary file must be encoded as UTF-8, with one entry per line, in the form `word[tab]lemma[tab]part-of-speech`, e.g., `wrote[tab]write[tab]VBD`. -`lemmatizerModel`:: (optional) The path of a language-specific OpenNLP lemmatizer model file. This path may be an absolute path, or path relative to the Solr config directory. +`lemmatizerModel`:: (optional) The path of a language-specific OpenNLP lemmatizer model file. See <> for more information. *Examples:* @@ -698,7 +698,7 @@ Solr can stem Catalan using the Snowball Porter Stemmer with an argument of `lan === Traditional Chinese -The default configuration of the <> is suitable for Traditional Chinese text. It follows the Word Break rules from the Unicode Text Segmentation algorithm for non-Chinese text, and uses a dictionary to segment Chinese words. To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See the `solr/contrib/analysis-extras/README.txt` for information on which jars you need to add to your `SOLR_HOME/lib`. +The default configuration of the <> is suitable for Traditional Chinese text. It follows the Word Break rules from the Unicode Text Segmentation algorithm for non-Chinese text, and uses a dictionary to segment Chinese words. To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See the `solr/contrib/analysis-extras/README.txt` for information on which jars you need to add. <> can also be used to tokenize Traditional Chinese text. Following the Word Break rules from the Unicode Text Segmentation algorithm, it produces one token per Chinese character. When combined with <>, overlapping bigrams of Chinese characters are formed. @@ -751,9 +751,9 @@ See the example under <>. === Simplified Chinese -For Simplified Chinese, Solr provides support for Chinese sentence and word segmentation with the <>. This component includes a large dictionary and segments Chinese text into words with the Hidden Markov Model. To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See the `solr/contrib/analysis-extras/README.txt` for information on which jars you need to add to your `SOLR_HOME/lib`. +For Simplified Chinese, Solr provides support for Chinese sentence and word segmentation with the <>. This component includes a large dictionary and segments Chinese text into words with the Hidden Markov Model. To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See the `solr/contrib/analysis-extras/README.txt` for information on which jars you need to add. -The default configuration of the <> is also suitable for Simplified Chinese text. It follows the Word Break rules from the Unicode Text Segmentation algorithm for non-Chinese text, and uses a dictionary to segment Chinese words. To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See the `solr/contrib/analysis-extras/README.txt` for information on which jars you need to add to your `SOLR_HOME/lib`. +The default configuration of the <> is also suitable for Simplified Chinese text. It follows the Word Break rules from the Unicode Text Segmentation algorithm for non-Chinese text, and uses a dictionary to segment Chinese words. To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See the `solr/contrib/analysis-extras/README.txt` for information on which jars you need to add. Also useful for Chinese analysis: @@ -786,7 +786,7 @@ Also useful for Chinese analysis: === HMM Chinese Tokenizer -For Simplified Chinese, Solr provides support for Chinese sentence and word segmentation with the `solr.HMMChineseTokenizerFactory` in the `analysis-extras` contrib module. This component includes a large dictionary and segments Chinese text into words with the Hidden Markov Model. To use this tokenizer, see `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add to your `solr_home/lib`. +For Simplified Chinese, Solr provides support for Chinese sentence and word segmentation with the `solr.HMMChineseTokenizerFactory` in the `analysis-extras` contrib module. This component includes a large dictionary and segments Chinese text into words with the Hidden Markov Model. To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add. *Factory class:* `solr.HMMChineseTokenizerFactory` @@ -1278,7 +1278,7 @@ Example: [[hebrew-lao-myanmar-khmer]] === Hebrew, Lao, Myanmar, Khmer -Lucene provides support, in addition to UAX#29 word break rules, for Hebrew's use of the double and single quote characters, and for segmenting Lao, Myanmar, and Khmer into syllables with the `solr.ICUTokenizerFactory` in the `analysis-extras` contrib module. To use this tokenizer, see `solr/contrib/analysis-extras/README.txt for` instructions on which jars you need to add to your `solr_home/lib`. +Lucene provides support, in addition to UAX#29 word break rules, for Hebrew's use of the double and single quote characters, and for segmenting Lao, Myanmar, and Khmer into syllables with the `solr.ICUTokenizerFactory` in the `analysis-extras` contrib module. To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See `solr/contrib/analysis-extras/README.txt for` instructions on which jars you need to add. See <> for more information. @@ -1423,7 +1423,7 @@ Solr includes support for normalizing Persian, and Lucene includes an example st === Polish -Solr provides support for Polish stemming with the `solr.StempelPolishStemFilterFactory`, and `solr.MorphologikFilterFactory` for lemmatization, in the `contrib/analysis-extras` module. The `solr.StempelPolishStemFilterFactory` component includes an algorithmic stemmer with tables for Polish. To use either of these filters, see `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add to your `solr_home/lib`. +Solr provides support for Polish stemming with the `solr.StempelPolishStemFilterFactory`, and `solr.MorphologikFilterFactory` for lemmatization, in the `contrib/analysis-extras` module. The `solr.StempelPolishStemFilterFactory` component includes an algorithmic stemmer with tables for Polish. To use either of these filters, you must add additional .jars to Solr's classpath (as described in the section <>). See `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add. *Factory class:* `solr.StempelPolishStemFilterFactory` and `solr.MorfologikFilterFactory` @@ -1750,7 +1750,7 @@ Solr includes support for stemming Turkish with the `solr.SnowballPorterFilterFa === Ukrainian -Solr provides support for Ukrainian lemmatization with the `solr.MorphologikFilterFactory`, in the `contrib/analysis-extras` module. To use this filter, see `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add to your `solr_home/lib`. +Solr provides support for Ukrainian lemmatization with the `solr.MorphologikFilterFactory`, in the `contrib/analysis-extras` module. To use this filter, you must add additional .jars to Solr's classpath (as described in the section <>). See `solr/contrib/analysis-extras/README.txt` for instructions on which jars you need to add. Lucene also includes an example Ukrainian stopword list, in the `lucene-analyzers-morfologik` jar. diff --git a/solr/solr-ref-guide/src/learning-to-rank.adoc b/solr/solr-ref-guide/src/learning-to-rank.adoc index 8fe3c3353a9..12fab3236fd 100644 --- a/solr/solr-ref-guide/src/learning-to-rank.adoc +++ b/solr/solr-ref-guide/src/learning-to-rank.adoc @@ -533,7 +533,7 @@ Assuming that you consider to use a large model placed at `/path/to/models/myMod } ---- -First, add the directory to Solr's resource paths by <>: +First, add the directory to Solr's resource paths with a <` directive>>: [source,xml] ---- diff --git a/solr/solr-ref-guide/src/lib-directives-in-solrconfig.adoc b/solr/solr-ref-guide/src/lib-directives-in-solrconfig.adoc deleted file mode 100644 index dc3b3193058..00000000000 --- a/solr/solr-ref-guide/src/lib-directives-in-solrconfig.adoc +++ /dev/null @@ -1,38 +0,0 @@ -= Lib Directives in SolrConfig -// Licensed to the Apache Software Foundation (ASF) under one -// or more contributor license agreements. See the NOTICE file -// distributed with this work for additional information -// regarding copyright ownership. The ASF licenses this file -// to you under the Apache License, Version 2.0 (the -// "License"); you may not use this file except in compliance -// with the License. You may obtain a copy of the License at -// -// http://www.apache.org/licenses/LICENSE-2.0 -// -// Unless required by applicable law or agreed to in writing, -// software distributed under the License is distributed on an -// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY -// KIND, either express or implied. See the License for the -// specific language governing permissions and limitations -// under the License. - -Solr allows loading plugins by defining `` directives in `solrconfig.xml`. - -The plugins are loaded in the order they appear in `solrconfig.xml`. If there are dependencies, list the lowest level dependency jar first. - -Regular expressions can be used to provide control loading jars with dependencies on other jars in the same directory. All directories are resolved as relative to the Solr `instanceDir`. - -[source,xml] ----- - - - - - - - - - - - ----- diff --git a/solr/solr-ref-guide/src/resource-and-plugin-loading.adoc b/solr/solr-ref-guide/src/resource-and-plugin-loading.adoc new file mode 100644 index 00000000000..60cd60f4791 --- /dev/null +++ b/solr/solr-ref-guide/src/resource-and-plugin-loading.adoc @@ -0,0 +1,86 @@ += Resource and Plugin Loading +// Licensed to the Apache Software Foundation (ASF) under one +// or more contributor license agreements. See the NOTICE file +// distributed with this work for additional information +// regarding copyright ownership. The ASF licenses this file +// to you under the Apache License, Version 2.0 (the +// "License"); you may not use this file except in compliance +// with the License. You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, +// software distributed under the License is distributed on an +// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +// KIND, either express or implied. See the License for the +// specific language governing permissions and limitations +// under the License. + +Solr components can be configured using *resources*: data stored in external files that may be referred to in a location-independent fashion. Examples include: files needed by schema components, e.g. a stopword list for <>; and machine-learned models for <>. + +Solr *plugins*, which can be configured in `solrconfig.xml`, are Java classes that are normally packaged in `.jar` files along with supporting classes and data. Solr ships with a number of built-in plugins, and can also be configured to use custom plugins. Example plugins are the <> and custom search components. + +Resources and plugins may be stored: + +* in ZooKeeper under a collection's configset node (SolrCloud only); +* on a filesystem accessible to Solr nodes; or +* in Solr's <> (SolrCloud only). + +NOTE: Schema components may not be stored as plugins in the Blob Store, and cannot access resources stored in the Blob Store. + +== Resource and Plugin Loading Sequence + +Under SolrCloud, resources and plugins to be loaded are first looked up in ZooKeeper under the collection's configset znode. If the resource or plugin is not found there, Solr will fall back to loading <>. + +Note that by default, Solr will not attempt to load resources and plugins from the Blob Store. To enable this, see the section <>. When loading from the Blob Store is enabled for a component, lookups occur only in the Blob Store, and never in ZooKeeper or on the filesystem. + +== Resources and Plugins in ConfigSets on ZooKeeper + +Resources and plugins may be uploaded to ZooKeeper as part of a configset, either via the <> or <>. + +To upload a plugin or resource to a configset already stored on ZooKeeper, you can use <>. + +CAUTION: By default, ZooKeeper's file size limit is 1MB. If your files are larger than this, you'll need to either <> or store them instead <>. + +== Resources and Plugins on the Filesystem + +Under standalone Solr, when looking up a plugin or resource to be loaded, Solr's resource loader will first look under the `/conf/` directory. If the plugin or resource is not found, the configured plugin and resource file paths are searched - see the section <> below. + +On core load, Solr's resource loader constructs a list of paths (subdirectories and jars), first under <>, and then under directories pointed to by <` directives in SolrConfig>>. + +When looking up a resource or plugin to be loaded, the paths on the list are searched in the order they were added. + +NOTE: Under SolrCloud, each node hosting a collection replica will need its own copy of plugins and resources to be loaded. + +To get Solr's resource loader to find resources either under subdirectories or in jar files that were created after Solr's resource path list was constructed, reload the collection (SolrCloud) or the core (standalone Solr). Restarting all affected Solr nodes also works. + +WARNING: Resource files *will not be loaded* if they are located directly under either `solr_home/lib` or a directory given by the `dir` attribute on a `` directive in SolrConfig. Resources are only searched for under subdirectories or in jar files found in those locations. + +=== solr_home/lib + +Each Solr node can have a directory named `lib/` under the <>. In order to use this directory to host resources or plugins, it must first be manually created. + +=== Lib Directives in SolrConfig + +Plugin and resource file paths are configurable via `` directives in `solrconfig.xml`. + +Loading occurs in the order `` directives appear in `solrconfig.xml`. If there are dependencies, list the lowest level dependency jar first. + +A regular expression supplied in the `` element's `regex` attribute value can be used to restrict which subdirectories and/or jar files are added to the Solr resource loader's list of search locations. If no regular expression is given, all direct subdirectory and jar children are included in the resource path list. All directories are resolved as relative to the Solr core's `instanceDir`. + +From an example SolrConfig: + +[source,xml] +---- + + + + + + + + + + + +---- diff --git a/solr/solr-ref-guide/src/tokenizers.adoc b/solr/solr-ref-guide/src/tokenizers.adoc index 82e730d5164..db32c784130 100644 --- a/solr/solr-ref-guide/src/tokenizers.adoc +++ b/solr/solr-ref-guide/src/tokenizers.adoc @@ -288,7 +288,7 @@ The default configuration for `solr.ICUTokenizerFactory` provides UAX#29 word br [IMPORTANT] ==== -To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See the `solr/contrib/analysis-extras/README.txt` for information on which jars you need to add to your `SOLR_HOME/lib`. +To use this tokenizer, you must add additional .jars to Solr's classpath (as described in the section <>). See the `solr/contrib/analysis-extras/README.txt` for information on which jars you need to add. ==== diff --git a/solr/solr-ref-guide/src/update-request-processors.adoc b/solr/solr-ref-guide/src/update-request-processors.adoc index 267ffbdabc7..27b999bc7eb 100644 --- a/solr/solr-ref-guide/src/update-request-processors.adoc +++ b/solr/solr-ref-guide/src/update-request-processors.adoc @@ -353,7 +353,7 @@ The {solr-javadocs}/solr-langid/index.html[`langid`] contrib provides:: The {solr-javadocs}/solr-analysis-extras/index.html[`analysis-extras`] contrib provides:: -{solr-javadocs}/solr-analysis-extras/org/apache/solr/update/processor/OpenNLPExtractNamedEntitiesUpdateProcessorFactory.html[OpenNLPExtractNamedEntitiesUpdateProcessorFactory]::: Update document(s) to be indexed with named entities extracted using an OpenNLP NER model. Note that in order to use model files larger than 1MB on SolrCloud, <>. +{solr-javadocs}/solr-analysis-extras/org/apache/solr/update/processor/OpenNLPExtractNamedEntitiesUpdateProcessorFactory.html[OpenNLPExtractNamedEntitiesUpdateProcessorFactory]::: Update document(s) to be indexed with named entities extracted using an OpenNLP NER model. Note that in order to use model files larger than 1MB on SolrCloud, you must either <> or <> on each node hosting a collection replica. === Update Processor Factories You Should _Not_ Modify or Remove