mirror of https://github.com/apache/lucene.git
LUCENE-2413: directory and package fixes
git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@955203 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
aa6afe1c4e
commit
5a661500c1
|
@ -23,7 +23,7 @@
|
|||
# topScoreDocUnordered - Like above, but allows out of order
|
||||
collector.class=coll:topScoreDocOrdered:topScoreDocUnordered:topScoreDocOrdered:topScoreDocUnordered
|
||||
|
||||
analyzer=org.apache.lucene.analysis.WhitespaceAnalyzer
|
||||
analyzer=org.apache.lucene.analysis.core.WhitespaceAnalyzer
|
||||
directory=FSDirectory
|
||||
#directory=RamDirectory
|
||||
|
||||
|
|
|
@ -23,7 +23,7 @@
|
|||
# topScoreDocUnordered - Like above, but allows out of order
|
||||
collector.class=coll:topScoreDocOrdered:topScoreDocUnordered:topScoreDocOrdered:topScoreDocUnordered
|
||||
|
||||
analyzer=org.apache.lucene.analysis.WhitespaceAnalyzer
|
||||
analyzer=org.apache.lucene.analysis.core.WhitespaceAnalyzer
|
||||
directory=FSDirectory
|
||||
#directory=RamDirectory
|
||||
|
||||
|
|
|
@ -29,7 +29,7 @@
|
|||
# ant run-task -Dtask.alg=conf/indexLineFile.alg
|
||||
#
|
||||
|
||||
analyzer=org.apache.lucene.analysis.SimpleAnalyzer
|
||||
analyzer=org.apache.lucene.analysis.core.SimpleAnalyzer
|
||||
|
||||
# Feed that knows how to process the line file format:
|
||||
content.source=org.apache.lucene.benchmark.byTask.feeds.LineDocSource
|
||||
|
|
|
@ -316,7 +316,7 @@ stop words and characters from the index. By stop words and characters I mean c
|
|||
words such as articles (a, an, the, etc.) and other strings that may have less value for searching
|
||||
(e.g. <b>'s</b>) . It should be noted that there are different rules for every language, and you
|
||||
should use the proper analyzer for each. Lucene currently provides Analyzers for a number of
|
||||
different languages (see the <span class="codefrag">*Analyzer.java</span> sources under <a href="http://svn.apache.org/repos/asf/lucene/dev/trunk/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/">contrib/analyzers/src/java/org/apache/lucene/analysis</a>).
|
||||
different languages (see the <span class="codefrag">*Analyzer.java</span> sources under <a href="http://svn.apache.org/repos/asf/lucene/dev/trunk/modules/analysis/common/src/java/org/apache/lucene/analysis/">modules/analysis/common/src/java/org/apache/lucene/analysis</a>).
|
||||
</p>
|
||||
<p>
|
||||
Looking further down in the file, you should see the <span class="codefrag">indexDocs()</span> code. This recursive
|
||||
|
|
|
@ -105,7 +105,7 @@ There are many post tokenization steps that can be done, including (but not limi
|
|||
<li>{@link org.apache.lucene.analysis.PerFieldAnalyzerWrapper} – Most Analyzers perform the same operation on all
|
||||
{@link org.apache.lucene.document.Field}s. The PerFieldAnalyzerWrapper can be used to associate a different Analyzer with different
|
||||
{@link org.apache.lucene.document.Field}s.</li>
|
||||
<li>The contrib/analyzers library located at the root of the Lucene distribution has a number of different Analyzer implementations to solve a variety
|
||||
<li>The modules/analysis library located at the root of the Lucene distribution has a number of different Analyzer implementations to solve a variety
|
||||
of different problems related to searching. Many of the Analyzers are designed to analyze non-English languages.</li>
|
||||
<li>The contrib/snowball library
|
||||
located at the root of the Lucene distribution has Analyzer and TokenFilter
|
||||
|
@ -168,7 +168,7 @@ There are many post tokenization steps that can be done, including (but not limi
|
|||
<h2>Implementing your own Analyzer</h2>
|
||||
<p>Creating your own Analyzer is straightforward. It usually involves either wrapping an existing Tokenizer and set of TokenFilters to create a new Analyzer
|
||||
or creating both the Analyzer and a Tokenizer or TokenFilter. Before pursuing this approach, you may find it worthwhile
|
||||
to explore the contrib/analyzers library and/or ask on the java-user@lucene.apache.org mailing list first to see if what you need already exists.
|
||||
to explore the modules/analysis library and/or ask on the java-user@lucene.apache.org mailing list first to see if what you need already exists.
|
||||
If you are still committed to creating your own Analyzer or TokenStream derivation (Tokenizer or TokenFilter) have a look at
|
||||
the source code of any one of the many samples located in this package.
|
||||
</p>
|
||||
|
|
|
@ -74,7 +74,7 @@ words such as articles (a, an, the, etc.) and other strings that may have less v
|
|||
(e.g. <b>'s</b>) . It should be noted that there are different rules for every language, and you
|
||||
should use the proper analyzer for each. Lucene currently provides Analyzers for a number of
|
||||
different languages (see the <code>*Analyzer.java</code> sources under <a
|
||||
href="http://svn.apache.org/repos/asf/lucene/dev/trunk/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/">contrib/analyzers/src/java/org/apache/lucene/analysis</a>).
|
||||
href="http://svn.apache.org/repos/asf/lucene/dev/trunk/modules/analysis/common/src/java/org/apache/lucene/analysis/">modules/analysis/common/src/java/org/apache/lucene/analysis</a>).
|
||||
</p>
|
||||
|
||||
<p>
|
||||
|
|
|
@ -42,7 +42,7 @@ import java.util.Set;
|
|||
* <li> As of 3.1, uses {@link TurkishLowerCaseFilter} for Turkish language.
|
||||
* </ul>
|
||||
* </p>
|
||||
* @deprecated Use the language-specific analyzer in contrib/analyzers instead.
|
||||
* @deprecated Use the language-specific analyzer in modules/analysis instead.
|
||||
* This analyzer will be removed in Lucene 4.0
|
||||
*/
|
||||
@Deprecated
|
||||
|
|
Loading…
Reference in New Issue