LUCENE-2413: directory and package fixes

git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@955203 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Robert Muir 2010-06-16 11:33:29 +00:00
parent aa6afe1c4e
commit 5a661500c1
7 changed files with 8 additions and 8 deletions

View File

@ -23,7 +23,7 @@
# topScoreDocUnordered - Like above, but allows out of order # topScoreDocUnordered - Like above, but allows out of order
collector.class=coll:topScoreDocOrdered:topScoreDocUnordered:topScoreDocOrdered:topScoreDocUnordered collector.class=coll:topScoreDocOrdered:topScoreDocUnordered:topScoreDocOrdered:topScoreDocUnordered
analyzer=org.apache.lucene.analysis.WhitespaceAnalyzer analyzer=org.apache.lucene.analysis.core.WhitespaceAnalyzer
directory=FSDirectory directory=FSDirectory
#directory=RamDirectory #directory=RamDirectory

View File

@ -23,7 +23,7 @@
# topScoreDocUnordered - Like above, but allows out of order # topScoreDocUnordered - Like above, but allows out of order
collector.class=coll:topScoreDocOrdered:topScoreDocUnordered:topScoreDocOrdered:topScoreDocUnordered collector.class=coll:topScoreDocOrdered:topScoreDocUnordered:topScoreDocOrdered:topScoreDocUnordered
analyzer=org.apache.lucene.analysis.WhitespaceAnalyzer analyzer=org.apache.lucene.analysis.core.WhitespaceAnalyzer
directory=FSDirectory directory=FSDirectory
#directory=RamDirectory #directory=RamDirectory

View File

@ -29,7 +29,7 @@
# ant run-task -Dtask.alg=conf/indexLineFile.alg # ant run-task -Dtask.alg=conf/indexLineFile.alg
# #
analyzer=org.apache.lucene.analysis.SimpleAnalyzer analyzer=org.apache.lucene.analysis.core.SimpleAnalyzer
# Feed that knows how to process the line file format: # Feed that knows how to process the line file format:
content.source=org.apache.lucene.benchmark.byTask.feeds.LineDocSource content.source=org.apache.lucene.benchmark.byTask.feeds.LineDocSource

View File

@ -316,7 +316,7 @@ stop words and characters from the index. By stop words and characters I mean c
words such as articles (a, an, the, etc.) and other strings that may have less value for searching words such as articles (a, an, the, etc.) and other strings that may have less value for searching
(e.g. <b>'s</b>) . It should be noted that there are different rules for every language, and you (e.g. <b>'s</b>) . It should be noted that there are different rules for every language, and you
should use the proper analyzer for each. Lucene currently provides Analyzers for a number of should use the proper analyzer for each. Lucene currently provides Analyzers for a number of
different languages (see the <span class="codefrag">*Analyzer.java</span> sources under <a href="http://svn.apache.org/repos/asf/lucene/dev/trunk/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/">contrib/analyzers/src/java/org/apache/lucene/analysis</a>). different languages (see the <span class="codefrag">*Analyzer.java</span> sources under <a href="http://svn.apache.org/repos/asf/lucene/dev/trunk/modules/analysis/common/src/java/org/apache/lucene/analysis/">modules/analysis/common/src/java/org/apache/lucene/analysis</a>).
</p> </p>
<p> <p>
Looking further down in the file, you should see the <span class="codefrag">indexDocs()</span> code. This recursive Looking further down in the file, you should see the <span class="codefrag">indexDocs()</span> code. This recursive

View File

@ -105,7 +105,7 @@ There are many post tokenization steps that can be done, including (but not limi
<li>{@link org.apache.lucene.analysis.PerFieldAnalyzerWrapper} &ndash; Most Analyzers perform the same operation on all <li>{@link org.apache.lucene.analysis.PerFieldAnalyzerWrapper} &ndash; Most Analyzers perform the same operation on all
{@link org.apache.lucene.document.Field}s. The PerFieldAnalyzerWrapper can be used to associate a different Analyzer with different {@link org.apache.lucene.document.Field}s. The PerFieldAnalyzerWrapper can be used to associate a different Analyzer with different
{@link org.apache.lucene.document.Field}s.</li> {@link org.apache.lucene.document.Field}s.</li>
<li>The contrib/analyzers library located at the root of the Lucene distribution has a number of different Analyzer implementations to solve a variety <li>The modules/analysis library located at the root of the Lucene distribution has a number of different Analyzer implementations to solve a variety
of different problems related to searching. Many of the Analyzers are designed to analyze non-English languages.</li> of different problems related to searching. Many of the Analyzers are designed to analyze non-English languages.</li>
<li>The contrib/snowball library <li>The contrib/snowball library
located at the root of the Lucene distribution has Analyzer and TokenFilter located at the root of the Lucene distribution has Analyzer and TokenFilter
@ -168,7 +168,7 @@ There are many post tokenization steps that can be done, including (but not limi
<h2>Implementing your own Analyzer</h2> <h2>Implementing your own Analyzer</h2>
<p>Creating your own Analyzer is straightforward. It usually involves either wrapping an existing Tokenizer and set of TokenFilters to create a new Analyzer <p>Creating your own Analyzer is straightforward. It usually involves either wrapping an existing Tokenizer and set of TokenFilters to create a new Analyzer
or creating both the Analyzer and a Tokenizer or TokenFilter. Before pursuing this approach, you may find it worthwhile or creating both the Analyzer and a Tokenizer or TokenFilter. Before pursuing this approach, you may find it worthwhile
to explore the contrib/analyzers library and/or ask on the java-user@lucene.apache.org mailing list first to see if what you need already exists. to explore the modules/analysis library and/or ask on the java-user@lucene.apache.org mailing list first to see if what you need already exists.
If you are still committed to creating your own Analyzer or TokenStream derivation (Tokenizer or TokenFilter) have a look at If you are still committed to creating your own Analyzer or TokenStream derivation (Tokenizer or TokenFilter) have a look at
the source code of any one of the many samples located in this package. the source code of any one of the many samples located in this package.
</p> </p>

View File

@ -74,7 +74,7 @@ words such as articles (a, an, the, etc.) and other strings that may have less v
(e.g. <b>'s</b>) . It should be noted that there are different rules for every language, and you (e.g. <b>'s</b>) . It should be noted that there are different rules for every language, and you
should use the proper analyzer for each. Lucene currently provides Analyzers for a number of should use the proper analyzer for each. Lucene currently provides Analyzers for a number of
different languages (see the <code>*Analyzer.java</code> sources under <a different languages (see the <code>*Analyzer.java</code> sources under <a
href="http://svn.apache.org/repos/asf/lucene/dev/trunk/lucene/contrib/analyzers/common/src/java/org/apache/lucene/analysis/">contrib/analyzers/src/java/org/apache/lucene/analysis</a>). href="http://svn.apache.org/repos/asf/lucene/dev/trunk/modules/analysis/common/src/java/org/apache/lucene/analysis/">modules/analysis/common/src/java/org/apache/lucene/analysis</a>).
</p> </p>
<p> <p>

View File

@ -42,7 +42,7 @@ import java.util.Set;
* <li> As of 3.1, uses {@link TurkishLowerCaseFilter} for Turkish language. * <li> As of 3.1, uses {@link TurkishLowerCaseFilter} for Turkish language.
* </ul> * </ul>
* </p> * </p>
* @deprecated Use the language-specific analyzer in contrib/analyzers instead. * @deprecated Use the language-specific analyzer in modules/analysis instead.
* This analyzer will be removed in Lucene 4.0 * This analyzer will be removed in Lucene 4.0
*/ */
@Deprecated @Deprecated