mirror of https://github.com/apache/lucene.git
fix numerous javadoc errors, many of which only show up at the private level
git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@808459 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
2e87fb12f1
commit
6c3d17a399
|
@ -176,7 +176,7 @@ public class HHMMSegmenter {
|
|||
/**
|
||||
* Get the character types for every character in a sentence.
|
||||
*
|
||||
* @see Utility#charType(char)
|
||||
* @see Utility#getCharType(char)
|
||||
* @param sentence input sentence
|
||||
* @return array of character types corresponding to character positions in the sentence
|
||||
*/
|
||||
|
|
|
@ -179,10 +179,6 @@ public class QueryScorer implements Scorer {
|
|||
}
|
||||
|
||||
/**
|
||||
* @param query
|
||||
* @param field
|
||||
* @param tokenStream
|
||||
* @param reader
|
||||
*/
|
||||
private void init(Query query, String field, IndexReader reader, boolean expandMultiTermQuery) {
|
||||
this.reader = reader;
|
||||
|
|
|
@ -124,7 +124,6 @@ public class NLS {
|
|||
}
|
||||
|
||||
/**
|
||||
* @param bundleName
|
||||
* @param clazz
|
||||
*/
|
||||
private static void load(Class<Object> clazz) {
|
||||
|
|
|
@ -98,7 +98,7 @@ public class EscapeQuerySyntaxImpl implements EscapeQuerySyntax {
|
|||
/**
|
||||
* replace with ignore case
|
||||
*
|
||||
* @param stringOrig
|
||||
* @param string
|
||||
* string to get replaced
|
||||
* @param sequence1
|
||||
* the old character sequence in lowercase
|
||||
|
@ -164,7 +164,7 @@ public class EscapeQuerySyntaxImpl implements EscapeQuerySyntax {
|
|||
/**
|
||||
* escape all tokens that are part of the parser syntax on a given string
|
||||
*
|
||||
* @param string
|
||||
* @param str
|
||||
* string to get replaced
|
||||
* @param locale
|
||||
* locale to be used when performing string compares
|
||||
|
|
|
@ -106,8 +106,6 @@ public class GroupQueryNodeProcessor implements QueryNodeProcessor {
|
|||
}
|
||||
|
||||
/**
|
||||
* @param node
|
||||
* @return
|
||||
*/
|
||||
private QueryNode applyModifier(QueryNode node, QueryNode parent) {
|
||||
|
||||
|
|
|
@ -99,9 +99,11 @@ public class CheckIndex {
|
|||
/** Directory index is in. */
|
||||
public Directory dir;
|
||||
|
||||
/** SegmentInfos instance containing only segments that
|
||||
* had no problems (this is used with the {@link
|
||||
* CheckIndex#fix} method to repair the index. */
|
||||
/**
|
||||
* SegmentInfos instance containing only segments that
|
||||
* had no problems (this is used with the {@link CheckIndex#fixIndex}
|
||||
* method to repair the index.
|
||||
*/
|
||||
SegmentInfos newSegments;
|
||||
|
||||
/** How many documents will be lost to bad segments. */
|
||||
|
|
|
@ -649,7 +649,6 @@ public class IndexWriter {
|
|||
* @see #release(SegmentReader)
|
||||
* @param info
|
||||
* @param doOpenStores
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public synchronized SegmentReader get(SegmentInfo info, boolean doOpenStores) throws IOException {
|
||||
|
@ -664,7 +663,6 @@ public class IndexWriter {
|
|||
* @param doOpenStores
|
||||
* @param readBufferSize
|
||||
* @param termsIndexDivisor
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public synchronized SegmentReader get(SegmentInfo info, boolean doOpenStores, int readBufferSize, int termsIndexDivisor) throws IOException {
|
||||
|
|
|
@ -21,11 +21,11 @@ package org.apache.lucene.index;
|
|||
/** This is the base class for an in-memory posting list,
|
||||
* keyed by a Token. {@link TermsHash} maintains a hash
|
||||
* table holding one instance of this per unique Token.
|
||||
* Consumers of TermsHash (@link TermsHashConsumer} must
|
||||
* Consumers of TermsHash ({@link TermsHashConsumer}) must
|
||||
* subclass this class with its own concrete class.
|
||||
* {@link FreqProxTermsWriter.RawPostingList} is the
|
||||
* subclass used for the freq/prox postings, and {@link
|
||||
* TermVectorsTermsWriter.PostingList} is the subclass
|
||||
* FreqProxTermsWriter.PostingList is a private inner class used
|
||||
* for the freq/prox postings, and
|
||||
* TermVectorsTermsWriter.PostingList is a private inner class
|
||||
* used to hold TermVectors postings. */
|
||||
|
||||
abstract class RawPostingList {
|
||||
|
|
|
@ -280,7 +280,7 @@ final class BooleanScorer extends Scorer {
|
|||
return false;
|
||||
}
|
||||
|
||||
/** @deprecated use {@link #score(Collector, int)} instead. */
|
||||
/** @deprecated use {@link #score(Collector, int, int)} instead. */
|
||||
protected boolean score(HitCollector hc, int max) throws IOException {
|
||||
return score(new HitCollectorWrapper(hc), max, docID());
|
||||
}
|
||||
|
|
|
@ -300,7 +300,7 @@ class BooleanScorer2 extends Scorer {
|
|||
* {@link HitCollector#collect(int, float)}.
|
||||
* @param max Do not score documents past this.
|
||||
* @return true if more matching documents may remain.
|
||||
* @deprecated use {@link #score(Collector, int)} instead.
|
||||
* @deprecated use {@link #score(Collector, int, int)} instead.
|
||||
*/
|
||||
protected boolean score(HitCollector hc, int max) throws IOException {
|
||||
return score(new HitCollectorWrapper(hc), max, docID());
|
||||
|
|
|
@ -136,7 +136,7 @@ class DisjunctionSumScorer extends Scorer {
|
|||
* {@link HitCollector#collect(int, float)}.
|
||||
* @param max Do not score documents past this.
|
||||
* @return true if more matching documents may remain.
|
||||
* @deprecated use {@link #score(Collector, int)} instead.
|
||||
* @deprecated use {@link #score(Collector, int, int)} instead.
|
||||
*/
|
||||
protected boolean score(HitCollector hc, int max) throws IOException {
|
||||
return score(new HitCollectorWrapper(hc), max, docID());
|
||||
|
|
|
@ -74,7 +74,7 @@ final class TermScorer extends Scorer {
|
|||
score(c, Integer.MAX_VALUE, nextDoc());
|
||||
}
|
||||
|
||||
/** @deprecated use {@link #score(Collector, int)} instead. */
|
||||
/** @deprecated use {@link #score(Collector, int, int)} instead. */
|
||||
protected boolean score(HitCollector c, int end) throws IOException {
|
||||
return score(new HitCollectorWrapper(c), end, doc);
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue