mirror of https://github.com/apache/lucene.git
LUCENE-2185: add @Deprecated annotations
git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@895342 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
cdac1f7113
commit
d22b7a98cd
|
@ -65,6 +65,7 @@ public final class ArabicAnalyzer extends StopwordAnalyzerBase {
|
||||||
* @deprecated use {@link WordlistLoader#getWordSet(File, String)} directly
|
* @deprecated use {@link WordlistLoader#getWordSet(File, String)} directly
|
||||||
*/
|
*/
|
||||||
// TODO make this private
|
// TODO make this private
|
||||||
|
@Deprecated
|
||||||
public static final String STOPWORDS_COMMENT = "#";
|
public static final String STOPWORDS_COMMENT = "#";
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -116,6 +117,7 @@ public final class ArabicAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #ArabicAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #ArabicAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public ArabicAnalyzer( Version matchVersion, String... stopwords ) {
|
public ArabicAnalyzer( Version matchVersion, String... stopwords ) {
|
||||||
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords ));
|
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords ));
|
||||||
}
|
}
|
||||||
|
@ -124,6 +126,7 @@ public final class ArabicAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #ArabicAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #ArabicAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public ArabicAnalyzer( Version matchVersion, Hashtable<?,?> stopwords ) {
|
public ArabicAnalyzer( Version matchVersion, Hashtable<?,?> stopwords ) {
|
||||||
this(matchVersion, stopwords.keySet());
|
this(matchVersion, stopwords.keySet());
|
||||||
}
|
}
|
||||||
|
@ -132,6 +135,7 @@ public final class ArabicAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words. Lines can be commented out using {@link #STOPWORDS_COMMENT}
|
* Builds an analyzer with the given stop words. Lines can be commented out using {@link #STOPWORDS_COMMENT}
|
||||||
* @deprecated use {@link #ArabicAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #ArabicAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public ArabicAnalyzer( Version matchVersion, File stopwords ) throws IOException {
|
public ArabicAnalyzer( Version matchVersion, File stopwords ) throws IOException {
|
||||||
this(matchVersion, WordlistLoader.getWordSet( stopwords, STOPWORDS_COMMENT));
|
this(matchVersion, WordlistLoader.getWordSet( stopwords, STOPWORDS_COMMENT));
|
||||||
}
|
}
|
||||||
|
|
|
@ -59,6 +59,7 @@ public final class BulgarianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* @deprecated use {@link WordlistLoader#getWordSet(File, String)} directly
|
* @deprecated use {@link WordlistLoader#getWordSet(File, String)} directly
|
||||||
*/
|
*/
|
||||||
//TODO make this private
|
//TODO make this private
|
||||||
|
@Deprecated
|
||||||
public static final String STOPWORDS_COMMENT = "#";
|
public static final String STOPWORDS_COMMENT = "#";
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -58,6 +58,7 @@ public final class BrazilianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* @deprecated use {@link #getDefaultStopSet()} instead
|
* @deprecated use {@link #getDefaultStopSet()} instead
|
||||||
*/
|
*/
|
||||||
// TODO make this private in 3.1
|
// TODO make this private in 3.1
|
||||||
|
@Deprecated
|
||||||
public final static String[] BRAZILIAN_STOP_WORDS = {
|
public final static String[] BRAZILIAN_STOP_WORDS = {
|
||||||
"a","ainda","alem","ambas","ambos","antes",
|
"a","ainda","alem","ambas","ambos","antes",
|
||||||
"ao","aonde","aos","apos","aquele","aqueles",
|
"ao","aonde","aos","apos","aquele","aqueles",
|
||||||
|
@ -138,6 +139,7 @@ public final class BrazilianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #BrazilianAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #BrazilianAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public BrazilianAnalyzer(Version matchVersion, String... stopwords) {
|
public BrazilianAnalyzer(Version matchVersion, String... stopwords) {
|
||||||
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
||||||
}
|
}
|
||||||
|
@ -146,6 +148,7 @@ public final class BrazilianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #BrazilianAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #BrazilianAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public BrazilianAnalyzer(Version matchVersion, Map<?,?> stopwords) {
|
public BrazilianAnalyzer(Version matchVersion, Map<?,?> stopwords) {
|
||||||
this(matchVersion, stopwords.keySet());
|
this(matchVersion, stopwords.keySet());
|
||||||
}
|
}
|
||||||
|
@ -154,6 +157,7 @@ public final class BrazilianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #BrazilianAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #BrazilianAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public BrazilianAnalyzer(Version matchVersion, File stopwords)
|
public BrazilianAnalyzer(Version matchVersion, File stopwords)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
this(matchVersion, WordlistLoader.getWordSet(stopwords));
|
this(matchVersion, WordlistLoader.getWordSet(stopwords));
|
||||||
|
@ -163,6 +167,7 @@ public final class BrazilianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an exclusionlist from an array of Strings.
|
* Builds an exclusionlist from an array of Strings.
|
||||||
* @deprecated use {@link #BrazilianAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #BrazilianAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable( String... exclusionlist ) {
|
public void setStemExclusionTable( String... exclusionlist ) {
|
||||||
excltable = StopFilter.makeStopSet( matchVersion, exclusionlist );
|
excltable = StopFilter.makeStopSet( matchVersion, exclusionlist );
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
@ -171,6 +176,7 @@ public final class BrazilianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an exclusionlist from a {@link Map}.
|
* Builds an exclusionlist from a {@link Map}.
|
||||||
* @deprecated use {@link #BrazilianAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #BrazilianAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable( Map<?,?> exclusionlist ) {
|
public void setStemExclusionTable( Map<?,?> exclusionlist ) {
|
||||||
excltable = new HashSet<Object>(exclusionlist.keySet());
|
excltable = new HashSet<Object>(exclusionlist.keySet());
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
@ -179,6 +185,7 @@ public final class BrazilianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an exclusionlist from the words contained in the given file.
|
* Builds an exclusionlist from the words contained in the given file.
|
||||||
* @deprecated use {@link #BrazilianAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #BrazilianAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable( File exclusionlist ) throws IOException {
|
public void setStemExclusionTable( File exclusionlist ) throws IOException {
|
||||||
excltable = WordlistLoader.getWordSet( exclusionlist );
|
excltable = WordlistLoader.getWordSet( exclusionlist );
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
|
|
@ -45,6 +45,7 @@ public final class CJKAnalyzer extends StopwordAnalyzerBase {
|
||||||
*/
|
*/
|
||||||
// TODO make this final in 3.1 -
|
// TODO make this final in 3.1 -
|
||||||
// this might be revised and merged with StopFilter stop words too
|
// this might be revised and merged with StopFilter stop words too
|
||||||
|
@Deprecated
|
||||||
public final static String[] STOP_WORDS = {
|
public final static String[] STOP_WORDS = {
|
||||||
"a", "and", "are", "as", "at", "be",
|
"a", "and", "are", "as", "at", "be",
|
||||||
"but", "by", "for", "if", "in",
|
"but", "by", "for", "if", "in",
|
||||||
|
@ -99,6 +100,7 @@ public final class CJKAnalyzer extends StopwordAnalyzerBase {
|
||||||
* @param stopWords stop word array
|
* @param stopWords stop word array
|
||||||
* @deprecated use {@link #CJKAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #CJKAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public CJKAnalyzer(Version matchVersion, String... stopWords) {
|
public CJKAnalyzer(Version matchVersion, String... stopWords) {
|
||||||
super(matchVersion, StopFilter.makeStopSet(matchVersion, stopWords));
|
super(matchVersion, StopFilter.makeStopSet(matchVersion, stopWords));
|
||||||
}
|
}
|
||||||
|
|
|
@ -80,6 +80,7 @@ public abstract class CompoundWordTokenFilterBase extends TokenFilter {
|
||||||
/**
|
/**
|
||||||
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, String[], int, int, int, boolean)} instead
|
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, String[], int, int, int, boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
protected CompoundWordTokenFilterBase(TokenStream input, String[] dictionary, int minWordSize, int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
protected CompoundWordTokenFilterBase(TokenStream input, String[] dictionary, int minWordSize, int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
||||||
this(Version.LUCENE_30, input, makeDictionary(dictionary),minWordSize,minSubwordSize,maxSubwordSize, onlyLongestMatch);
|
this(Version.LUCENE_30, input, makeDictionary(dictionary),minWordSize,minSubwordSize,maxSubwordSize, onlyLongestMatch);
|
||||||
}
|
}
|
||||||
|
@ -87,6 +88,7 @@ public abstract class CompoundWordTokenFilterBase extends TokenFilter {
|
||||||
/**
|
/**
|
||||||
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, String[], boolean)} instead
|
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, String[], boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
protected CompoundWordTokenFilterBase(TokenStream input, String[] dictionary, boolean onlyLongestMatch) {
|
protected CompoundWordTokenFilterBase(TokenStream input, String[] dictionary, boolean onlyLongestMatch) {
|
||||||
this(Version.LUCENE_30, input, makeDictionary(dictionary),DEFAULT_MIN_WORD_SIZE,DEFAULT_MIN_SUBWORD_SIZE,DEFAULT_MAX_SUBWORD_SIZE, onlyLongestMatch);
|
this(Version.LUCENE_30, input, makeDictionary(dictionary),DEFAULT_MIN_WORD_SIZE,DEFAULT_MIN_SUBWORD_SIZE,DEFAULT_MAX_SUBWORD_SIZE, onlyLongestMatch);
|
||||||
}
|
}
|
||||||
|
@ -94,6 +96,7 @@ public abstract class CompoundWordTokenFilterBase extends TokenFilter {
|
||||||
/**
|
/**
|
||||||
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, Set, boolean)} instead
|
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, Set, boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
protected CompoundWordTokenFilterBase(TokenStream input, Set<?> dictionary, boolean onlyLongestMatch) {
|
protected CompoundWordTokenFilterBase(TokenStream input, Set<?> dictionary, boolean onlyLongestMatch) {
|
||||||
this(Version.LUCENE_30, input, dictionary,DEFAULT_MIN_WORD_SIZE,DEFAULT_MIN_SUBWORD_SIZE,DEFAULT_MAX_SUBWORD_SIZE, onlyLongestMatch);
|
this(Version.LUCENE_30, input, dictionary,DEFAULT_MIN_WORD_SIZE,DEFAULT_MIN_SUBWORD_SIZE,DEFAULT_MAX_SUBWORD_SIZE, onlyLongestMatch);
|
||||||
}
|
}
|
||||||
|
@ -101,6 +104,7 @@ public abstract class CompoundWordTokenFilterBase extends TokenFilter {
|
||||||
/**
|
/**
|
||||||
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, String[])} instead
|
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, String[])} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
protected CompoundWordTokenFilterBase(TokenStream input, String[] dictionary) {
|
protected CompoundWordTokenFilterBase(TokenStream input, String[] dictionary) {
|
||||||
this(Version.LUCENE_30, input, makeDictionary(dictionary),DEFAULT_MIN_WORD_SIZE,DEFAULT_MIN_SUBWORD_SIZE,DEFAULT_MAX_SUBWORD_SIZE, false);
|
this(Version.LUCENE_30, input, makeDictionary(dictionary),DEFAULT_MIN_WORD_SIZE,DEFAULT_MIN_SUBWORD_SIZE,DEFAULT_MAX_SUBWORD_SIZE, false);
|
||||||
}
|
}
|
||||||
|
@ -108,6 +112,7 @@ public abstract class CompoundWordTokenFilterBase extends TokenFilter {
|
||||||
/**
|
/**
|
||||||
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, Set)} instead
|
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
protected CompoundWordTokenFilterBase(TokenStream input, Set<?> dictionary) {
|
protected CompoundWordTokenFilterBase(TokenStream input, Set<?> dictionary) {
|
||||||
this(Version.LUCENE_30, input, dictionary,DEFAULT_MIN_WORD_SIZE,DEFAULT_MIN_SUBWORD_SIZE,DEFAULT_MAX_SUBWORD_SIZE, false);
|
this(Version.LUCENE_30, input, dictionary,DEFAULT_MIN_WORD_SIZE,DEFAULT_MIN_SUBWORD_SIZE,DEFAULT_MAX_SUBWORD_SIZE, false);
|
||||||
}
|
}
|
||||||
|
@ -115,6 +120,7 @@ public abstract class CompoundWordTokenFilterBase extends TokenFilter {
|
||||||
/**
|
/**
|
||||||
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, Set, int, int, int, boolean)} instead
|
* @deprecated use {@link #CompoundWordTokenFilterBase(Version, TokenStream, Set, int, int, int, boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
protected CompoundWordTokenFilterBase(TokenStream input, Set<?> dictionary, int minWordSize, int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
protected CompoundWordTokenFilterBase(TokenStream input, Set<?> dictionary, int minWordSize, int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
||||||
this(Version.LUCENE_30, input, dictionary, minWordSize, minSubwordSize, maxSubwordSize, onlyLongestMatch);
|
this(Version.LUCENE_30, input, dictionary, minWordSize, minSubwordSize, maxSubwordSize, onlyLongestMatch);
|
||||||
}
|
}
|
||||||
|
|
|
@ -46,6 +46,7 @@ public class DictionaryCompoundWordTokenFilter extends CompoundWordTokenFilterBa
|
||||||
* @param onlyLongestMatch Add only the longest matching subword to the stream
|
* @param onlyLongestMatch Add only the longest matching subword to the stream
|
||||||
* @deprecated use {@link #DictionaryCompoundWordTokenFilter(Version, TokenStream, String[], int, int, int, boolean)} instead
|
* @deprecated use {@link #DictionaryCompoundWordTokenFilter(Version, TokenStream, String[], int, int, int, boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public DictionaryCompoundWordTokenFilter(TokenStream input, String[] dictionary,
|
public DictionaryCompoundWordTokenFilter(TokenStream input, String[] dictionary,
|
||||||
int minWordSize, int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
int minWordSize, int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
||||||
super(Version.LUCENE_30, input, dictionary, minWordSize, minSubwordSize, maxSubwordSize, onlyLongestMatch);
|
super(Version.LUCENE_30, input, dictionary, minWordSize, minSubwordSize, maxSubwordSize, onlyLongestMatch);
|
||||||
|
@ -58,6 +59,7 @@ public class DictionaryCompoundWordTokenFilter extends CompoundWordTokenFilterBa
|
||||||
* @param dictionary the word dictionary to match against
|
* @param dictionary the word dictionary to match against
|
||||||
* @deprecated use {@link #DictionaryCompoundWordTokenFilter(Version, TokenStream, String[])} instead
|
* @deprecated use {@link #DictionaryCompoundWordTokenFilter(Version, TokenStream, String[])} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public DictionaryCompoundWordTokenFilter(TokenStream input, String[] dictionary) {
|
public DictionaryCompoundWordTokenFilter(TokenStream input, String[] dictionary) {
|
||||||
super(Version.LUCENE_30, input, dictionary);
|
super(Version.LUCENE_30, input, dictionary);
|
||||||
}
|
}
|
||||||
|
@ -70,6 +72,7 @@ public class DictionaryCompoundWordTokenFilter extends CompoundWordTokenFilterBa
|
||||||
* lower case strings.
|
* lower case strings.
|
||||||
* @deprecated use {@link #DictionaryCompoundWordTokenFilter(Version, TokenStream, Set)} instead
|
* @deprecated use {@link #DictionaryCompoundWordTokenFilter(Version, TokenStream, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public DictionaryCompoundWordTokenFilter(TokenStream input, Set dictionary) {
|
public DictionaryCompoundWordTokenFilter(TokenStream input, Set dictionary) {
|
||||||
super(Version.LUCENE_30, input, dictionary);
|
super(Version.LUCENE_30, input, dictionary);
|
||||||
}
|
}
|
||||||
|
@ -86,6 +89,7 @@ public class DictionaryCompoundWordTokenFilter extends CompoundWordTokenFilterBa
|
||||||
* @param onlyLongestMatch Add only the longest matching subword to the stream
|
* @param onlyLongestMatch Add only the longest matching subword to the stream
|
||||||
* @deprecated use {@link #DictionaryCompoundWordTokenFilter(Version, TokenStream, Set, int, int, int, boolean)} instead
|
* @deprecated use {@link #DictionaryCompoundWordTokenFilter(Version, TokenStream, Set, int, int, int, boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public DictionaryCompoundWordTokenFilter(TokenStream input, Set dictionary,
|
public DictionaryCompoundWordTokenFilter(TokenStream input, Set dictionary,
|
||||||
int minWordSize, int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
int minWordSize, int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
||||||
super(Version.LUCENE_30, input, dictionary, minWordSize, minSubwordSize, maxSubwordSize, onlyLongestMatch);
|
super(Version.LUCENE_30, input, dictionary, minWordSize, minSubwordSize, maxSubwordSize, onlyLongestMatch);
|
||||||
|
|
|
@ -167,6 +167,7 @@ public class HyphenationCompoundWordTokenFilter extends
|
||||||
* @param onlyLongestMatch Add only the longest matching subword to the stream
|
* @param onlyLongestMatch Add only the longest matching subword to the stream
|
||||||
* @deprecated use {@link #HyphenationCompoundWordTokenFilter(Version, TokenStream, HyphenationTree, String[], int, int, int, boolean)} instead.
|
* @deprecated use {@link #HyphenationCompoundWordTokenFilter(Version, TokenStream, HyphenationTree, String[], int, int, int, boolean)} instead.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public HyphenationCompoundWordTokenFilter(TokenStream input,
|
public HyphenationCompoundWordTokenFilter(TokenStream input,
|
||||||
HyphenationTree hyphenator, String[] dictionary, int minWordSize,
|
HyphenationTree hyphenator, String[] dictionary, int minWordSize,
|
||||||
int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
||||||
|
@ -182,6 +183,7 @@ public class HyphenationCompoundWordTokenFilter extends
|
||||||
* @param dictionary the word dictionary to match against
|
* @param dictionary the word dictionary to match against
|
||||||
* @deprecated use {@link #HyphenationCompoundWordTokenFilter(Version, TokenStream, HyphenationTree, String[])} instead.
|
* @deprecated use {@link #HyphenationCompoundWordTokenFilter(Version, TokenStream, HyphenationTree, String[])} instead.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public HyphenationCompoundWordTokenFilter(TokenStream input,
|
public HyphenationCompoundWordTokenFilter(TokenStream input,
|
||||||
HyphenationTree hyphenator, String[] dictionary) {
|
HyphenationTree hyphenator, String[] dictionary) {
|
||||||
this(Version.LUCENE_30, input, hyphenator, makeDictionary(dictionary), DEFAULT_MIN_WORD_SIZE,
|
this(Version.LUCENE_30, input, hyphenator, makeDictionary(dictionary), DEFAULT_MIN_WORD_SIZE,
|
||||||
|
@ -197,6 +199,7 @@ public class HyphenationCompoundWordTokenFilter extends
|
||||||
* lower case strings.
|
* lower case strings.
|
||||||
* @deprecated use {@link #HyphenationCompoundWordTokenFilter(Version, TokenStream, HyphenationTree, Set)} instead.
|
* @deprecated use {@link #HyphenationCompoundWordTokenFilter(Version, TokenStream, HyphenationTree, Set)} instead.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public HyphenationCompoundWordTokenFilter(TokenStream input,
|
public HyphenationCompoundWordTokenFilter(TokenStream input,
|
||||||
HyphenationTree hyphenator, Set dictionary) {
|
HyphenationTree hyphenator, Set dictionary) {
|
||||||
this(Version.LUCENE_30, input, hyphenator, dictionary, DEFAULT_MIN_WORD_SIZE,
|
this(Version.LUCENE_30, input, hyphenator, dictionary, DEFAULT_MIN_WORD_SIZE,
|
||||||
|
@ -218,6 +221,7 @@ public class HyphenationCompoundWordTokenFilter extends
|
||||||
* @param onlyLongestMatch Add only the longest matching subword to the stream
|
* @param onlyLongestMatch Add only the longest matching subword to the stream
|
||||||
* @deprecated use {@link #HyphenationCompoundWordTokenFilter(Version, TokenStream, HyphenationTree, Set, int, int, int, boolean)} instead.
|
* @deprecated use {@link #HyphenationCompoundWordTokenFilter(Version, TokenStream, HyphenationTree, Set, int, int, int, boolean)} instead.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public HyphenationCompoundWordTokenFilter(TokenStream input,
|
public HyphenationCompoundWordTokenFilter(TokenStream input,
|
||||||
HyphenationTree hyphenator, Set dictionary, int minWordSize,
|
HyphenationTree hyphenator, Set dictionary, int minWordSize,
|
||||||
int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
int minSubwordSize, int maxSubwordSize, boolean onlyLongestMatch) {
|
||||||
|
|
|
@ -62,6 +62,7 @@ public final class CzechAnalyzer extends ReusableAnalyzerBase {
|
||||||
* @deprecated use {@link #getDefaultStopSet()} instead
|
* @deprecated use {@link #getDefaultStopSet()} instead
|
||||||
*/
|
*/
|
||||||
// TODO make this private in 3.1
|
// TODO make this private in 3.1
|
||||||
|
@Deprecated
|
||||||
public final static String[] CZECH_STOP_WORDS = {
|
public final static String[] CZECH_STOP_WORDS = {
|
||||||
"a","s","k","o","i","u","v","z","dnes","cz","t\u00edmto","bude\u0161","budem",
|
"a","s","k","o","i","u","v","z","dnes","cz","t\u00edmto","bude\u0161","budem",
|
||||||
"byli","jse\u0161","m\u016fj","sv\u00fdm","ta","tomto","tohle","tuto","tyto",
|
"byli","jse\u0161","m\u016fj","sv\u00fdm","ta","tomto","tohle","tuto","tyto",
|
||||||
|
@ -136,6 +137,7 @@ public final class CzechAnalyzer extends ReusableAnalyzerBase {
|
||||||
* @param stopwords a stopword set
|
* @param stopwords a stopword set
|
||||||
* @deprecated use {@link #CzechAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #CzechAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public CzechAnalyzer(Version matchVersion, String... stopwords) {
|
public CzechAnalyzer(Version matchVersion, String... stopwords) {
|
||||||
this(matchVersion, StopFilter.makeStopSet( matchVersion, stopwords ));
|
this(matchVersion, StopFilter.makeStopSet( matchVersion, stopwords ));
|
||||||
}
|
}
|
||||||
|
@ -148,6 +150,7 @@ public final class CzechAnalyzer extends ReusableAnalyzerBase {
|
||||||
* @param stopwords a stopword set
|
* @param stopwords a stopword set
|
||||||
* @deprecated use {@link #CzechAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #CzechAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public CzechAnalyzer(Version matchVersion, HashSet<?> stopwords) {
|
public CzechAnalyzer(Version matchVersion, HashSet<?> stopwords) {
|
||||||
this(matchVersion, (Set<?>)stopwords);
|
this(matchVersion, (Set<?>)stopwords);
|
||||||
}
|
}
|
||||||
|
@ -160,6 +163,7 @@ public final class CzechAnalyzer extends ReusableAnalyzerBase {
|
||||||
* @param stopwords a file containing stopwords
|
* @param stopwords a file containing stopwords
|
||||||
* @deprecated use {@link #CzechAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #CzechAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public CzechAnalyzer(Version matchVersion, File stopwords ) throws IOException {
|
public CzechAnalyzer(Version matchVersion, File stopwords ) throws IOException {
|
||||||
this(matchVersion, (Set<?>)WordlistLoader.getWordSet( stopwords ));
|
this(matchVersion, (Set<?>)WordlistLoader.getWordSet( stopwords ));
|
||||||
}
|
}
|
||||||
|
@ -172,6 +176,7 @@ public final class CzechAnalyzer extends ReusableAnalyzerBase {
|
||||||
* and {@link #CzechAnalyzer(Version, Set)} instead
|
* and {@link #CzechAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
// TODO extend StopwordAnalyzerBase once this method is gone!
|
// TODO extend StopwordAnalyzerBase once this method is gone!
|
||||||
|
@Deprecated
|
||||||
public void loadStopWords( InputStream wordfile, String encoding ) {
|
public void loadStopWords( InputStream wordfile, String encoding ) {
|
||||||
setPreviousTokenStream(null); // force a new stopfilter to be created
|
setPreviousTokenStream(null); // force a new stopfilter to be created
|
||||||
if ( wordfile == null ) {
|
if ( wordfile == null ) {
|
||||||
|
|
|
@ -60,6 +60,7 @@ public final class GermanAnalyzer extends StopwordAnalyzerBase {
|
||||||
* @deprecated use {@link #getDefaultStopSet()} instead
|
* @deprecated use {@link #getDefaultStopSet()} instead
|
||||||
*/
|
*/
|
||||||
//TODO make this private in 3.1
|
//TODO make this private in 3.1
|
||||||
|
@Deprecated
|
||||||
public final static String[] GERMAN_STOP_WORDS = {
|
public final static String[] GERMAN_STOP_WORDS = {
|
||||||
"einer", "eine", "eines", "einem", "einen",
|
"einer", "eine", "eines", "einem", "einen",
|
||||||
"der", "die", "das", "dass", "daß",
|
"der", "die", "das", "dass", "daß",
|
||||||
|
@ -137,6 +138,7 @@ public final class GermanAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #GermanAnalyzer(Version, Set)}
|
* @deprecated use {@link #GermanAnalyzer(Version, Set)}
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public GermanAnalyzer(Version matchVersion, String... stopwords) {
|
public GermanAnalyzer(Version matchVersion, String... stopwords) {
|
||||||
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
||||||
}
|
}
|
||||||
|
@ -145,6 +147,7 @@ public final class GermanAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #GermanAnalyzer(Version, Set)}
|
* @deprecated use {@link #GermanAnalyzer(Version, Set)}
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public GermanAnalyzer(Version matchVersion, Map<?,?> stopwords) {
|
public GermanAnalyzer(Version matchVersion, Map<?,?> stopwords) {
|
||||||
this(matchVersion, stopwords.keySet());
|
this(matchVersion, stopwords.keySet());
|
||||||
|
|
||||||
|
@ -154,6 +157,7 @@ public final class GermanAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #GermanAnalyzer(Version, Set)}
|
* @deprecated use {@link #GermanAnalyzer(Version, Set)}
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public GermanAnalyzer(Version matchVersion, File stopwords) throws IOException {
|
public GermanAnalyzer(Version matchVersion, File stopwords) throws IOException {
|
||||||
this(matchVersion, WordlistLoader.getWordSet(stopwords));
|
this(matchVersion, WordlistLoader.getWordSet(stopwords));
|
||||||
}
|
}
|
||||||
|
@ -162,6 +166,7 @@ public final class GermanAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an exclusionlist from an array of Strings.
|
* Builds an exclusionlist from an array of Strings.
|
||||||
* @deprecated use {@link #GermanAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #GermanAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable(String[] exclusionlist) {
|
public void setStemExclusionTable(String[] exclusionlist) {
|
||||||
exclusionSet = StopFilter.makeStopSet(matchVersion, exclusionlist);
|
exclusionSet = StopFilter.makeStopSet(matchVersion, exclusionlist);
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
@ -171,6 +176,7 @@ public final class GermanAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an exclusionlist from a {@link Map}
|
* Builds an exclusionlist from a {@link Map}
|
||||||
* @deprecated use {@link #GermanAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #GermanAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable(Map<?,?> exclusionlist) {
|
public void setStemExclusionTable(Map<?,?> exclusionlist) {
|
||||||
exclusionSet = new HashSet<Object>(exclusionlist.keySet());
|
exclusionSet = new HashSet<Object>(exclusionlist.keySet());
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
@ -180,6 +186,7 @@ public final class GermanAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an exclusionlist from the words contained in the given file.
|
* Builds an exclusionlist from the words contained in the given file.
|
||||||
* @deprecated use {@link #GermanAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #GermanAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable(File exclusionlist) throws IOException {
|
public void setStemExclusionTable(File exclusionlist) throws IOException {
|
||||||
exclusionSet = WordlistLoader.getWordSet(exclusionlist);
|
exclusionSet = WordlistLoader.getWordSet(exclusionlist);
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
|
|
@ -95,6 +95,7 @@ public final class GreekAnalyzer extends StopwordAnalyzerBase
|
||||||
* @param stopwords Array of stopwords to use.
|
* @param stopwords Array of stopwords to use.
|
||||||
* @deprecated use {@link #GreekAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #GreekAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public GreekAnalyzer(Version matchVersion, String... stopwords)
|
public GreekAnalyzer(Version matchVersion, String... stopwords)
|
||||||
{
|
{
|
||||||
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
||||||
|
@ -104,6 +105,7 @@ public final class GreekAnalyzer extends StopwordAnalyzerBase
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #GreekAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #GreekAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public GreekAnalyzer(Version matchVersion, Map<?,?> stopwords)
|
public GreekAnalyzer(Version matchVersion, Map<?,?> stopwords)
|
||||||
{
|
{
|
||||||
this(matchVersion, stopwords.keySet());
|
this(matchVersion, stopwords.keySet());
|
||||||
|
|
|
@ -111,6 +111,7 @@ public final class PersianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #PersianAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #PersianAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public PersianAnalyzer(Version matchVersion, String... stopwords) {
|
public PersianAnalyzer(Version matchVersion, String... stopwords) {
|
||||||
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
||||||
}
|
}
|
||||||
|
@ -119,6 +120,7 @@ public final class PersianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #PersianAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #PersianAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public PersianAnalyzer(Version matchVersion, Hashtable<?, ?> stopwords) {
|
public PersianAnalyzer(Version matchVersion, Hashtable<?, ?> stopwords) {
|
||||||
this(matchVersion, stopwords.keySet());
|
this(matchVersion, stopwords.keySet());
|
||||||
}
|
}
|
||||||
|
@ -128,6 +130,7 @@ public final class PersianAnalyzer extends StopwordAnalyzerBase {
|
||||||
* using {@link #STOPWORDS_COMMENT}
|
* using {@link #STOPWORDS_COMMENT}
|
||||||
* @deprecated use {@link #PersianAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #PersianAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public PersianAnalyzer(Version matchVersion, File stopwords) throws IOException {
|
public PersianAnalyzer(Version matchVersion, File stopwords) throws IOException {
|
||||||
this(matchVersion, WordlistLoader.getWordSet(stopwords, STOPWORDS_COMMENT));
|
this(matchVersion, WordlistLoader.getWordSet(stopwords, STOPWORDS_COMMENT));
|
||||||
}
|
}
|
||||||
|
|
|
@ -50,6 +50,7 @@ public final class ElisionFilter extends TokenFilter {
|
||||||
* @param articles a set of articles
|
* @param articles a set of articles
|
||||||
* @deprecated use {@link #ElisionFilter(Version, TokenStream, Set)} instead
|
* @deprecated use {@link #ElisionFilter(Version, TokenStream, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setArticles(Version matchVersion, Set<?> articles) {
|
public void setArticles(Version matchVersion, Set<?> articles) {
|
||||||
this.articles = CharArraySet.unmodifiableSet(
|
this.articles = CharArraySet.unmodifiableSet(
|
||||||
CharArraySet.copy(matchVersion, articles));
|
CharArraySet.copy(matchVersion, articles));
|
||||||
|
@ -60,6 +61,7 @@ public final class ElisionFilter extends TokenFilter {
|
||||||
* @param articles a set of articles
|
* @param articles a set of articles
|
||||||
* @deprecated use {@link #setArticles(Version, Set)} instead
|
* @deprecated use {@link #setArticles(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setArticles(Set<?> articles) {
|
public void setArticles(Set<?> articles) {
|
||||||
setArticles(Version.LUCENE_CURRENT, articles);
|
setArticles(Version.LUCENE_CURRENT, articles);
|
||||||
}
|
}
|
||||||
|
@ -74,6 +76,7 @@ public final class ElisionFilter extends TokenFilter {
|
||||||
* Constructs an elision filter with standard stop words
|
* Constructs an elision filter with standard stop words
|
||||||
* @deprecated use {@link #ElisionFilter(Version, TokenStream)} instead
|
* @deprecated use {@link #ElisionFilter(Version, TokenStream)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
protected ElisionFilter(TokenStream input) {
|
protected ElisionFilter(TokenStream input) {
|
||||||
this(Version.LUCENE_30, input);
|
this(Version.LUCENE_30, input);
|
||||||
}
|
}
|
||||||
|
@ -82,6 +85,7 @@ public final class ElisionFilter extends TokenFilter {
|
||||||
* Constructs an elision filter with a Set of stop words
|
* Constructs an elision filter with a Set of stop words
|
||||||
* @deprecated use {@link #ElisionFilter(Version, TokenStream, Set)} instead
|
* @deprecated use {@link #ElisionFilter(Version, TokenStream, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public ElisionFilter(TokenStream input, Set<?> articles) {
|
public ElisionFilter(TokenStream input, Set<?> articles) {
|
||||||
this(Version.LUCENE_30, input, articles);
|
this(Version.LUCENE_30, input, articles);
|
||||||
}
|
}
|
||||||
|
@ -103,6 +107,7 @@ public final class ElisionFilter extends TokenFilter {
|
||||||
* Constructs an elision filter with an array of stop words
|
* Constructs an elision filter with an array of stop words
|
||||||
* @deprecated use {@link #ElisionFilter(Version, TokenStream, Set)} instead
|
* @deprecated use {@link #ElisionFilter(Version, TokenStream, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public ElisionFilter(TokenStream input, String[] articles) {
|
public ElisionFilter(TokenStream input, String[] articles) {
|
||||||
this(Version.LUCENE_CURRENT, input,
|
this(Version.LUCENE_CURRENT, input,
|
||||||
new CharArraySet(Version.LUCENE_CURRENT,
|
new CharArraySet(Version.LUCENE_CURRENT,
|
||||||
|
|
|
@ -68,6 +68,7 @@ public final class FrenchAnalyzer extends StopwordAnalyzerBase {
|
||||||
* @deprecated use {@link #getDefaultStopSet()} instead
|
* @deprecated use {@link #getDefaultStopSet()} instead
|
||||||
*/
|
*/
|
||||||
// TODO make this private in 3.1
|
// TODO make this private in 3.1
|
||||||
|
@Deprecated
|
||||||
public final static String[] FRENCH_STOP_WORDS = {
|
public final static String[] FRENCH_STOP_WORDS = {
|
||||||
"a", "afin", "ai", "ainsi", "après", "attendu", "au", "aujourd", "auquel", "aussi",
|
"a", "afin", "ai", "ainsi", "après", "attendu", "au", "aujourd", "auquel", "aussi",
|
||||||
"autre", "autres", "aux", "auxquelles", "auxquels", "avait", "avant", "avec", "avoir",
|
"autre", "autres", "aux", "auxquelles", "auxquels", "avait", "avant", "avec", "avoir",
|
||||||
|
@ -154,6 +155,7 @@ public final class FrenchAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #FrenchAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #FrenchAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public FrenchAnalyzer(Version matchVersion, String... stopwords) {
|
public FrenchAnalyzer(Version matchVersion, String... stopwords) {
|
||||||
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
||||||
}
|
}
|
||||||
|
@ -163,6 +165,7 @@ public final class FrenchAnalyzer extends StopwordAnalyzerBase {
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
* @deprecated use {@link #FrenchAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #FrenchAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public FrenchAnalyzer(Version matchVersion, File stopwords) throws IOException {
|
public FrenchAnalyzer(Version matchVersion, File stopwords) throws IOException {
|
||||||
this(matchVersion, WordlistLoader.getWordSet(stopwords));
|
this(matchVersion, WordlistLoader.getWordSet(stopwords));
|
||||||
}
|
}
|
||||||
|
@ -171,6 +174,7 @@ public final class FrenchAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an exclusionlist from an array of Strings.
|
* Builds an exclusionlist from an array of Strings.
|
||||||
* @deprecated use {@link #FrenchAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #FrenchAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable(String... exclusionlist) {
|
public void setStemExclusionTable(String... exclusionlist) {
|
||||||
excltable = StopFilter.makeStopSet(matchVersion, exclusionlist);
|
excltable = StopFilter.makeStopSet(matchVersion, exclusionlist);
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
@ -180,6 +184,7 @@ public final class FrenchAnalyzer extends StopwordAnalyzerBase {
|
||||||
* Builds an exclusionlist from a Map.
|
* Builds an exclusionlist from a Map.
|
||||||
* @deprecated use {@link #FrenchAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #FrenchAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable(Map<?,?> exclusionlist) {
|
public void setStemExclusionTable(Map<?,?> exclusionlist) {
|
||||||
excltable = new HashSet<Object>(exclusionlist.keySet());
|
excltable = new HashSet<Object>(exclusionlist.keySet());
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
@ -190,6 +195,7 @@ public final class FrenchAnalyzer extends StopwordAnalyzerBase {
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
* @deprecated use {@link #FrenchAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #FrenchAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable(File exclusionlist) throws IOException {
|
public void setStemExclusionTable(File exclusionlist) throws IOException {
|
||||||
excltable = new HashSet<Object>(WordlistLoader.getWordSet(exclusionlist));
|
excltable = new HashSet<Object>(WordlistLoader.getWordSet(exclusionlist));
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
|
|
@ -57,6 +57,7 @@ public final class DutchAnalyzer extends Analyzer {
|
||||||
* List of typical Dutch stopwords.
|
* List of typical Dutch stopwords.
|
||||||
* @deprecated use {@link #getDefaultStopSet()} instead
|
* @deprecated use {@link #getDefaultStopSet()} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public final static String[] DUTCH_STOP_WORDS =
|
public final static String[] DUTCH_STOP_WORDS =
|
||||||
{
|
{
|
||||||
"de", "en", "van", "ik", "te", "dat", "die", "in", "een",
|
"de", "en", "van", "ik", "te", "dat", "die", "in", "een",
|
||||||
|
@ -128,6 +129,7 @@ public final class DutchAnalyzer extends Analyzer {
|
||||||
* @param stopwords
|
* @param stopwords
|
||||||
* @deprecated use {@link #DutchAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #DutchAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public DutchAnalyzer(Version matchVersion, String... stopwords) {
|
public DutchAnalyzer(Version matchVersion, String... stopwords) {
|
||||||
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
||||||
}
|
}
|
||||||
|
@ -138,6 +140,7 @@ public final class DutchAnalyzer extends Analyzer {
|
||||||
* @param stopwords
|
* @param stopwords
|
||||||
* @deprecated use {@link #DutchAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #DutchAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public DutchAnalyzer(Version matchVersion, HashSet<?> stopwords) {
|
public DutchAnalyzer(Version matchVersion, HashSet<?> stopwords) {
|
||||||
this(matchVersion, (Set<?>)stopwords);
|
this(matchVersion, (Set<?>)stopwords);
|
||||||
}
|
}
|
||||||
|
@ -148,6 +151,7 @@ public final class DutchAnalyzer extends Analyzer {
|
||||||
* @param stopwords
|
* @param stopwords
|
||||||
* @deprecated use {@link #DutchAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #DutchAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public DutchAnalyzer(Version matchVersion, File stopwords) {
|
public DutchAnalyzer(Version matchVersion, File stopwords) {
|
||||||
// this is completely broken!
|
// this is completely broken!
|
||||||
try {
|
try {
|
||||||
|
@ -165,6 +169,7 @@ public final class DutchAnalyzer extends Analyzer {
|
||||||
* @param exclusionlist
|
* @param exclusionlist
|
||||||
* @deprecated use {@link #DutchAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #DutchAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable(String... exclusionlist) {
|
public void setStemExclusionTable(String... exclusionlist) {
|
||||||
excltable = StopFilter.makeStopSet(matchVersion, exclusionlist);
|
excltable = StopFilter.makeStopSet(matchVersion, exclusionlist);
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
@ -174,6 +179,7 @@ public final class DutchAnalyzer extends Analyzer {
|
||||||
* Builds an exclusionlist from a Hashtable.
|
* Builds an exclusionlist from a Hashtable.
|
||||||
* @deprecated use {@link #DutchAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #DutchAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable(HashSet<?> exclusionlist) {
|
public void setStemExclusionTable(HashSet<?> exclusionlist) {
|
||||||
excltable = exclusionlist;
|
excltable = exclusionlist;
|
||||||
setPreviousTokenStream(null); // force a new stemmer to be created
|
setPreviousTokenStream(null); // force a new stemmer to be created
|
||||||
|
@ -183,6 +189,7 @@ public final class DutchAnalyzer extends Analyzer {
|
||||||
* Builds an exclusionlist from the words contained in the given file.
|
* Builds an exclusionlist from the words contained in the given file.
|
||||||
* @deprecated use {@link #DutchAnalyzer(Version, Set, Set)} instead
|
* @deprecated use {@link #DutchAnalyzer(Version, Set, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setStemExclusionTable(File exclusionlist) {
|
public void setStemExclusionTable(File exclusionlist) {
|
||||||
try {
|
try {
|
||||||
excltable = org.apache.lucene.analysis.WordlistLoader.getWordSet(exclusionlist);
|
excltable = org.apache.lucene.analysis.WordlistLoader.getWordSet(exclusionlist);
|
||||||
|
|
|
@ -78,6 +78,7 @@ public final class ReverseStringFilter extends TokenFilter {
|
||||||
* @deprecated use {@link #ReverseStringFilter(Version, TokenStream)}
|
* @deprecated use {@link #ReverseStringFilter(Version, TokenStream)}
|
||||||
* instead. This constructor will be removed in Lucene 4.0
|
* instead. This constructor will be removed in Lucene 4.0
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public ReverseStringFilter(TokenStream in) {
|
public ReverseStringFilter(TokenStream in) {
|
||||||
this(in, NOMARKER);
|
this(in, NOMARKER);
|
||||||
}
|
}
|
||||||
|
@ -95,6 +96,7 @@ public final class ReverseStringFilter extends TokenFilter {
|
||||||
* @deprecated use {@link #ReverseStringFilter(Version, TokenStream, char)}
|
* @deprecated use {@link #ReverseStringFilter(Version, TokenStream, char)}
|
||||||
* instead. This constructor will be removed in Lucene 4.0
|
* instead. This constructor will be removed in Lucene 4.0
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public ReverseStringFilter(TokenStream in, char marker) {
|
public ReverseStringFilter(TokenStream in, char marker) {
|
||||||
this(Version.LUCENE_30, in, marker);
|
this(Version.LUCENE_30, in, marker);
|
||||||
}
|
}
|
||||||
|
@ -157,6 +159,7 @@ public final class ReverseStringFilter extends TokenFilter {
|
||||||
* @deprecated use {@link #reverse(Version, String)} instead. This method
|
* @deprecated use {@link #reverse(Version, String)} instead. This method
|
||||||
* will be removed in Lucene 4.0
|
* will be removed in Lucene 4.0
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static String reverse( final String input ){
|
public static String reverse( final String input ){
|
||||||
return reverse(Version.LUCENE_30, input);
|
return reverse(Version.LUCENE_30, input);
|
||||||
}
|
}
|
||||||
|
@ -180,6 +183,7 @@ public final class ReverseStringFilter extends TokenFilter {
|
||||||
* @deprecated use {@link #reverse(Version, char[])} instead. This
|
* @deprecated use {@link #reverse(Version, char[])} instead. This
|
||||||
* method will be removed in Lucene 4.0
|
* method will be removed in Lucene 4.0
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static void reverse( final char[] buffer ){
|
public static void reverse( final char[] buffer ){
|
||||||
reverse( buffer, 0, buffer.length );
|
reverse( buffer, 0, buffer.length );
|
||||||
}
|
}
|
||||||
|
@ -202,6 +206,7 @@ public final class ReverseStringFilter extends TokenFilter {
|
||||||
* @deprecated use {@link #reverse(Version, char[], int)} instead. This
|
* @deprecated use {@link #reverse(Version, char[], int)} instead. This
|
||||||
* method will be removed in Lucene 4.0
|
* method will be removed in Lucene 4.0
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static void reverse( final char[] buffer, final int len ){
|
public static void reverse( final char[] buffer, final int len ){
|
||||||
reverse( buffer, 0, len );
|
reverse( buffer, 0, len );
|
||||||
}
|
}
|
||||||
|
@ -229,6 +234,7 @@ public final class ReverseStringFilter extends TokenFilter {
|
||||||
* @deprecated use {@link #reverse(Version, char[], int, int)} instead. This
|
* @deprecated use {@link #reverse(Version, char[], int, int)} instead. This
|
||||||
* method will be removed in Lucene 4.0
|
* method will be removed in Lucene 4.0
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static void reverse(char[] buffer, int start, int len ) {
|
public static void reverse(char[] buffer, int start, int len ) {
|
||||||
reverseUnicode3(buffer, start, len);
|
reverseUnicode3(buffer, start, len);
|
||||||
}
|
}
|
||||||
|
@ -236,6 +242,7 @@ public final class ReverseStringFilter extends TokenFilter {
|
||||||
/**
|
/**
|
||||||
* @deprecated Remove this when support for 3.0 indexes is no longer needed.
|
* @deprecated Remove this when support for 3.0 indexes is no longer needed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
private static void reverseUnicode3( char[] buffer, int start, int len ){
|
private static void reverseUnicode3( char[] buffer, int start, int len ){
|
||||||
if( len <= 1 ) return;
|
if( len <= 1 ) return;
|
||||||
int num = len>>1;
|
int num = len>>1;
|
||||||
|
|
|
@ -72,6 +72,7 @@ public final class RussianAnalyzer extends StopwordAnalyzerBase
|
||||||
* Builds an analyzer with the given stop words.
|
* Builds an analyzer with the given stop words.
|
||||||
* @deprecated use {@link #RussianAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #RussianAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public RussianAnalyzer(Version matchVersion, String... stopwords) {
|
public RussianAnalyzer(Version matchVersion, String... stopwords) {
|
||||||
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
this(matchVersion, StopFilter.makeStopSet(matchVersion, stopwords));
|
||||||
}
|
}
|
||||||
|
@ -93,6 +94,7 @@ public final class RussianAnalyzer extends StopwordAnalyzerBase
|
||||||
* TODO: create a Set version of this ctor
|
* TODO: create a Set version of this ctor
|
||||||
* @deprecated use {@link #RussianAnalyzer(Version, Set)} instead
|
* @deprecated use {@link #RussianAnalyzer(Version, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public RussianAnalyzer(Version matchVersion, Map<?,?> stopwords)
|
public RussianAnalyzer(Version matchVersion, Map<?,?> stopwords)
|
||||||
{
|
{
|
||||||
this(matchVersion, stopwords.keySet());
|
this(matchVersion, stopwords.keySet());
|
||||||
|
|
|
@ -29,6 +29,7 @@ import org.apache.lucene.analysis.tokenattributes.TermAttribute;
|
||||||
* @deprecated Use {@link LowerCaseFilter} instead, which has the same
|
* @deprecated Use {@link LowerCaseFilter} instead, which has the same
|
||||||
* functionality. This filter will be removed in Lucene 4.0
|
* functionality. This filter will be removed in Lucene 4.0
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public final class RussianLowerCaseFilter extends TokenFilter
|
public final class RussianLowerCaseFilter extends TokenFilter
|
||||||
{
|
{
|
||||||
private TermAttribute termAtt;
|
private TermAttribute termAtt;
|
||||||
|
|
|
@ -40,6 +40,7 @@ public class TestCzechAnalyzer extends BaseTokenStreamTestCase {
|
||||||
/**
|
/**
|
||||||
* @deprecated Remove this test when support for 3.0 indexes is no longer needed.
|
* @deprecated Remove this test when support for 3.0 indexes is no longer needed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void testStopWordLegacy() throws Exception {
|
public void testStopWordLegacy() throws Exception {
|
||||||
assertAnalyzesTo(new CzechAnalyzer(Version.LUCENE_30), "Pokud mluvime o volnem",
|
assertAnalyzesTo(new CzechAnalyzer(Version.LUCENE_30), "Pokud mluvime o volnem",
|
||||||
new String[] { "mluvime", "volnem" });
|
new String[] { "mluvime", "volnem" });
|
||||||
|
@ -53,6 +54,7 @@ public class TestCzechAnalyzer extends BaseTokenStreamTestCase {
|
||||||
/**
|
/**
|
||||||
* @deprecated Remove this test when support for 3.0 indexes is no longer needed.
|
* @deprecated Remove this test when support for 3.0 indexes is no longer needed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void testReusableTokenStreamLegacy() throws Exception {
|
public void testReusableTokenStreamLegacy() throws Exception {
|
||||||
Analyzer analyzer = new CzechAnalyzer(Version.LUCENE_30);
|
Analyzer analyzer = new CzechAnalyzer(Version.LUCENE_30);
|
||||||
assertAnalyzesToReuse(analyzer, "Pokud mluvime o volnem", new String[] { "mluvime", "volnem" });
|
assertAnalyzesToReuse(analyzer, "Pokud mluvime o volnem", new String[] { "mluvime", "volnem" });
|
||||||
|
@ -69,6 +71,7 @@ public class TestCzechAnalyzer extends BaseTokenStreamTestCase {
|
||||||
* An input stream that always throws IOException for testing.
|
* An input stream that always throws IOException for testing.
|
||||||
* @deprecated Remove this class when the loadStopWords method is removed.
|
* @deprecated Remove this class when the loadStopWords method is removed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
private class UnreliableInputStream extends InputStream {
|
private class UnreliableInputStream extends InputStream {
|
||||||
@Override
|
@Override
|
||||||
public int read() throws IOException {
|
public int read() throws IOException {
|
||||||
|
@ -82,6 +85,7 @@ public class TestCzechAnalyzer extends BaseTokenStreamTestCase {
|
||||||
* this would cause a NPE when it is time to create the StopFilter.
|
* this would cause a NPE when it is time to create the StopFilter.
|
||||||
* @deprecated Remove this test when the loadStopWords method is removed.
|
* @deprecated Remove this test when the loadStopWords method is removed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void testInvalidStopWordFile() throws Exception {
|
public void testInvalidStopWordFile() throws Exception {
|
||||||
CzechAnalyzer cz = new CzechAnalyzer(Version.LUCENE_30);
|
CzechAnalyzer cz = new CzechAnalyzer(Version.LUCENE_30);
|
||||||
cz.loadStopWords(new UnreliableInputStream(), "UTF-8");
|
cz.loadStopWords(new UnreliableInputStream(), "UTF-8");
|
||||||
|
@ -94,6 +98,7 @@ public class TestCzechAnalyzer extends BaseTokenStreamTestCase {
|
||||||
* when using reusable token streams.
|
* when using reusable token streams.
|
||||||
* @deprecated Remove this test when the loadStopWords method is removed.
|
* @deprecated Remove this test when the loadStopWords method is removed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void testStopWordFileReuse() throws Exception {
|
public void testStopWordFileReuse() throws Exception {
|
||||||
CzechAnalyzer cz = new CzechAnalyzer(Version.LUCENE_30);
|
CzechAnalyzer cz = new CzechAnalyzer(Version.LUCENE_30);
|
||||||
assertAnalyzesToReuse(cz, "Česká Republika",
|
assertAnalyzesToReuse(cz, "Česká Republika",
|
||||||
|
|
|
@ -41,6 +41,7 @@ public class TextFragment
|
||||||
* @deprecated Use {@link #TextFragment(CharSequence, int, int)} instead.
|
* @deprecated Use {@link #TextFragment(CharSequence, int, int)} instead.
|
||||||
* This constructor will be removed in Lucene 4.0
|
* This constructor will be removed in Lucene 4.0
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public TextFragment(StringBuffer markedUpText,int textStartPos, int fragNum)
|
public TextFragment(StringBuffer markedUpText,int textStartPos, int fragNum)
|
||||||
{
|
{
|
||||||
this.markedUpText=markedUpText;
|
this.markedUpText=markedUpText;
|
||||||
|
|
|
@ -156,6 +156,7 @@ public class ChainedFilter extends Filter
|
||||||
* switch to a different DocIdSet implementation yourself.
|
* switch to a different DocIdSet implementation yourself.
|
||||||
* This method will be removed in Lucene 4.0
|
* This method will be removed in Lucene 4.0
|
||||||
**/
|
**/
|
||||||
|
@Deprecated
|
||||||
protected final DocIdSet finalResult(OpenBitSetDISI result, int maxDocs) {
|
protected final DocIdSet finalResult(OpenBitSetDISI result, int maxDocs) {
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
|
@ -117,6 +117,7 @@ public class BooleanFilter extends Filter
|
||||||
* switch to a different DocIdSet implementation yourself.
|
* switch to a different DocIdSet implementation yourself.
|
||||||
* This method will be removed in Lucene 4.0
|
* This method will be removed in Lucene 4.0
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
protected final DocIdSet finalResult(OpenBitSetDISI result, int maxDocs) {
|
protected final DocIdSet finalResult(OpenBitSetDISI result, int maxDocs) {
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
|
@ -36,6 +36,7 @@ import org.apache.lucene.search.Query;
|
||||||
* used along the transition from the old query parser to the new
|
* used along the transition from the old query parser to the new
|
||||||
* one
|
* one
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public class MultiFieldQueryParserWrapper extends QueryParserWrapper {
|
public class MultiFieldQueryParserWrapper extends QueryParserWrapper {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -64,6 +64,7 @@ import org.apache.lucene.search.Query;
|
||||||
* used along the transition from the old query parser to the new
|
* used along the transition from the old query parser to the new
|
||||||
* one
|
* one
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public class QueryParserWrapper {
|
public class QueryParserWrapper {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -55,6 +55,7 @@ public final class SnowballAnalyzer extends Analyzer {
|
||||||
* Builds the named analyzer with the given stop words.
|
* Builds the named analyzer with the given stop words.
|
||||||
* @deprecated Use {@link #SnowballAnalyzer(Version, String, Set)} instead.
|
* @deprecated Use {@link #SnowballAnalyzer(Version, String, Set)} instead.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public SnowballAnalyzer(Version matchVersion, String name, String[] stopWords) {
|
public SnowballAnalyzer(Version matchVersion, String name, String[] stopWords) {
|
||||||
this(matchVersion, name);
|
this(matchVersion, name);
|
||||||
stopSet = StopFilter.makeStopSet(matchVersion, stopWords);
|
stopSet = StopFilter.makeStopSet(matchVersion, stopWords);
|
||||||
|
|
|
@ -27,6 +27,7 @@ package org.apache.lucene.spatial.geometry.shape;
|
||||||
* @deprecated This has been replaced with more accurate
|
* @deprecated This has been replaced with more accurate
|
||||||
* math in {@link LLRect}. This class will be removed in a future release.
|
* math in {@link LLRect}. This class will be removed in a future release.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public class DistanceApproximation
|
public class DistanceApproximation
|
||||||
{
|
{
|
||||||
private double m_testLat;
|
private double m_testLat;
|
||||||
|
|
|
@ -43,6 +43,7 @@ public class HTMLParser implements HTMLParserConstants {
|
||||||
/**
|
/**
|
||||||
* @deprecated Use HTMLParser(FileInputStream) instead
|
* @deprecated Use HTMLParser(FileInputStream) instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public HTMLParser(File file) throws FileNotFoundException {
|
public HTMLParser(File file) throws FileNotFoundException {
|
||||||
this(new FileInputStream(file));
|
this(new FileInputStream(file));
|
||||||
}
|
}
|
||||||
|
|
|
@ -70,6 +70,7 @@ public class HTMLParser {
|
||||||
/**
|
/**
|
||||||
* @deprecated Use HTMLParser(FileInputStream) instead
|
* @deprecated Use HTMLParser(FileInputStream) instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public HTMLParser(File file) throws FileNotFoundException {
|
public HTMLParser(File file) throws FileNotFoundException {
|
||||||
this(new FileInputStream(file));
|
this(new FileInputStream(file));
|
||||||
}
|
}
|
||||||
|
|
|
@ -85,11 +85,13 @@ public abstract class Analyzer implements Closeable {
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @deprecated */
|
/** @deprecated */
|
||||||
|
@Deprecated
|
||||||
protected boolean overridesTokenStreamMethod = false;
|
protected boolean overridesTokenStreamMethod = false;
|
||||||
|
|
||||||
/** @deprecated This is only present to preserve
|
/** @deprecated This is only present to preserve
|
||||||
* back-compat of classes that subclass a core analyzer
|
* back-compat of classes that subclass a core analyzer
|
||||||
* and override tokenStream but not reusableTokenStream */
|
* and override tokenStream but not reusableTokenStream */
|
||||||
|
@Deprecated
|
||||||
protected void setOverridesTokenStreamMethod(Class<? extends Analyzer> baseClass) {
|
protected void setOverridesTokenStreamMethod(Class<? extends Analyzer> baseClass) {
|
||||||
try {
|
try {
|
||||||
Method m = this.getClass().getMethod("tokenStream", String.class, Reader.class);
|
Method m = this.getClass().getMethod("tokenStream", String.class, Reader.class);
|
||||||
|
|
|
@ -117,6 +117,7 @@ public class CharArraySet extends AbstractSet<Object> {
|
||||||
* otherwise <code>true</code>.
|
* otherwise <code>true</code>.
|
||||||
* @deprecated use {@link #CharArraySet(Version, int, boolean)} instead
|
* @deprecated use {@link #CharArraySet(Version, int, boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public CharArraySet(int startSize, boolean ignoreCase) {
|
public CharArraySet(int startSize, boolean ignoreCase) {
|
||||||
this(Version.LUCENE_30, startSize, ignoreCase);
|
this(Version.LUCENE_30, startSize, ignoreCase);
|
||||||
}
|
}
|
||||||
|
@ -131,6 +132,7 @@ public class CharArraySet extends AbstractSet<Object> {
|
||||||
* otherwise <code>true</code>.
|
* otherwise <code>true</code>.
|
||||||
* @deprecated use {@link #CharArraySet(Version, Collection, boolean)} instead
|
* @deprecated use {@link #CharArraySet(Version, Collection, boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public CharArraySet(Collection<? extends Object> c, boolean ignoreCase) {
|
public CharArraySet(Collection<? extends Object> c, boolean ignoreCase) {
|
||||||
this(Version.LUCENE_30, c.size(), ignoreCase);
|
this(Version.LUCENE_30, c.size(), ignoreCase);
|
||||||
addAll(c);
|
addAll(c);
|
||||||
|
@ -381,6 +383,7 @@ public class CharArraySet extends AbstractSet<Object> {
|
||||||
* preserved.
|
* preserved.
|
||||||
* @deprecated use {@link #copy(Version, Set)} instead.
|
* @deprecated use {@link #copy(Version, Set)} instead.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static CharArraySet copy(final Set<?> set) {
|
public static CharArraySet copy(final Set<?> set) {
|
||||||
if(set == EMPTY_SET)
|
if(set == EMPTY_SET)
|
||||||
return EMPTY_SET;
|
return EMPTY_SET;
|
||||||
|
|
|
@ -31,6 +31,7 @@ import org.apache.lucene.analysis.tokenattributes.TermAttribute;
|
||||||
* This class is included for use with existing
|
* This class is included for use with existing
|
||||||
* indexes and will be removed in a future release (possibly Lucene 4.0).
|
* indexes and will be removed in a future release (possibly Lucene 4.0).
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public final class ISOLatin1AccentFilter extends TokenFilter {
|
public final class ISOLatin1AccentFilter extends TokenFilter {
|
||||||
public ISOLatin1AccentFilter(TokenStream input) {
|
public ISOLatin1AccentFilter(TokenStream input) {
|
||||||
super(input);
|
super(input);
|
||||||
|
|
|
@ -50,6 +50,7 @@ public final class LowerCaseFilter extends TokenFilter {
|
||||||
/**
|
/**
|
||||||
* @deprecated Use {@link #LowerCaseFilter(Version, TokenStream)} instead.
|
* @deprecated Use {@link #LowerCaseFilter(Version, TokenStream)} instead.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public LowerCaseFilter(TokenStream in) {
|
public LowerCaseFilter(TokenStream in) {
|
||||||
this(Version.LUCENE_30, in);
|
this(Version.LUCENE_30, in);
|
||||||
}
|
}
|
||||||
|
|
|
@ -64,6 +64,7 @@ public final class StopFilter extends TokenFilter {
|
||||||
* @param ignoreCase if true, all words are lower cased first
|
* @param ignoreCase if true, all words are lower cased first
|
||||||
* @deprecated use {@link #StopFilter(Version, TokenStream, Set, boolean)} instead
|
* @deprecated use {@link #StopFilter(Version, TokenStream, Set, boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public StopFilter(boolean enablePositionIncrements, TokenStream input, Set<?> stopWords, boolean ignoreCase)
|
public StopFilter(boolean enablePositionIncrements, TokenStream input, Set<?> stopWords, boolean ignoreCase)
|
||||||
{
|
{
|
||||||
this(Version.LUCENE_30, enablePositionIncrements, input, stopWords, ignoreCase);
|
this(Version.LUCENE_30, enablePositionIncrements, input, stopWords, ignoreCase);
|
||||||
|
@ -117,6 +118,7 @@ public final class StopFilter extends TokenFilter {
|
||||||
* @see #makeStopSet(Version, java.lang.String[])
|
* @see #makeStopSet(Version, java.lang.String[])
|
||||||
* @deprecated use {@link #StopFilter(Version, TokenStream, Set)} instead
|
* @deprecated use {@link #StopFilter(Version, TokenStream, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public StopFilter(boolean enablePositionIncrements, TokenStream in, Set<?> stopWords) {
|
public StopFilter(boolean enablePositionIncrements, TokenStream in, Set<?> stopWords) {
|
||||||
this(Version.LUCENE_CURRENT, enablePositionIncrements, in, stopWords, false);
|
this(Version.LUCENE_CURRENT, enablePositionIncrements, in, stopWords, false);
|
||||||
}
|
}
|
||||||
|
@ -148,6 +150,7 @@ public final class StopFilter extends TokenFilter {
|
||||||
* @see #makeStopSet(Version, java.lang.String[], boolean) passing false to ignoreCase
|
* @see #makeStopSet(Version, java.lang.String[], boolean) passing false to ignoreCase
|
||||||
* @deprecated use {@link #makeStopSet(Version, String...)} instead
|
* @deprecated use {@link #makeStopSet(Version, String...)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static final Set<Object> makeStopSet(String... stopWords) {
|
public static final Set<Object> makeStopSet(String... stopWords) {
|
||||||
return makeStopSet(Version.LUCENE_30, stopWords, false);
|
return makeStopSet(Version.LUCENE_30, stopWords, false);
|
||||||
}
|
}
|
||||||
|
@ -176,6 +179,7 @@ public final class StopFilter extends TokenFilter {
|
||||||
* @see #makeStopSet(Version, java.lang.String[], boolean) passing false to ignoreCase
|
* @see #makeStopSet(Version, java.lang.String[], boolean) passing false to ignoreCase
|
||||||
* @deprecated use {@link #makeStopSet(Version, List)} instead
|
* @deprecated use {@link #makeStopSet(Version, List)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static final Set<Object> makeStopSet(List<?> stopWords) {
|
public static final Set<Object> makeStopSet(List<?> stopWords) {
|
||||||
return makeStopSet(Version.LUCENE_30, stopWords, false);
|
return makeStopSet(Version.LUCENE_30, stopWords, false);
|
||||||
}
|
}
|
||||||
|
@ -202,6 +206,7 @@ public final class StopFilter extends TokenFilter {
|
||||||
* @return a Set containing the words
|
* @return a Set containing the words
|
||||||
* @deprecated use {@link #makeStopSet(Version, String[], boolean)} instead;
|
* @deprecated use {@link #makeStopSet(Version, String[], boolean)} instead;
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static final Set<Object> makeStopSet(String[] stopWords, boolean ignoreCase) {
|
public static final Set<Object> makeStopSet(String[] stopWords, boolean ignoreCase) {
|
||||||
return makeStopSet(Version.LUCENE_30, stopWords, ignoreCase);
|
return makeStopSet(Version.LUCENE_30, stopWords, ignoreCase);
|
||||||
}
|
}
|
||||||
|
@ -226,6 +231,7 @@ public final class StopFilter extends TokenFilter {
|
||||||
* @return A Set ({@link CharArraySet}) containing the words
|
* @return A Set ({@link CharArraySet}) containing the words
|
||||||
* @deprecated use {@link #makeStopSet(Version, List, boolean)} instead
|
* @deprecated use {@link #makeStopSet(Version, List, boolean)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static final Set<Object> makeStopSet(List<?> stopWords, boolean ignoreCase){
|
public static final Set<Object> makeStopSet(List<?> stopWords, boolean ignoreCase){
|
||||||
return makeStopSet(Version.LUCENE_30, stopWords, ignoreCase);
|
return makeStopSet(Version.LUCENE_30, stopWords, ignoreCase);
|
||||||
}
|
}
|
||||||
|
@ -271,6 +277,7 @@ public final class StopFilter extends TokenFilter {
|
||||||
* or later, it returns true.
|
* or later, it returns true.
|
||||||
* @deprecated use {@link #StopFilter(Version, TokenStream, Set)} instead
|
* @deprecated use {@link #StopFilter(Version, TokenStream, Set)} instead
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static boolean getEnablePositionIncrementsVersionDefault(Version matchVersion) {
|
public static boolean getEnablePositionIncrementsVersionDefault(Version matchVersion) {
|
||||||
return matchVersion.onOrAfter(Version.LUCENE_29);
|
return matchVersion.onOrAfter(Version.LUCENE_29);
|
||||||
}
|
}
|
||||||
|
|
|
@ -71,6 +71,7 @@ public final class StandardTokenizer extends Tokenizer {
|
||||||
* @deprecated this solves a bug where HOSTs that end with '.' are identified
|
* @deprecated this solves a bug where HOSTs that end with '.' are identified
|
||||||
* as ACRONYMs.
|
* as ACRONYMs.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static final int ACRONYM_DEP = 8;
|
public static final int ACRONYM_DEP = 8;
|
||||||
|
|
||||||
/** String token types that correspond to token type int constants */
|
/** String token types that correspond to token type int constants */
|
||||||
|
@ -227,6 +228,7 @@ public final class StandardTokenizer extends Tokenizer {
|
||||||
*
|
*
|
||||||
* @deprecated Remove in 3.X and make true the only valid value
|
* @deprecated Remove in 3.X and make true the only valid value
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public boolean isReplaceInvalidAcronym() {
|
public boolean isReplaceInvalidAcronym() {
|
||||||
return replaceInvalidAcronym;
|
return replaceInvalidAcronym;
|
||||||
}
|
}
|
||||||
|
@ -238,6 +240,7 @@ public final class StandardTokenizer extends Tokenizer {
|
||||||
*
|
*
|
||||||
* See https://issues.apache.org/jira/browse/LUCENE-1068
|
* See https://issues.apache.org/jira/browse/LUCENE-1068
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void setReplaceInvalidAcronym(boolean replaceInvalidAcronym) {
|
public void setReplaceInvalidAcronym(boolean replaceInvalidAcronym) {
|
||||||
this.replaceInvalidAcronym = replaceInvalidAcronym;
|
this.replaceInvalidAcronym = replaceInvalidAcronym;
|
||||||
}
|
}
|
||||||
|
|
|
@ -354,6 +354,7 @@ public static final int CJ = StandardTokenizer.CJ;
|
||||||
* @deprecated this solves a bug where HOSTs that end with '.' are identified
|
* @deprecated this solves a bug where HOSTs that end with '.' are identified
|
||||||
* as ACRONYMs.
|
* as ACRONYMs.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static final int ACRONYM_DEP = StandardTokenizer.ACRONYM_DEP;
|
public static final int ACRONYM_DEP = StandardTokenizer.ACRONYM_DEP;
|
||||||
|
|
||||||
public static final String [] TOKEN_TYPES = StandardTokenizer.TOKEN_TYPES;
|
public static final String [] TOKEN_TYPES = StandardTokenizer.TOKEN_TYPES;
|
||||||
|
|
|
@ -55,6 +55,7 @@ public static final int CJ = StandardTokenizer.CJ;
|
||||||
* @deprecated this solves a bug where HOSTs that end with '.' are identified
|
* @deprecated this solves a bug where HOSTs that end with '.' are identified
|
||||||
* as ACRONYMs.
|
* as ACRONYMs.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static final int ACRONYM_DEP = StandardTokenizer.ACRONYM_DEP;
|
public static final int ACRONYM_DEP = StandardTokenizer.ACRONYM_DEP;
|
||||||
|
|
||||||
public static final String [] TOKEN_TYPES = StandardTokenizer.TOKEN_TYPES;
|
public static final String [] TOKEN_TYPES = StandardTokenizer.TOKEN_TYPES;
|
||||||
|
|
|
@ -57,6 +57,7 @@ import java.util.Calendar; // for javadoc
|
||||||
* This class is included for use with existing
|
* This class is included for use with existing
|
||||||
* indices and will be removed in a future release (possibly Lucene 4.0).
|
* indices and will be removed in a future release (possibly Lucene 4.0).
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public class DateField {
|
public class DateField {
|
||||||
|
|
||||||
private DateField() {}
|
private DateField() {}
|
||||||
|
|
|
@ -44,6 +44,7 @@ import org.apache.lucene.util.NumericUtils; // for javadocs
|
||||||
* This class is included for use with existing
|
* This class is included for use with existing
|
||||||
* indices and will be removed in a future release (possibly Lucene 4.0).
|
* indices and will be removed in a future release (possibly Lucene 4.0).
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public class NumberTools {
|
public class NumberTools {
|
||||||
|
|
||||||
private static final int RADIX = 36;
|
private static final int RADIX = 36;
|
||||||
|
|
|
@ -412,6 +412,7 @@ final class FieldsReader implements Cloneable {
|
||||||
private int toRead;
|
private int toRead;
|
||||||
private long pointer;
|
private long pointer;
|
||||||
/** @deprecated Only kept for backward-compatbility with <3.0 indexes. Will be removed in 4.0. */
|
/** @deprecated Only kept for backward-compatbility with <3.0 indexes. Will be removed in 4.0. */
|
||||||
|
@Deprecated
|
||||||
private boolean isCompressed;
|
private boolean isCompressed;
|
||||||
|
|
||||||
public LazyField(String name, Field.Store store, int toRead, long pointer, boolean isBinary, boolean isCompressed) {
|
public LazyField(String name, Field.Store store, int toRead, long pointer, boolean isBinary, boolean isCompressed) {
|
||||||
|
|
|
@ -32,6 +32,7 @@ final class FieldsWriter
|
||||||
static final byte FIELD_IS_BINARY = 0x2;
|
static final byte FIELD_IS_BINARY = 0x2;
|
||||||
|
|
||||||
/** @deprecated Kept for backwards-compatibility with <3.0 indexes; will be removed in 4.0 */
|
/** @deprecated Kept for backwards-compatibility with <3.0 indexes; will be removed in 4.0 */
|
||||||
|
@Deprecated
|
||||||
static final byte FIELD_IS_COMPRESSED = 0x4;
|
static final byte FIELD_IS_COMPRESSED = 0x4;
|
||||||
|
|
||||||
// Original format
|
// Original format
|
||||||
|
|
|
@ -1276,6 +1276,7 @@ public class SegmentReader extends IndexReader implements Cloneable {
|
||||||
* We do it with R/W access for the tests (BW compatibility)
|
* We do it with R/W access for the tests (BW compatibility)
|
||||||
* @deprecated Remove this when tests are fixed!
|
* @deprecated Remove this when tests are fixed!
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
static SegmentReader getOnlySegmentReader(Directory dir) throws IOException {
|
static SegmentReader getOnlySegmentReader(Directory dir) throws IOException {
|
||||||
return getOnlySegmentReader(IndexReader.open(dir,false));
|
return getOnlySegmentReader(IndexReader.open(dir,false));
|
||||||
}
|
}
|
||||||
|
|
|
@ -373,6 +373,7 @@ public abstract class FSDirectory extends Directory {
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @deprecated Use {@link #getDirectory} instead. */
|
/** @deprecated Use {@link #getDirectory} instead. */
|
||||||
|
@Deprecated
|
||||||
public File getFile() {
|
public File getFile() {
|
||||||
return getDirectory();
|
return getDirectory();
|
||||||
}
|
}
|
||||||
|
|
|
@ -146,6 +146,7 @@ public abstract class IndexInput implements Cloneable,Closeable {
|
||||||
* instead, and construct the string
|
* instead, and construct the string
|
||||||
* from those utf8 bytes
|
* from those utf8 bytes
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void readChars(char[] buffer, int start, int length)
|
public void readChars(char[] buffer, int start, int length)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
final int end = start + length;
|
final int end = start + length;
|
||||||
|
@ -174,6 +175,7 @@ public abstract class IndexInput implements Cloneable,Closeable {
|
||||||
* @deprecated this method operates on old "modified utf8" encoded
|
* @deprecated this method operates on old "modified utf8" encoded
|
||||||
* strings
|
* strings
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void skipChars(int length) throws IOException{
|
public void skipChars(int length) throws IOException{
|
||||||
for (int i = 0; i < length; i++) {
|
for (int i = 0; i < length; i++) {
|
||||||
byte b = readByte();
|
byte b = readByte();
|
||||||
|
|
|
@ -114,6 +114,7 @@ public abstract class IndexOutput implements Closeable {
|
||||||
* @deprecated -- please pre-convert to utf8 bytes
|
* @deprecated -- please pre-convert to utf8 bytes
|
||||||
* instead or use {@link #writeString}
|
* instead or use {@link #writeString}
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void writeChars(String s, int start, int length)
|
public void writeChars(String s, int start, int length)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
final int end = start + length;
|
final int end = start + length;
|
||||||
|
@ -139,6 +140,7 @@ public abstract class IndexOutput implements Closeable {
|
||||||
* @param length the number of characters in the sequence
|
* @param length the number of characters in the sequence
|
||||||
* @deprecated -- please pre-convert to utf8 bytes instead or use {@link #writeString}
|
* @deprecated -- please pre-convert to utf8 bytes instead or use {@link #writeString}
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void writeChars(char[] s, int start, int length)
|
public void writeChars(char[] s, int start, int length)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
final int end = start + length;
|
final int end = start + length;
|
||||||
|
|
|
@ -39,6 +39,7 @@ public class NoLockFactory extends LockFactory {
|
||||||
* @see #getNoLockFactory()
|
* @see #getNoLockFactory()
|
||||||
*/
|
*/
|
||||||
// make private in 4.0!
|
// make private in 4.0!
|
||||||
|
@Deprecated
|
||||||
public NoLockFactory() {}
|
public NoLockFactory() {}
|
||||||
|
|
||||||
public static NoLockFactory getNoLockFactory() {
|
public static NoLockFactory getNoLockFactory() {
|
||||||
|
|
|
@ -27,6 +27,7 @@ import java.util.Map;
|
||||||
* A serializable Enum class.
|
* A serializable Enum class.
|
||||||
* @deprecated Use Java 5 enum, will be removed in a later Lucene 3.x release.
|
* @deprecated Use Java 5 enum, will be removed in a later Lucene 3.x release.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
@SuppressWarnings("serial")
|
@SuppressWarnings("serial")
|
||||||
public abstract class Parameter implements Serializable
|
public abstract class Parameter implements Serializable
|
||||||
{
|
{
|
||||||
|
|
|
@ -28,6 +28,7 @@ import java.util.Map;
|
||||||
* @deprecated Lucene's internal use of this class has now
|
* @deprecated Lucene's internal use of this class has now
|
||||||
* switched to {@link DoubleBarrelLRUCache}.
|
* switched to {@link DoubleBarrelLRUCache}.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public class SimpleLRUCache<K,V> extends SimpleMapCache<K,V> {
|
public class SimpleLRUCache<K,V> extends SimpleMapCache<K,V> {
|
||||||
private final static float LOADFACTOR = 0.75f;
|
private final static float LOADFACTOR = 0.75f;
|
||||||
|
|
||||||
|
|
|
@ -29,6 +29,7 @@ import java.util.Set;
|
||||||
* @deprecated Lucene's internal use of this class has now
|
* @deprecated Lucene's internal use of this class has now
|
||||||
* switched to {@link DoubleBarrelLRUCache}.
|
* switched to {@link DoubleBarrelLRUCache}.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public class SimpleMapCache<K,V> extends Cache<K,V> {
|
public class SimpleMapCache<K,V> extends Cache<K,V> {
|
||||||
protected Map<K,V> map;
|
protected Map<K,V> map;
|
||||||
|
|
||||||
|
|
|
@ -154,6 +154,7 @@ public class TestAnalyzers extends BaseTokenStreamTestCase {
|
||||||
* @deprecated remove this when lucene 3.0 "broken unicode 4" support
|
* @deprecated remove this when lucene 3.0 "broken unicode 4" support
|
||||||
* is no longer needed.
|
* is no longer needed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
private static class LowerCaseWhitespaceAnalyzerBWComp extends Analyzer {
|
private static class LowerCaseWhitespaceAnalyzerBWComp extends Analyzer {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -215,6 +216,7 @@ public class TestAnalyzers extends BaseTokenStreamTestCase {
|
||||||
* @deprecated remove this test when lucene 3.0 "broken unicode 4" support
|
* @deprecated remove this test when lucene 3.0 "broken unicode 4" support
|
||||||
* is no longer needed.
|
* is no longer needed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void testLowerCaseFilterBWComp() throws IOException {
|
public void testLowerCaseFilterBWComp() throws IOException {
|
||||||
Analyzer a = new LowerCaseWhitespaceAnalyzerBWComp();
|
Analyzer a = new LowerCaseWhitespaceAnalyzerBWComp();
|
||||||
// BMP
|
// BMP
|
||||||
|
|
|
@ -254,6 +254,7 @@ public class TestCharArraySet extends LuceneTestCase {
|
||||||
* @deprecated remove this test when lucene 3.0 "broken unicode 4" support is
|
* @deprecated remove this test when lucene 3.0 "broken unicode 4" support is
|
||||||
* no longer needed.
|
* no longer needed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void testSupplementaryCharsBWCompat() {
|
public void testSupplementaryCharsBWCompat() {
|
||||||
String missing = "Term %s is missing in the set";
|
String missing = "Term %s is missing in the set";
|
||||||
String falsePos = "Term %s is in the set but shouldn't";
|
String falsePos = "Term %s is in the set but shouldn't";
|
||||||
|
@ -285,6 +286,7 @@ public class TestCharArraySet extends LuceneTestCase {
|
||||||
* @deprecated remove this test when lucene 3.0 "broken unicode 4" support is
|
* @deprecated remove this test when lucene 3.0 "broken unicode 4" support is
|
||||||
* no longer needed.
|
* no longer needed.
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public void testSingleHighSurrogateBWComapt() {
|
public void testSingleHighSurrogateBWComapt() {
|
||||||
String missing = "Term %s is missing in the set";
|
String missing = "Term %s is missing in the set";
|
||||||
String falsePos = "Term %s is in the set but shouldn't";
|
String falsePos = "Term %s is in the set but shouldn't";
|
||||||
|
|
|
@ -456,6 +456,7 @@ public class CheckHits {
|
||||||
* @deprecated
|
* @deprecated
|
||||||
* @see CheckHits#EXPLAIN_SCORE_TOLERANCE_DELTA
|
* @see CheckHits#EXPLAIN_SCORE_TOLERANCE_DELTA
|
||||||
*/
|
*/
|
||||||
|
@Deprecated
|
||||||
public static float SCORE_TOLERANCE_DELTA = 0.00005f;
|
public static float SCORE_TOLERANCE_DELTA = 0.00005f;
|
||||||
|
|
||||||
Query q;
|
Query q;
|
||||||
|
|
|
@ -84,6 +84,7 @@ public class _TestUtil {
|
||||||
/** Use only for testing.
|
/** Use only for testing.
|
||||||
* @deprecated -- in 3.0 we can use Arrays.toString
|
* @deprecated -- in 3.0 we can use Arrays.toString
|
||||||
* instead */
|
* instead */
|
||||||
|
@Deprecated
|
||||||
public static String arrayToString(int[] array) {
|
public static String arrayToString(int[] array) {
|
||||||
StringBuilder buf = new StringBuilder();
|
StringBuilder buf = new StringBuilder();
|
||||||
buf.append("[");
|
buf.append("[");
|
||||||
|
@ -100,6 +101,7 @@ public class _TestUtil {
|
||||||
/** Use only for testing.
|
/** Use only for testing.
|
||||||
* @deprecated -- in 3.0 we can use Arrays.toString
|
* @deprecated -- in 3.0 we can use Arrays.toString
|
||||||
* instead */
|
* instead */
|
||||||
|
@Deprecated
|
||||||
public static String arrayToString(Object[] array) {
|
public static String arrayToString(Object[] array) {
|
||||||
StringBuilder buf = new StringBuilder();
|
StringBuilder buf = new StringBuilder();
|
||||||
buf.append("[");
|
buf.append("[");
|
||||||
|
|
|
@ -18,6 +18,7 @@ package org.apache.lucene.util.cache;
|
||||||
*/
|
*/
|
||||||
|
|
||||||
/** @deprecated */
|
/** @deprecated */
|
||||||
|
@Deprecated
|
||||||
public class TestSimpleLRUCache extends BaseTestLRU {
|
public class TestSimpleLRUCache extends BaseTestLRU {
|
||||||
public void testLRUCache() throws Exception {
|
public void testLRUCache() throws Exception {
|
||||||
final int n = 100;
|
final int n = 100;
|
||||||
|
|
Loading…
Reference in New Issue