From 7d5844740e70b9bc523cfe1ce7a3abd891197092 Mon Sep 17 00:00:00 2001 From: Robert Muir Date: Sun, 10 Jan 2010 21:09:58 +0000 Subject: [PATCH] LUCENE-2200: final classes had non-overriding protected members git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@897707 13f79535-47bb-0310-9956-ffa450edef68 --- CHANGES.txt | 4 ++++ .../analysis/ar/ArabicNormalizationFilter.java | 4 ++-- .../analysis/ngram/EdgeNGramTokenFilter.java | 15 ++++----------- .../payloads/DelimitedPayloadTokenFilter.java | 15 ++++----------- .../analysis/shingle/ShingleAnalyzerWrapper.java | 6 +++--- .../payloads/DelimitedPayloadTokenFilterTest.java | 8 ++++++-- .../org/apache/lucene/analysis/CharReader.java | 2 +- .../org/apache/lucene/index/FieldsReader.java | 2 +- 8 files changed, 25 insertions(+), 31 deletions(-) diff --git a/CHANGES.txt b/CHANGES.txt index ed04ac34c4c..81b071b9807 100644 --- a/CHANGES.txt +++ b/CHANGES.txt @@ -58,6 +58,10 @@ API Changes which reader the int doc, passed to customScore, refers to. (Paul chez Jamespot via Mike McCandless) +* LUCENE-2200: Several final classes had non-overriding protected + members. These were converted to private and unused protected + constructors removed. (Steven Rowe via Robert Muir) + Bug fixes * LUCENE-2092: BooleanQuery was ignoring disableCoord in its hashCode diff --git a/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ArabicNormalizationFilter.java b/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ArabicNormalizationFilter.java index 937911bae40..8a74505bd7d 100644 --- a/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ArabicNormalizationFilter.java +++ b/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ArabicNormalizationFilter.java @@ -30,8 +30,8 @@ import org.apache.lucene.analysis.tokenattributes.TermAttribute; public final class ArabicNormalizationFilter extends TokenFilter { - protected ArabicNormalizer normalizer = null; - private TermAttribute termAtt; + private final ArabicNormalizer normalizer; + private final TermAttribute termAtt; public ArabicNormalizationFilter(TokenStream input) { super(input); diff --git a/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilter.java b/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilter.java index eceef3f7d7b..faa81f593c4 100644 --- a/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilter.java +++ b/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilter.java @@ -64,22 +64,15 @@ public final class EdgeNGramTokenFilter extends TokenFilter { } } - private int minGram; - private int maxGram; + private final int minGram; + private final int maxGram; private Side side; private char[] curTermBuffer; private int curTermLength; private int curGramSize; - private TermAttribute termAtt; - private OffsetAttribute offsetAtt; - - - protected EdgeNGramTokenFilter(TokenStream input) { - super(input); - this.termAtt = addAttribute(TermAttribute.class); - this.offsetAtt = addAttribute(OffsetAttribute.class); - } + private final TermAttribute termAtt; + private final OffsetAttribute offsetAtt; /** * Creates EdgeNGramTokenFilter that can generate n-grams in the sizes of the given range diff --git a/contrib/analyzers/common/src/java/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilter.java b/contrib/analyzers/common/src/java/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilter.java index 89f939faa39..b032f76a8d8 100644 --- a/contrib/analyzers/common/src/java/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilter.java +++ b/contrib/analyzers/common/src/java/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilter.java @@ -38,17 +38,10 @@ import org.apache.lucene.analysis.tokenattributes.TermAttribute; */ public final class DelimitedPayloadTokenFilter extends TokenFilter { public static final char DEFAULT_DELIMITER = '|'; - protected char delimiter = DEFAULT_DELIMITER; - protected TermAttribute termAtt; - protected PayloadAttribute payAtt; - protected PayloadEncoder encoder; - - /** - * Construct a token stream filtering the given input. - */ - protected DelimitedPayloadTokenFilter(TokenStream input) { - this(input, DEFAULT_DELIMITER, new IdentityEncoder()); - } + private final char delimiter; + private final TermAttribute termAtt; + private final PayloadAttribute payAtt; + private final PayloadEncoder encoder; public DelimitedPayloadTokenFilter(TokenStream input, char delimiter, PayloadEncoder encoder) { diff --git a/contrib/analyzers/common/src/java/org/apache/lucene/analysis/shingle/ShingleAnalyzerWrapper.java b/contrib/analyzers/common/src/java/org/apache/lucene/analysis/shingle/ShingleAnalyzerWrapper.java index 9da0efccb88..2aa6cd4ea2a 100644 --- a/contrib/analyzers/common/src/java/org/apache/lucene/analysis/shingle/ShingleAnalyzerWrapper.java +++ b/contrib/analyzers/common/src/java/org/apache/lucene/analysis/shingle/ShingleAnalyzerWrapper.java @@ -33,9 +33,9 @@ import org.apache.lucene.util.Version; */ public final class ShingleAnalyzerWrapper extends Analyzer { - protected Analyzer defaultAnalyzer; - protected int maxShingleSize = 2; - protected boolean outputUnigrams = true; + private final Analyzer defaultAnalyzer; + private int maxShingleSize = 2; + private boolean outputUnigrams = true; public ShingleAnalyzerWrapper(Analyzer defaultAnalyzer) { super(); diff --git a/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilterTest.java b/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilterTest.java index e96316b8ab0..9754ca45efe 100644 --- a/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilterTest.java +++ b/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilterTest.java @@ -34,7 +34,9 @@ public class DelimitedPayloadTokenFilterTest extends LuceneTestCase { public void testPayloads() throws Exception { String test = "The quick|JJ red|JJ fox|NN jumped|VB over the lazy|JJ brown|JJ dogs|NN"; - DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter(new WhitespaceTokenizer(new StringReader(test))); + DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter + (new WhitespaceTokenizer(new StringReader(test)), + DelimitedPayloadTokenFilter.DEFAULT_DELIMITER, new IdentityEncoder()); TermAttribute termAtt = filter.getAttribute(TermAttribute.class); PayloadAttribute payAtt = filter.getAttribute(PayloadAttribute.class); assertTermEquals("The", filter, termAtt, payAtt, null); @@ -53,7 +55,9 @@ public class DelimitedPayloadTokenFilterTest extends LuceneTestCase { public void testNext() throws Exception { String test = "The quick|JJ red|JJ fox|NN jumped|VB over the lazy|JJ brown|JJ dogs|NN"; - DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter(new WhitespaceTokenizer(new StringReader(test))); + DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter + (new WhitespaceTokenizer(new StringReader(test)), + DelimitedPayloadTokenFilter.DEFAULT_DELIMITER, new IdentityEncoder()); assertTermEquals("The", filter, null); assertTermEquals("quick", filter, "JJ".getBytes("UTF-8")); assertTermEquals("red", filter, "JJ".getBytes("UTF-8")); diff --git a/src/java/org/apache/lucene/analysis/CharReader.java b/src/java/org/apache/lucene/analysis/CharReader.java index 2d070482ba2..a83637f455c 100644 --- a/src/java/org/apache/lucene/analysis/CharReader.java +++ b/src/java/org/apache/lucene/analysis/CharReader.java @@ -28,7 +28,7 @@ import java.io.Reader; */ public final class CharReader extends CharStream { - protected Reader input; + private final Reader input; public static CharStream get(Reader input) { return input instanceof CharStream ? diff --git a/src/java/org/apache/lucene/index/FieldsReader.java b/src/java/org/apache/lucene/index/FieldsReader.java index 7162b6c984c..ce05c9b3225 100644 --- a/src/java/org/apache/lucene/index/FieldsReader.java +++ b/src/java/org/apache/lucene/index/FieldsReader.java @@ -164,7 +164,7 @@ final class FieldsReader implements Cloneable { /** * @throws AlreadyClosedException if this FieldsReader is closed */ - protected final void ensureOpen() throws AlreadyClosedException { + private void ensureOpen() throws AlreadyClosedException { if (closed) { throw new AlreadyClosedException("this FieldsReader is closed"); }