From d1d9781458d5df38f914e650e10e5f5e5c775ece Mon Sep 17 00:00:00 2001 From: Michael McCandless Date: Fri, 4 Dec 2009 13:07:47 +0000 Subject: [PATCH] LUCENE-2065: use generics throughout unit tests git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@887181 13f79535-47bb-0310-9956-ffa450edef68 --- CHANGES.txt | 3 + .../index/PositionBasedTermVectorMapper.java | 2 +- .../lucene/search/CachingWrapperFilter.java | 2 +- .../org/apache/lucene/util/AttributeImpl.java | 2 +- .../lucene/util/FieldCacheSanityChecker.java | 4 +- src/test/org/apache/lucene/TestDemo.java | 1 - .../lucene/TestSearchForDuplicates.java | 1 - .../lucene/TestSnapshotDeletionPolicy.java | 11 ++-- .../analysis/BaseTokenStreamTestCase.java | 1 - .../analysis/TestASCIIFoldingFilter.java | 6 +- .../analysis/TestMappingCharFilter.java | 1 - .../lucene/analysis/TestStandardAnalyzer.java | 6 +- .../lucene/analysis/TestStopAnalyzer.java | 8 +-- .../lucene/analysis/TestStopFilter.java | 24 ++++---- .../analysis/TestTeeSinkTokenFilter.java | 3 +- .../TestSimpleAttributeImpls.java | 2 +- .../apache/lucene/document/TestDateTools.java | 1 - .../lucene/document/TestNumberTools.java | 2 +- .../org/apache/lucene/index/DocHelper.java | 27 ++++----- .../apache/lucene/index/TestAtomicUpdate.java | 2 - .../index/TestBackwardsCompatibility.java | 12 ++-- .../apache/lucene/index/TestByteSlices.java | 4 +- .../apache/lucene/index/TestCheckIndex.java | 4 +- .../lucene/index/TestDeletionPolicy.java | 51 +++++++--------- src/test/org/apache/lucene/index/TestDoc.java | 12 ++-- .../apache/lucene/index/TestFieldsReader.java | 19 +++--- .../lucene/index/TestIndexFileDeleter.java | 20 +++---- .../apache/lucene/index/TestIndexReader.java | 60 +++++++++---------- .../index/TestIndexReaderCloneNorms.java | 20 +++---- .../lucene/index/TestIndexReaderReopen.java | 36 ++++++----- .../apache/lucene/index/TestIndexWriter.java | 17 +++--- .../index/TestIndexWriterExceptions.java | 2 +- .../lucene/index/TestIndexWriterReader.java | 14 ++--- .../org/apache/lucene/index/TestLazyBug.java | 8 +-- .../org/apache/lucene/index/TestNorms.java | 20 +++---- .../lucene/index/TestParallelReader.java | 2 +- .../org/apache/lucene/index/TestPayloads.java | 10 ++-- .../TestPositionBasedTermVectorMapper.java | 13 ++-- .../lucene/index/TestSegmentMerger.java | 2 +- .../lucene/index/TestSegmentReader.java | 15 +++-- .../lucene/index/TestStressIndexing.java | 3 +- .../lucene/index/TestStressIndexing2.java | 42 ++++++------- .../lucene/index/TestTermVectorsReader.java | 29 ++++----- .../lucene/index/TestTransactionRollback.java | 35 ++++++----- .../apache/lucene/index/TestTransactions.java | 2 +- .../lucene/index/TestWordlistLoader.java | 8 +-- .../lucene/queryParser/TestMultiAnalyzer.java | 1 - .../TestMultiFieldQueryParser.java | 3 +- .../lucene/queryParser/TestQueryParser.java | 4 +- .../search/CachingWrapperFilterHelper.java | 5 +- .../org/apache/lucene/search/CheckHits.java | 14 ++--- .../lucene/search/JustCompileSearch.java | 4 +- .../org/apache/lucene/search/QueryUtils.java | 4 +- .../TestComplexExplanationsOfNonMatches.java | 18 ------ .../lucene/search/TestCustomSearcherSort.java | 6 +- .../apache/lucene/search/TestDocIdSet.java | 6 +- .../search/TestElevationComparator.java | 8 +-- .../search/TestFieldCacheRangeFilter.java | 13 ++-- .../search/TestFieldCacheTermsFilter.java | 12 ++-- .../lucene/search/TestMultiPhraseQuery.java | 10 ++-- .../lucene/search/TestMultiSearcher.java | 6 +- .../lucene/search/TestPhrasePrefixQuery.java | 6 +- .../TestPositiveScoresOnlyCollector.java | 2 +- .../TestSimpleExplanationsOfNonMatches.java | 17 ------ .../org/apache/lucene/search/TestSort.java | 20 +++---- .../lucene/search/TestSpanQueryFilter.java | 6 +- .../lucene/search/TestTermRangeQuery.java | 1 - .../apache/lucene/search/TestTermScorer.java | 6 +- .../apache/lucene/search/TestTermVectors.java | 26 ++++---- .../apache/lucene/search/TestThreadSafe.java | 5 +- .../lucene/search/TestTopDocsCollector.java | 24 ++++---- .../search/TestTopScoreDocCollector.java | 2 +- .../search/function/TestCustomScoreQuery.java | 30 +++++----- .../lucene/search/function/TestDocValues.java | 1 - .../search/function/TestFieldScoreQuery.java | 2 +- .../search/payloads/TestPayloadNearQuery.java | 3 - .../search/spans/JustCompileSearchSpans.java | 4 +- .../spans/TestFieldMaskingSpanQuery.java | 2 +- .../lucene/search/spans/TestPayloadSpans.java | 53 +++++++--------- .../search/spans/TestSpanExplanations.java | 17 ------ .../apache/lucene/store/MockRAMDirectory.java | 14 ++--- .../lucene/store/MockRAMInputStream.java | 2 +- .../lucene/store/TestBufferedIndexInput.java | 8 +-- .../lucene/store/TestFileSwitchDirectory.java | 2 +- .../apache/lucene/store/TestHugeRamFile.java | 4 +- .../apache/lucene/store/TestLockFactory.java | 6 +- .../apache/lucene/util/LocalizedTestCase.java | 6 +- .../apache/lucene/util/LuceneTestCase.java | 4 +- .../lucene/util/TestCloseableThreadLocal.java | 7 +-- .../apache/lucene/util/TestNumericUtils.java | 19 +++--- .../apache/lucene/util/cache/BaseTestLRU.java | 2 +- .../util/cache/TestDoubleBarrelLRUCache.java | 2 +- 92 files changed, 433 insertions(+), 553 deletions(-) diff --git a/CHANGES.txt b/CHANGES.txt index 4ff1b7204e7..69fb293229a 100644 --- a/CHANGES.txt +++ b/CHANGES.txt @@ -66,6 +66,9 @@ Test Cases * LUCENE-1844: Speed up the unit tests (Mark Miller, Erick Erickson, Mike McCandless) +* LUCENE-2065: Use Java 5 generics throughout our unit tests. (Kay + Kay via Mike McCandless) + ======================= Release 3.0.0 2009-11-25 ======================= Changes in backwards compatibility policy diff --git a/src/java/org/apache/lucene/index/PositionBasedTermVectorMapper.java b/src/java/org/apache/lucene/index/PositionBasedTermVectorMapper.java index 73d4e0a7c31..59b570cd972 100644 --- a/src/java/org/apache/lucene/index/PositionBasedTermVectorMapper.java +++ b/src/java/org/apache/lucene/index/PositionBasedTermVectorMapper.java @@ -110,7 +110,7 @@ public class PositionBasedTermVectorMapper extends TermVectorMapper{ * * @return A map between field names and a Map. The sub-Map key is the position as the integer, the value is {@link org.apache.lucene.index.PositionBasedTermVectorMapper.TVPositionInfo}. */ - public Map> getFieldToTerms() { + public Map> getFieldToTerms() { return fieldToTerms; } diff --git a/src/java/org/apache/lucene/search/CachingWrapperFilter.java b/src/java/org/apache/lucene/search/CachingWrapperFilter.java index 9b2956162c4..0b4e191b950 100644 --- a/src/java/org/apache/lucene/search/CachingWrapperFilter.java +++ b/src/java/org/apache/lucene/search/CachingWrapperFilter.java @@ -35,7 +35,7 @@ public class CachingWrapperFilter extends Filter { /** * A transient Filter cache (package private because of test) */ - transient Map cache; + transient Map cache; private final ReentrantLock lock = new ReentrantLock(); diff --git a/src/java/org/apache/lucene/util/AttributeImpl.java b/src/java/org/apache/lucene/util/AttributeImpl.java index 774e7968753..c8bf649b6bf 100644 --- a/src/java/org/apache/lucene/util/AttributeImpl.java +++ b/src/java/org/apache/lucene/util/AttributeImpl.java @@ -51,7 +51,7 @@ public abstract class AttributeImpl implements Cloneable, Serializable, Attribut @Override public String toString() { StringBuilder buffer = new StringBuilder(); - Class clazz = this.getClass(); + Class clazz = this.getClass(); Field[] fields = clazz.getDeclaredFields(); try { for (int i = 0; i < fields.length; i++) { diff --git a/src/java/org/apache/lucene/util/FieldCacheSanityChecker.java b/src/java/org/apache/lucene/util/FieldCacheSanityChecker.java index be802a4c753..75c65c12069 100644 --- a/src/java/org/apache/lucene/util/FieldCacheSanityChecker.java +++ b/src/java/org/apache/lucene/util/FieldCacheSanityChecker.java @@ -212,7 +212,7 @@ public final class FieldCacheSanityChecker { if (seen.contains(rf)) continue; - List kids = getAllDecendentReaderKeys(rf.readerKey); + List kids = getAllDecendentReaderKeys(rf.readerKey); for (Object kidKey : kids) { ReaderField kid = new ReaderField(kidKey, rf.fieldName); @@ -270,7 +270,7 @@ public final class FieldCacheSanityChecker { * the hierarchy of subReaders building up a list of the objects * returned by obj.getFieldCacheKey() */ - private List getAllDecendentReaderKeys(Object seed) { + private List getAllDecendentReaderKeys(Object seed) { List all = new ArrayList(17); // will grow as we iter all.add(seed); for (int i = 0; i < all.size(); i++) { diff --git a/src/test/org/apache/lucene/TestDemo.java b/src/test/org/apache/lucene/TestDemo.java index c0a8d898489..f74f63c649e 100644 --- a/src/test/org/apache/lucene/TestDemo.java +++ b/src/test/org/apache/lucene/TestDemo.java @@ -33,7 +33,6 @@ import org.apache.lucene.store.Directory; import org.apache.lucene.store.RAMDirectory; import org.apache.lucene.util.LuceneTestCase; import org.apache.lucene.util.Version; -import org.apache.lucene.util._TestUtil; /** * A very simple demo used in the API documentation (src/java/overview.html). diff --git a/src/test/org/apache/lucene/TestSearchForDuplicates.java b/src/test/org/apache/lucene/TestSearchForDuplicates.java index 829a4db2d29..d3f5e29ec8e 100644 --- a/src/test/org/apache/lucene/TestSearchForDuplicates.java +++ b/src/test/org/apache/lucene/TestSearchForDuplicates.java @@ -30,7 +30,6 @@ import org.apache.lucene.queryParser.*; import org.apache.lucene.util.Version; import org.apache.lucene.util.LuceneTestCase; -import org.apache.lucene.util.Version; import junit.framework.TestSuite; import junit.textui.TestRunner; diff --git a/src/test/org/apache/lucene/TestSnapshotDeletionPolicy.java b/src/test/org/apache/lucene/TestSnapshotDeletionPolicy.java index fd8bacab70a..7714cde46f9 100644 --- a/src/test/org/apache/lucene/TestSnapshotDeletionPolicy.java +++ b/src/test/org/apache/lucene/TestSnapshotDeletionPolicy.java @@ -19,7 +19,6 @@ package org.apache.lucene; * limitations under the License. */ -import java.util.Iterator; import java.util.Collection; import java.io.File; import java.io.IOException; @@ -79,7 +78,7 @@ public class TestSnapshotDeletionPolicy extends LuceneTestCase writer.commit(); } } - IndexCommit cp = (IndexCommit) dp.snapshot(); + IndexCommit cp = dp.snapshot(); copyFiles(dir, cp); writer.close(); copyFiles(dir, cp); @@ -181,7 +180,7 @@ public class TestSnapshotDeletionPolicy extends LuceneTestCase public void backupIndex(Directory dir, SnapshotDeletionPolicy dp) throws Exception { // To backup an index we first take a snapshot: try { - copyFiles(dir, (IndexCommit) dp.snapshot()); + copyFiles(dir, dp.snapshot()); } finally { // Make sure to release the snapshot, otherwise these // files will never be deleted during this IndexWriter @@ -195,10 +194,8 @@ public class TestSnapshotDeletionPolicy extends LuceneTestCase // While we hold the snapshot, and nomatter how long // we take to do the backup, the IndexWriter will // never delete the files in the snapshot: - Collection files = cp.getFileNames(); - Iterator it = files.iterator(); - while(it.hasNext()) { - final String fileName = (String) it.next(); + Collection files = cp.getFileNames(); + for (final String fileName : files) { // NOTE: in a real backup you would not use // readFile; you would need to use something else // that copies the file to a backup location. This diff --git a/src/test/org/apache/lucene/analysis/BaseTokenStreamTestCase.java b/src/test/org/apache/lucene/analysis/BaseTokenStreamTestCase.java index caf8992152f..5d03b369f6e 100644 --- a/src/test/org/apache/lucene/analysis/BaseTokenStreamTestCase.java +++ b/src/test/org/apache/lucene/analysis/BaseTokenStreamTestCase.java @@ -17,7 +17,6 @@ package org.apache.lucene.analysis; * limitations under the License. */ -import java.util.Set; import java.io.StringReader; import java.io.IOException; diff --git a/src/test/org/apache/lucene/analysis/TestASCIIFoldingFilter.java b/src/test/org/apache/lucene/analysis/TestASCIIFoldingFilter.java index 4f83525e4a1..4c5b53c81a0 100644 --- a/src/test/org/apache/lucene/analysis/TestASCIIFoldingFilter.java +++ b/src/test/org/apache/lucene/analysis/TestASCIIFoldingFilter.java @@ -1871,7 +1871,7 @@ public class TestASCIIFoldingFilter extends BaseTokenStreamTestCase { }; // Construct input text and expected output tokens - List expectedOutputTokens = new ArrayList(); + List expectedOutputTokens = new ArrayList(); StringBuilder inputText = new StringBuilder(); for (int n = 0 ; n < foldings.length ; n += 2) { if (n > 0) { @@ -1892,9 +1892,9 @@ public class TestASCIIFoldingFilter extends BaseTokenStreamTestCase { TokenStream stream = new WhitespaceTokenizer(new StringReader(inputText.toString())); ASCIIFoldingFilter filter = new ASCIIFoldingFilter(stream); TermAttribute termAtt = filter.getAttribute(TermAttribute.class); - Iterator expectedIter = expectedOutputTokens.iterator(); + Iterator expectedIter = expectedOutputTokens.iterator(); while (expectedIter.hasNext()) {; - assertTermEquals((String)expectedIter.next(), filter, termAtt); + assertTermEquals(expectedIter.next(), filter, termAtt); } assertFalse(filter.incrementToken()); } diff --git a/src/test/org/apache/lucene/analysis/TestMappingCharFilter.java b/src/test/org/apache/lucene/analysis/TestMappingCharFilter.java index 1d02c05b488..44bd0a7558a 100644 --- a/src/test/org/apache/lucene/analysis/TestMappingCharFilter.java +++ b/src/test/org/apache/lucene/analysis/TestMappingCharFilter.java @@ -18,7 +18,6 @@ package org.apache.lucene.analysis; import java.io.StringReader; -import java.util.List; public class TestMappingCharFilter extends BaseTokenStreamTestCase { diff --git a/src/test/org/apache/lucene/analysis/TestStandardAnalyzer.java b/src/test/org/apache/lucene/analysis/TestStandardAnalyzer.java index bc49153ecdf..878a7a42650 100644 --- a/src/test/org/apache/lucene/analysis/TestStandardAnalyzer.java +++ b/src/test/org/apache/lucene/analysis/TestStandardAnalyzer.java @@ -1,13 +1,9 @@ package org.apache.lucene.analysis; import org.apache.lucene.analysis.standard.StandardAnalyzer; -import org.apache.lucene.analysis.tokenattributes.OffsetAttribute; -import org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute; -import org.apache.lucene.analysis.tokenattributes.TermAttribute; -import org.apache.lucene.analysis.tokenattributes.TypeAttribute; + import org.apache.lucene.util.Version; -import java.io.StringReader; /** * Copyright 2004 The Apache Software Foundation diff --git a/src/test/org/apache/lucene/analysis/TestStopAnalyzer.java b/src/test/org/apache/lucene/analysis/TestStopAnalyzer.java index fc59a7c7431..e7171c7dc90 100644 --- a/src/test/org/apache/lucene/analysis/TestStopAnalyzer.java +++ b/src/test/org/apache/lucene/analysis/TestStopAnalyzer.java @@ -30,7 +30,7 @@ import java.util.HashSet; public class TestStopAnalyzer extends BaseTokenStreamTestCase { private StopAnalyzer stop = new StopAnalyzer(Version.LUCENE_CURRENT); - private Set inValidTokens = new HashSet(); + private Set inValidTokens = new HashSet(); public TestStopAnalyzer(String s) { super(s); @@ -40,7 +40,7 @@ public class TestStopAnalyzer extends BaseTokenStreamTestCase { protected void setUp() throws Exception { super.setUp(); - Iterator it = StopAnalyzer.ENGLISH_STOP_WORDS_SET.iterator(); + Iterator it = StopAnalyzer.ENGLISH_STOP_WORDS_SET.iterator(); while(it.hasNext()) { inValidTokens.add(it.next()); } @@ -59,7 +59,7 @@ public class TestStopAnalyzer extends BaseTokenStreamTestCase { } public void testStopList() throws IOException { - Set stopWordsSet = new HashSet(); + Set stopWordsSet = new HashSet(); stopWordsSet.add("good"); stopWordsSet.add("test"); stopWordsSet.add("analyzer"); @@ -78,7 +78,7 @@ public class TestStopAnalyzer extends BaseTokenStreamTestCase { } public void testStopListPositions() throws IOException { - Set stopWordsSet = new HashSet(); + Set stopWordsSet = new HashSet(); stopWordsSet.add("good"); stopWordsSet.add("test"); stopWordsSet.add("analyzer"); diff --git a/src/test/org/apache/lucene/analysis/TestStopFilter.java b/src/test/org/apache/lucene/analysis/TestStopFilter.java index e6a84069306..81172998979 100644 --- a/src/test/org/apache/lucene/analysis/TestStopFilter.java +++ b/src/test/org/apache/lucene/analysis/TestStopFilter.java @@ -37,7 +37,7 @@ public class TestStopFilter extends BaseTokenStreamTestCase { public void testExactCase() throws IOException { StringReader reader = new StringReader("Now is The Time"); - Set stopWords = new HashSet(Arrays.asList("is", "the", "Time")); + Set stopWords = new HashSet(Arrays.asList("is", "the", "Time")); TokenStream stream = new StopFilter(Version.LUCENE_CURRENT, new WhitespaceTokenizer(reader), stopWords, false); final TermAttribute termAtt = stream.getAttribute(TermAttribute.class); assertTrue(stream.incrementToken()); @@ -49,7 +49,7 @@ public class TestStopFilter extends BaseTokenStreamTestCase { public void testIgnoreCase() throws IOException { StringReader reader = new StringReader("Now is The Time"); - Set stopWords = new HashSet(Arrays.asList( "is", "the", "Time" )); + Set stopWords = new HashSet(Arrays.asList( "is", "the", "Time" )); TokenStream stream = new StopFilter(Version.LUCENE_CURRENT, new WhitespaceTokenizer(reader), stopWords, true); final TermAttribute termAtt = stream.getAttribute(TermAttribute.class); assertTrue(stream.incrementToken()); @@ -60,7 +60,7 @@ public class TestStopFilter extends BaseTokenStreamTestCase { public void testStopFilt() throws IOException { StringReader reader = new StringReader("Now is The Time"); String[] stopWords = new String[] { "is", "the", "Time" }; - Set stopSet = StopFilter.makeStopSet(Version.LUCENE_CURRENT, stopWords); + Set stopSet = StopFilter.makeStopSet(Version.LUCENE_CURRENT, stopWords); TokenStream stream = new StopFilter(Version.LUCENE_CURRENT, new WhitespaceTokenizer(reader), stopSet); final TermAttribute termAtt = stream.getAttribute(TermAttribute.class); assertTrue(stream.incrementToken()); @@ -75,16 +75,16 @@ public class TestStopFilter extends BaseTokenStreamTestCase { */ public void testStopPositons() throws IOException { StringBuilder sb = new StringBuilder(); - ArrayList a = new ArrayList(); + ArrayList a = new ArrayList(); for (int i=0; i<20; i++) { String w = English.intToEnglish(i).trim(); sb.append(w).append(" "); if (i%3 != 0) a.add(w); } log(sb.toString()); - String stopWords[] = (String[]) a.toArray(new String[0]); + String stopWords[] = a.toArray(new String[0]); for (int i=0; i stopSet = StopFilter.makeStopSet(Version.LUCENE_CURRENT, stopWords); // with increments StringReader reader = new StringReader(sb.toString()); StopFilter stpf = new StopFilter(Version.LUCENE_24, new WhitespaceTokenizer(reader), stopSet); @@ -94,8 +94,8 @@ public class TestStopFilter extends BaseTokenStreamTestCase { stpf = new StopFilter(Version.LUCENE_CURRENT, new WhitespaceTokenizer(reader), stopSet); doTestStopPositons(stpf,false); // with increments, concatenating two stop filters - ArrayList a0 = new ArrayList(); - ArrayList a1 = new ArrayList(); + ArrayList a0 = new ArrayList(); + ArrayList a1 = new ArrayList(); for (int i=0; i stopSet0 = StopFilter.makeStopSet(Version.LUCENE_CURRENT, stopWords0); + Set stopSet1 = StopFilter.makeStopSet(Version.LUCENE_CURRENT, stopWords1); reader = new StringReader(sb.toString()); StopFilter stpf0 = new StopFilter(Version.LUCENE_CURRENT, new WhitespaceTokenizer(reader), stopSet0); // first part of the set stpf0.setEnablePositionIncrements(true); diff --git a/src/test/org/apache/lucene/analysis/TestTeeSinkTokenFilter.java b/src/test/org/apache/lucene/analysis/TestTeeSinkTokenFilter.java index 71293a2b3c1..e20199ca2e2 100644 --- a/src/test/org/apache/lucene/analysis/TestTeeSinkTokenFilter.java +++ b/src/test/org/apache/lucene/analysis/TestTeeSinkTokenFilter.java @@ -26,8 +26,7 @@ import org.apache.lucene.util.Version; import java.io.IOException; import java.io.StringReader; -import java.util.ArrayList; -import java.util.List; + /** * tests for the TestTeeSinkTokenFilter diff --git a/src/test/org/apache/lucene/analysis/tokenattributes/TestSimpleAttributeImpls.java b/src/test/org/apache/lucene/analysis/tokenattributes/TestSimpleAttributeImpls.java index 9a7d34ee3e8..5d173537d2b 100644 --- a/src/test/org/apache/lucene/analysis/tokenattributes/TestSimpleAttributeImpls.java +++ b/src/test/org/apache/lucene/analysis/tokenattributes/TestSimpleAttributeImpls.java @@ -126,7 +126,7 @@ public class TestSimpleAttributeImpls extends LuceneTestCase { } public static final AttributeImpl assertCopyIsEqual(AttributeImpl att) throws Exception { - AttributeImpl copy = (AttributeImpl) att.getClass().newInstance(); + AttributeImpl copy = att.getClass().newInstance(); att.copyTo(copy); assertEquals("Copied instance must be equal", att, copy); assertEquals("Copied instance's hashcode must be equal", att.hashCode(), copy.hashCode()); diff --git a/src/test/org/apache/lucene/document/TestDateTools.java b/src/test/org/apache/lucene/document/TestDateTools.java index b64ce1166bc..e846288f40b 100644 --- a/src/test/org/apache/lucene/document/TestDateTools.java +++ b/src/test/org/apache/lucene/document/TestDateTools.java @@ -9,7 +9,6 @@ import java.util.TimeZone; import java.util.Locale; import org.apache.lucene.util.LocalizedTestCase; -import org.apache.lucene.util.LuceneTestCase; /** * Licensed to the Apache Software Foundation (ASF) under one or more diff --git a/src/test/org/apache/lucene/document/TestNumberTools.java b/src/test/org/apache/lucene/document/TestNumberTools.java index e457b338617..12734d91cb8 100644 --- a/src/test/org/apache/lucene/document/TestNumberTools.java +++ b/src/test/org/apache/lucene/document/TestNumberTools.java @@ -29,7 +29,7 @@ public class TestNumberTools extends LuceneTestCase { } public void testMax() { - // make sure the constants convert to their equivelents + // make sure the constants convert to their equivalents assertEquals(Long.MAX_VALUE, NumberTools .stringToLong(NumberTools.MAX_STRING_VALUE)); assertEquals(NumberTools.MAX_STRING_VALUE, NumberTools diff --git a/src/test/org/apache/lucene/index/DocHelper.java b/src/test/org/apache/lucene/index/DocHelper.java index 5f72b9e1fba..93ef7accdf9 100644 --- a/src/test/org/apache/lucene/index/DocHelper.java +++ b/src/test/org/apache/lucene/index/DocHelper.java @@ -109,7 +109,7 @@ class DocHelper { - public static Map nameValues = null; + public static Map nameValues = null; // ordered list of all the fields... // could use LinkedHashMap for this purpose if Java1.4 is OK @@ -130,17 +130,16 @@ class DocHelper { largeLazyField//placeholder for large field, since this is null. It must always be last }; - // Map - public static Map all=new HashMap(); - public static Map indexed=new HashMap(); - public static Map stored=new HashMap(); - public static Map unstored=new HashMap(); - public static Map unindexed=new HashMap(); - public static Map termvector=new HashMap(); - public static Map notermvector=new HashMap(); - public static Map lazy= new HashMap(); - public static Map noNorms=new HashMap(); - public static Map noTf=new HashMap(); + public static Map all =new HashMap(); + public static Map indexed =new HashMap(); + public static Map stored =new HashMap(); + public static Map unstored=new HashMap(); + public static Map unindexed=new HashMap(); + public static Map termvector=new HashMap(); + public static Map notermvector=new HashMap(); + public static Map lazy= new HashMap(); + public static Map noNorms=new HashMap(); + public static Map noTf=new HashMap(); static { //Initialize the large Lazy Field @@ -175,14 +174,14 @@ class DocHelper { } - private static void add(Map map, Fieldable field) { + private static void add(Map map, Fieldable field) { map.put(field.name(), field); } static { - nameValues = new HashMap(); + nameValues = new HashMap(); nameValues.put(TEXT_FIELD_1_KEY, FIELD_1_TEXT); nameValues.put(TEXT_FIELD_2_KEY, FIELD_2_TEXT); nameValues.put(TEXT_FIELD_3_KEY, FIELD_3_TEXT); diff --git a/src/test/org/apache/lucene/index/TestAtomicUpdate.java b/src/test/org/apache/lucene/index/TestAtomicUpdate.java index edee6a265f2..f8b7d0e4b3b 100644 --- a/src/test/org/apache/lucene/index/TestAtomicUpdate.java +++ b/src/test/org/apache/lucene/index/TestAtomicUpdate.java @@ -20,8 +20,6 @@ import org.apache.lucene.util.*; import org.apache.lucene.store.*; import org.apache.lucene.document.*; import org.apache.lucene.analysis.*; -import org.apache.lucene.search.*; -import org.apache.lucene.queryParser.*; import java.util.Random; import java.io.File; diff --git a/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java b/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java index 07a7668f4ea..0e80ec9bbdb 100644 --- a/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java +++ b/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java @@ -286,7 +286,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase for(int i=0;i<35;i++) { if (!reader.isDeleted(i)) { Document d = reader.document(i); - List fields = d.getFields(); + List fields = d.getFields(); if (!oldName.startsWith("19.") && !oldName.startsWith("20.") && !oldName.startsWith("21.") && @@ -295,19 +295,19 @@ public class TestBackwardsCompatibility extends LuceneTestCase if (d.getField("content3") == null) { final int numFields = oldName.startsWith("29.") ? 7 : 5; assertEquals(numFields, fields.size()); - Field f = (Field) d.getField("id"); + Field f = d.getField("id"); assertEquals(""+i, f.stringValue()); - f = (Field) d.getField("utf8"); + f = d.getField("utf8"); assertEquals("Lu\uD834\uDD1Ece\uD834\uDD60ne \u0000 \u2620 ab\ud917\udc17cd", f.stringValue()); - f = (Field) d.getField("autf8"); + f = d.getField("autf8"); assertEquals("Lu\uD834\uDD1Ece\uD834\uDD60ne \u0000 \u2620 ab\ud917\udc17cd", f.stringValue()); - f = (Field) d.getField("content2"); + f = d.getField("content2"); assertEquals("here is more content with aaa aaa aaa", f.stringValue()); - f = (Field) d.getField("fie\u2C77ld"); + f = d.getField("fie\u2C77ld"); assertEquals("field with non-ascii name", f.stringValue()); } } diff --git a/src/test/org/apache/lucene/index/TestByteSlices.java b/src/test/org/apache/lucene/index/TestByteSlices.java index 8f6f943dc94..06b601500e2 100644 --- a/src/test/org/apache/lucene/index/TestByteSlices.java +++ b/src/test/org/apache/lucene/index/TestByteSlices.java @@ -21,7 +21,7 @@ import org.apache.lucene.util.LuceneTestCase; public class TestByteSlices extends LuceneTestCase { private static class ByteBlockAllocator extends ByteBlockPool.Allocator { - ArrayList freeByteBlocks = new ArrayList(); + ArrayList freeByteBlocks = new ArrayList(); /* Allocate another byte[] from the shared pool */ @Override @@ -31,7 +31,7 @@ public class TestByteSlices extends LuceneTestCase { if (0 == size) b = new byte[DocumentsWriter.BYTE_BLOCK_SIZE]; else - b = (byte[]) freeByteBlocks.remove(size-1); + b = freeByteBlocks.remove(size-1); return b; } diff --git a/src/test/org/apache/lucene/index/TestCheckIndex.java b/src/test/org/apache/lucene/index/TestCheckIndex.java index 237ccffdd50..ce50f4a7c1f 100644 --- a/src/test/org/apache/lucene/index/TestCheckIndex.java +++ b/src/test/org/apache/lucene/index/TestCheckIndex.java @@ -58,7 +58,7 @@ public class TestCheckIndex extends LuceneTestCase { fail(); } - final CheckIndex.Status.SegmentInfoStatus seg = (CheckIndex.Status.SegmentInfoStatus) indexStatus.segmentInfos.get(0); + final CheckIndex.Status.SegmentInfoStatus seg = indexStatus.segmentInfos.get(0); assertTrue(seg.openReaderPassed); assertNotNull(seg.diagnostics); @@ -84,7 +84,7 @@ public class TestCheckIndex extends LuceneTestCase { assertEquals(18, seg.termVectorStatus.totVectors); assertTrue(seg.diagnostics.size() > 0); - final List onlySegments = new ArrayList(); + final List onlySegments = new ArrayList(); onlySegments.add("_0"); assertTrue(checker.checkIndex(onlySegments).clean == true); diff --git a/src/test/org/apache/lucene/index/TestDeletionPolicy.java b/src/test/org/apache/lucene/index/TestDeletionPolicy.java index c4751ff9817..02f36eab2c2 100644 --- a/src/test/org/apache/lucene/index/TestDeletionPolicy.java +++ b/src/test/org/apache/lucene/index/TestDeletionPolicy.java @@ -19,7 +19,6 @@ package org.apache.lucene.index; import java.io.IOException; import java.util.HashSet; -import java.util.Iterator; import java.util.List; import java.util.Set; import java.util.Collection; @@ -43,14 +42,14 @@ import org.apache.lucene.util.LuceneTestCase; public class TestDeletionPolicy extends LuceneTestCase { - private void verifyCommitOrder(List commits) throws IOException { - final IndexCommit firstCommit = ((IndexCommit) commits.get(0)); + private void verifyCommitOrder(List commits) throws IOException { + final IndexCommit firstCommit = commits.get(0); long last = SegmentInfos.generationFromSegmentsFileName(firstCommit.getSegmentsFileName()); assertEquals(last, firstCommit.getGeneration()); long lastVersion = firstCommit.getVersion(); long lastTimestamp = firstCommit.getTimestamp(); for(int i=1;i commits) throws IOException { verifyCommitOrder(commits); numOnInit++; } - public void onCommit(List commits) throws IOException { - IndexCommit lastCommit = (IndexCommit) commits.get(commits.size()-1); + public void onCommit(List commits) throws IOException { + IndexCommit lastCommit = commits.get(commits.size()-1); IndexReader r = IndexReader.open(dir, true); assertEquals("lastCommit.isOptimized()=" + lastCommit.isOptimized() + " vs IndexReader.isOptimized=" + r.isOptimized(), r.isOptimized(), lastCommit.isOptimized()); r.close(); @@ -89,18 +88,16 @@ public class TestDeletionPolicy extends LuceneTestCase class KeepNoneOnInitDeletionPolicy implements IndexDeletionPolicy { int numOnInit; int numOnCommit; - public void onInit(List commits) throws IOException { + public void onInit(List commits) throws IOException { verifyCommitOrder(commits); numOnInit++; // On init, delete all commit points: - Iterator it = commits.iterator(); - while(it.hasNext()) { - final IndexCommit commit = (IndexCommit) it.next(); + for (final IndexCommit commit : commits) { commit.delete(); assertTrue(commit.isDeleted()); } } - public void onCommit(List commits) throws IOException { + public void onCommit(List commits) throws IOException { verifyCommitOrder(commits); int size = commits.size(); // Delete all but last one: @@ -116,25 +113,25 @@ public class TestDeletionPolicy extends LuceneTestCase int numOnCommit; int numToKeep; int numDelete; - Set seen = new HashSet(); + Set seen = new HashSet(); public KeepLastNDeletionPolicy(int numToKeep) { this.numToKeep = numToKeep; } - public void onInit(List commits) throws IOException { + public void onInit(List commits) throws IOException { verifyCommitOrder(commits); numOnInit++; // do no deletions on init doDeletes(commits, false); } - public void onCommit(List commits) throws IOException { + public void onCommit(List commits) throws IOException { verifyCommitOrder(commits); doDeletes(commits, true); } - private void doDeletes(List commits, boolean isCommit) { + private void doDeletes(List commits, boolean isCommit) { // Assert that we really are only called for each new // commit: @@ -169,23 +166,21 @@ public class TestDeletionPolicy extends LuceneTestCase this.expirationTimeSeconds = seconds; } - public void onInit(List commits) throws IOException { + public void onInit(List commits) throws IOException { verifyCommitOrder(commits); onCommit(commits); } - public void onCommit(List commits) throws IOException { + public void onCommit(List commits) throws IOException { verifyCommitOrder(commits); - IndexCommit lastCommit = (IndexCommit) commits.get(commits.size()-1); + IndexCommit lastCommit = commits.get(commits.size()-1); // Any commit older than expireTime should be deleted: double expireTime = dir.fileModified(lastCommit.getSegmentsFileName())/1000.0 - expirationTimeSeconds; - Iterator it = commits.iterator(); - while(it.hasNext()) { - IndexCommit commit = (IndexCommit) it.next(); + for (final IndexCommit commit : commits) { double modTime = dir.fileModified(commit.getSegmentsFileName())/1000.0; if (commit != lastCommit && modTime < expireTime) { commit.delete(); @@ -297,14 +292,12 @@ public class TestDeletionPolicy extends LuceneTestCase assertEquals(2, policy.numOnCommit); // Test listCommits - Collection commits = IndexReader.listCommits(dir); + Collection commits = IndexReader.listCommits(dir); // 1 from opening writer + 2 from closing writer assertEquals(3, commits.size()); - Iterator it = commits.iterator(); // Make sure we can open a reader on each commit: - while(it.hasNext()) { - IndexCommit commit = (IndexCommit) it.next(); + for (final IndexCommit commit : commits) { IndexReader r = IndexReader.open(commit, null, false); r.close(); } @@ -356,12 +349,10 @@ public class TestDeletionPolicy extends LuceneTestCase } writer.close(); - Collection commits = IndexReader.listCommits(dir); + Collection commits = IndexReader.listCommits(dir); assertEquals(6, commits.size()); IndexCommit lastCommit = null; - Iterator it = commits.iterator(); - while(it.hasNext()) { - IndexCommit commit = (IndexCommit) it.next(); + for (final IndexCommit commit : commits) { if (lastCommit == null || commit.getGeneration() > lastCommit.getGeneration()) lastCommit = commit; } diff --git a/src/test/org/apache/lucene/index/TestDoc.java b/src/test/org/apache/lucene/index/TestDoc.java index 6b94ff9fd7f..302873a2659 100644 --- a/src/test/org/apache/lucene/index/TestDoc.java +++ b/src/test/org/apache/lucene/index/TestDoc.java @@ -22,7 +22,7 @@ import java.io.FileWriter; import java.io.IOException; import java.io.PrintWriter; import java.io.StringWriter; -import java.util.Iterator; + import java.util.LinkedList; import java.util.List; @@ -48,7 +48,7 @@ public class TestDoc extends LuceneTestCase { private File workDir; private File indexDir; - private LinkedList files; + private LinkedList files; /** Set the test case. This test case needs @@ -66,7 +66,7 @@ public class TestDoc extends LuceneTestCase { Directory directory = FSDirectory.open(indexDir); directory.close(); - files = new LinkedList(); + files = new LinkedList(); files.add(createOutput("test.txt", "This is the first test file" )); @@ -188,9 +188,9 @@ public class TestDoc extends LuceneTestCase { merger.closeReaders(); if (useCompoundFile) { - List filesToDelete = merger.createCompoundFile(merged + ".cfs"); - for (Iterator iter = filesToDelete.iterator(); iter.hasNext();) - si1.dir.deleteFile((String) iter.next()); + List filesToDelete = merger.createCompoundFile(merged + ".cfs"); + for (final String fileToDelete : filesToDelete) + si1.dir.deleteFile(fileToDelete); } return new SegmentInfo(merged, si1.docCount + si2.docCount, si1.dir, useCompoundFile, true); diff --git a/src/test/org/apache/lucene/index/TestFieldsReader.java b/src/test/org/apache/lucene/index/TestFieldsReader.java index ef8876040f4..20996362b2d 100644 --- a/src/test/org/apache/lucene/index/TestFieldsReader.java +++ b/src/test/org/apache/lucene/index/TestFieldsReader.java @@ -100,10 +100,10 @@ public class TestFieldsReader extends LuceneTestCase { FieldsReader reader = new FieldsReader(dir, TEST_SEGMENT_NAME, fieldInfos); assertTrue(reader != null); assertTrue(reader.size() == 1); - Set loadFieldNames = new HashSet(); + Set loadFieldNames = new HashSet(); loadFieldNames.add(DocHelper.TEXT_FIELD_1_KEY); loadFieldNames.add(DocHelper.TEXT_FIELD_UTF1_KEY); - Set lazyFieldNames = new HashSet(); + Set lazyFieldNames = new HashSet(); //new String[]{DocHelper.LARGE_LAZY_FIELD_KEY, DocHelper.LAZY_FIELD_KEY, DocHelper.LAZY_FIELD_BINARY_KEY}; lazyFieldNames.add(DocHelper.LARGE_LAZY_FIELD_KEY); lazyFieldNames.add(DocHelper.LAZY_FIELD_KEY); @@ -150,10 +150,10 @@ public class TestFieldsReader extends LuceneTestCase { FieldsReader reader = new FieldsReader(dir, TEST_SEGMENT_NAME, fieldInfos); assertTrue(reader != null); assertTrue(reader.size() == 1); - Set loadFieldNames = new HashSet(); + Set loadFieldNames = new HashSet(); loadFieldNames.add(DocHelper.TEXT_FIELD_1_KEY); loadFieldNames.add(DocHelper.TEXT_FIELD_UTF1_KEY); - Set lazyFieldNames = new HashSet(); + Set lazyFieldNames = new HashSet(); lazyFieldNames.add(DocHelper.LARGE_LAZY_FIELD_KEY); lazyFieldNames.add(DocHelper.LAZY_FIELD_KEY); lazyFieldNames.add(DocHelper.LAZY_FIELD_BINARY_KEY); @@ -183,9 +183,10 @@ public class TestFieldsReader extends LuceneTestCase { Document doc = reader.doc(0, fieldSelector); assertTrue("doc is null and it shouldn't be", doc != null); int count = 0; - List l = doc.getFields(); - for (Iterator iter = l.iterator(); iter.hasNext();) { - Field field = (Field) iter.next(); + List l = doc.getFields(); + for (final Fieldable fieldable : l ) { + Field field = (Field) fieldable; + assertTrue("field is null and it shouldn't be", field != null); String sv = field.stringValue(); assertTrue("sv is null and it shouldn't be", sv != null); @@ -220,9 +221,9 @@ public class TestFieldsReader extends LuceneTestCase { long lazyTime = 0; long regularTime = 0; int length = 50; - Set lazyFieldNames = new HashSet(); + Set lazyFieldNames = new HashSet(); lazyFieldNames.add(DocHelper.LARGE_LAZY_FIELD_KEY); - SetBasedFieldSelector fieldSelector = new SetBasedFieldSelector(Collections.EMPTY_SET, lazyFieldNames); + SetBasedFieldSelector fieldSelector = new SetBasedFieldSelector(Collections. emptySet(), lazyFieldNames); for (int i = 0; i < length; i++) { reader = new FieldsReader(tmpDir, TEST_SEGMENT_NAME, fieldInfos); diff --git a/src/test/org/apache/lucene/index/TestIndexFileDeleter.java b/src/test/org/apache/lucene/index/TestIndexFileDeleter.java index 3f9b716715c..f95133911d9 100644 --- a/src/test/org/apache/lucene/index/TestIndexFileDeleter.java +++ b/src/test/org/apache/lucene/index/TestIndexFileDeleter.java @@ -28,7 +28,6 @@ import org.apache.lucene.document.Document; import org.apache.lucene.document.Field; import java.io.*; import java.util.*; -import java.util.zip.*; /* Verify we can read the pre-2.1 file format, do searches @@ -155,33 +154,34 @@ public class TestIndexFileDeleter extends LuceneTestCase Arrays.sort(files); Arrays.sort(files2); - Set dif = difFiles(files, files2); + Set dif = difFiles(files, files2); if (!Arrays.equals(files, files2)) { fail("IndexFileDeleter failed to delete unreferenced extra files: should have deleted " + (filesPre.length-files.length) + " files but only deleted " + (filesPre.length - files2.length) + "; expected files:\n " + asString(files) + "\n actual files:\n " + asString(files2)+"\ndif: "+dif); } } - private static Set difFiles(String[] files1, String[] files2) { - Set set1 = new HashSet(); - Set set2 = new HashSet(); - Set extra = new HashSet(); + private static Set difFiles(String[] files1, String[] files2) { + Set set1 = new HashSet(); + Set set2 = new HashSet(); + Set extra = new HashSet(); + for (int x=0; x < files1.length; x++) { set1.add(files1[x]); } for (int x=0; x < files2.length; x++) { set2.add(files2[x]); } - Iterator i1 = set1.iterator(); + Iterator i1 = set1.iterator(); while (i1.hasNext()) { - Object o = i1.next(); + String o = i1.next(); if (!set2.contains(o)) { extra.add(o); } } - Iterator i2 = set2.iterator(); + Iterator i2 = set2.iterator(); while (i2.hasNext()) { - Object o = i2.next(); + String o = i2.next(); if (!set1.contains(o)) { extra.add(o); } diff --git a/src/test/org/apache/lucene/index/TestIndexReader.java b/src/test/org/apache/lucene/index/TestIndexReader.java index 0fe0bfdb123..47e149f3473 100644 --- a/src/test/org/apache/lucene/index/TestIndexReader.java +++ b/src/test/org/apache/lucene/index/TestIndexReader.java @@ -25,9 +25,11 @@ import java.util.Arrays; import java.util.Collection; import java.util.HashSet; import java.util.Iterator; +import java.util.List; import java.util.Map; import java.util.HashMap; import java.util.Set; +import java.util.SortedSet; import junit.framework.TestSuite; import junit.textui.TestRunner; @@ -72,7 +74,7 @@ public class TestIndexReader extends LuceneTestCase public void testCommitUserData() throws Exception { RAMDirectory d = new MockRAMDirectory(); - Map commitUserData = new HashMap(); + Map commitUserData = new HashMap(); commitUserData.put("foo", "fighters"); // set up writer @@ -156,7 +158,7 @@ public class TestIndexReader extends LuceneTestCase writer.close(); // set up reader IndexReader reader = IndexReader.open(d, false); - Collection fieldNames = reader.getFieldNames(IndexReader.FieldOption.ALL); + Collection fieldNames = reader.getFieldNames(IndexReader.FieldOption.ALL); assertTrue(fieldNames.contains("keyword")); assertTrue(fieldNames.contains("text")); assertTrue(fieldNames.contains("unindexed")); @@ -260,12 +262,12 @@ public class TestIndexReader extends LuceneTestCase IndexReader reader = IndexReader.open(d, false); FieldSortedTermVectorMapper mapper = new FieldSortedTermVectorMapper(new TermVectorEntryFreqSortedComparator()); reader.getTermFreqVector(0, mapper); - Map map = mapper.getFieldToTerms(); + Map> map = mapper.getFieldToTerms(); assertTrue("map is null and it shouldn't be", map != null); assertTrue("map Size: " + map.size() + " is not: " + 4, map.size() == 4); - Set set = (Set) map.get("termvector"); - for (Iterator iterator = set.iterator(); iterator.hasNext();) { - TermVectorEntry entry = (TermVectorEntry) iterator.next(); + Set set = map.get("termvector"); + for (Iterator iterator = set.iterator(); iterator.hasNext();) { + TermVectorEntry entry = iterator.next(); assertTrue("entry is null and it shouldn't be", entry != null); System.out.println("Entry: " + entry); } @@ -380,9 +382,9 @@ public class TestIndexReader extends LuceneTestCase for (int i = 0; i < bin.length; i++) { assertEquals(bin[i], data1[i + b1.getBinaryOffset()]); } - Set lazyFields = new HashSet(); + Set lazyFields = new HashSet(); lazyFields.add("bin1"); - FieldSelector sel = new SetBasedFieldSelector(new HashSet(), lazyFields); + FieldSelector sel = new SetBasedFieldSelector(new HashSet(), lazyFields); doc = reader.document(reader.maxDoc() - 1, sel); Fieldable[] fieldables = doc.getFieldables("bin1"); assertNotNull(fieldables); @@ -1340,19 +1342,19 @@ public class TestIndexReader extends LuceneTestCase assertEquals("Only one index is optimized.", index1.isOptimized(), index2.isOptimized()); // check field names - Collection fields1 = index1.getFieldNames(FieldOption.ALL); - Collection fields2 = index1.getFieldNames(FieldOption.ALL); + Collection fields1 = index1.getFieldNames(FieldOption.ALL); + Collection fields2 = index1.getFieldNames(FieldOption.ALL); assertEquals("IndexReaders have different numbers of fields.", fields1.size(), fields2.size()); - Iterator it1 = fields1.iterator(); - Iterator it2 = fields1.iterator(); + Iterator it1 = fields1.iterator(); + Iterator it2 = fields1.iterator(); while (it1.hasNext()) { - assertEquals("Different field names.", (String) it1.next(), (String) it2.next()); + assertEquals("Different field names.", it1.next(), it2.next()); } // check norms it1 = fields1.iterator(); while (it1.hasNext()) { - String curField = (String) it1.next(); + String curField = it1.next(); byte[] norms1 = index1.norms(curField); byte[] norms2 = index2.norms(curField); if (norms1 != null && norms2 != null) @@ -1378,14 +1380,14 @@ public class TestIndexReader extends LuceneTestCase if (!index1.isDeleted(i)) { Document doc1 = index1.document(i); Document doc2 = index2.document(i); - fields1 = doc1.getFields(); - fields2 = doc2.getFields(); - assertEquals("Different numbers of fields for doc " + i + ".", fields1.size(), fields2.size()); - it1 = fields1.iterator(); - it2 = fields2.iterator(); - while (it1.hasNext()) { - Field curField1 = (Field) it1.next(); - Field curField2 = (Field) it2.next(); + List fieldable1 = doc1.getFields(); + List fieldable2 = doc2.getFields(); + assertEquals("Different numbers of fields for doc " + i + ".", fieldable1.size(), fieldable2.size()); + Iterator itField1 = fieldable1.iterator(); + Iterator itField2 = fieldable2.iterator(); + while (itField1.hasNext()) { + Field curField1 = (Field) itField1.next(); + Field curField2 = (Field) itField2.next(); assertEquals("Different fields names for doc " + i + ".", curField1.name(), curField2.name()); assertEquals("Different field values for doc " + i + ".", curField1.stringValue(), curField2.stringValue()); } @@ -1587,15 +1589,11 @@ public class TestIndexReader extends LuceneTestCase writer.addDocument(createDocument("a")); writer.close(); - Collection commits = IndexReader.listCommits(dir); - Iterator it = commits.iterator(); - while(it.hasNext()) { - IndexCommit commit = (IndexCommit) it.next(); - Collection files = commit.getFileNames(); - HashSet seen = new HashSet(); - Iterator it2 = files.iterator(); - while(it2.hasNext()) { - String fileName = (String) it2.next(); + Collection commits = IndexReader.listCommits(dir); + for (final IndexCommit commit : commits) { + Collection files = commit.getFileNames(); + HashSet seen = new HashSet(); + for (final String fileName : files) { assertTrue("file " + fileName + " was duplicated", !seen.contains(fileName)); seen.add(fileName); } diff --git a/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java b/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java index 577d4925a60..a1446bc0b3a 100644 --- a/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java +++ b/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java @@ -55,9 +55,9 @@ public class TestIndexReaderCloneNorms extends LuceneTestCase { private int numDocNorms; - private ArrayList norms; + private ArrayList norms; - private ArrayList modifiedNorms; + private ArrayList modifiedNorms; private float lastNorm = 0; @@ -91,19 +91,19 @@ public class TestIndexReaderCloneNorms extends LuceneTestCase { Directory dir1 = FSDirectory.open(indexDir1); IndexWriter.unlock(dir1); - norms = new ArrayList(); - modifiedNorms = new ArrayList(); + norms = new ArrayList(); + modifiedNorms = new ArrayList(); createIndex(dir1); doTestNorms(dir1); // test with a single index: index2 - ArrayList norms1 = norms; - ArrayList modifiedNorms1 = modifiedNorms; + ArrayList norms1 = norms; + ArrayList modifiedNorms1 = modifiedNorms; int numDocNorms1 = numDocNorms; - norms = new ArrayList(); - modifiedNorms = new ArrayList(); + norms = new ArrayList(); + modifiedNorms = new ArrayList(); numDocNorms = 0; File indexDir2 = new File(tempDir, "lucenetestindex2"); @@ -282,10 +282,10 @@ public class TestIndexReaderCloneNorms extends LuceneTestCase { String field = "f" + i; byte b[] = ir.norms(field); assertEquals("number of norms mismatches", numDocNorms, b.length); - ArrayList storedNorms = (i == 1 ? modifiedNorms : norms); + ArrayList storedNorms = (i == 1 ? modifiedNorms : norms); for (int j = 0; j < b.length; j++) { float norm = Similarity.getDefault().decodeNormValue(b[j]); - float norm1 = ((Float) storedNorms.get(j)).floatValue(); + float norm1 = storedNorms.get(j).floatValue(); assertEquals("stored norm value of " + field + " for doc " + j + " is " + norm + " - a mismatch!", norm, norm1, 0.000001); } diff --git a/src/test/org/apache/lucene/index/TestIndexReaderReopen.java b/src/test/org/apache/lucene/index/TestIndexReaderReopen.java index c5206ebfc98..cf8e7686b8d 100644 --- a/src/test/org/apache/lucene/index/TestIndexReaderReopen.java +++ b/src/test/org/apache/lucene/index/TestIndexReaderReopen.java @@ -20,9 +20,10 @@ package org.apache.lucene.index; import java.io.File; import java.io.IOException; import java.util.ArrayList; +import java.util.Collection; import java.util.Collections; import java.util.HashSet; -import java.util.Iterator; + import java.util.List; import java.util.Random; import java.util.Map; @@ -732,13 +733,13 @@ public class TestIndexReaderReopen extends LuceneTestCase { } }; - final List readers = Collections.synchronizedList(new ArrayList()); + final List readers = Collections.synchronizedList(new ArrayList()); IndexReader firstReader = IndexReader.open(dir, false); IndexReader reader = firstReader; final Random rnd = newRandom(); ReaderThread[] threads = new ReaderThread[n]; - final Set readersToClose = Collections.synchronizedSet(new HashSet()); + final Set readersToClose = Collections.synchronizedSet(new HashSet()); for (int i = 0; i < n; i++) { if (i % 2 == 0) { @@ -806,7 +807,7 @@ public class TestIndexReaderReopen extends LuceneTestCase { while (!stopped) { int numReaders = readers.size(); if (numReaders > 0) { - ReaderCouple c = (ReaderCouple) readers.get(rnd.nextInt(numReaders)); + ReaderCouple c = readers.get(rnd.nextInt(numReaders)); TestIndexReader.assertIndexEquals(c.newReader, c.refreshedReader); } @@ -845,17 +846,15 @@ public class TestIndexReaderReopen extends LuceneTestCase { } - Iterator it = readersToClose.iterator(); - while (it.hasNext()) { - ((IndexReader) it.next()).close(); + for (final IndexReader readerToClose : readersToClose) { + readerToClose.close(); } firstReader.close(); reader.close(); - it = readersToClose.iterator(); - while (it.hasNext()) { - assertReaderClosed((IndexReader) it.next(), true, true); + for (final IndexReader readerToClose : readersToClose) { + assertReaderClosed(readerToClose, true, true); } assertReaderClosed(reader, true, true); @@ -1185,9 +1184,9 @@ public class TestIndexReaderReopen extends LuceneTestCase { } private static class KeepAllCommits implements IndexDeletionPolicy { - public void onInit(List commits) { + public void onInit(List commits) { } - public void onCommit(List commits) { + public void onCommit(List commits) { } } @@ -1198,13 +1197,13 @@ public class TestIndexReaderReopen extends LuceneTestCase { Document doc = new Document(); doc.add(new Field("id", ""+i, Field.Store.NO, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); - Map data = new HashMap(); + Map data = new HashMap(); data.put("index", i+""); writer.commit(data); } for(int i=0;i<4;i++) { writer.deleteDocuments(new Term("id", ""+i)); - Map data = new HashMap(); + Map data = new HashMap(); data.put("index", (4+i)+""); writer.commit(data); } @@ -1214,9 +1213,8 @@ public class TestIndexReaderReopen extends LuceneTestCase { assertEquals(0, r.numDocs()); assertEquals(4, r.maxDoc()); - Iterator it = IndexReader.listCommits(dir).iterator(); - while(it.hasNext()) { - IndexCommit commit = (IndexCommit) it.next(); + Collection commits = IndexReader.listCommits(dir); + for (final IndexCommit commit : commits) { IndexReader r2 = r.reopen(commit); assertTrue(r2 != r); @@ -1228,13 +1226,13 @@ public class TestIndexReaderReopen extends LuceneTestCase { // expected } - final Map s = commit.getUserData(); + final Map s = commit.getUserData(); final int v; if (s.size() == 0) { // First commit created by IW v = -1; } else { - v = Integer.parseInt((String) s.get("index")); + v = Integer.parseInt(s.get("index")); } if (v < 4) { assertEquals(1+v, r2.numDocs()); diff --git a/src/test/org/apache/lucene/index/TestIndexWriter.java b/src/test/org/apache/lucene/index/TestIndexWriter.java index d2965f10fb4..55ce6a06ad2 100644 --- a/src/test/org/apache/lucene/index/TestIndexWriter.java +++ b/src/test/org/apache/lucene/index/TestIndexWriter.java @@ -48,6 +48,7 @@ import org.apache.lucene.analysis.tokenattributes.TermAttribute; import org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute; import org.apache.lucene.document.Document; import org.apache.lucene.document.Field; +import org.apache.lucene.document.Fieldable; import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.search.PhraseQuery; import org.apache.lucene.search.Query; @@ -2131,7 +2132,7 @@ public class TestIndexWriter extends LuceneTestCase { writer.setMergeFactor(2); final IndexWriter finalWriter = writer; - final ArrayList failure = new ArrayList(); + final ArrayList failure = new ArrayList(); Thread t1 = new Thread() { @Override public void run() { @@ -2160,7 +2161,7 @@ public class TestIndexWriter extends LuceneTestCase { }; if (failure.size() > 0) - throw (Throwable) failure.get(0); + throw failure.get(0); t1.start(); @@ -3475,14 +3476,14 @@ public class TestIndexWriter extends LuceneTestCase { final TermAttribute termAtt = addAttribute(TermAttribute.class); final PositionIncrementAttribute posIncrAtt = addAttribute(PositionIncrementAttribute.class); - final Iterator tokens = Arrays.asList(new String[]{"a","b","c"}).iterator(); + final Iterator tokens = Arrays.asList(new String[]{"a","b","c"}).iterator(); boolean first = true; @Override public boolean incrementToken() { if (!tokens.hasNext()) return false; clearAttributes(); - termAtt.setTermBuffer((String) tokens.next()); + termAtt.setTermBuffer( tokens.next()); posIncrAtt.setPositionIncrement(first ? 0 : 1); first = false; return true; @@ -3643,7 +3644,7 @@ public class TestIndexWriter extends LuceneTestCase { Directory dir, dir2; final static int NUM_INIT_DOCS = 17; IndexWriter writer2; - final List failures = new ArrayList(); + final List failures = new ArrayList(); volatile boolean didClose; final IndexReader[] readers; final int NUM_COPY; @@ -3992,7 +3993,7 @@ public class TestIndexWriter extends LuceneTestCase { w.setMaxBufferedDocs(2); for(int j=0;j<17;j++) addDoc(w); - Map data = new HashMap(); + Map data = new HashMap(); data.put("label", "test1"); w.commit(data); w.close(); @@ -4040,7 +4041,7 @@ public class TestIndexWriter extends LuceneTestCase { // LUCENE-1429 public void testOutOfMemoryErrorCausesCloseToFail() throws Exception { - final List thrown = new ArrayList(); + final List thrown = new ArrayList(); final IndexWriter writer = new IndexWriter(new MockRAMDirectory(), new StandardAnalyzer(org.apache.lucene.util.Version.LUCENE_CURRENT), IndexWriter.MaxFieldLength.UNLIMITED) { @Override @@ -4562,7 +4563,7 @@ public class TestIndexWriter extends LuceneTestCase { w.addDocument(doc); IndexReader r = w.getReader(); doc = r.document(0); - Iterator it = doc.getFields().iterator(); + Iterator it = doc.getFields().iterator(); assertTrue(it.hasNext()); Field f = (Field) it.next(); assertEquals(f.name(), "zzz"); diff --git a/src/test/org/apache/lucene/index/TestIndexWriterExceptions.java b/src/test/org/apache/lucene/index/TestIndexWriterExceptions.java index be3f7806742..dfc8b05eee1 100644 --- a/src/test/org/apache/lucene/index/TestIndexWriterExceptions.java +++ b/src/test/org/apache/lucene/index/TestIndexWriterExceptions.java @@ -109,7 +109,7 @@ public class TestIndexWriterExceptions extends LuceneTestCase { } } - ThreadLocal doFail = new ThreadLocal(); + ThreadLocal doFail = new ThreadLocal(); public class MockIndexWriter extends IndexWriter { Random r = new java.util.Random(17); diff --git a/src/test/org/apache/lucene/index/TestIndexWriterReader.java b/src/test/org/apache/lucene/index/TestIndexWriterReader.java index ddd97dfda73..cccf31ae098 100644 --- a/src/test/org/apache/lucene/index/TestIndexWriterReader.java +++ b/src/test/org/apache/lucene/index/TestIndexWriterReader.java @@ -306,10 +306,10 @@ public class TestIndexWriterReader extends LuceneTestCase { final static int NUM_THREADS = 5; final Thread[] threads = new Thread[NUM_THREADS]; IndexWriter mainWriter; - List deletedTerms = new ArrayList(); - LinkedList toDeleteTerms = new LinkedList(); + List deletedTerms = new ArrayList(); + LinkedList toDeleteTerms = new LinkedList(); Random random; - final List failures = new ArrayList(); + final List failures = new ArrayList(); public DeleteThreads(IndexWriter mainWriter) throws IOException { this.mainWriter = mainWriter; @@ -326,7 +326,7 @@ public class TestIndexWriterReader extends LuceneTestCase { Term getDeleteTerm() { synchronized (toDeleteTerms) { - return (Term)toDeleteTerms.removeFirst(); + return toDeleteTerms.removeFirst(); } } @@ -373,7 +373,7 @@ public class TestIndexWriterReader extends LuceneTestCase { int numDirs; final Thread[] threads = new Thread[NUM_THREADS]; IndexWriter mainWriter; - final List failures = new ArrayList(); + final List failures = new ArrayList(); IndexReader[] readers; boolean didClose = false; HeavyAtomicInt count = new HeavyAtomicInt(0); @@ -723,7 +723,7 @@ public class TestIndexWriterReader extends LuceneTestCase { final float SECONDS = 0.5f; final long endTime = (long) (System.currentTimeMillis() + 1000.*SECONDS); - final List excs = Collections.synchronizedList(new ArrayList()); + final List excs = Collections.synchronizedList(new ArrayList()); final Thread[] threads = new Thread[NUM_THREAD]; for(int i=0;i excs = Collections.synchronizedList(new ArrayList()); final Thread[] threads = new Thread[NUM_THREAD]; for(int i=0;i dataset = new HashSet(Arrays.asList(data)); private static String MAGIC_FIELD = "f"+(NUM_FIELDS/3); @@ -93,11 +93,11 @@ public class TestLazyBug extends LuceneTestCase { Document d = reader.document(docs[i], SELECTOR); d.get(MAGIC_FIELD); - List fields = d.getFields(); - for (Iterator fi = fields.iterator(); fi.hasNext(); ) { + List fields = d.getFields(); + for (Iterator fi = fields.iterator(); fi.hasNext(); ) { Fieldable f=null; try { - f = (Fieldable) fi.next(); + f = fi.next(); String fname = f.name(); String fval = f.stringValue(); assertNotNull(docs[i]+" FIELD: "+fname, fval); diff --git a/src/test/org/apache/lucene/index/TestNorms.java b/src/test/org/apache/lucene/index/TestNorms.java index fa544fb5eb4..9d8b2cca645 100755 --- a/src/test/org/apache/lucene/index/TestNorms.java +++ b/src/test/org/apache/lucene/index/TestNorms.java @@ -52,8 +52,8 @@ public class TestNorms extends LuceneTestCase { private Similarity similarityOne; private Analyzer anlzr; private int numDocNorms; - private ArrayList norms; - private ArrayList modifiedNorms; + private ArrayList norms; + private ArrayList modifiedNorms; private float lastNorm = 0; private float normDelta = (float) 0.001; @@ -85,19 +85,19 @@ public class TestNorms extends LuceneTestCase { File indexDir1 = new File(tempDir, "lucenetestindex1"); Directory dir1 = FSDirectory.open(indexDir1); - norms = new ArrayList(); - modifiedNorms = new ArrayList(); + norms = new ArrayList(); + modifiedNorms = new ArrayList(); createIndex(dir1); doTestNorms(dir1); // test with a single index: index2 - ArrayList norms1 = norms; - ArrayList modifiedNorms1 = modifiedNorms; + ArrayList norms1 = norms; + ArrayList modifiedNorms1 = modifiedNorms; int numDocNorms1 = numDocNorms; - norms = new ArrayList(); - modifiedNorms = new ArrayList(); + norms = new ArrayList(); + modifiedNorms = new ArrayList(); numDocNorms = 0; File indexDir2 = new File(tempDir, "lucenetestindex2"); @@ -187,10 +187,10 @@ public class TestNorms extends LuceneTestCase { String field = "f"+i; byte b[] = ir.norms(field); assertEquals("number of norms mismatches",numDocNorms,b.length); - ArrayList storedNorms = (i==1 ? modifiedNorms : norms); + ArrayList storedNorms = (i==1 ? modifiedNorms : norms); for (int j = 0; j < b.length; j++) { float norm = similarityOne.decodeNormValue(b[j]); - float norm1 = ((Float)storedNorms.get(j)).floatValue(); + float norm1 = storedNorms.get(j).floatValue(); assertEquals("stored norm value of "+field+" for doc "+j+" is "+norm+" - a mismatch!", norm, norm1, 0.000001); } } diff --git a/src/test/org/apache/lucene/index/TestParallelReader.java b/src/test/org/apache/lucene/index/TestParallelReader.java index 573fc8af85f..27ada396792 100644 --- a/src/test/org/apache/lucene/index/TestParallelReader.java +++ b/src/test/org/apache/lucene/index/TestParallelReader.java @@ -71,7 +71,7 @@ public class TestParallelReader extends LuceneTestCase { ParallelReader pr = new ParallelReader(); pr.add(IndexReader.open(dir1, false)); pr.add(IndexReader.open(dir2, false)); - Collection fieldNames = pr.getFieldNames(IndexReader.FieldOption.ALL); + Collection fieldNames = pr.getFieldNames(IndexReader.FieldOption.ALL); assertEquals(4, fieldNames.size()); assertTrue(fieldNames.contains("f1")); assertTrue(fieldNames.contains("f2")); diff --git a/src/test/org/apache/lucene/index/TestPayloads.java b/src/test/org/apache/lucene/index/TestPayloads.java index 356baa28804..ae13b86d8dd 100644 --- a/src/test/org/apache/lucene/index/TestPayloads.java +++ b/src/test/org/apache/lucene/index/TestPayloads.java @@ -382,7 +382,7 @@ public class TestPayloads extends LuceneTestCase { * This Analyzer uses an WhitespaceTokenizer and PayloadFilter. */ private static class PayloadAnalyzer extends Analyzer { - Map fieldToData = new HashMap(); + Map fieldToData = new HashMap(); void setPayloadData(String field, byte[] data, int offset, int length) { fieldToData.put(field, new PayloadData(0, data, offset, length)); @@ -394,7 +394,7 @@ public class TestPayloads extends LuceneTestCase { @Override public TokenStream tokenStream(String fieldName, Reader reader) { - PayloadData payload = (PayloadData) fieldToData.get(fieldName); + PayloadData payload = fieldToData.get(fieldName); TokenStream ts = new WhitespaceTokenizer(reader); if (payload != null) { if (payload.numFieldInstancesToSkip == 0) { @@ -550,10 +550,10 @@ public class TestPayloads extends LuceneTestCase { } private static class ByteArrayPool { - private List pool; + private List pool; ByteArrayPool(int capacity, int size) { - pool = new ArrayList(); + pool = new ArrayList(); for (int i = 0; i < capacity; i++) { pool.add(new byte[size]); } @@ -572,7 +572,7 @@ public class TestPayloads extends LuceneTestCase { } synchronized byte[] get() { - return (byte[]) pool.remove(0); + return pool.remove(0); } synchronized void release(byte[] b) { diff --git a/src/test/org/apache/lucene/index/TestPositionBasedTermVectorMapper.java b/src/test/org/apache/lucene/index/TestPositionBasedTermVectorMapper.java index 457fc9241ca..9dc45ac197c 100644 --- a/src/test/org/apache/lucene/index/TestPositionBasedTermVectorMapper.java +++ b/src/test/org/apache/lucene/index/TestPositionBasedTermVectorMapper.java @@ -19,7 +19,6 @@ import org.apache.lucene.util.LuceneTestCase; import java.io.IOException; import java.util.BitSet; -import java.util.Iterator; import java.util.Map; public class TestPositionBasedTermVectorMapper extends LuceneTestCase { @@ -69,19 +68,19 @@ public class TestPositionBasedTermVectorMapper extends LuceneTestCase { mapper.map(token, 1, null, thePositions[i]); } - Map map = mapper.getFieldToTerms(); + Map> map = mapper.getFieldToTerms(); assertTrue("map is null and it shouldn't be", map != null); assertTrue("map Size: " + map.size() + " is not: " + 1, map.size() == 1); - Map positions = (Map) map.get("test"); + Map positions = map.get("test"); assertTrue("thePositions is null and it shouldn't be", positions != null); assertTrue("thePositions Size: " + positions.size() + " is not: " + numPositions, positions.size() == numPositions); BitSet bits = new BitSet(numPositions); - for (Iterator iterator = positions.entrySet().iterator(); iterator.hasNext();) { - Map.Entry entry = (Map.Entry) iterator.next(); - PositionBasedTermVectorMapper.TVPositionInfo info = (PositionBasedTermVectorMapper.TVPositionInfo) entry.getValue(); + for (Map.Entry entry : positions.entrySet()) { + + PositionBasedTermVectorMapper.TVPositionInfo info = entry.getValue(); assertTrue("info is null and it shouldn't be", info != null); - int pos = ((Integer) entry.getKey()).intValue(); + int pos = entry.getKey().intValue(); bits.set(pos); assertTrue(info.getPosition() + " does not equal: " + pos, info.getPosition() == pos); assertTrue("info.getOffsets() is null and it shouldn't be", info.getOffsets() != null); diff --git a/src/test/org/apache/lucene/index/TestSegmentMerger.java b/src/test/org/apache/lucene/index/TestSegmentMerger.java index 7a6ac7e3778..6643c956698 100644 --- a/src/test/org/apache/lucene/index/TestSegmentMerger.java +++ b/src/test/org/apache/lucene/index/TestSegmentMerger.java @@ -85,7 +85,7 @@ public class TestSegmentMerger extends LuceneTestCase { assertTrue(termDocs != null); assertTrue(termDocs.next() == true); - Collection stored = mergedReader.getFieldNames(IndexReader.FieldOption.INDEXED_WITH_TERMVECTOR); + Collection stored = mergedReader.getFieldNames(IndexReader.FieldOption.INDEXED_WITH_TERMVECTOR); assertTrue(stored != null); //System.out.println("stored size: " + stored.size()); assertTrue("We do not have 3 fields that were indexed with term vector",stored.size() == 3); diff --git a/src/test/org/apache/lucene/index/TestSegmentReader.java b/src/test/org/apache/lucene/index/TestSegmentReader.java index f9cc0ef53d1..fff8d9cba15 100644 --- a/src/test/org/apache/lucene/index/TestSegmentReader.java +++ b/src/test/org/apache/lucene/index/TestSegmentReader.java @@ -62,9 +62,8 @@ public class TestSegmentReader extends LuceneTestCase { //There are 2 unstored fields on the document that are not preserved across writing assertTrue(DocHelper.numFields(result) == DocHelper.numFields(testDoc) - DocHelper.unstored.size()); - List fields = result.getFields(); - for (Iterator iter = fields.iterator(); iter.hasNext();) { - Fieldable field = (Fieldable) iter.next(); + List fields = result.getFields(); + for (final Fieldable field : fields ) { assertTrue(field != null); assertTrue(DocHelper.nameValues.containsKey(field.name())); } @@ -84,19 +83,19 @@ public class TestSegmentReader extends LuceneTestCase { } public void testGetFieldNameVariations() { - Collection result = reader.getFieldNames(IndexReader.FieldOption.ALL); + Collection result = reader.getFieldNames(IndexReader.FieldOption.ALL); assertTrue(result != null); assertTrue(result.size() == DocHelper.all.size()); - for (Iterator iter = result.iterator(); iter.hasNext();) { - String s = (String) iter.next(); + for (Iterator iter = result.iterator(); iter.hasNext();) { + String s = iter.next(); //System.out.println("Name: " + s); assertTrue(DocHelper.nameValues.containsKey(s) == true || s.equals("")); } result = reader.getFieldNames(IndexReader.FieldOption.INDEXED); assertTrue(result != null); assertTrue(result.size() == DocHelper.indexed.size()); - for (Iterator iter = result.iterator(); iter.hasNext();) { - String s = (String) iter.next(); + for (Iterator iter = result.iterator(); iter.hasNext();) { + String s = iter.next(); assertTrue(DocHelper.indexed.containsKey(s) == true || s.equals("")); } diff --git a/src/test/org/apache/lucene/index/TestStressIndexing.java b/src/test/org/apache/lucene/index/TestStressIndexing.java index 6e16eb164ee..ac149738c4c 100644 --- a/src/test/org/apache/lucene/index/TestStressIndexing.java +++ b/src/test/org/apache/lucene/index/TestStressIndexing.java @@ -21,7 +21,6 @@ import org.apache.lucene.store.*; import org.apache.lucene.document.*; import org.apache.lucene.analysis.*; import org.apache.lucene.search.*; -import org.apache.lucene.queryParser.*; import java.util.Random; import java.io.File; @@ -155,7 +154,7 @@ public class TestStressIndexing extends LuceneTestCase { modifier.close(); for(int i=0;i docs = indexRandom(10, 10, 100, dir1); indexSerial(docs, dir2); // verifying verify @@ -97,7 +97,7 @@ public class TestStressIndexing2 extends LuceneTestCase { int range=r.nextInt(20)+1; Directory dir1 = new MockRAMDirectory(); Directory dir2 = new MockRAMDirectory(); - Map docs = indexRandom(nThreads, iter, range, dir1); + Map docs = indexRandom(nThreads, iter, range, dir1); indexSerial(docs, dir2); verifyEquals(dir1, dir2, "id"); } @@ -106,9 +106,9 @@ public class TestStressIndexing2 extends LuceneTestCase { static Term idTerm = new Term("id",""); IndexingThread[] threads; - static Comparator fieldNameComparator = new Comparator() { - public int compare(Object o1, Object o2) { - return ((Fieldable)o1).name().compareTo(((Fieldable)o2).name()); + static Comparator fieldNameComparator = new Comparator() { + public int compare(Fieldable o1, Fieldable o2) { + return o1.name().compareTo(o2.name()); } }; @@ -117,12 +117,12 @@ public class TestStressIndexing2 extends LuceneTestCase { // everything. public static class DocsAndWriter { - Map docs; + Map docs; IndexWriter writer; } public DocsAndWriter indexRandomIWReader(int nThreads, int iterations, int range, Directory dir) throws IOException, InterruptedException { - Map docs = new HashMap(); + Map docs = new HashMap(); IndexWriter w = new MockIndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.UNLIMITED); w.setUseCompoundFile(false); @@ -172,8 +172,8 @@ public class TestStressIndexing2 extends LuceneTestCase { return dw; } - public Map indexRandom(int nThreads, int iterations, int range, Directory dir) throws IOException, InterruptedException { - Map docs = new HashMap(); + public Map indexRandom(int nThreads, int iterations, int range, Directory dir) throws IOException, InterruptedException { + Map docs = new HashMap(); for(int iter=0;iter<3;iter++) { IndexWriter w = new MockIndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.UNLIMITED); w.setUseCompoundFile(false); @@ -217,14 +217,14 @@ public class TestStressIndexing2 extends LuceneTestCase { } - public static void indexSerial(Map docs, Directory dir) throws IOException { + public static void indexSerial(Map docs, Directory dir) throws IOException { IndexWriter w = new IndexWriter(dir, new WhitespaceAnalyzer(), IndexWriter.MaxFieldLength.UNLIMITED); // index all docs in a single thread - Iterator iter = docs.values().iterator(); + Iterator iter = docs.values().iterator(); while (iter.hasNext()) { - Document d = (Document)iter.next(); - ArrayList fields = new ArrayList(); + Document d = iter.next(); + ArrayList fields = new ArrayList(); fields.addAll(d.getFields()); // put fields in same order each time Collections.sort(fields, fieldNameComparator); @@ -232,7 +232,7 @@ public class TestStressIndexing2 extends LuceneTestCase { Document d1 = new Document(); d1.setBoost(d.getBoost()); for (int i=0; i ff1 = d1.getFields(); + List ff2 = d2.getFields(); Collections.sort(ff1, fieldNameComparator); Collections.sort(ff2, fieldNameComparator); @@ -405,8 +405,8 @@ public class TestStressIndexing2 extends LuceneTestCase { for (int i=0; i + Map docs = new HashMap(); Random r; public int nextInt(int lim) { @@ -561,7 +561,7 @@ public class TestStressIndexing2 extends LuceneTestCase { public void indexDoc() throws IOException { Document d = new Document(); - ArrayList fields = new ArrayList(); + ArrayList fields = new ArrayList(); String idString = getIdString(); Field idField = new Field(idTerm.field(), idString, Field.Store.YES, Field.Index.NOT_ANALYZED_NO_NORMS); fields.add(idField); @@ -609,7 +609,7 @@ public class TestStressIndexing2 extends LuceneTestCase { } for (int i=0; i set = mapper.getTermVectorEntrySet(); assertTrue("set is null and it shouldn't be", set != null); //three fields, 4 terms, all terms are the same assertTrue("set Size: " + set.size() + " is not: " + 4, set.size() == 4); //Check offsets and positions - for (Iterator iterator = set.iterator(); iterator.hasNext();) { - TermVectorEntry tve = (TermVectorEntry) iterator.next(); + for (Iterator iterator = set.iterator(); iterator.hasNext();) { + TermVectorEntry tve = iterator.next(); assertTrue("tve is null and it shouldn't be", tve != null); assertTrue("tve.getOffsets() is null and it shouldn't be", tve.getOffsets() != null); assertTrue("tve.getPositions() is null and it shouldn't be", tve.getPositions() != null); @@ -278,8 +277,8 @@ public class TestTermVectorsReader extends LuceneTestCase { //three fields, 4 terms, all terms are the same assertTrue("set Size: " + set.size() + " is not: " + 4, set.size() == 4); //Should have offsets and positions b/c we are munging all the fields together - for (Iterator iterator = set.iterator(); iterator.hasNext();) { - TermVectorEntry tve = (TermVectorEntry) iterator.next(); + for (Iterator iterator = set.iterator(); iterator.hasNext();) { + TermVectorEntry tve = iterator.next(); assertTrue("tve is null and it shouldn't be", tve != null); assertTrue("tve.getOffsets() is null and it shouldn't be", tve.getOffsets() != null); assertTrue("tve.getPositions() is null and it shouldn't be", tve.getPositions() != null); @@ -289,14 +288,12 @@ public class TestTermVectorsReader extends LuceneTestCase { FieldSortedTermVectorMapper fsMapper = new FieldSortedTermVectorMapper(new TermVectorEntryFreqSortedComparator()); reader.get(0, fsMapper); - Map map = fsMapper.getFieldToTerms(); + Map> map = fsMapper.getFieldToTerms(); assertTrue("map Size: " + map.size() + " is not: " + testFields.length, map.size() == testFields.length); - for (Iterator iterator = map.entrySet().iterator(); iterator.hasNext();) { - Map.Entry entry = (Map.Entry) iterator.next(); - SortedSet sortedSet = (SortedSet) entry.getValue(); + for (Map.Entry> entry : map.entrySet()) { + SortedSet sortedSet = entry.getValue(); assertTrue("sortedSet Size: " + sortedSet.size() + " is not: " + 4, sortedSet.size() == 4); - for (Iterator inner = sortedSet.iterator(); inner.hasNext();) { - TermVectorEntry tve = (TermVectorEntry) inner.next(); + for (final TermVectorEntry tve : sortedSet) { assertTrue("tve is null and it shouldn't be", tve != null); //Check offsets and positions. assertTrue("tve is null and it shouldn't be", tve != null); @@ -320,12 +317,10 @@ public class TestTermVectorsReader extends LuceneTestCase { reader.get(0, fsMapper); map = fsMapper.getFieldToTerms(); assertTrue("map Size: " + map.size() + " is not: " + testFields.length, map.size() == testFields.length); - for (Iterator iterator = map.entrySet().iterator(); iterator.hasNext();) { - Map.Entry entry = (Map.Entry) iterator.next(); - SortedSet sortedSet = (SortedSet) entry.getValue(); + for (final Map.Entry> entry : map.entrySet()) { + SortedSet sortedSet = entry.getValue(); assertTrue("sortedSet Size: " + sortedSet.size() + " is not: " + 4, sortedSet.size() == 4); - for (Iterator inner = sortedSet.iterator(); inner.hasNext();) { - TermVectorEntry tve = (TermVectorEntry) inner.next(); + for (final TermVectorEntry tve : sortedSet) { assertTrue("tve is null and it shouldn't be", tve != null); //Check offsets and positions. assertTrue("tve is null and it shouldn't be", tve != null); diff --git a/src/test/org/apache/lucene/index/TestTransactionRollback.java b/src/test/org/apache/lucene/index/TestTransactionRollback.java index 34d51c6a3bd..59e47ba3d75 100644 --- a/src/test/org/apache/lucene/index/TestTransactionRollback.java +++ b/src/test/org/apache/lucene/index/TestTransactionRollback.java @@ -54,12 +54,12 @@ public class TestTransactionRollback extends LuceneTestCase { // System.out.println("Attempting to rollback to "+id); String ids="-"+id; IndexCommit last=null; - Collection commits = IndexReader.listCommits(dir); - for (Iterator iterator = commits.iterator(); iterator.hasNext();) { - IndexCommit commit = (IndexCommit) iterator.next(); - Map ud=commit.getUserData(); + Collection commits = IndexReader.listCommits(dir); + for (Iterator iterator = commits.iterator(); iterator.hasNext();) { + IndexCommit commit = iterator.next(); + Map ud=commit.getUserData(); if (ud.size() > 0) - if (((String) ud.get("index")).endsWith(ids)) + if (ud.get("index").endsWith(ids)) last=commit; } @@ -68,7 +68,7 @@ public class TestTransactionRollback extends LuceneTestCase { IndexWriter w = new IndexWriter(dir, new WhitespaceAnalyzer(), new RollbackDeletionPolicy(id), MaxFieldLength.UNLIMITED, last); - Map data = new HashMap(); + Map data = new HashMap(); data.put("index", "Rolled back to 1-"+id); w.commit(data); w.close(); @@ -135,7 +135,7 @@ public class TestTransactionRollback extends LuceneTestCase { w.addDocument(doc); if (currentRecordId%10 == 0) { - Map data = new HashMap(); + Map data = new HashMap(); data.put("index", "records 1-"+currentRecordId); w.commit(data); } @@ -152,18 +152,17 @@ public class TestTransactionRollback extends LuceneTestCase { this.rollbackPoint = rollbackPoint; } - public void onCommit(List commits) throws IOException { + public void onCommit(List commits) throws IOException { } - public void onInit(List commits) throws IOException { - for (Iterator iterator = commits.iterator(); iterator.hasNext();) { - IndexCommit commit = (IndexCommit) iterator.next(); - Map userData=commit.getUserData(); + public void onInit(List commits) throws IOException { + for (final IndexCommit commit : commits) { + Map userData=commit.getUserData(); if (userData.size() > 0) { // Label for a commit point is "Records 1-30" // This code reads the last id ("30" in this example) and deletes it // if it is after the desired rollback point - String x = (String) userData.get("index"); + String x = userData.get("index"); String lastVal = x.substring(x.lastIndexOf("-")+1); int last = Integer.parseInt(lastVal); if (last>rollbackPoint) { @@ -186,10 +185,10 @@ public class TestTransactionRollback extends LuceneTestCase { class DeleteLastCommitPolicy implements IndexDeletionPolicy { - public void onCommit(List commits) throws IOException {} + public void onCommit(List commits) throws IOException {} - public void onInit(List commits) throws IOException { - ((IndexCommit) commits.get(commits.size()-1)).delete(); + public void onInit(List commits) throws IOException { + commits.get(commits.size()-1).delete(); } } @@ -208,7 +207,7 @@ public class TestTransactionRollback extends LuceneTestCase { // Keeps all commit points (used to build index) class KeepAllDeletionPolicy implements IndexDeletionPolicy { - public void onCommit(List commits) throws IOException {} - public void onInit(List commits) throws IOException {} + public void onCommit(List commits) throws IOException {} + public void onInit(List commits) throws IOException {} } } diff --git a/src/test/org/apache/lucene/index/TestTransactions.java b/src/test/org/apache/lucene/index/TestTransactions.java index e4fa6565922..7bbffe24ff7 100644 --- a/src/test/org/apache/lucene/index/TestTransactions.java +++ b/src/test/org/apache/lucene/index/TestTransactions.java @@ -217,6 +217,6 @@ public class TestTransactions extends LuceneTestCase threads[i].join(); for(int i=0;i wordSet1 = WordlistLoader.getWordSet(new StringReader(s)); checkSet(wordSet1); - HashSet wordSet2 = WordlistLoader.getWordSet(new BufferedReader(new StringReader(s))); + HashSet wordSet2 = WordlistLoader.getWordSet(new BufferedReader(new StringReader(s))); checkSet(wordSet2); } public void testComments() throws Exception { String s = "ONE\n two \nthree\n#comment"; - HashSet wordSet1 = WordlistLoader.getWordSet(new StringReader(s), "#"); + HashSet wordSet1 = WordlistLoader.getWordSet(new StringReader(s), "#"); checkSet(wordSet1); assertFalse(wordSet1.contains("#comment")); assertFalse(wordSet1.contains("comment")); } - private void checkSet(HashSet wordset) { + private void checkSet(HashSet wordset) { assertEquals(3, wordset.size()); assertTrue(wordset.contains("ONE")); // case is not modified assertTrue(wordset.contains("two")); // surrounding whitespace is removed diff --git a/src/test/org/apache/lucene/queryParser/TestMultiAnalyzer.java b/src/test/org/apache/lucene/queryParser/TestMultiAnalyzer.java index 8fb64758efa..cfe45893a6d 100644 --- a/src/test/org/apache/lucene/queryParser/TestMultiAnalyzer.java +++ b/src/test/org/apache/lucene/queryParser/TestMultiAnalyzer.java @@ -17,7 +17,6 @@ package org.apache.lucene.queryParser; * limitations under the License. */ -import java.io.IOException; import java.io.Reader; import org.apache.lucene.analysis.Analyzer; diff --git a/src/test/org/apache/lucene/queryParser/TestMultiFieldQueryParser.java b/src/test/org/apache/lucene/queryParser/TestMultiFieldQueryParser.java index eebf3bc0697..adb5ea9df50 100644 --- a/src/test/org/apache/lucene/queryParser/TestMultiFieldQueryParser.java +++ b/src/test/org/apache/lucene/queryParser/TestMultiFieldQueryParser.java @@ -130,7 +130,7 @@ public class TestMultiFieldQueryParser extends LuceneTestCase { } public void testBoostsSimple() throws Exception { - Map boosts = new HashMap(); + Map boosts = new HashMap(); boosts.put("b", Float.valueOf(5)); boosts.put("t", Float.valueOf(10)); String[] fields = {"b", "t"}; @@ -218,7 +218,6 @@ public class TestMultiFieldQueryParser extends LuceneTestCase { String[] fields = {"b", "t"}; //int[] flags = {MultiFieldQueryParser.REQUIRED_FIELD, MultiFieldQueryParser.PROHIBITED_FIELD}; BooleanClause.Occur[] flags = {BooleanClause.Occur.MUST, BooleanClause.Occur.MUST_NOT}; - MultiFieldQueryParser parser = new MultiFieldQueryParser(Version.LUCENE_CURRENT, fields, new StandardAnalyzer(org.apache.lucene.util.Version.LUCENE_CURRENT)); Query q = MultiFieldQueryParser.parse(Version.LUCENE_CURRENT, "one", fields, flags, new StandardAnalyzer(org.apache.lucene.util.Version.LUCENE_CURRENT));//, fields, flags, new StandardAnalyzer()); assertEquals("+b:one -t:one", q.toString()); diff --git a/src/test/org/apache/lucene/queryParser/TestQueryParser.java b/src/test/org/apache/lucene/queryParser/TestQueryParser.java index ea76813b74d..a13a0c6ad25 100644 --- a/src/test/org/apache/lucene/queryParser/TestQueryParser.java +++ b/src/test/org/apache/lucene/queryParser/TestQueryParser.java @@ -72,7 +72,7 @@ import org.apache.lucene.util.Version; public class TestQueryParser extends LocalizedTestCase { public TestQueryParser(String name) { - super(name, new HashSet(Arrays.asList( + super(name, new HashSet(Arrays.asList( "testLegacyDateRange", "testDateRange", "testCJK", "testNumber", "testFarsiRangeCollating", "testLocalDateFormat" @@ -798,7 +798,7 @@ public class TestQueryParser extends LocalizedTestCase { public void testBoost() throws Exception { - Set stopWords = new HashSet(1); + Set stopWords = new HashSet(1); stopWords.add("on"); StandardAnalyzer oneStopAnalyzer = new StandardAnalyzer(org.apache.lucene.util.Version.LUCENE_CURRENT, stopWords); QueryParser qp = new QueryParser(Version.LUCENE_CURRENT, "field", oneStopAnalyzer); diff --git a/src/test/org/apache/lucene/search/CachingWrapperFilterHelper.java b/src/test/org/apache/lucene/search/CachingWrapperFilterHelper.java index b8d14f4b2a6..15607400fab 100644 --- a/src/test/org/apache/lucene/search/CachingWrapperFilterHelper.java +++ b/src/test/org/apache/lucene/search/CachingWrapperFilterHelper.java @@ -18,7 +18,6 @@ package org.apache.lucene.search; */ import java.io.IOException; -import java.util.BitSet; import java.util.WeakHashMap; import junit.framework.TestCase; @@ -46,11 +45,11 @@ public class CachingWrapperFilterHelper extends CachingWrapperFilter { @Override public DocIdSet getDocIdSet(IndexReader reader) throws IOException { if (cache == null) { - cache = new WeakHashMap(); + cache = new WeakHashMap(); } synchronized (cache) { // check cache - DocIdSet cached = (DocIdSet) cache.get(reader); + DocIdSet cached = cache.get(reader); if (shouldHaveCache) { TestCase.assertNotNull("Cache should have data ", cached); } else { diff --git a/src/test/org/apache/lucene/search/CheckHits.java b/src/test/org/apache/lucene/search/CheckHits.java index 6f42684233a..99d1124bbc4 100644 --- a/src/test/org/apache/lucene/search/CheckHits.java +++ b/src/test/org/apache/lucene/search/CheckHits.java @@ -45,7 +45,7 @@ public class CheckHits { throws IOException { String d = q.toString(defaultFieldName); - Set ignore = new TreeSet(); + Set ignore = new TreeSet(); for (int i = 0; i < results.length; i++) { ignore.add(Integer.valueOf(results[i])); } @@ -85,11 +85,11 @@ public class CheckHits { QueryUtils.check(query,searcher); - Set correct = new TreeSet(); + Set correct = new TreeSet(); for (int i = 0; i < results.length; i++) { correct.add(Integer.valueOf(results[i])); } - final Set actual = new TreeSet(); + final Set actual = new TreeSet(); final Collector c = new SetCollector(actual); searcher.search(query, c); @@ -117,8 +117,8 @@ public class CheckHits { } public static class SetCollector extends Collector { - final Set bag; - public SetCollector(Set bag) { + final Set bag; + public SetCollector(Set bag) { this.bag = bag; } private int base = 0; @@ -161,12 +161,12 @@ public class CheckHits { ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs; - Set correct = new TreeSet(); + Set correct = new TreeSet(); for (int i = 0; i < results.length; i++) { correct.add(Integer.valueOf(results[i])); } - Set actual = new TreeSet(); + Set actual = new TreeSet(); for (int i = 0; i < hits.length; i++) { actual.add(Integer.valueOf(hits[i].doc)); } diff --git a/src/test/org/apache/lucene/search/JustCompileSearch.java b/src/test/org/apache/lucene/search/JustCompileSearch.java index 0ecf9d69a19..5055165e35e 100644 --- a/src/test/org/apache/lucene/search/JustCompileSearch.java +++ b/src/test/org/apache/lucene/search/JustCompileSearch.java @@ -419,9 +419,9 @@ final class JustCompileSearch { } } - static final class JustCompileTopDocsCollector extends TopDocsCollector { + static final class JustCompileTopDocsCollector extends TopDocsCollector { - protected JustCompileTopDocsCollector(PriorityQueue pq) { + protected JustCompileTopDocsCollector(PriorityQueue pq) { super(pq); } diff --git a/src/test/org/apache/lucene/search/QueryUtils.java b/src/test/org/apache/lucene/search/QueryUtils.java index 3279ca9329b..d608dd7a1aa 100644 --- a/src/test/org/apache/lucene/search/QueryUtils.java +++ b/src/test/org/apache/lucene/search/QueryUtils.java @@ -352,7 +352,7 @@ public class QueryUtils { List readerList = new ArrayList(); ReaderUtil.gatherSubReaders(readerList, s.getIndexReader()); - IndexReader[] readers = (IndexReader[]) readerList.toArray(new IndexReader[0]); + IndexReader[] readers = readerList.toArray(new IndexReader[0]); for(int i = 0; i < readers.length; i++) { IndexReader reader = readers[i]; Weight w = q.weight(s); @@ -413,7 +413,7 @@ public class QueryUtils { List readerList = new ArrayList(); ReaderUtil.gatherSubReaders(readerList, s.getIndexReader()); - IndexReader[] readers = (IndexReader[]) readerList.toArray(new IndexReader[0]); + IndexReader[] readers = readerList.toArray(new IndexReader[0]); for(int i = 0; i < readers.length; i++) { IndexReader reader = readers[i]; Weight w = q.weight(s); diff --git a/src/test/org/apache/lucene/search/TestComplexExplanationsOfNonMatches.java b/src/test/org/apache/lucene/search/TestComplexExplanationsOfNonMatches.java index e35d2934b85..2ca3324958a 100644 --- a/src/test/org/apache/lucene/search/TestComplexExplanationsOfNonMatches.java +++ b/src/test/org/apache/lucene/search/TestComplexExplanationsOfNonMatches.java @@ -18,24 +18,6 @@ package org.apache.lucene.search; */ -import org.apache.lucene.store.RAMDirectory; - -import org.apache.lucene.index.IndexWriter; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.Term; - -import org.apache.lucene.analysis.WhitespaceAnalyzer; - -import org.apache.lucene.document.Document; -import org.apache.lucene.document.Field; - -import org.apache.lucene.queryParser.QueryParser; -import org.apache.lucene.queryParser.ParseException; - -import junit.framework.TestCase; - -import java.util.Random; -import java.util.BitSet; /** * subclass of TestSimpleExplanations that verifies non matches. diff --git a/src/test/org/apache/lucene/search/TestCustomSearcherSort.java b/src/test/org/apache/lucene/search/TestCustomSearcherSort.java index 7a6a4bd740c..792d0c88425 100644 --- a/src/test/org/apache/lucene/search/TestCustomSearcherSort.java +++ b/src/test/org/apache/lucene/search/TestCustomSearcherSort.java @@ -153,7 +153,7 @@ implements Serializable { // make a query without sorting first ScoreDoc[] hitsByRank = searcher.search(query, null, 1000).scoreDocs; checkHits(hitsByRank, "Sort by rank: "); // check for duplicates - Map resultMap = new TreeMap(); + Map resultMap = new TreeMap(); // store hits in TreeMap - TreeMap does not allow duplicates; existing entries are silently overwritten for(int hitid=0;hitid idMap = new TreeMap(); for(int docnum=0;docnum fcrf; result = search.search(q,fcrf = FieldCacheRangeFilter.newStringRange("id",minIP,maxIP,T,T), numDocs).scoreDocs; assertTrue(fcrf.getDocIdSet(reader.getSequentialSubReaders()[0]).isCacheable()); assertEquals("find all", numDocs, result.length); @@ -213,7 +212,7 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter { Query q = new TermQuery(new Term("body","body")); // test id, bounded on both ends - FieldCacheRangeFilter fcrf; + FieldCacheRangeFilter fcrf; result = search.search(q,fcrf=FieldCacheRangeFilter.newShortRange("id",minIdO,maxIdO,T,T), numDocs).scoreDocs; assertTrue(fcrf.getDocIdSet(reader.getSequentialSubReaders()[0]).isCacheable()); assertEquals("find all", numDocs, result.length); @@ -305,7 +304,7 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter { // test id, bounded on both ends - FieldCacheRangeFilter fcrf; + FieldCacheRangeFilter fcrf; result = search.search(q,fcrf=FieldCacheRangeFilter.newIntRange("id",minIdO,maxIdO,T,T), numDocs).scoreDocs; assertTrue(fcrf.getDocIdSet(reader.getSequentialSubReaders()[0]).isCacheable()); assertEquals("find all", numDocs, result.length); @@ -397,7 +396,7 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter { // test id, bounded on both ends - FieldCacheRangeFilter fcrf; + FieldCacheRangeFilter fcrf; result = search.search(q,fcrf=FieldCacheRangeFilter.newLongRange("id",minIdO,maxIdO,T,T), numDocs).scoreDocs; assertTrue(fcrf.getDocIdSet(reader.getSequentialSubReaders()[0]).isCacheable()); assertEquals("find all", numDocs, result.length); @@ -550,7 +549,7 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter { assertTrue(reader.hasDeletions()); ScoreDoc[] result; - FieldCacheRangeFilter fcrf; + FieldCacheRangeFilter fcrf; Query q = new TermQuery(new Term("body","body")); result = search.search(q,fcrf=FieldCacheRangeFilter.newByteRange("id",Byte.valueOf((byte) -20),Byte.valueOf((byte) 20),T,T), 100).scoreDocs; diff --git a/src/test/org/apache/lucene/search/TestFieldCacheTermsFilter.java b/src/test/org/apache/lucene/search/TestFieldCacheTermsFilter.java index b550bb0a8b5..48886101c7a 100644 --- a/src/test/org/apache/lucene/search/TestFieldCacheTermsFilter.java +++ b/src/test/org/apache/lucene/search/TestFieldCacheTermsFilter.java @@ -54,20 +54,20 @@ public class TestFieldCacheTermsFilter extends LuceneTestCase { ScoreDoc[] results; MatchAllDocsQuery q = new MatchAllDocsQuery(); - List terms = new ArrayList(); + List terms = new ArrayList(); terms.add("5"); - results = searcher.search(q, new FieldCacheTermsFilter(fieldName, (String[]) terms.toArray(new String[0])), numDocs).scoreDocs; + results = searcher.search(q, new FieldCacheTermsFilter(fieldName, terms.toArray(new String[0])), numDocs).scoreDocs; assertEquals("Must match nothing", 0, results.length); - terms = new ArrayList(); + terms = new ArrayList(); terms.add("10"); - results = searcher.search(q, new FieldCacheTermsFilter(fieldName, (String[]) terms.toArray(new String[0])), numDocs).scoreDocs; + results = searcher.search(q, new FieldCacheTermsFilter(fieldName, terms.toArray(new String[0])), numDocs).scoreDocs; assertEquals("Must match 1", 1, results.length); - terms = new ArrayList(); + terms = new ArrayList(); terms.add("10"); terms.add("20"); - results = searcher.search(q, new FieldCacheTermsFilter(fieldName, (String[]) terms.toArray(new String[0])), numDocs).scoreDocs; + results = searcher.search(q, new FieldCacheTermsFilter(fieldName, terms.toArray(new String[0])), numDocs).scoreDocs; assertEquals("Must match 2", 2, results.length); reader.close(); diff --git a/src/test/org/apache/lucene/search/TestMultiPhraseQuery.java b/src/test/org/apache/lucene/search/TestMultiPhraseQuery.java index c92af488b25..0c000375b37 100644 --- a/src/test/org/apache/lucene/search/TestMultiPhraseQuery.java +++ b/src/test/org/apache/lucene/search/TestMultiPhraseQuery.java @@ -17,7 +17,6 @@ package org.apache.lucene.search; * limitations under the License. */ -import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.index.Term; import org.apache.lucene.index.TermEnum; import org.apache.lucene.index.IndexReader; @@ -31,7 +30,6 @@ import org.apache.lucene.document.Field; import org.apache.lucene.util.LuceneTestCase; import java.io.IOException; -import java.util.HashSet; import java.util.LinkedList; import java.util.Collections; @@ -68,7 +66,7 @@ public class TestMultiPhraseQuery extends LuceneTestCase query1.add(new Term("body", "blueberry")); query2.add(new Term("body", "strawberry")); - LinkedList termsWithPrefix = new LinkedList(); + LinkedList termsWithPrefix = new LinkedList(); IndexReader ir = IndexReader.open(indexStore, true); // this TermEnum gives "piccadilly", "pie" and "pizza". @@ -81,9 +79,9 @@ public class TestMultiPhraseQuery extends LuceneTestCase } } while (te.next()); - query1.add((Term[])termsWithPrefix.toArray(new Term[0])); + query1.add(termsWithPrefix.toArray(new Term[0])); assertEquals("body:\"blueberry (piccadilly pie pizza)\"", query1.toString()); - query2.add((Term[])termsWithPrefix.toArray(new Term[0])); + query2.add(termsWithPrefix.toArray(new Term[0])); assertEquals("body:\"strawberry (piccadilly pie pizza)\"", query2.toString()); ScoreDoc[] result; @@ -103,7 +101,7 @@ public class TestMultiPhraseQuery extends LuceneTestCase termsWithPrefix.add(te.term()); } } while (te.next()); - query3.add((Term[])termsWithPrefix.toArray(new Term[0])); + query3.add(termsWithPrefix.toArray(new Term[0])); query3.add(new Term("body", "pizza")); result = searcher.search(query3, null, 1000).scoreDocs; diff --git a/src/test/org/apache/lucene/search/TestMultiSearcher.java b/src/test/org/apache/lucene/search/TestMultiSearcher.java index cac6ba444a5..687fc62108b 100644 --- a/src/test/org/apache/lucene/search/TestMultiSearcher.java +++ b/src/test/org/apache/lucene/search/TestMultiSearcher.java @@ -253,9 +253,9 @@ public class TestMultiSearcher extends LuceneTestCase assertTrue("document.getFields() Size: " + document.getFields().size() + " is not: " + 2, document.getFields().size() == 2); //Should be one document from each directory //they both have two fields, contents and other - Set ftl = new HashSet(); + Set ftl = new HashSet(); ftl.add("other"); - SetBasedFieldSelector fs = new SetBasedFieldSelector(ftl, Collections.EMPTY_SET); + SetBasedFieldSelector fs = new SetBasedFieldSelector(ftl, Collections. emptySet()); document = searcher.doc(hits[0].doc, fs); assertTrue("document is null and it shouldn't be", document != null); assertTrue("document.getFields() Size: " + document.getFields().size() + " is not: " + 1, document.getFields().size() == 1); @@ -265,7 +265,7 @@ public class TestMultiSearcher extends LuceneTestCase assertTrue("value is null and it shouldn't be", value != null); ftl.clear(); ftl.add("contents"); - fs = new SetBasedFieldSelector(ftl, Collections.EMPTY_SET); + fs = new SetBasedFieldSelector(ftl, Collections. emptySet()); document = searcher.doc(hits[1].doc, fs); value = document.get("contents"); assertTrue("value is null and it shouldn't be", value != null); diff --git a/src/test/org/apache/lucene/search/TestPhrasePrefixQuery.java b/src/test/org/apache/lucene/search/TestPhrasePrefixQuery.java index e92a6cc1a78..924ba857f5a 100644 --- a/src/test/org/apache/lucene/search/TestPhrasePrefixQuery.java +++ b/src/test/org/apache/lucene/search/TestPhrasePrefixQuery.java @@ -76,7 +76,7 @@ public class TestPhrasePrefixQuery query1.add(new Term("body", "blueberry")); query2.add(new Term("body", "strawberry")); - LinkedList termsWithPrefix = new LinkedList(); + LinkedList termsWithPrefix = new LinkedList(); IndexReader ir = IndexReader.open(indexStore, true); // this TermEnum gives "piccadilly", "pie" and "pizza". @@ -89,8 +89,8 @@ public class TestPhrasePrefixQuery } } while (te.next()); - query1.add((Term[])termsWithPrefix.toArray(new Term[0])); - query2.add((Term[])termsWithPrefix.toArray(new Term[0])); + query1.add(termsWithPrefix.toArray(new Term[0])); + query2.add(termsWithPrefix.toArray(new Term[0])); ScoreDoc[] result; result = searcher.search(query1, null, 1000).scoreDocs; diff --git a/src/test/org/apache/lucene/search/TestPositiveScoresOnlyCollector.java b/src/test/org/apache/lucene/search/TestPositiveScoresOnlyCollector.java index 0aa2fe32559..28045441dce 100644 --- a/src/test/org/apache/lucene/search/TestPositiveScoresOnlyCollector.java +++ b/src/test/org/apache/lucene/search/TestPositiveScoresOnlyCollector.java @@ -66,7 +66,7 @@ public class TestPositiveScoresOnlyCollector extends LuceneTestCase { } Scorer s = new SimpleScorer(); - TopDocsCollector tdc = TopScoreDocCollector.create(scores.length, true); + TopDocsCollector tdc = TopScoreDocCollector.create(scores.length, true); Collector c = new PositiveScoresOnlyCollector(tdc); c.setScorer(s); while (s.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { diff --git a/src/test/org/apache/lucene/search/TestSimpleExplanationsOfNonMatches.java b/src/test/org/apache/lucene/search/TestSimpleExplanationsOfNonMatches.java index 6ead574dd31..ca1e3a73faf 100644 --- a/src/test/org/apache/lucene/search/TestSimpleExplanationsOfNonMatches.java +++ b/src/test/org/apache/lucene/search/TestSimpleExplanationsOfNonMatches.java @@ -18,24 +18,7 @@ package org.apache.lucene.search; */ -import org.apache.lucene.store.RAMDirectory; -import org.apache.lucene.index.IndexWriter; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.Term; - -import org.apache.lucene.analysis.WhitespaceAnalyzer; - -import org.apache.lucene.document.Document; -import org.apache.lucene.document.Field; - -import org.apache.lucene.queryParser.QueryParser; -import org.apache.lucene.queryParser.ParseException; - -import junit.framework.TestCase; - -import java.util.Random; -import java.util.BitSet; /** * subclass of TestSimpleExplanations that verifies non matches. diff --git a/src/test/org/apache/lucene/search/TestSort.java b/src/test/org/apache/lucene/search/TestSort.java index dc318926eb0..217608210c8 100644 --- a/src/test/org/apache/lucene/search/TestSort.java +++ b/src/test/org/apache/lucene/search/TestSort.java @@ -634,9 +634,9 @@ public class TestSort extends LuceneTestCase implements Serializable { public void testNormalizedScores() throws Exception { // capture relevancy scores - HashMap scoresX = getScores (full.search (queryX, null, 1000).scoreDocs, full); - HashMap scoresY = getScores (full.search (queryY, null, 1000).scoreDocs, full); - HashMap scoresA = getScores (full.search (queryA, null, 1000).scoreDocs, full); + HashMap scoresX = getScores (full.search (queryX, null, 1000).scoreDocs, full); + HashMap scoresY = getScores (full.search (queryY, null, 1000).scoreDocs, full); + HashMap scoresA = getScores (full.search (queryA, null, 1000).scoreDocs, full); // we'll test searching locally, remote and multi @@ -977,9 +977,9 @@ public class TestSort extends LuceneTestCase implements Serializable { assertEquals (expectedResult, buff.toString()); } - private HashMap getScores (ScoreDoc[] hits, Searcher searcher) + private HashMap getScores (ScoreDoc[] hits, Searcher searcher) throws IOException { - HashMap scoreMap = new HashMap(); + HashMap scoreMap = new HashMap(); int n = hits.length; for (int i=0; i void assertSameValues (HashMap m1, HashMap m2) { int n = m1.size(); int m = m2.size(); assertEquals (n, m); - Iterator iter = m1.keySet().iterator(); + Iterator iter = m1.keySet().iterator(); while (iter.hasNext()) { - Object key = iter.next(); - Object o1 = m1.get(key); - Object o2 = m2.get(key); + K key = iter.next(); + V o1 = m1.get(key); + V o2 = m2.get(key); if (o1 instanceof Float) { assertEquals(((Float)o1).floatValue(), ((Float)o2).floatValue(), 1e-6); } else { diff --git a/src/test/org/apache/lucene/search/TestSpanQueryFilter.java b/src/test/org/apache/lucene/search/TestSpanQueryFilter.java index 4b2104e70ee..14ddc80fd50 100644 --- a/src/test/org/apache/lucene/search/TestSpanQueryFilter.java +++ b/src/test/org/apache/lucene/search/TestSpanQueryFilter.java @@ -16,7 +16,6 @@ package org.apache.lucene.search; * limitations under the License. */ -import java.util.Iterator; import java.util.List; import org.apache.lucene.analysis.SimpleAnalyzer; @@ -58,12 +57,11 @@ public class TestSpanQueryFilter extends LuceneTestCase { DocIdSet docIdSet = result.getDocIdSet(); assertTrue("docIdSet is null and it shouldn't be", docIdSet != null); assertContainsDocId("docIdSet doesn't contain docId 10", docIdSet, 10); - List spans = result.getPositions(); + List spans = result.getPositions(); assertTrue("spans is null and it shouldn't be", spans != null); int size = getDocIdSetSize(docIdSet); assertTrue("spans Size: " + spans.size() + " is not: " + size, spans.size() == size); - for (Iterator iterator = spans.iterator(); iterator.hasNext();) { - SpanFilterResult.PositionInfo info = (SpanFilterResult.PositionInfo) iterator.next(); + for (final SpanFilterResult.PositionInfo info: spans) { assertTrue("info is null and it shouldn't be", info != null); //The doc should indicate the bit is on assertContainsDocId("docIdSet doesn't contain docId " + info.getDoc(), docIdSet, info.getDoc()); diff --git a/src/test/org/apache/lucene/search/TestTermRangeQuery.java b/src/test/org/apache/lucene/search/TestTermRangeQuery.java index f36d02664f5..25205374326 100644 --- a/src/test/org/apache/lucene/search/TestTermRangeQuery.java +++ b/src/test/org/apache/lucene/search/TestTermRangeQuery.java @@ -21,7 +21,6 @@ import org.apache.lucene.analysis.WhitespaceAnalyzer; import org.apache.lucene.document.Document; import org.apache.lucene.document.Field; import org.apache.lucene.index.IndexWriter; -import org.apache.lucene.index.Term; import org.apache.lucene.store.RAMDirectory; import org.apache.lucene.analysis.Analyzer; import org.apache.lucene.analysis.TokenStream; diff --git a/src/test/org/apache/lucene/search/TestTermScorer.java b/src/test/org/apache/lucene/search/TestTermScorer.java index 1dae462694f..6cfd105d95b 100644 --- a/src/test/org/apache/lucene/search/TestTermScorer.java +++ b/src/test/org/apache/lucene/search/TestTermScorer.java @@ -76,7 +76,7 @@ public class TestTermScorer extends LuceneTestCase indexReader.termDocs(allTerm), indexSearcher.getSimilarity(), indexReader.norms(FIELD)); //we have 2 documents with the term all in them, one document for all the other values - final List docs = new ArrayList(); + final List docs = new ArrayList(); //must call next first @@ -107,8 +107,8 @@ public class TestTermScorer extends LuceneTestCase } }); assertTrue("docs Size: " + docs.size() + " is not: " + 2, docs.size() == 2); - TestHit doc0 = (TestHit) docs.get(0); - TestHit doc5 = (TestHit) docs.get(1); + TestHit doc0 = docs.get(0); + TestHit doc5 = docs.get(1); //The scores should be the same assertTrue(doc0.score + " does not equal: " + doc5.score, doc0.score == doc5.score); /* diff --git a/src/test/org/apache/lucene/search/TestTermVectors.java b/src/test/org/apache/lucene/search/TestTermVectors.java index 1ffbee4b1dd..9a97ca9d87b 100644 --- a/src/test/org/apache/lucene/search/TestTermVectors.java +++ b/src/test/org/apache/lucene/search/TestTermVectors.java @@ -28,7 +28,6 @@ import org.apache.lucene.util.English; import java.io.IOException; import java.util.HashMap; -import java.util.Iterator; import java.util.Map; import java.util.SortedSet; @@ -167,7 +166,6 @@ public class TestTermVectors extends LuceneTestCase { } else{ try{ - TermPositionVector posVec = (TermPositionVector)vector[0]; assertTrue(false); } catch(ClassCastException ignore){ @@ -208,7 +206,7 @@ public class TestTermVectors extends LuceneTestCase { String test2 = "computer in a computer lab"; //5 terms String test3 = "a chocolate lab grows old"; //5 terms String test4 = "eating chocolate with a chocolate lab in an old chocolate colored computer lab"; //13 terms - Map test4Map = new HashMap(); + Map test4Map = new HashMap(); test4Map.put("chocolate", Integer.valueOf(3)); test4Map.put("lab", Integer.valueOf(2)); test4Map.put("eating", Integer.valueOf(1)); @@ -246,7 +244,7 @@ public class TestTermVectors extends LuceneTestCase { TermDocs termDocs = knownSearcher.reader.termDocs(); //System.out.println("Terms: " + termEnum.size() + " Orig Len: " + termArray.length); - Similarity sim = knownSearcher.getSimilarity(); + //Similarity sim = knownSearcher.getSimilarity(); while (termEnum.next() == true) { Term term = termEnum.term(); @@ -258,11 +256,11 @@ public class TestTermVectors extends LuceneTestCase { int freq = termDocs.freq(); //System.out.println("Doc Id: " + docId + " freq " + freq); TermFreqVector vector = knownSearcher.reader.getTermFreqVector(docId, "field"); - float tf = sim.tf(freq); - float idf = sim.idf(knownSearcher.docFreq(term), knownSearcher.maxDoc()); + //float tf = sim.tf(freq); + //float idf = sim.idf(knownSearcher.docFreq(term), knownSearcher.maxDoc()); //float qNorm = sim.queryNorm() //This is fine since we don't have stop words - float lNorm = sim.lengthNorm("field", vector.getTerms().length); + //float lNorm = sim.lengthNorm("field", vector.getTerms().length); //float coord = sim.coord() //System.out.println("TF: " + tf + " IDF: " + idf + " LenNorm: " + lNorm); assertTrue(vector != null); @@ -283,7 +281,6 @@ public class TestTermVectors extends LuceneTestCase { ScoreDoc[] hits = knownSearcher.search(query, null, 1000).scoreDocs; //doc 3 should be the first hit b/c it is the shortest match assertTrue(hits.length == 3); - float score = hits[0].score; /*System.out.println("Hit 0: " + hits.id(0) + " Score: " + hits.score(0) + " String: " + hits.doc(0).toString()); System.out.println("Explain: " + knownSearcher.explain(query, hits.id(0))); System.out.println("Hit 1: " + hits.id(1) + " Score: " + hits.score(1) + " String: " + hits.doc(1).toString()); @@ -304,21 +301,20 @@ public class TestTermVectors extends LuceneTestCase { //System.out.println("Term: " + term); int freq = freqs[i]; assertTrue(test4.indexOf(term) != -1); - Integer freqInt = (Integer)test4Map.get(term); + Integer freqInt = test4Map.get(term); assertTrue(freqInt != null); assertTrue(freqInt.intValue() == freq); } SortedTermVectorMapper mapper = new SortedTermVectorMapper(new TermVectorEntryFreqSortedComparator()); knownSearcher.reader.getTermFreqVector(hits[1].doc, mapper); - SortedSet vectorEntrySet = mapper.getTermVectorEntrySet(); + SortedSet vectorEntrySet = mapper.getTermVectorEntrySet(); assertTrue("mapper.getTermVectorEntrySet() Size: " + vectorEntrySet.size() + " is not: " + 10, vectorEntrySet.size() == 10); TermVectorEntry last = null; - for (Iterator iterator = vectorEntrySet.iterator(); iterator.hasNext();) { - TermVectorEntry tve = (TermVectorEntry) iterator.next(); + for (final TermVectorEntry tve : vectorEntrySet) { if (tve != null && last != null) { assertTrue("terms are not properly sorted", last.getFrequency() >= tve.getFrequency()); - Integer expectedFreq = (Integer) test4Map.get(tve.getTerm()); + Integer expectedFreq = test4Map.get(tve.getTerm()); //we expect double the expectedFreq, since there are two fields with the exact same text and we are collapsing all fields assertTrue("Frequency is not correct:", tve.getFrequency() == 2*expectedFreq.intValue()); } @@ -328,9 +324,9 @@ public class TestTermVectors extends LuceneTestCase { FieldSortedTermVectorMapper fieldMapper = new FieldSortedTermVectorMapper(new TermVectorEntryFreqSortedComparator()); knownSearcher.reader.getTermFreqVector(hits[1].doc, fieldMapper); - Map map = fieldMapper.getFieldToTerms(); + Map> map = fieldMapper.getFieldToTerms(); assertTrue("map Size: " + map.size() + " is not: " + 2, map.size() == 2); - vectorEntrySet = (SortedSet) map.get("field"); + vectorEntrySet = map.get("field"); assertTrue("vectorEntrySet is null and it shouldn't be", vectorEntrySet != null); assertTrue("vectorEntrySet Size: " + vectorEntrySet.size() + " is not: " + 10, vectorEntrySet.size() == 10); knownSearcher.close(); diff --git a/src/test/org/apache/lucene/search/TestThreadSafe.java b/src/test/org/apache/lucene/search/TestThreadSafe.java index ef10ee43c17..dbe9ed572d5 100755 --- a/src/test/org/apache/lucene/search/TestThreadSafe.java +++ b/src/test/org/apache/lucene/search/TestThreadSafe.java @@ -85,9 +85,8 @@ public class TestThreadSafe extends LuceneTestCase { } ); - List fields = doc.getFields(); - for (int i=0; i fields = doc.getFields(); + for (final Fieldable f : fields ) { validateField(f); } diff --git a/src/test/org/apache/lucene/search/TestTopDocsCollector.java b/src/test/org/apache/lucene/search/TestTopDocsCollector.java index 063392b8cdb..2983bd570d3 100644 --- a/src/test/org/apache/lucene/search/TestTopDocsCollector.java +++ b/src/test/org/apache/lucene/search/TestTopDocsCollector.java @@ -30,7 +30,7 @@ import org.apache.lucene.util.LuceneTestCase; public class TestTopDocsCollector extends LuceneTestCase { - private static final class MyTopsDocCollector extends TopDocsCollector { + private static final class MyTopsDocCollector extends TopDocsCollector { private int idx = 0; private int base = 0; @@ -50,7 +50,7 @@ public class TestTopDocsCollector extends LuceneTestCase { maxScore = results[0].score; } else { for (int i = pq.size(); i > 1; i--) { pq.pop(); } - maxScore = ((ScoreDoc) pq.pop()).score; + maxScore = pq.pop().score; } return new TopDocs(totalHits, results, maxScore); @@ -94,10 +94,10 @@ public class TestTopDocsCollector extends LuceneTestCase { private Directory dir = new RAMDirectory(); - private TopDocsCollector doSearch(int numResults) throws IOException { + private TopDocsCollector doSearch(int numResults) throws IOException { Query q = new MatchAllDocsQuery(); IndexSearcher searcher = new IndexSearcher(dir, true); - TopDocsCollector tdc = new MyTopsDocCollector(numResults); + TopDocsCollector tdc = new MyTopsDocCollector(numResults); searcher.search(q, tdc); searcher.close(); return tdc; @@ -125,7 +125,7 @@ public class TestTopDocsCollector extends LuceneTestCase { public void testInvalidArguments() throws Exception { int numResults = 5; - TopDocsCollector tdc = doSearch(numResults); + TopDocsCollector tdc = doSearch(numResults); // start < 0 assertEquals(0, tdc.topDocs(-1).scoreDocs.length); @@ -145,17 +145,17 @@ public class TestTopDocsCollector extends LuceneTestCase { } public void testZeroResults() throws Exception { - TopDocsCollector tdc = new MyTopsDocCollector(5); + TopDocsCollector tdc = new MyTopsDocCollector(5); assertEquals(0, tdc.topDocs(0, 1).scoreDocs.length); } public void testFirstResultsPage() throws Exception { - TopDocsCollector tdc = doSearch(15); + TopDocsCollector tdc = doSearch(15); assertEquals(10, tdc.topDocs(0, 10).scoreDocs.length); } public void testSecondResultsPages() throws Exception { - TopDocsCollector tdc = doSearch(15); + TopDocsCollector tdc = doSearch(15); // ask for more results than are available assertEquals(5, tdc.topDocs(10, 10).scoreDocs.length); @@ -169,12 +169,12 @@ public class TestTopDocsCollector extends LuceneTestCase { } public void testGetAllResults() throws Exception { - TopDocsCollector tdc = doSearch(15); + TopDocsCollector tdc = doSearch(15); assertEquals(15, tdc.topDocs().scoreDocs.length); } public void testGetResultsFromStart() throws Exception { - TopDocsCollector tdc = doSearch(15); + TopDocsCollector tdc = doSearch(15); // should bring all results assertEquals(15, tdc.topDocs(0).scoreDocs.length); @@ -185,7 +185,7 @@ public class TestTopDocsCollector extends LuceneTestCase { public void testMaxScore() throws Exception { // ask for all results - TopDocsCollector tdc = doSearch(15); + TopDocsCollector tdc = doSearch(15); TopDocs td = tdc.topDocs(); assertEquals(MAX_SCORE, td.getMaxScore(), 0f); @@ -198,7 +198,7 @@ public class TestTopDocsCollector extends LuceneTestCase { // This does not test the PQ's correctness, but whether topDocs() // implementations return the results in decreasing score order. public void testResultsOrder() throws Exception { - TopDocsCollector tdc = doSearch(15); + TopDocsCollector tdc = doSearch(15); ScoreDoc[] sd = tdc.topDocs().scoreDocs; assertEquals(MAX_SCORE, sd[0].score, 0f); diff --git a/src/test/org/apache/lucene/search/TestTopScoreDocCollector.java b/src/test/org/apache/lucene/search/TestTopScoreDocCollector.java index 11954d4e892..d4dd7e707a1 100644 --- a/src/test/org/apache/lucene/search/TestTopScoreDocCollector.java +++ b/src/test/org/apache/lucene/search/TestTopScoreDocCollector.java @@ -59,7 +59,7 @@ public class TestTopScoreDocCollector extends LuceneTestCase { bq.setMinimumNumberShouldMatch(1); IndexSearcher searcher = new IndexSearcher(dir, true); for (int i = 0; i < inOrder.length; i++) { - TopDocsCollector tdc = TopScoreDocCollector.create(3, inOrder[i]); + TopDocsCollector tdc = TopScoreDocCollector.create(3, inOrder[i]); assertEquals("org.apache.lucene.search.TopScoreDocCollector$" + actualTSDCClass[i], tdc.getClass().getName()); searcher.search(new MatchAllDocsQuery(), tdc); diff --git a/src/test/org/apache/lucene/search/function/TestCustomScoreQuery.java b/src/test/org/apache/lucene/search/function/TestCustomScoreQuery.java index fe75b657777..7f32b11e99d 100755 --- a/src/test/org/apache/lucene/search/function/TestCustomScoreQuery.java +++ b/src/test/org/apache/lucene/search/function/TestCustomScoreQuery.java @@ -19,7 +19,6 @@ package org.apache.lucene.search.function; import java.io.IOException; import java.util.HashMap; -import java.util.Iterator; import org.apache.lucene.index.CorruptIndexException; import org.apache.lucene.queryParser.QueryParser; @@ -184,11 +183,11 @@ public class TestCustomScoreQuery extends FunctionTestSetup { TopDocs td5CustomMulAdd = s.search(q5CustomMulAdd,null,1000); // put results in map so we can verify the scores although they have changed - HashMap h1 = topDocsToMap(td1); - HashMap h2CustomNeutral = topDocsToMap(td2CustomNeutral); - HashMap h3CustomMul = topDocsToMap(td3CustomMul); - HashMap h4CustomAdd = topDocsToMap(td4CustomAdd); - HashMap h5CustomMulAdd = topDocsToMap(td5CustomMulAdd); + HashMap h1 = topDocsToMap(td1); + HashMap h2CustomNeutral = topDocsToMap(td2CustomNeutral); + HashMap h3CustomMul = topDocsToMap(td3CustomMul); + HashMap h4CustomAdd = topDocsToMap(td4CustomAdd); + HashMap h5CustomMulAdd = topDocsToMap(td5CustomMulAdd); verifyResults(boost, s, h1, h2CustomNeutral, h3CustomMul, h4CustomAdd, h5CustomMulAdd, @@ -197,7 +196,7 @@ public class TestCustomScoreQuery extends FunctionTestSetup { // verify results are as expected. private void verifyResults(float boost, IndexSearcher s, - HashMap h1, HashMap h2customNeutral, HashMap h3CustomMul, HashMap h4CustomAdd, HashMap h5CustomMulAdd, + HashMap h1, HashMap h2customNeutral, HashMap h3CustomMul, HashMap h4CustomAdd, HashMap h5CustomMulAdd, Query q1, Query q2, Query q3, Query q4, Query q5) throws Exception { // verify numbers of matches @@ -214,8 +213,7 @@ public class TestCustomScoreQuery extends FunctionTestSetup { QueryUtils.check(q5,s); // verify scores ratios - for (Iterator it = h1.keySet().iterator(); it.hasNext();) { - Integer x = (Integer) it.next(); + for (final Integer x : h1.keySet()) { int doc = x.intValue(); log("doc = "+doc); @@ -224,22 +222,22 @@ public class TestCustomScoreQuery extends FunctionTestSetup { log("fieldScore = "+fieldScore); assertTrue("fieldScore should not be 0",fieldScore>0); - float score1 = ((Float)h1.get(x)).floatValue(); + float score1 = h1.get(x).floatValue(); logResult("score1=", s, q1, doc, score1); - float score2 = ((Float)h2customNeutral.get(x)).floatValue(); + float score2 = h2customNeutral.get(x).floatValue(); logResult("score2=", s, q2, doc, score2); assertEquals("same score (just boosted) for neutral", boost * score1, score2, TEST_SCORE_TOLERANCE_DELTA); - float score3 = ((Float)h3CustomMul.get(x)).floatValue(); + float score3 = h3CustomMul.get(x).floatValue(); logResult("score3=", s, q3, doc, score3); assertEquals("new score for custom mul", boost * fieldScore * score1, score3, TEST_SCORE_TOLERANCE_DELTA); - float score4 = ((Float)h4CustomAdd.get(x)).floatValue(); + float score4 = h4CustomAdd.get(x).floatValue(); logResult("score4=", s, q4, doc, score4); assertEquals("new score for custom add", boost * (fieldScore + score1), score4, TEST_SCORE_TOLERANCE_DELTA); - float score5 = ((Float)h5CustomMulAdd.get(x)).floatValue(); + float score5 = h5CustomMulAdd.get(x).floatValue(); logResult("score5=", s, q5, doc, score5); assertEquals("new score for custom mul add", boost * fieldScore * (score1 + fieldScore), score5, TEST_SCORE_TOLERANCE_DELTA); } @@ -253,8 +251,8 @@ public class TestCustomScoreQuery extends FunctionTestSetup { // since custom scoring modifies the order of docs, map results // by doc ids so that we can later compare/verify them - private HashMap topDocsToMap(TopDocs td) { - HashMap h = new HashMap(); + private HashMap topDocsToMap(TopDocs td) { + HashMap h = new HashMap(); for (int i=0; i expectedArrayTypes = new HashMap(); expectedArrayTypes.put(FieldScoreQuery.Type.BYTE, new byte[0]); expectedArrayTypes.put(FieldScoreQuery.Type.SHORT, new short[0]); expectedArrayTypes.put(FieldScoreQuery.Type.INT, new int[0]); diff --git a/src/test/org/apache/lucene/search/payloads/TestPayloadNearQuery.java b/src/test/org/apache/lucene/search/payloads/TestPayloadNearQuery.java index 11986c2aacb..1a9b38c3346 100644 --- a/src/test/org/apache/lucene/search/payloads/TestPayloadNearQuery.java +++ b/src/test/org/apache/lucene/search/payloads/TestPayloadNearQuery.java @@ -21,7 +21,6 @@ import java.util.Collection; import org.apache.lucene.analysis.Analyzer; import org.apache.lucene.analysis.LowerCaseTokenizer; -import org.apache.lucene.analysis.Token; import org.apache.lucene.analysis.TokenFilter; import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.tokenattributes.PayloadAttribute; @@ -91,7 +90,6 @@ public class TestPayloadNearQuery extends LuceneTestCase { } private PayloadNearQuery newPhraseQuery (String fieldName, String phrase, boolean inOrder) { - int n; String[] words = phrase.split("[\\s]+"); SpanQuery clauses[] = new SpanQuery[words.length]; for (int i=0;i getPayload() throws IOException { throw new UnsupportedOperationException(UNSUPPORTED_MSG); } @@ -96,7 +96,7 @@ final class JustCompileSearchSpans { static final class JustCompilePayloadSpans extends Spans { @Override - public Collection getPayload() throws IOException { + public Collection getPayload() throws IOException { throw new UnsupportedOperationException(UNSUPPORTED_MSG); } diff --git a/src/test/org/apache/lucene/search/spans/TestFieldMaskingSpanQuery.java b/src/test/org/apache/lucene/search/spans/TestFieldMaskingSpanQuery.java index 6dc100843f5..4e71d5dfa78 100644 --- a/src/test/org/apache/lucene/search/spans/TestFieldMaskingSpanQuery.java +++ b/src/test/org/apache/lucene/search/spans/TestFieldMaskingSpanQuery.java @@ -168,7 +168,7 @@ public class TestFieldMaskingSpanQuery extends LuceneTestCase { QueryUtils.checkEqual(q, qr); - HashSet set = new HashSet(); + HashSet set = new HashSet(); qr.extractTerms(set); assertEquals(2, set.size()); } diff --git a/src/test/org/apache/lucene/search/spans/TestPayloadSpans.java b/src/test/org/apache/lucene/search/spans/TestPayloadSpans.java index 468f69bbf7f..d896c307ccc 100644 --- a/src/test/org/apache/lucene/search/spans/TestPayloadSpans.java +++ b/src/test/org/apache/lucene/search/spans/TestPayloadSpans.java @@ -21,7 +21,6 @@ import java.io.Reader; import java.io.StringReader; import java.util.Collection; import java.util.HashSet; -import java.util.Iterator; import java.util.Set; import org.apache.lucene.analysis.Analyzer; @@ -271,13 +270,13 @@ public class TestPayloadSpans extends LuceneTestCase { Spans spans = snq.getSpans(is.getIndexReader()); TopDocs topDocs = is.search(snq, 1); - Set payloadSet = new HashSet(); + Set payloadSet = new HashSet(); for (int i = 0; i < topDocs.scoreDocs.length; i++) { while (spans.next()) { - Collection payloads = spans.getPayload(); + Collection payloads = spans.getPayload(); - for (Iterator it = payloads.iterator(); it.hasNext();) { - payloadSet.add(new String((byte[]) it.next())); + for (final byte [] payload : payloads) { + payloadSet.add(new String(payload)); } } } @@ -305,12 +304,12 @@ public class TestPayloadSpans extends LuceneTestCase { Spans spans = snq.getSpans(is.getIndexReader()); TopDocs topDocs = is.search(snq, 1); - Set payloadSet = new HashSet(); + Set payloadSet = new HashSet(); for (int i = 0; i < topDocs.scoreDocs.length; i++) { while (spans.next()) { - Collection payloads = spans.getPayload(); - for (Iterator it = payloads.iterator(); it.hasNext();) { - payloadSet.add(new String((byte[]) it.next())); + Collection payloads = spans.getPayload(); + for (final byte[] payload : payloads) { + payloadSet.add(new String(payload)); } } } @@ -338,22 +337,21 @@ public class TestPayloadSpans extends LuceneTestCase { Spans spans = snq.getSpans(is.getIndexReader()); TopDocs topDocs = is.search(snq, 1); - Set payloadSet = new HashSet(); + Set payloadSet = new HashSet(); for (int i = 0; i < topDocs.scoreDocs.length; i++) { while (spans.next()) { - Collection payloads = spans.getPayload(); + Collection payloads = spans.getPayload(); - for (Iterator it = payloads.iterator(); it.hasNext();) { - payloadSet.add(new String((byte[]) it.next())); + for (final byte [] payload : payloads) { + payloadSet.add(new String(payload)); } } } assertEquals(2, payloadSet.size()); if(DEBUG) { - Iterator pit = payloadSet.iterator(); - while (pit.hasNext()) { - System.out.println("match:" + pit.next()); - } + for (final String payload : payloadSet) + System.out.println("match:" + payload); + } assertTrue(payloadSet.contains("a:Noise:10")); assertTrue(payloadSet.contains("k:Noise:11")); @@ -375,12 +373,10 @@ public class TestPayloadSpans extends LuceneTestCase { IndexReader reader = searcher.getIndexReader(); PayloadSpanUtil psu = new PayloadSpanUtil(reader); - Collection payloads = psu.getPayloadsForQuery(new TermQuery(new Term(PayloadHelper.FIELD, "rr"))); + Collection payloads = psu.getPayloadsForQuery(new TermQuery(new Term(PayloadHelper.FIELD, "rr"))); if(DEBUG) System.out.println("Num payloads:" + payloads.size()); - Iterator it = payloads.iterator(); - while(it.hasNext()) { - byte[] bytes = (byte[]) it.next(); + for (final byte [] bytes : payloads) { if(DEBUG) System.out.println(new String(bytes)); } @@ -405,10 +401,9 @@ public class TestPayloadSpans extends LuceneTestCase { } //See payload helper, for the PayloadHelper.FIELD field, there is a single byte payload at every token if (spans.isPayloadAvailable()) { - Collection payload = spans.getPayload(); + Collection payload = spans.getPayload(); assertTrue("payload Size: " + payload.size() + " is not: " + expectedNumPayloads, payload.size() == expectedNumPayloads); - for (Iterator iterator = payload.iterator(); iterator.hasNext();) { - byte[] thePayload = (byte[]) iterator.next(); + for (final byte [] thePayload : payload) { assertTrue("payload[0] Size: " + thePayload.length + " is not: " + expectedPayloadLength, thePayload.length == expectedPayloadLength); assertTrue(thePayload[0] + " does not equal: " + expectedFirstByte, thePayload[0] == expectedFirstByte); @@ -450,12 +445,10 @@ public class TestPayloadSpans extends LuceneTestCase { if(DEBUG) System.out.println("\nSpans Dump --"); if (spans.isPayloadAvailable()) { - Collection payload = spans.getPayload(); + Collection payload = spans.getPayload(); if(DEBUG) System.out.println("payloads for span:" + payload.size()); - Iterator it = payload.iterator(); - while(it.hasNext()) { - byte[] bytes = (byte[]) it.next(); + for (final byte [] bytes : payload) { if(DEBUG) System.out.println("doc:" + spans.doc() + " s:" + spans.start() + " e:" + spans.end() + " " + new String(bytes)); @@ -484,8 +477,8 @@ public class TestPayloadSpans extends LuceneTestCase { class PayloadFilter extends TokenFilter { String fieldName; int numSeen = 0; - Set entities = new HashSet(); - Set nopayload = new HashSet(); + Set entities = new HashSet(); + Set nopayload = new HashSet(); int pos; PayloadAttribute payloadAtt; TermAttribute termAtt; diff --git a/src/test/org/apache/lucene/search/spans/TestSpanExplanations.java b/src/test/org/apache/lucene/search/spans/TestSpanExplanations.java index 4ab5d95ebc9..c5da9d5533f 100644 --- a/src/test/org/apache/lucene/search/spans/TestSpanExplanations.java +++ b/src/test/org/apache/lucene/search/spans/TestSpanExplanations.java @@ -18,24 +18,7 @@ package org.apache.lucene.search.spans; */ import org.apache.lucene.search.*; -import org.apache.lucene.store.RAMDirectory; -import org.apache.lucene.index.IndexWriter; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.Term; - -import org.apache.lucene.analysis.WhitespaceAnalyzer; - -import org.apache.lucene.document.Document; -import org.apache.lucene.document.Field; - -import org.apache.lucene.queryParser.QueryParser; -import org.apache.lucene.queryParser.ParseException; - -import junit.framework.TestCase; - -import java.util.Random; -import java.util.BitSet; /** * TestExplanations subclass focusing on span queries diff --git a/src/test/org/apache/lucene/store/MockRAMDirectory.java b/src/test/org/apache/lucene/store/MockRAMDirectory.java index 65ba8686d71..a15a94598a1 100644 --- a/src/test/org/apache/lucene/store/MockRAMDirectory.java +++ b/src/test/org/apache/lucene/store/MockRAMDirectory.java @@ -56,9 +56,9 @@ public class MockRAMDirectory extends RAMDirectory { if (openFiles == null) openFiles = new HashMap(); if (createdFiles == null) - createdFiles = new HashSet(); + createdFiles = new HashSet(); if (unSyncedFiles == null) - unSyncedFiles = new HashSet(); + unSyncedFiles = new HashSet(); } public MockRAMDirectory() { @@ -89,9 +89,9 @@ public class MockRAMDirectory extends RAMDirectory { * unsynced files. */ public synchronized void crash() throws IOException { crashed = true; - openFiles = new HashMap(); + openFiles = new HashMap(); Iterator it = unSyncedFiles.iterator(); - unSyncedFiles = new HashSet(); + unSyncedFiles = new HashSet(); int count = 0; while(it.hasNext()) { String name = it.next(); @@ -264,7 +264,7 @@ public class MockRAMDirectory extends RAMDirectory { @Override public synchronized void close() { if (openFiles == null) { - openFiles = new HashMap(); + openFiles = new HashMap(); } if (noDeleteOpenFile && openFiles.size() > 0) { // RuntimeException instead of IOException because @@ -311,7 +311,7 @@ public class MockRAMDirectory extends RAMDirectory { } } - ArrayList failures; + ArrayList failures; /** * add a Failure object to the list of objects to be evaluated @@ -319,7 +319,7 @@ public class MockRAMDirectory extends RAMDirectory { */ synchronized public void failOn(Failure fail) { if (failures == null) { - failures = new ArrayList(); + failures = new ArrayList(); } failures.add(fail); } diff --git a/src/test/org/apache/lucene/store/MockRAMInputStream.java b/src/test/org/apache/lucene/store/MockRAMInputStream.java index 56b5bc880c9..b7606c94f15 100644 --- a/src/test/org/apache/lucene/store/MockRAMInputStream.java +++ b/src/test/org/apache/lucene/store/MockRAMInputStream.java @@ -45,7 +45,7 @@ public class MockRAMInputStream extends RAMInputStream { // all clones get closed: if (!isClone) { synchronized(dir) { - Integer v = (Integer) dir.openFiles.get(name); + Integer v = dir.openFiles.get(name); // Could be null when MockRAMDirectory.crash() was called if (v != null) { if (v.intValue() == 1) { diff --git a/src/test/org/apache/lucene/store/TestBufferedIndexInput.java b/src/test/org/apache/lucene/store/TestBufferedIndexInput.java index def32daddb6..5f4f97a8bc0 100755 --- a/src/test/org/apache/lucene/store/TestBufferedIndexInput.java +++ b/src/test/org/apache/lucene/store/TestBufferedIndexInput.java @@ -22,7 +22,6 @@ import java.io.FileOutputStream; import java.io.IOException; import java.io.OutputStream; import java.util.ArrayList; -import java.util.Iterator; import java.util.List; import java.util.Random; @@ -287,7 +286,7 @@ public class TestBufferedIndexInput extends LuceneTestCase { private static class MockFSDirectory extends Directory { - List allIndexInputs = new ArrayList(); + List allIndexInputs = new ArrayList(); Random rand; @@ -305,10 +304,9 @@ public class TestBufferedIndexInput extends LuceneTestCase { } public void tweakBufferSizes() { - Iterator it = allIndexInputs.iterator(); //int count = 0; - while(it.hasNext()) { - BufferedIndexInput bii = (BufferedIndexInput) it.next(); + for (final IndexInput ip : allIndexInputs) { + BufferedIndexInput bii = (BufferedIndexInput) ip; int bufferSize = 1024+(int) Math.abs(rand.nextInt() % 32768); bii.setBufferSize(bufferSize); //count++; diff --git a/src/test/org/apache/lucene/store/TestFileSwitchDirectory.java b/src/test/org/apache/lucene/store/TestFileSwitchDirectory.java index b2fec010acf..645f5627c0f 100644 --- a/src/test/org/apache/lucene/store/TestFileSwitchDirectory.java +++ b/src/test/org/apache/lucene/store/TestFileSwitchDirectory.java @@ -33,7 +33,7 @@ public class TestFileSwitchDirectory extends LuceneTestCase { * @throws IOException */ public void testBasic() throws IOException { - Set fileExtensions = new HashSet(); + Set fileExtensions = new HashSet(); fileExtensions.add("fdt"); fileExtensions.add("fdx"); diff --git a/src/test/org/apache/lucene/store/TestHugeRamFile.java b/src/test/org/apache/lucene/store/TestHugeRamFile.java index 33c9891080e..e3b9a31f10b 100755 --- a/src/test/org/apache/lucene/store/TestHugeRamFile.java +++ b/src/test/org/apache/lucene/store/TestHugeRamFile.java @@ -31,13 +31,13 @@ public class TestHugeRamFile extends LuceneTestCase { * buffers under maxint. */ private static class DenseRAMFile extends RAMFile { private long capacity = 0; - private HashMap singleBuffers = new HashMap(); + private HashMap singleBuffers = new HashMap(); @Override byte[] newBuffer(int size) { capacity += size; if (capacity <= MAX_VALUE) { // below maxint we reuse buffers - byte buf[] = (byte[]) singleBuffers.get(Integer.valueOf(size)); + byte buf[] = singleBuffers.get(Integer.valueOf(size)); if (buf==null) { buf = new byte[size]; //System.out.println("allocate: "+size); diff --git a/src/test/org/apache/lucene/store/TestLockFactory.java b/src/test/org/apache/lucene/store/TestLockFactory.java index 75215840b09..535e5c142b8 100755 --- a/src/test/org/apache/lucene/store/TestLockFactory.java +++ b/src/test/org/apache/lucene/store/TestLockFactory.java @@ -21,7 +21,6 @@ import java.io.File; import java.io.IOException; import java.util.Collections; import java.util.HashMap; -import java.util.Iterator; import java.util.Map; import org.apache.lucene.analysis.WhitespaceAnalyzer; @@ -63,8 +62,7 @@ public class TestLockFactory extends LuceneTestCase { assertTrue("# calls to makeLock is 0 (after instantiating IndexWriter)", lf.makeLockCount >= 1); - for(Iterator e = lf.locksCreated.keySet().iterator(); e.hasNext();) { - String lockName = (String) e.next(); + for(final String lockName : lf.locksCreated.keySet()) { MockLockFactory.MockLock lock = (MockLockFactory.MockLock) lf.locksCreated.get(lockName); assertTrue("# calls to Lock.obtain is 0 (after instantiating IndexWriter)", lock.lockAttempts > 0); @@ -341,7 +339,7 @@ public class TestLockFactory extends LuceneTestCase { public class MockLockFactory extends LockFactory { public boolean lockPrefixSet; - public Map locksCreated = Collections.synchronizedMap(new HashMap()); + public Map locksCreated = Collections.synchronizedMap(new HashMap()); public int makeLockCount = 0; @Override diff --git a/src/test/org/apache/lucene/util/LocalizedTestCase.java b/src/test/org/apache/lucene/util/LocalizedTestCase.java index 54f84de5fa0..92fb2722351 100644 --- a/src/test/org/apache/lucene/util/LocalizedTestCase.java +++ b/src/test/org/apache/lucene/util/LocalizedTestCase.java @@ -43,7 +43,7 @@ public abstract class LocalizedTestCase extends LuceneTestCase { /** * An optional limited set of testcases that will run under different Locales. */ - private final Set testWithDifferentLocales; + private final Set testWithDifferentLocales; public LocalizedTestCase() { super(); @@ -55,12 +55,12 @@ public abstract class LocalizedTestCase extends LuceneTestCase { testWithDifferentLocales = null; } - public LocalizedTestCase(Set testWithDifferentLocales) { + public LocalizedTestCase(Set testWithDifferentLocales) { super(); this.testWithDifferentLocales = testWithDifferentLocales; } - public LocalizedTestCase(String name, Set testWithDifferentLocales) { + public LocalizedTestCase(String name, Set testWithDifferentLocales) { super(name); this.testWithDifferentLocales = testWithDifferentLocales; } diff --git a/src/test/org/apache/lucene/util/LuceneTestCase.java b/src/test/org/apache/lucene/util/LuceneTestCase.java index 06deb58bde1..f695955d0a9 100644 --- a/src/test/org/apache/lucene/util/LuceneTestCase.java +++ b/src/test/org/apache/lucene/util/LuceneTestCase.java @@ -151,7 +151,7 @@ public abstract class LuceneTestCase extends TestCase { * @param iter Each next() is toString()ed and logged on it's own line. If iter is null this is logged differnetly then an empty iterator. * @param stream Stream to log messages to. */ - public static void dumpIterator(String label, Iterator iter, + public static void dumpIterator(String label, Iterator iter, PrintStream stream) { stream.println("*** BEGIN "+label+" ***"); if (null == iter) { @@ -170,7 +170,7 @@ public abstract class LuceneTestCase extends TestCase { */ public static void dumpArray(String label, Object[] objs, PrintStream stream) { - Iterator iter = (null == objs) ? null : Arrays.asList(objs).iterator(); + Iterator iter = (null == objs) ? null : Arrays.asList(objs).iterator(); dumpIterator(label, iter, stream); } diff --git a/src/test/org/apache/lucene/util/TestCloseableThreadLocal.java b/src/test/org/apache/lucene/util/TestCloseableThreadLocal.java index 031d343cab4..9b708103d8e 100644 --- a/src/test/org/apache/lucene/util/TestCloseableThreadLocal.java +++ b/src/test/org/apache/lucene/util/TestCloseableThreadLocal.java @@ -29,7 +29,7 @@ public class TestCloseableThreadLocal extends LuceneTestCase { public void testNullValue() throws Exception { // Tests that null can be set as a valid value (LUCENE-1805). This // previously failed in get(). - CloseableThreadLocal ctl = new CloseableThreadLocal(); + CloseableThreadLocal ctl = new CloseableThreadLocal(); ctl.set(null); assertNull(ctl.get()); } @@ -37,12 +37,11 @@ public class TestCloseableThreadLocal extends LuceneTestCase { public void testDefaultValueWithoutSetting() throws Exception { // LUCENE-1805: make sure default get returns null, // twice in a row - CloseableThreadLocal ctl = new CloseableThreadLocal(); - assertNull(ctl.get()); + CloseableThreadLocal ctl = new CloseableThreadLocal(); assertNull(ctl.get()); } - public class InitValueThreadLocal extends CloseableThreadLocal { + public class InitValueThreadLocal extends CloseableThreadLocal { @Override protected Object initialValue() { return TEST_VALUE; diff --git a/src/test/org/apache/lucene/util/TestNumericUtils.java b/src/test/org/apache/lucene/util/TestNumericUtils.java index 6a85a1d887e..8e82720e140 100644 --- a/src/test/org/apache/lucene/util/TestNumericUtils.java +++ b/src/test/org/apache/lucene/util/TestNumericUtils.java @@ -17,9 +17,6 @@ package org.apache.lucene.util; * limitations under the License. */ -import org.apache.lucene.util.LuceneTestCase; -import org.apache.lucene.util.OpenBitSet; - import java.util.Arrays; import java.util.Collections; import java.util.Iterator; @@ -174,7 +171,7 @@ public class TestNumericUtils extends LuceneTestCase { /** Note: The neededBounds iterator must be unsigned (easier understanding what's happening) */ protected void assertLongRangeSplit(final long lower, final long upper, int precisionStep, - final boolean useBitSet, final Iterator neededBounds + final boolean useBitSet, final Iterator neededBounds ) throws Exception { final OpenBitSet bits=useBitSet ? new OpenBitSet(upper-lower+1) : null; @@ -189,8 +186,8 @@ public class TestNumericUtils extends LuceneTestCase { min ^= 0x8000000000000000L; max ^= 0x8000000000000000L; //System.out.println("Long.valueOf(0x"+Long.toHexString(min>>>shift)+"L),Long.valueOf(0x"+Long.toHexString(max>>>shift)+"L),"); - assertEquals( "inner min bound", ((Long)neededBounds.next()).longValue(), min>>>shift); - assertEquals( "inner max bound", ((Long)neededBounds.next()).longValue(), max>>>shift); + assertEquals( "inner min bound", neededBounds.next().longValue(), min>>>shift); + assertEquals( "inner max bound", neededBounds.next().longValue(), max>>>shift); } }, precisionStep, lower, upper); @@ -246,7 +243,7 @@ public class TestNumericUtils extends LuceneTestCase { }).iterator()); // a inverse range should produce no sub-ranges - assertLongRangeSplit(9500L, -5000L, 4, false, Collections.EMPTY_LIST.iterator()); + assertLongRangeSplit(9500L, -5000L, 4, false, Collections. emptyList().iterator()); // a 0-length range should reproduce the range itsself assertLongRangeSplit(9500L, 9500L, 4, false, Arrays.asList(new Long[]{ @@ -256,7 +253,7 @@ public class TestNumericUtils extends LuceneTestCase { /** Note: The neededBounds iterator must be unsigned (easier understanding what's happening) */ protected void assertIntRangeSplit(final int lower, final int upper, int precisionStep, - final boolean useBitSet, final Iterator neededBounds + final boolean useBitSet, final Iterator neededBounds ) throws Exception { final OpenBitSet bits=useBitSet ? new OpenBitSet(upper-lower+1) : null; @@ -271,8 +268,8 @@ public class TestNumericUtils extends LuceneTestCase { min ^= 0x80000000; max ^= 0x80000000; //System.out.println("Integer.valueOf(0x"+Integer.toHexString(min>>>shift)+"),Integer.valueOf(0x"+Integer.toHexString(max>>>shift)+"),"); - assertEquals( "inner min bound", ((Integer)neededBounds.next()).intValue(), min>>>shift); - assertEquals( "inner max bound", ((Integer)neededBounds.next()).intValue(), max>>>shift); + assertEquals( "inner min bound", neededBounds.next().intValue(), min>>>shift); + assertEquals( "inner max bound", neededBounds.next().intValue(), max>>>shift); } }, precisionStep, lower, upper); @@ -328,7 +325,7 @@ public class TestNumericUtils extends LuceneTestCase { }).iterator()); // a inverse range should produce no sub-ranges - assertIntRangeSplit(9500, -5000, 4, false, Collections.EMPTY_LIST.iterator()); + assertIntRangeSplit(9500, -5000, 4, false, Collections. emptyList().iterator()); // a 0-length range should reproduce the range itsself assertIntRangeSplit(9500, 9500, 4, false, Arrays.asList(new Integer[]{ diff --git a/src/test/org/apache/lucene/util/cache/BaseTestLRU.java b/src/test/org/apache/lucene/util/cache/BaseTestLRU.java index 69341a2ae65..f6f7d1978ce 100644 --- a/src/test/org/apache/lucene/util/cache/BaseTestLRU.java +++ b/src/test/org/apache/lucene/util/cache/BaseTestLRU.java @@ -21,7 +21,7 @@ import org.apache.lucene.util.LuceneTestCase; public class BaseTestLRU extends LuceneTestCase { - protected void testCache(Cache cache, int n) throws Exception { + protected void testCache(Cache cache, int n) throws Exception { Object dummy = new Object(); for (int i = 0; i < n; i++) { diff --git a/src/test/org/apache/lucene/util/cache/TestDoubleBarrelLRUCache.java b/src/test/org/apache/lucene/util/cache/TestDoubleBarrelLRUCache.java index 045501284cf..a4ca7736f3b 100644 --- a/src/test/org/apache/lucene/util/cache/TestDoubleBarrelLRUCache.java +++ b/src/test/org/apache/lucene/util/cache/TestDoubleBarrelLRUCache.java @@ -21,7 +21,7 @@ public class TestDoubleBarrelLRUCache extends BaseTestLRU { public void testLRUCache() throws Exception { final int n = 100; - testCache(new DoubleBarrelLRUCache(n), n); + testCache(new DoubleBarrelLRUCache(n), n); } private class CacheThread extends Thread {