LUCENE-3640: Remove IndexSearcher.close

git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@1213117 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Robert Muir 2011-12-12 00:21:40 +00:00
parent 3c00042ca0
commit 3899e18ca3
157 changed files with 194 additions and 542 deletions

View File

@ -222,6 +222,10 @@ Changes in backwards compatibility policy
FilterIndexReader, overriding FilterIndexReader.norms(). To persist the FilterIndexReader, overriding FilterIndexReader.norms(). To persist the
changes on disk, copy the FilteredIndexReader to a new index using changes on disk, copy the FilteredIndexReader to a new index using
IndexWriter.addIndexes(). (Uwe Schindler, Robert Muir) IndexWriter.addIndexes(). (Uwe Schindler, Robert Muir)
* LUCENE-3640: Removed IndexSearcher.close(), because IndexSearcher no longer
takes a Directory and no longer "manages" IndexReaders, it is a no-op.
(Robert Muir)
Changes in Runtime Behavior Changes in Runtime Behavior

View File

@ -131,7 +131,6 @@ public class SearchFiles {
break; break;
} }
} }
searcher.close();
reader.close(); reader.close();
} }

View File

@ -72,26 +72,21 @@ public class HighlighterPhraseTest extends LuceneTestCase {
try { try {
assertEquals(1, indexReader.numDocs()); assertEquals(1, indexReader.numDocs());
final IndexSearcher indexSearcher = newSearcher(indexReader); final IndexSearcher indexSearcher = newSearcher(indexReader);
try { final PhraseQuery phraseQuery = new PhraseQuery();
final PhraseQuery phraseQuery = new PhraseQuery(); phraseQuery.add(new Term(FIELD, "fox"));
phraseQuery.add(new Term(FIELD, "fox")); phraseQuery.add(new Term(FIELD, "jumped"));
phraseQuery.add(new Term(FIELD, "jumped")); phraseQuery.setSlop(0);
phraseQuery.setSlop(0); TopDocs hits = indexSearcher.search(phraseQuery, 1);
TopDocs hits = indexSearcher.search(phraseQuery, 1); assertEquals(1, hits.totalHits);
assertEquals(1, hits.totalHits); final Highlighter highlighter = new Highlighter(
final Highlighter highlighter = new Highlighter( new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(), new QueryScorer(phraseQuery));
new QueryScorer(phraseQuery));
final TokenStream tokenStream = TokenSources final TokenStream tokenStream = TokenSources
.getTokenStream(indexReader.getTermVector( .getTokenStream(indexReader.getTermVector(
0, FIELD), false); 0, FIELD), false);
assertEquals(highlighter.getBestFragment(new TokenStreamConcurrent(), assertEquals(highlighter.getBestFragment(new TokenStreamConcurrent(),
TEXT), highlighter.getBestFragment(tokenStream, TEXT)); TEXT), highlighter.getBestFragment(tokenStream, TEXT));
} finally {
indexSearcher.close();
}
} finally { } finally {
indexReader.close(); indexReader.close();
directory.close(); directory.close();
@ -120,52 +115,48 @@ public class HighlighterPhraseTest extends LuceneTestCase {
try { try {
assertEquals(1, indexReader.numDocs()); assertEquals(1, indexReader.numDocs());
final IndexSearcher indexSearcher = newSearcher(indexReader); final IndexSearcher indexSearcher = newSearcher(indexReader);
try { final Query phraseQuery = new SpanNearQuery(new SpanQuery[] {
final Query phraseQuery = new SpanNearQuery(new SpanQuery[] { new SpanTermQuery(new Term(FIELD, "fox")),
new SpanTermQuery(new Term(FIELD, "fox")), new SpanTermQuery(new Term(FIELD, "jumped")) }, 0, true);
new SpanTermQuery(new Term(FIELD, "jumped")) }, 0, true); final FixedBitSet bitset = new FixedBitSet(indexReader.maxDoc());
final FixedBitSet bitset = new FixedBitSet(indexReader.maxDoc()); indexSearcher.search(phraseQuery, new Collector() {
indexSearcher.search(phraseQuery, new Collector() { private int baseDoc;
private int baseDoc;
@Override @Override
public boolean acceptsDocsOutOfOrder() { public boolean acceptsDocsOutOfOrder() {
return true; return true;
}
@Override
public void collect(int i) throws IOException {
bitset.set(this.baseDoc + i);
}
@Override
public void setNextReader(AtomicReaderContext context)
throws IOException {
this.baseDoc = context.docBase;
}
@Override
public void setScorer(org.apache.lucene.search.Scorer scorer)
throws IOException {
// Do Nothing
}
});
assertEquals(1, bitset.cardinality());
final int maxDoc = indexReader.maxDoc();
final Highlighter highlighter = new Highlighter(
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new QueryScorer(phraseQuery));
for (int position = bitset.nextSetBit(0); position >= 0 && position < maxDoc-1; position = bitset
.nextSetBit(position + 1)) {
assertEquals(0, position);
final TokenStream tokenStream = TokenSources.getTokenStream(
indexReader.getTermVector(position,
FIELD), false);
assertEquals(highlighter.getBestFragment(new TokenStreamConcurrent(),
TEXT), highlighter.getBestFragment(tokenStream, TEXT));
} }
} finally {
indexSearcher.close(); @Override
public void collect(int i) throws IOException {
bitset.set(this.baseDoc + i);
}
@Override
public void setNextReader(AtomicReaderContext context)
throws IOException {
this.baseDoc = context.docBase;
}
@Override
public void setScorer(org.apache.lucene.search.Scorer scorer)
throws IOException {
// Do Nothing
}
});
assertEquals(1, bitset.cardinality());
final int maxDoc = indexReader.maxDoc();
final Highlighter highlighter = new Highlighter(
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new QueryScorer(phraseQuery));
for (int position = bitset.nextSetBit(0); position >= 0 && position < maxDoc-1; position = bitset
.nextSetBit(position + 1)) {
assertEquals(0, position);
final TokenStream tokenStream = TokenSources.getTokenStream(
indexReader.getTermVector(position,
FIELD), false);
assertEquals(highlighter.getBestFragment(new TokenStreamConcurrent(),
TEXT), highlighter.getBestFragment(tokenStream, TEXT));
} }
} finally { } finally {
indexReader.close(); indexReader.close();
@ -195,25 +186,21 @@ public class HighlighterPhraseTest extends LuceneTestCase {
try { try {
assertEquals(1, indexReader.numDocs()); assertEquals(1, indexReader.numDocs());
final IndexSearcher indexSearcher = newSearcher(indexReader); final IndexSearcher indexSearcher = newSearcher(indexReader);
try { final PhraseQuery phraseQuery = new PhraseQuery();
final PhraseQuery phraseQuery = new PhraseQuery(); phraseQuery.add(new Term(FIELD, "did"));
phraseQuery.add(new Term(FIELD, "did")); phraseQuery.add(new Term(FIELD, "jump"));
phraseQuery.add(new Term(FIELD, "jump")); phraseQuery.setSlop(0);
phraseQuery.setSlop(0); TopDocs hits = indexSearcher.search(phraseQuery, 1);
TopDocs hits = indexSearcher.search(phraseQuery, 1); assertEquals(0, hits.totalHits);
assertEquals(0, hits.totalHits); final Highlighter highlighter = new Highlighter(
final Highlighter highlighter = new Highlighter( new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(), new QueryScorer(phraseQuery));
new QueryScorer(phraseQuery)); final TokenStream tokenStream = TokenSources
final TokenStream tokenStream = TokenSources .getTokenStream(indexReader.getTermVector(
.getTokenStream(indexReader.getTermVector( 0, FIELD), false);
0, FIELD), false); assertEquals(
assertEquals( highlighter.getBestFragment(new TokenStreamSparse(), TEXT),
highlighter.getBestFragment(new TokenStreamSparse(), TEXT), highlighter.getBestFragment(tokenStream, TEXT));
highlighter.getBestFragment(tokenStream, TEXT));
} finally {
indexSearcher.close();
}
} finally { } finally {
indexReader.close(); indexReader.close();
directory.close(); directory.close();
@ -241,23 +228,19 @@ public class HighlighterPhraseTest extends LuceneTestCase {
try { try {
assertEquals(1, indexReader.numDocs()); assertEquals(1, indexReader.numDocs());
final IndexSearcher indexSearcher = newSearcher(indexReader); final IndexSearcher indexSearcher = newSearcher(indexReader);
try { final PhraseQuery phraseQuery = new PhraseQuery();
final PhraseQuery phraseQuery = new PhraseQuery(); phraseQuery.add(new Term(FIELD, "did"));
phraseQuery.add(new Term(FIELD, "did")); phraseQuery.add(new Term(FIELD, "jump"));
phraseQuery.add(new Term(FIELD, "jump")); phraseQuery.setSlop(1);
phraseQuery.setSlop(1); TopDocs hits = indexSearcher.search(phraseQuery, 1);
TopDocs hits = indexSearcher.search(phraseQuery, 1); assertEquals(1, hits.totalHits);
assertEquals(1, hits.totalHits); final Highlighter highlighter = new Highlighter(
final Highlighter highlighter = new Highlighter( new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(), new QueryScorer(phraseQuery));
new QueryScorer(phraseQuery)); final TokenStream tokenStream = TokenSources.getTokenStream(
final TokenStream tokenStream = TokenSources.getTokenStream( indexReader.getTermVector(0, FIELD), true);
indexReader.getTermVector(0, FIELD), true); assertEquals("the fox <B>did</B> not <B>jump</B>", highlighter
assertEquals("the fox <B>did</B> not <B>jump</B>", highlighter .getBestFragment(tokenStream, TEXT));
.getBestFragment(tokenStream, TEXT));
} finally {
indexSearcher.close();
}
} finally { } finally {
indexReader.close(); indexReader.close();
directory.close(); directory.close();
@ -285,25 +268,21 @@ public class HighlighterPhraseTest extends LuceneTestCase {
try { try {
assertEquals(1, indexReader.numDocs()); assertEquals(1, indexReader.numDocs());
final IndexSearcher indexSearcher = newSearcher(indexReader); final IndexSearcher indexSearcher = newSearcher(indexReader);
try { final Query phraseQuery = new SpanNearQuery(new SpanQuery[] {
final Query phraseQuery = new SpanNearQuery(new SpanQuery[] { new SpanTermQuery(new Term(FIELD, "did")),
new SpanTermQuery(new Term(FIELD, "did")), new SpanTermQuery(new Term(FIELD, "jump")) }, 0, true);
new SpanTermQuery(new Term(FIELD, "jump")) }, 0, true);
TopDocs hits = indexSearcher.search(phraseQuery, 1); TopDocs hits = indexSearcher.search(phraseQuery, 1);
assertEquals(0, hits.totalHits); assertEquals(0, hits.totalHits);
final Highlighter highlighter = new Highlighter( final Highlighter highlighter = new Highlighter(
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(), new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new QueryScorer(phraseQuery)); new QueryScorer(phraseQuery));
final TokenStream tokenStream = TokenSources final TokenStream tokenStream = TokenSources
.getTokenStream(indexReader.getTermVector( .getTokenStream(indexReader.getTermVector(
0, FIELD), false); 0, FIELD), false);
assertEquals( assertEquals(
highlighter.getBestFragment(new TokenStreamSparse(), TEXT), highlighter.getBestFragment(new TokenStreamSparse(), TEXT),
highlighter.getBestFragment(tokenStream, TEXT)); highlighter.getBestFragment(tokenStream, TEXT));
} finally {
indexSearcher.close();
}
} finally { } finally {
indexReader.close(); indexReader.close();
directory.close(); directory.close();

View File

@ -109,7 +109,6 @@ public class HighlighterTest extends BaseTokenStreamTestCase implements Formatte
if (VERBOSE) System.out.println(fragment); if (VERBOSE) System.out.println(fragment);
} }
searcher.close();
} }
public void testHighlightingWithDefaultField() throws Exception { public void testHighlightingWithDefaultField() throws Exception {
@ -1271,7 +1270,6 @@ public class HighlighterTest extends BaseTokenStreamTestCase implements Formatte
public void run() throws Exception { public void run() throws Exception {
numHighlights = 0; numHighlights = 0;
// test to show how rewritten query can still be used // test to show how rewritten query can still be used
if (searcher != null) searcher.close();
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
Analyzer analyzer = new MockAnalyzer(random, MockTokenizer.SIMPLE, true, MockTokenFilter.ENGLISH_STOPSET, true); Analyzer analyzer = new MockAnalyzer(random, MockTokenizer.SIMPLE, true, MockTokenFilter.ENGLISH_STOPSET, true);
@ -1664,7 +1662,6 @@ public class HighlighterTest extends BaseTokenStreamTestCase implements Formatte
if (VERBOSE) System.out.println("result:" + result); if (VERBOSE) System.out.println("result:" + result);
assertEquals("more <B>random</B> words for second field", result); assertEquals("more <B>random</B> words for second field", result);
} }
searcher.close();
reader.close(); reader.close();
} }
@ -1703,7 +1700,6 @@ public class HighlighterTest extends BaseTokenStreamTestCase implements Formatte
} }
public void doSearching(Query unReWrittenQuery) throws Exception { public void doSearching(Query unReWrittenQuery) throws Exception {
if (searcher != null) searcher.close();
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
// for any multi-term queries to work (prefix, wildcard, range,fuzzy etc) // for any multi-term queries to work (prefix, wildcard, range,fuzzy etc)
// you must use a rewritten query! // you must use a rewritten query!
@ -1769,7 +1765,6 @@ public class HighlighterTest extends BaseTokenStreamTestCase implements Formatte
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
if (searcher != null) searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
ramDir.close(); ramDir.close();

View File

@ -114,31 +114,27 @@ public class TokenSourcesTest extends LuceneTestCase {
indexWriter.close(); indexWriter.close();
} }
final IndexReader indexReader = IndexReader.open(directory); final IndexReader indexReader = IndexReader.open(directory);
assertEquals(1, indexReader.numDocs());
final IndexSearcher indexSearcher = newSearcher(indexReader);
try { try {
assertEquals(1, indexReader.numDocs()); final DisjunctionMaxQuery query = new DisjunctionMaxQuery(1);
final IndexSearcher indexSearcher = newSearcher(indexReader); query.add(new SpanTermQuery(new Term(FIELD, "{fox}")));
try { query.add(new SpanTermQuery(new Term(FIELD, "fox")));
final DisjunctionMaxQuery query = new DisjunctionMaxQuery(1);
query.add(new SpanTermQuery(new Term(FIELD, "{fox}")));
query.add(new SpanTermQuery(new Term(FIELD, "fox")));
// final Query phraseQuery = new SpanNearQuery(new SpanQuery[] { // final Query phraseQuery = new SpanNearQuery(new SpanQuery[] {
// new SpanTermQuery(new Term(FIELD, "{fox}")), // new SpanTermQuery(new Term(FIELD, "{fox}")),
// new SpanTermQuery(new Term(FIELD, "fox")) }, 0, true); // new SpanTermQuery(new Term(FIELD, "fox")) }, 0, true);
TopDocs hits = indexSearcher.search(query, 1); TopDocs hits = indexSearcher.search(query, 1);
assertEquals(1, hits.totalHits); assertEquals(1, hits.totalHits);
final Highlighter highlighter = new Highlighter( final Highlighter highlighter = new Highlighter(
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(), new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new QueryScorer(query)); new QueryScorer(query));
final TokenStream tokenStream = TokenSources final TokenStream tokenStream = TokenSources
.getTokenStream( .getTokenStream(
indexReader.getTermVector(0, FIELD), indexReader.getTermVector(0, FIELD),
false); false);
assertEquals("<B>the fox</B> did not jump", assertEquals("<B>the fox</B> did not jump",
highlighter.getBestFragment(tokenStream, TEXT)); highlighter.getBestFragment(tokenStream, TEXT));
} finally {
indexSearcher.close();
}
} finally { } finally {
indexReader.close(); indexReader.close();
directory.close(); directory.close();
@ -166,28 +162,24 @@ public class TokenSourcesTest extends LuceneTestCase {
try { try {
assertEquals(1, indexReader.numDocs()); assertEquals(1, indexReader.numDocs());
final IndexSearcher indexSearcher = newSearcher(indexReader); final IndexSearcher indexSearcher = newSearcher(indexReader);
try { final DisjunctionMaxQuery query = new DisjunctionMaxQuery(1);
final DisjunctionMaxQuery query = new DisjunctionMaxQuery(1); query.add(new SpanTermQuery(new Term(FIELD, "{fox}")));
query.add(new SpanTermQuery(new Term(FIELD, "{fox}"))); query.add(new SpanTermQuery(new Term(FIELD, "fox")));
query.add(new SpanTermQuery(new Term(FIELD, "fox"))); // final Query phraseQuery = new SpanNearQuery(new SpanQuery[] {
// final Query phraseQuery = new SpanNearQuery(new SpanQuery[] { // new SpanTermQuery(new Term(FIELD, "{fox}")),
// new SpanTermQuery(new Term(FIELD, "{fox}")), // new SpanTermQuery(new Term(FIELD, "fox")) }, 0, true);
// new SpanTermQuery(new Term(FIELD, "fox")) }, 0, true);
TopDocs hits = indexSearcher.search(query, 1); TopDocs hits = indexSearcher.search(query, 1);
assertEquals(1, hits.totalHits); assertEquals(1, hits.totalHits);
final Highlighter highlighter = new Highlighter( final Highlighter highlighter = new Highlighter(
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(), new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new QueryScorer(query)); new QueryScorer(query));
final TokenStream tokenStream = TokenSources final TokenStream tokenStream = TokenSources
.getTokenStream( .getTokenStream(
indexReader.getTermVector(0, FIELD), indexReader.getTermVector(0, FIELD),
false); false);
assertEquals("<B>the fox</B> did not jump", assertEquals("<B>the fox</B> did not jump",
highlighter.getBestFragment(tokenStream, TEXT)); highlighter.getBestFragment(tokenStream, TEXT));
} finally {
indexSearcher.close();
}
} finally { } finally {
indexReader.close(); indexReader.close();
directory.close(); directory.close();
@ -214,28 +206,24 @@ public class TokenSourcesTest extends LuceneTestCase {
try { try {
assertEquals(1, indexReader.numDocs()); assertEquals(1, indexReader.numDocs());
final IndexSearcher indexSearcher = newSearcher(indexReader); final IndexSearcher indexSearcher = newSearcher(indexReader);
try { // final DisjunctionMaxQuery query = new DisjunctionMaxQuery(1);
// final DisjunctionMaxQuery query = new DisjunctionMaxQuery(1); // query.add(new SpanTermQuery(new Term(FIELD, "{fox}")));
// query.add(new SpanTermQuery(new Term(FIELD, "{fox}"))); // query.add(new SpanTermQuery(new Term(FIELD, "fox")));
// query.add(new SpanTermQuery(new Term(FIELD, "fox"))); final Query phraseQuery = new SpanNearQuery(new SpanQuery[] {
final Query phraseQuery = new SpanNearQuery(new SpanQuery[] { new SpanTermQuery(new Term(FIELD, "the")),
new SpanTermQuery(new Term(FIELD, "the")), new SpanTermQuery(new Term(FIELD, "fox"))}, 0, true);
new SpanTermQuery(new Term(FIELD, "fox"))}, 0, true);
TopDocs hits = indexSearcher.search(phraseQuery, 1); TopDocs hits = indexSearcher.search(phraseQuery, 1);
assertEquals(1, hits.totalHits); assertEquals(1, hits.totalHits);
final Highlighter highlighter = new Highlighter( final Highlighter highlighter = new Highlighter(
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(), new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new QueryScorer(phraseQuery)); new QueryScorer(phraseQuery));
final TokenStream tokenStream = TokenSources final TokenStream tokenStream = TokenSources
.getTokenStream( .getTokenStream(
indexReader.getTermVector(0, FIELD), indexReader.getTermVector(0, FIELD),
false); false);
assertEquals("<B>the fox</B> did not jump", assertEquals("<B>the fox</B> did not jump",
highlighter.getBestFragment(tokenStream, TEXT)); highlighter.getBestFragment(tokenStream, TEXT));
} finally {
indexSearcher.close();
}
} finally { } finally {
indexReader.close(); indexReader.close();
directory.close(); directory.close();
@ -263,28 +251,24 @@ public class TokenSourcesTest extends LuceneTestCase {
try { try {
assertEquals(1, indexReader.numDocs()); assertEquals(1, indexReader.numDocs());
final IndexSearcher indexSearcher = newSearcher(indexReader); final IndexSearcher indexSearcher = newSearcher(indexReader);
try { // final DisjunctionMaxQuery query = new DisjunctionMaxQuery(1);
// final DisjunctionMaxQuery query = new DisjunctionMaxQuery(1); // query.add(new SpanTermQuery(new Term(FIELD, "the")));
// query.add(new SpanTermQuery(new Term(FIELD, "the"))); // query.add(new SpanTermQuery(new Term(FIELD, "fox")));
// query.add(new SpanTermQuery(new Term(FIELD, "fox"))); final Query phraseQuery = new SpanNearQuery(new SpanQuery[] {
final Query phraseQuery = new SpanNearQuery(new SpanQuery[] { new SpanTermQuery(new Term(FIELD, "the")),
new SpanTermQuery(new Term(FIELD, "the")), new SpanTermQuery(new Term(FIELD, "fox"))}, 0, true);
new SpanTermQuery(new Term(FIELD, "fox"))}, 0, true);
TopDocs hits = indexSearcher.search(phraseQuery, 1); TopDocs hits = indexSearcher.search(phraseQuery, 1);
assertEquals(1, hits.totalHits); assertEquals(1, hits.totalHits);
final Highlighter highlighter = new Highlighter( final Highlighter highlighter = new Highlighter(
new SimpleHTMLFormatter(), new SimpleHTMLEncoder(), new SimpleHTMLFormatter(), new SimpleHTMLEncoder(),
new QueryScorer(phraseQuery)); new QueryScorer(phraseQuery));
final TokenStream tokenStream = TokenSources final TokenStream tokenStream = TokenSources
.getTokenStream( .getTokenStream(
indexReader.getTermVector(0, FIELD), indexReader.getTermVector(0, FIELD),
false); false);
assertEquals("<B>the fox</B> did not jump", assertEquals("<B>the fox</B> did not jump",
highlighter.getBestFragment(tokenStream, TEXT)); highlighter.getBestFragment(tokenStream, TEXT));
} finally {
indexSearcher.close();
}
} finally { } finally {
indexReader.close(); indexReader.close();
directory.close(); directory.close();

View File

@ -143,9 +143,7 @@ public class MemoryIndexTest extends BaseTokenStreamTestCase {
TopDocs memDocs = mem.search(qp.parse(query), 1); TopDocs memDocs = mem.search(qp.parse(query), 1);
assertEquals(ramDocs.totalHits, memDocs.totalHits); assertEquals(ramDocs.totalHits, memDocs.totalHits);
} }
ram.close();
reader.close(); reader.close();
mem.close();
} }
/** /**

View File

@ -69,7 +69,6 @@ public class DuplicateFilterTest extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
reader.close(); reader.close();
searcher.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();
} }

View File

@ -60,7 +60,6 @@ public class FuzzyLikeThisQueryTest extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();

View File

@ -71,7 +71,6 @@ public class TestSlowCollationMethods extends LuceneTestCase {
@AfterClass @AfterClass
public static void afterClass() throws Exception { public static void afterClass() throws Exception {
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
collator = null; collator = null;

View File

@ -56,7 +56,6 @@ public class TestRegexQuery extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();

View File

@ -80,7 +80,6 @@ public class TestSpanRegexQuery extends LuceneTestCase {
// true); // true);
int numHits = searcher.search(sfq, null, 1000).totalHits; int numHits = searcher.search(sfq, null, 1000).totalHits;
assertEquals(1, numHits); assertEquals(1, numHits);
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }

View File

@ -260,7 +260,6 @@ public class TestCartesian extends LuceneTestCase {
assertTrue(geo_distance >= lastDistance); assertTrue(geo_distance >= lastDistance);
lastDistance = geo_distance; lastDistance = geo_distance;
} }
searcher.close();
reader.close(); reader.close();
} }
@ -332,7 +331,6 @@ public class TestCartesian extends LuceneTestCase {
assertTrue(geo_distance >= lastDistance); assertTrue(geo_distance >= lastDistance);
lastDistance = geo_distance; lastDistance = geo_distance;
} }
searcher.close();
reader.close(); reader.close();
} }
@ -404,7 +402,6 @@ public class TestCartesian extends LuceneTestCase {
lastDistance = geo_distance; lastDistance = geo_distance;
} }
} }
searcher.close();
reader.close(); reader.close();
} }
@ -476,7 +473,6 @@ public class TestCartesian extends LuceneTestCase {
} }
} }
searcher.close();
reader.close(); reader.close();
} }
} }

View File

@ -42,8 +42,6 @@ import org.apache.lucene.index.Terms;
import org.apache.lucene.search.similarities.DefaultSimilarityProvider; import org.apache.lucene.search.similarities.DefaultSimilarityProvider;
import org.apache.lucene.search.similarities.SimilarityProvider; import org.apache.lucene.search.similarities.SimilarityProvider;
import org.apache.lucene.store.NIOFSDirectory; // javadoc import org.apache.lucene.store.NIOFSDirectory; // javadoc
import org.apache.lucene.util.Bits;
import org.apache.lucene.util.BytesRef;
import org.apache.lucene.util.ReaderUtil; import org.apache.lucene.util.ReaderUtil;
import org.apache.lucene.util.TermContext; import org.apache.lucene.util.TermContext;
import org.apache.lucene.util.ThreadInterruptedException; import org.apache.lucene.util.ThreadInterruptedException;
@ -73,7 +71,7 @@ import org.apache.lucene.util.ThreadInterruptedException;
* synchronize on the <code>IndexSearcher</code> instance; * synchronize on the <code>IndexSearcher</code> instance;
* use your own (non-Lucene) objects instead.</p> * use your own (non-Lucene) objects instead.</p>
*/ */
public class IndexSearcher implements Closeable { public class IndexSearcher {
final IndexReader reader; // package private for testing! final IndexReader reader; // package private for testing!
// NOTE: these members might change in incompatible ways // NOTE: these members might change in incompatible ways
@ -198,10 +196,6 @@ public class IndexSearcher implements Closeable {
public SimilarityProvider getSimilarityProvider() { public SimilarityProvider getSimilarityProvider() {
return similarityProvider; return similarityProvider;
} }
@Override
public void close() throws IOException {
}
/** @lucene.internal */ /** @lucene.internal */
protected Query wrapFilter(Query query, Filter filter) { protected Query wrapFilter(Query query, Filter filter) {

View File

@ -102,7 +102,6 @@ public abstract class CollationTestBase extends LuceneTestCase {
(query, new TermRangeFilter("content", secondBeg, secondEnd, true, true), 1).scoreDocs; (query, new TermRangeFilter("content", secondBeg, secondEnd, true, true), 1).scoreDocs;
assertEquals("The index Term should be included.", 1, result.length); assertEquals("The index Term should be included.", 1, result.length);
searcher.close();
reader.close(); reader.close();
} }
@ -131,7 +130,6 @@ public abstract class CollationTestBase extends LuceneTestCase {
query = new TermRangeQuery("content", secondBeg, secondEnd, true, true); query = new TermRangeQuery("content", secondBeg, secondEnd, true, true);
hits = searcher.search(query, null, 1000).scoreDocs; hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals("The index Term should be included.", 1, hits.length); assertEquals("The index Term should be included.", 1, hits.length);
searcher.close();
reader.close(); reader.close();
} }
@ -164,7 +162,6 @@ public abstract class CollationTestBase extends LuceneTestCase {
("content", secondBeg, secondEnd, true, true); ("content", secondBeg, secondEnd, true, true);
result = search.search(csrq, null, 1000).scoreDocs; result = search.search(csrq, null, 1000).scoreDocs;
assertEquals("The index Term should be included.", 1, result.length); assertEquals("The index Term should be included.", 1, result.length);
search.close();
} }
// Test using various international locales with accented characters (which // Test using various international locales with accented characters (which
@ -240,7 +237,6 @@ public abstract class CollationTestBase extends LuceneTestCase {
sort.setSort(new SortField("Denmark", SortField.Type.STRING)); sort.setSort(new SortField("Denmark", SortField.Type.STRING));
assertMatches(searcher, queryY, sort, dkResult); assertMatches(searcher, queryY, sort, dkResult);
searcher.close();
reader.close(); reader.close();
} }

View File

@ -472,7 +472,6 @@ public abstract class ThreadedIndexingAndSearchingTestCase extends LuceneTestCas
IndexSearcher searcher = newSearcher(reader); IndexSearcher searcher = newSearcher(reader);
sum += searcher.search(new TermQuery(new Term("body", "united")), 10).totalHits; sum += searcher.search(new TermQuery(new Term("body", "united")), 10).totalHits;
searcher.close();
if (VERBOSE) { if (VERBOSE) {
System.out.println("TEST: warm visited " + sum + " fields"); System.out.println("TEST: warm visited " + sum + " fields");

View File

@ -116,7 +116,6 @@ public class CheckHits {
query.toString(defaultFieldName), query.toString(defaultFieldName),
correct, actual); correct, actual);
FieldCache.DEFAULT.purge(s.getIndexReader()); // our wrapping can create insanity otherwise FieldCache.DEFAULT.purge(s.getIndexReader()); // our wrapping can create insanity otherwise
s.close();
} }
} }

View File

@ -115,13 +115,10 @@ public class QueryUtils {
IndexSearcher wrapped; IndexSearcher wrapped;
check(random, q1, wrapped = wrapUnderlyingReader(random, s, -1), false); check(random, q1, wrapped = wrapUnderlyingReader(random, s, -1), false);
FieldCache.DEFAULT.purge(wrapped.getIndexReader()); // // our wrapping can create insanity otherwise FieldCache.DEFAULT.purge(wrapped.getIndexReader()); // // our wrapping can create insanity otherwise
wrapped.close();
check(random, q1, wrapped = wrapUnderlyingReader(random, s, 0), false); check(random, q1, wrapped = wrapUnderlyingReader(random, s, 0), false);
FieldCache.DEFAULT.purge(wrapped.getIndexReader()); // // our wrapping can create insanity otherwise FieldCache.DEFAULT.purge(wrapped.getIndexReader()); // // our wrapping can create insanity otherwise
wrapped.close();
check(random, q1, wrapped = wrapUnderlyingReader(random, s, +1), false); check(random, q1, wrapped = wrapUnderlyingReader(random, s, +1), false);
FieldCache.DEFAULT.purge(wrapped.getIndexReader()); // // our wrapping can create insanity otherwise FieldCache.DEFAULT.purge(wrapped.getIndexReader()); // // our wrapping can create insanity otherwise
wrapped.close();
} }
checkExplanations(q1,s); checkExplanations(q1,s);
@ -309,7 +306,6 @@ public class QueryUtils {
Assert.assertFalse("query's last doc was "+ lastDoc[0] +" but skipTo("+(lastDoc[0]+1)+") got to "+scorer.docID(),more); Assert.assertFalse("query's last doc was "+ lastDoc[0] +" but skipTo("+(lastDoc[0]+1)+") got to "+scorer.docID(),more);
} }
leafPtr++; leafPtr++;
indexSearcher.close();
} }
lastReader[0] = context.reader; lastReader[0] = context.reader;
assert readerContextArray[leafPtr].reader == context.reader; assert readerContextArray[leafPtr].reader == context.reader;
@ -335,7 +331,6 @@ public class QueryUtils {
boolean more = scorer.advance(lastDoc[0] + 1) != DocIdSetIterator.NO_MORE_DOCS; boolean more = scorer.advance(lastDoc[0] + 1) != DocIdSetIterator.NO_MORE_DOCS;
Assert.assertFalse("query's last doc was "+ lastDoc[0] +" but skipTo("+(lastDoc[0]+1)+") got to "+scorer.docID(),more); Assert.assertFalse("query's last doc was "+ lastDoc[0] +" but skipTo("+(lastDoc[0]+1)+") got to "+scorer.docID(),more);
} }
indexSearcher.close();
} }
} }
} }
@ -394,7 +389,6 @@ public class QueryUtils {
boolean more = scorer.advance(lastDoc[0] + 1) != DocIdSetIterator.NO_MORE_DOCS; boolean more = scorer.advance(lastDoc[0] + 1) != DocIdSetIterator.NO_MORE_DOCS;
Assert.assertFalse("query's last doc was "+ lastDoc[0] +" but skipTo("+(lastDoc[0]+1)+") got to "+scorer.docID(),more); Assert.assertFalse("query's last doc was "+ lastDoc[0] +" but skipTo("+(lastDoc[0]+1)+") got to "+scorer.docID(),more);
} }
indexSearcher.close();
leafPtr++; leafPtr++;
} }
@ -419,7 +413,6 @@ public class QueryUtils {
boolean more = scorer.advance(lastDoc[0] + 1) != DocIdSetIterator.NO_MORE_DOCS; boolean more = scorer.advance(lastDoc[0] + 1) != DocIdSetIterator.NO_MORE_DOCS;
Assert.assertFalse("query's last doc was "+ lastDoc[0] +" but skipTo("+(lastDoc[0]+1)+") got to "+scorer.docID(),more); Assert.assertFalse("query's last doc was "+ lastDoc[0] +" but skipTo("+(lastDoc[0]+1)+") got to "+scorer.docID(),more);
} }
indexSearcher.close();
} }
} }
} }

View File

@ -37,6 +37,7 @@ import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.document.Field; import org.apache.lucene.document.Field;
import org.apache.lucene.document.FieldType; import org.apache.lucene.document.FieldType;
import org.apache.lucene.index.*; import org.apache.lucene.index.*;
import org.apache.lucene.index.IndexReader.ReaderFinishedListener;
import org.apache.lucene.index.codecs.Codec; import org.apache.lucene.index.codecs.Codec;
import org.apache.lucene.index.codecs.PostingsFormat; import org.apache.lucene.index.codecs.PostingsFormat;
import org.apache.lucene.index.codecs.appending.AppendingCodec; import org.apache.lucene.index.codecs.appending.AppendingCodec;
@ -731,7 +732,8 @@ public abstract class LuceneTestCase extends Assert {
rogueThreads.put(t, true); rogueThreads.put(t, true);
rogueCount++; rogueCount++;
if (t.getName().startsWith("LuceneTestCase")) { if (t.getName().startsWith("LuceneTestCase")) {
System.err.println("PLEASE CLOSE YOUR INDEXSEARCHERS IN YOUR TEST!!!!"); // TODO: should we fail here now? really test should be failing?
System.err.println("PLEASE CLOSE YOUR INDEXREADERS IN YOUR TEST!!!!");
continue; continue;
} else { } else {
// wait on the thread to die of natural causes // wait on the thread to die of natural causes
@ -1228,23 +1230,25 @@ public abstract class LuceneTestCase extends Assert {
final ExecutorService ex = (random.nextBoolean()) ? null final ExecutorService ex = (random.nextBoolean()) ? null
: Executors.newFixedThreadPool(threads = _TestUtil.nextInt(random, 1, 8), : Executors.newFixedThreadPool(threads = _TestUtil.nextInt(random, 1, 8),
new NamedThreadFactory("LuceneTestCase")); new NamedThreadFactory("LuceneTestCase"));
if (ex != null && VERBOSE) { if (ex != null) {
if (VERBOSE) {
System.out.println("NOTE: newSearcher using ExecutorService with " + threads + " threads"); System.out.println("NOTE: newSearcher using ExecutorService with " + threads + " threads");
}
final IndexReader r0 = r;
r.addReaderFinishedListener(new ReaderFinishedListener() {
@Override
public void finished(IndexReader reader) {
// readerFinishedListener bogusly calls us with other random readers
// so we must check that its *actually* the one we registered it on.
if (reader == r0) {
shutdownExecutorService(ex);
}
}
});
} }
IndexSearcher ret = random.nextBoolean() ? IndexSearcher ret = random.nextBoolean()
new AssertingIndexSearcher(random, r, ex) { ? new AssertingIndexSearcher(random, r, ex)
@Override : new AssertingIndexSearcher(random, r.getTopReaderContext(), ex);
public void close() throws IOException {
super.close();
shutdownExecutorService(ex);
}
} : new AssertingIndexSearcher(random, r.getTopReaderContext(), ex) {
@Override
public void close() throws IOException {
super.close();
shutdownExecutorService(ex);
}
};
ret.setSimilarityProvider(similarityProvider); ret.setSimilarityProvider(similarityProvider);
return ret; return ret;
} }

View File

@ -73,7 +73,6 @@ public class TestDemo extends LuceneTestCase {
phraseQuery.add(new Term("fieldname", "be")); phraseQuery.add(new Term("fieldname", "be"));
assertEquals(1, isearcher.search(phraseQuery, null, 1).totalHits); assertEquals(1, isearcher.search(phraseQuery, null, 1).totalHits);
isearcher.close();
ireader.close(); ireader.close();
directory.close(); directory.close();
} }

View File

@ -96,7 +96,6 @@ public class TestExternalCodecs extends LuceneTestCase {
assertEquals(NUM_DOCS-1, s.search(new TermQuery(new Term("field1", "standard")), 1).totalHits); assertEquals(NUM_DOCS-1, s.search(new TermQuery(new Term("field1", "standard")), 1).totalHits);
assertEquals(NUM_DOCS-1, s.search(new TermQuery(new Term("field2", "pulsing")), 1).totalHits); assertEquals(NUM_DOCS-1, s.search(new TermQuery(new Term("field2", "pulsing")), 1).totalHits);
r.close(); r.close();
s.close();
if (VERBOSE) { if (VERBOSE) {
System.out.println("\nTEST: now delete 2nd doc"); System.out.println("\nTEST: now delete 2nd doc");
@ -114,7 +113,6 @@ public class TestExternalCodecs extends LuceneTestCase {
assertEquals(0, s.search(new TermQuery(new Term("id", "44")), 1).totalHits); assertEquals(0, s.search(new TermQuery(new Term("id", "44")), 1).totalHits);
r.close(); r.close();
s.close();
w.close(); w.close();

View File

@ -122,7 +122,6 @@ public class TestSearch extends LuceneTestCase {
out.println(i + " " + hits[i].score + " " + d.get("contents")); out.println(i + " " + hits[i].score + " " + d.get("contents"));
} }
} }
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }

View File

@ -115,8 +115,6 @@ public class TestSearchForDuplicates extends LuceneTestCase {
printHits(out, hits, searcher); printHits(out, hits, searcher);
checkHits(hits, MAX_DOCS, searcher); checkHits(hits, MAX_DOCS, searcher);
searcher.close();
// try a new search with OR // try a new search with OR
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
hits = null; hits = null;
@ -130,7 +128,6 @@ public class TestSearchForDuplicates extends LuceneTestCase {
printHits(out, hits, searcher); printHits(out, hits, searcher);
checkHits(hits, MAX_DOCS, searcher); checkHits(hits, MAX_DOCS, searcher);
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }

View File

@ -174,7 +174,6 @@ public class TestDocument extends LuceneTestCase {
doAssert(searcher.doc(hits[0].doc), true); doAssert(searcher.doc(hits[0].doc), true);
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
} }
@ -270,7 +269,6 @@ public class TestDocument extends LuceneTestCase {
else fail("unexpected id field"); else fail("unexpected id field");
} }
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
assertEquals("did not see all IDs", 7, result); assertEquals("did not see all IDs", 7, result);

View File

@ -336,7 +336,6 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
hits = searcher.search(new TermQuery(new Term("utf8", "ab\ud917\udc17cd")), null, 1000).scoreDocs; hits = searcher.search(new TermQuery(new Term("utf8", "ab\ud917\udc17cd")), null, 1000).scoreDocs;
assertEquals(34, hits.length); assertEquals(34, hits.length);
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
} }
@ -374,7 +373,6 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
Document d = searcher.getIndexReader().document(hits[0].doc); Document d = searcher.getIndexReader().document(hits[0].doc);
assertEquals("wrong first document", "21", d.get("id")); assertEquals("wrong first document", "21", d.get("id"));
doTestHits(hits, 44, searcher.getIndexReader()); doTestHits(hits, 44, searcher.getIndexReader());
searcher.close();
reader.close(); reader.close();
// fully merge // fully merge
@ -389,7 +387,6 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
d = searcher.doc(hits[0].doc); d = searcher.doc(hits[0].doc);
doTestHits(hits, 44, searcher.getIndexReader()); doTestHits(hits, 44, searcher.getIndexReader());
assertEquals("wrong first document", "21", d.get("id")); assertEquals("wrong first document", "21", d.get("id"));
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
@ -406,7 +403,6 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
assertEquals("wrong number of hits", 34, hits.length); assertEquals("wrong number of hits", 34, hits.length);
Document d = searcher.doc(hits[0].doc); Document d = searcher.doc(hits[0].doc);
assertEquals("wrong first document", "21", d.get("id")); assertEquals("wrong first document", "21", d.get("id"));
searcher.close();
reader.close(); reader.close();
// fully merge // fully merge
@ -419,7 +415,6 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
hits = searcher.search(new TermQuery(new Term("content", "aaa")), null, 1000).scoreDocs; hits = searcher.search(new TermQuery(new Term("content", "aaa")), null, 1000).scoreDocs;
assertEquals("wrong number of hits", 34, hits.length); assertEquals("wrong number of hits", 34, hits.length);
doTestHits(hits, 34, searcher.getIndexReader()); doTestHits(hits, 34, searcher.getIndexReader());
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
@ -664,7 +659,6 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
assertTrue("value in id bounds", val >= 0L && val < 35L); assertTrue("value in id bounds", val >= 0L && val < 35L);
} }
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
_TestUtil.rmDir(oldIndexDir); _TestUtil.rmDir(oldIndexDir);

View File

@ -70,7 +70,6 @@ public class TestBinaryTerms extends LuceneTestCase {
assertEquals("" + i, is.doc(docs.scoreDocs[0].doc).get("id")); assertEquals("" + i, is.doc(docs.scoreDocs[0].doc).get("id"));
} }
is.close();
ir.close(); ir.close();
dir.close(); dir.close();
} }

View File

@ -393,7 +393,6 @@ public class TestCodecs extends LuceneTestCase {
return searcher.search(q, null, n).scoreDocs; return searcher.search(q, null, n).scoreDocs;
} }
finally { finally {
searcher.close();
reader.close(); reader.close();
} }
} }

View File

@ -126,7 +126,6 @@ public class TestCrashCausesCorruptIndex extends LuceneTestCase {
TopDocs topDocs = indexSearcher.search(new TermQuery(new Term(TEXT_FIELD, "fleas")), 10); TopDocs topDocs = indexSearcher.search(new TermQuery(new Term(TEXT_FIELD, "fleas")), 10);
assertNotNull(topDocs); assertNotNull(topDocs);
assertEquals(expectedTotalHits, topDocs.totalHits); assertEquals(expectedTotalHits, topDocs.totalHits);
indexSearcher.close();
indexReader.close(); indexReader.close();
realDirectory.close(); realDirectory.close();
} }

View File

@ -660,7 +660,6 @@ public class TestDeletionPolicy extends LuceneTestCase {
ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals(16, hits.length); assertEquals(16, hits.length);
reader.close(); reader.close();
searcher.close();
writer = new IndexWriter(dir, newIndexWriterConfig( writer = new IndexWriter(dir, newIndexWriterConfig(
TEST_VERSION_CURRENT, new MockAnalyzer(random)) TEST_VERSION_CURRENT, new MockAnalyzer(random))
@ -685,7 +684,6 @@ public class TestDeletionPolicy extends LuceneTestCase {
dir.deleteFile(IndexFileNames.SEGMENTS_GEN); dir.deleteFile(IndexFileNames.SEGMENTS_GEN);
int expectedCount = 0; int expectedCount = 0;
searcher.close();
rwReader.close(); rwReader.close();
for(int i=0;i<N+1;i++) { for(int i=0;i<N+1;i++) {
@ -697,7 +695,6 @@ public class TestDeletionPolicy extends LuceneTestCase {
searcher = newSearcher(reader); searcher = newSearcher(reader);
hits = searcher.search(query, null, 1000).scoreDocs; hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals(expectedCount, hits.length); assertEquals(expectedCount, hits.length);
searcher.close();
if (expectedCount == 0) { if (expectedCount == 0) {
expectedCount = 16; expectedCount = 16;
} else if (expectedCount == 16) { } else if (expectedCount == 16) {

View File

@ -75,7 +75,6 @@ public class TestForTooMuchCloning extends LuceneTestCase {
final int queryCloneCount = dir.getInputCloneCount() - cloneCount; final int queryCloneCount = dir.getInputCloneCount() - cloneCount;
//System.out.println("query clone count=" + queryCloneCount); //System.out.println("query clone count=" + queryCloneCount);
assertTrue("too many calls to IndexInput.clone during TermRangeQuery: " + queryCloneCount, queryCloneCount < 50); assertTrue("too many calls to IndexInput.clone during TermRangeQuery: " + queryCloneCount, queryCloneCount < 50);
s.close();
r.close(); r.close();
dir.close(); dir.close();
} }

View File

@ -515,7 +515,6 @@ public class TestIndexReaderReopen extends LuceneTestCase {
if (hits.length > 0) { if (hits.length > 0) {
searcher.doc(hits[0].doc); searcher.doc(hits[0].doc);
} }
searcher.close();
if (refreshed != r) { if (refreshed != r) {
refreshed.close(); refreshed.close();
} }

View File

@ -456,7 +456,6 @@ public class TestIndexWriter extends LuceneTestCase {
IndexSearcher searcher = new IndexSearcher(reader); IndexSearcher searcher = new IndexSearcher(reader);
ScoreDoc[] hits = searcher.search(new TermQuery(new Term("field", "aaa")), null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(new TermQuery(new Term("field", "aaa")), null, 1000).scoreDocs;
assertEquals(300, hits.length); assertEquals(300, hits.length);
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
@ -488,7 +487,6 @@ public class TestIndexWriter extends LuceneTestCase {
IndexSearcher searcher = new IndexSearcher(reader); IndexSearcher searcher = new IndexSearcher(reader);
ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals(10, hits.length); assertEquals(10, hits.length);
searcher.close();
reader.close(); reader.close();
writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)) writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))
@ -511,7 +509,6 @@ public class TestIndexWriter extends LuceneTestCase {
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals(27, hits.length); assertEquals(27, hits.length);
searcher.close();
reader.close(); reader.close();
reader = IndexReader.open(dir); reader = IndexReader.open(dir);
@ -590,7 +587,6 @@ public class TestIndexWriter extends LuceneTestCase {
IndexSearcher searcher = new IndexSearcher(reader); IndexSearcher searcher = new IndexSearcher(reader);
ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("did not get right number of hits", 100, hits.length); assertEquals("did not get right number of hits", 100, hits.length);
searcher.close();
reader.close(); reader.close();
writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)) writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))
@ -1013,7 +1009,6 @@ public class TestIndexWriter extends LuceneTestCase {
assertEquals(0, tps.nextPosition()); assertEquals(0, tps.nextPosition());
w.close(); w.close();
s.close();
r.close(); r.close();
dir.close(); dir.close();
} }
@ -1708,7 +1703,6 @@ public class TestIndexWriter extends LuceneTestCase {
assertEquals("doc " + testID + ", field f" + fieldCount + " is wrong", docExp.get("f"+i), doc.get("f"+i)); assertEquals("doc " + testID + ", field f" + fieldCount + " is wrong", docExp.get("f"+i), doc.get("f"+i));
} }
} }
s.close();
r.close(); r.close();
w.forceMerge(1); w.forceMerge(1);
} }

View File

@ -55,7 +55,6 @@ public class TestIndexWriterCommit extends LuceneTestCase {
IndexSearcher searcher = new IndexSearcher(reader); IndexSearcher searcher = new IndexSearcher(reader);
ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("first number of hits", 14, hits.length); assertEquals("first number of hits", 14, hits.length);
searcher.close();
reader.close(); reader.close();
reader = IndexReader.open(dir); reader = IndexReader.open(dir);
@ -69,7 +68,6 @@ public class TestIndexWriterCommit extends LuceneTestCase {
searcher = new IndexSearcher(r); searcher = new IndexSearcher(r);
hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("reader incorrectly sees changes from writer", 14, hits.length); assertEquals("reader incorrectly sees changes from writer", 14, hits.length);
searcher.close();
r.close(); r.close();
assertTrue("reader should have still been current", reader.isCurrent()); assertTrue("reader should have still been current", reader.isCurrent());
} }
@ -82,7 +80,6 @@ public class TestIndexWriterCommit extends LuceneTestCase {
searcher = new IndexSearcher(r); searcher = new IndexSearcher(r);
hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("reader did not see changes after writer was closed", 47, hits.length); assertEquals("reader did not see changes after writer was closed", 47, hits.length);
searcher.close();
r.close(); r.close();
reader.close(); reader.close();
dir.close(); dir.close();
@ -109,7 +106,6 @@ public class TestIndexWriterCommit extends LuceneTestCase {
IndexSearcher searcher = new IndexSearcher(reader); IndexSearcher searcher = new IndexSearcher(reader);
ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("first number of hits", 14, hits.length); assertEquals("first number of hits", 14, hits.length);
searcher.close();
reader.close(); reader.close();
writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)) writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))
@ -124,7 +120,6 @@ public class TestIndexWriterCommit extends LuceneTestCase {
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("reader incorrectly sees changes from writer", 14, hits.length); assertEquals("reader incorrectly sees changes from writer", 14, hits.length);
searcher.close();
reader.close(); reader.close();
// Now, close the writer: // Now, close the writer:
@ -136,7 +131,6 @@ public class TestIndexWriterCommit extends LuceneTestCase {
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("saw changes after writer.abort", 14, hits.length); assertEquals("saw changes after writer.abort", 14, hits.length);
searcher.close();
reader.close(); reader.close();
// Now make sure we can re-open the index, add docs, // Now make sure we can re-open the index, add docs,
@ -156,7 +150,6 @@ public class TestIndexWriterCommit extends LuceneTestCase {
searcher = new IndexSearcher(r); searcher = new IndexSearcher(r);
hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("reader incorrectly sees changes from writer", 14, hits.length); assertEquals("reader incorrectly sees changes from writer", 14, hits.length);
searcher.close();
r.close(); r.close();
} }
@ -165,7 +158,6 @@ public class TestIndexWriterCommit extends LuceneTestCase {
searcher = new IndexSearcher(r); searcher = new IndexSearcher(r);
hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("didn't see changes after close", 218, hits.length); assertEquals("didn't see changes after close", 218, hits.length);
searcher.close();
r.close(); r.close();
dir.close(); dir.close();

View File

@ -403,7 +403,6 @@ public class TestIndexWriterDelete extends LuceneTestCase {
IndexReader reader = IndexReader.open(dir); IndexReader reader = IndexReader.open(dir);
IndexSearcher searcher = new IndexSearcher(reader); IndexSearcher searcher = new IndexSearcher(reader);
int hitCount = searcher.search(new TermQuery(term), null, 1000).totalHits; int hitCount = searcher.search(new TermQuery(term), null, 1000).totalHits;
searcher.close();
reader.close(); reader.close();
return hitCount; return hitCount;
} }
@ -612,7 +611,6 @@ public class TestIndexWriterDelete extends LuceneTestCase {
+ result2 + " instead of expected " + START_COUNT + " or " + END_COUNT); + result2 + " instead of expected " + START_COUNT + " or " + END_COUNT);
} }
} }
searcher.close();
newReader.close(); newReader.close();
if (result2 == END_COUNT) { if (result2 == END_COUNT) {
break; break;

View File

@ -195,7 +195,6 @@ public class TestIndexWriterOnDiskFull extends LuceneTestCase {
IndexSearcher searcher = newSearcher(reader); IndexSearcher searcher = newSearcher(reader);
ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(new TermQuery(searchTerm), null, 1000).scoreDocs;
assertEquals("first number of hits", 57, hits.length); assertEquals("first number of hits", 57, hits.length);
searcher.close();
reader.close(); reader.close();
// Iterate with larger and larger amounts of free // Iterate with larger and larger amounts of free
@ -395,7 +394,6 @@ public class TestIndexWriterOnDiskFull extends LuceneTestCase {
} }
} }
searcher.close();
reader.close(); reader.close();
if (VERBOSE) { if (VERBOSE) {
System.out.println(" count is " + result); System.out.println(" count is " + result);

View File

@ -695,7 +695,6 @@ public class TestIndexWriterReader extends LuceneTestCase {
Query q = new TermQuery(new Term("indexname", "test")); Query q = new TermQuery(new Term("indexname", "test"));
IndexSearcher searcher = newSearcher(r); IndexSearcher searcher = newSearcher(r);
assertEquals(100, searcher.search(q, 10).totalHits); assertEquals(100, searcher.search(q, 10).totalHits);
searcher.close();
try { try {
IndexReader.openIfChanged(r); IndexReader.openIfChanged(r);
fail("failed to hit AlreadyClosedException"); fail("failed to hit AlreadyClosedException");
@ -761,7 +760,6 @@ public class TestIndexWriterReader extends LuceneTestCase {
Query q = new TermQuery(new Term("indexname", "test")); Query q = new TermQuery(new Term("indexname", "test"));
IndexSearcher searcher = newSearcher(r); IndexSearcher searcher = newSearcher(r);
final int count = searcher.search(q, 10).totalHits; final int count = searcher.search(q, 10).totalHits;
searcher.close();
assertTrue(count >= lastCount); assertTrue(count >= lastCount);
lastCount = count; lastCount = count;
} }
@ -778,7 +776,6 @@ public class TestIndexWriterReader extends LuceneTestCase {
Query q = new TermQuery(new Term("indexname", "test")); Query q = new TermQuery(new Term("indexname", "test"));
IndexSearcher searcher = newSearcher(r); IndexSearcher searcher = newSearcher(r);
final int count = searcher.search(q, 10).totalHits; final int count = searcher.search(q, 10).totalHits;
searcher.close();
assertTrue(count >= lastCount); assertTrue(count >= lastCount);
assertEquals(0, excs.size()); assertEquals(0, excs.size());
@ -850,7 +847,6 @@ public class TestIndexWriterReader extends LuceneTestCase {
Query q = new TermQuery(new Term("indexname", "test")); Query q = new TermQuery(new Term("indexname", "test"));
IndexSearcher searcher = newSearcher(r); IndexSearcher searcher = newSearcher(r);
sum += searcher.search(q, 10).totalHits; sum += searcher.search(q, 10).totalHits;
searcher.close();
} }
for(int i=0;i<numThreads;i++) { for(int i=0;i<numThreads;i++) {
@ -865,7 +861,6 @@ public class TestIndexWriterReader extends LuceneTestCase {
Query q = new TermQuery(new Term("indexname", "test")); Query q = new TermQuery(new Term("indexname", "test"));
IndexSearcher searcher = newSearcher(r); IndexSearcher searcher = newSearcher(r);
sum += searcher.search(q, 10).totalHits; sum += searcher.search(q, 10).totalHits;
searcher.close();
assertTrue("no documents found at all", sum > 0); assertTrue("no documents found at all", sum > 0);
assertEquals(0, excs.size()); assertEquals(0, excs.size());
@ -954,7 +949,6 @@ public class TestIndexWriterReader extends LuceneTestCase {
TopDocs hits = s.search(new TermQuery(new Term("foo", "bar")), 10); TopDocs hits = s.search(new TermQuery(new Term("foo", "bar")), 10);
assertEquals(20, hits.totalHits); assertEquals(20, hits.totalHits);
didWarm.set(true); didWarm.set(true);
s.close();
} }
}). }).
setMergePolicy(newLogMergePolicy(10)) setMergePolicy(newLogMergePolicy(10))

View File

@ -138,9 +138,7 @@ public class TestLazyProxSkipping extends LuceneTestCase {
// test whether only the minimum amount of seeks() // test whether only the minimum amount of seeks()
// are performed // are performed
performTest(5); performTest(5);
searcher.close();
performTest(10); performTest(10);
searcher.close();
} }
public void testSeek() throws IOException { public void testSeek() throws IOException {

View File

@ -100,7 +100,6 @@ public class TestNRTThreads extends ThreadedIndexingAndSearchingTestCase {
if (s != fixedSearcher) { if (s != fixedSearcher) {
// Final searcher: // Final searcher:
s.getIndexReader().close(); s.getIndexReader().close();
s.close();
} }
} }

View File

@ -400,8 +400,7 @@ public class TestOmitTf extends LuceneTestCase {
} }
}); });
assertEquals(15, CountingHitCollector.getCount()); assertEquals(15, CountingHitCollector.getCount());
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
} }

View File

@ -46,9 +46,7 @@ public class TestParallelReader extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
single.getIndexReader().close(); single.getIndexReader().close();
single.close();
parallel.getIndexReader().close(); parallel.getIndexReader().close();
parallel.close();
dir.close(); dir.close();
dir1.close(); dir1.close();
dir2.close(); dir2.close();

View File

@ -61,7 +61,6 @@ public class TestReaderClosed extends LuceneTestCase {
public void test() throws Exception { public void test() throws Exception {
TermRangeQuery query = TermRangeQuery.newStringRange("field", "a", "z", true, true); TermRangeQuery query = TermRangeQuery.newStringRange("field", "a", "z", true, true);
searcher.search(query, 5); searcher.search(query, 5);
searcher.close();
reader.close(); reader.close();
try { try {
searcher.search(query, 5); searcher.search(query, 5);

View File

@ -105,7 +105,6 @@ public class TestStressIndexing extends LuceneTestCase {
for (int i=0; i<100; i++) { for (int i=0; i<100; i++) {
IndexReader ir = IndexReader.open(directory); IndexReader ir = IndexReader.open(directory);
IndexSearcher is = new IndexSearcher(ir); IndexSearcher is = new IndexSearcher(ir);
is.close();
ir.close(); ir.close();
} }
count += 100; count += 100;

View File

@ -84,7 +84,6 @@ public class TestTermsEnum2 extends LuceneTestCase {
} }
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
super.tearDown(); super.tearDown();

View File

@ -136,7 +136,6 @@ public class TestTermInfosReaderIndex extends LuceneTestCase {
TopDocs topDocs = searcher.search(query, 10); TopDocs topDocs = searcher.search(query, 10);
assertTrue(topDocs.totalHits > 0); assertTrue(topDocs.totalHits > 0);
} }
searcher.close();
} }
private List<Term> sample(IndexReader reader, int size) throws IOException { private List<Term> sample(IndexReader reader, int size) throws IOException {

View File

@ -194,7 +194,6 @@ public class TestPerFieldPostingsFormat extends LuceneTestCase {
IndexSearcher searcher = newSearcher(reader); IndexSearcher searcher = newSearcher(reader);
TopDocs search = searcher.search(new TermQuery(t), num + 10); TopDocs search = searcher.search(new TermQuery(t), num + 10);
assertEquals(num, search.totalHits); assertEquals(num, search.totalHits);
searcher.close();
reader.close(); reader.close();
} }

View File

@ -67,7 +67,6 @@ public class TestAutomatonQuery extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();

View File

@ -86,7 +86,6 @@ public class TestAutomatonQueryUnicode extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();

View File

@ -101,12 +101,10 @@ public class TestBoolean2 extends LuceneTestCase {
@AfterClass @AfterClass
public static void afterClass() throws Exception { public static void afterClass() throws Exception {
searcher.close();
reader.close(); reader.close();
littleReader.close(); littleReader.close();
dir2.close(); dir2.close();
directory.close(); directory.close();
bigSearcher.close();
searcher = null; searcher = null;
reader = null; reader = null;
littleReader = null; littleReader = null;

View File

@ -73,7 +73,6 @@ public class TestBooleanMinShouldMatch extends LuceneTestCase {
@AfterClass @AfterClass
public static void afterClass() throws Exception { public static void afterClass() throws Exception {
s.close();
s = null; s = null;
r.close(); r.close();
r = null; r = null;

View File

@ -158,7 +158,6 @@ public class TestBooleanOr extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
super.tearDown(); super.tearDown();

View File

@ -140,7 +140,6 @@ public class TestBooleanQuery extends LuceneTestCase {
dmq.add(pq); dmq.add(pq);
assertEquals(1, s.search(dmq, 10).totalHits); assertEquals(1, s.search(dmq, 10).totalHits);
s.close();
r.close(); r.close();
w.close(); w.close();
dir.close(); dir.close();

View File

@ -62,7 +62,6 @@ public class TestBooleanScorer extends LuceneTestCase
IndexSearcher indexSearcher = newSearcher(ir); IndexSearcher indexSearcher = newSearcher(ir);
ScoreDoc[] hits = indexSearcher.search(query, null, 1000).scoreDocs; ScoreDoc[] hits = indexSearcher.search(query, null, 1000).scoreDocs;
assertEquals("Number of matched documents", 2, hits.length); assertEquals("Number of matched documents", 2, hits.length);
indexSearcher.close();
ir.close(); ir.close();
directory.close(); directory.close();
} }
@ -122,7 +121,6 @@ public class TestBooleanScorer extends LuceneTestCase
assertEquals("should have only 1 hit", 1, hits.size()); assertEquals("should have only 1 hit", 1, hits.size());
assertEquals("hit should have been docID=3000", 3000, hits.get(0).intValue()); assertEquals("hit should have been docID=3000", 3000, hits.get(0).intValue());
searcher.close();
ir.close(); ir.close();
directory.close(); directory.close();
} }
@ -176,7 +174,6 @@ public class TestBooleanScorer extends LuceneTestCase
assertEquals(1, count[0]); assertEquals(1, count[0]);
s.close();
r.close(); r.close();
d.close(); d.close();
} }

View File

@ -179,7 +179,6 @@ public class TestCachingWrapperFilter extends LuceneTestCase {
writer.addDocument(doc); writer.addDocument(doc);
reader = refreshReader(reader); reader = refreshReader(reader);
searcher.close();
searcher = newSearcher(reader, false); searcher = newSearcher(reader, false);
TopDocs docs = searcher.search(new MatchAllDocsQuery(), 1); TopDocs docs = searcher.search(new MatchAllDocsQuery(), 1);
@ -205,7 +204,6 @@ public class TestCachingWrapperFilter extends LuceneTestCase {
writer.addDocument(doc); writer.addDocument(doc);
reader = refreshReader(reader); reader = refreshReader(reader);
searcher.close();
searcher = newSearcher(reader, false); searcher = newSearcher(reader, false);
docs = searcher.search(new MatchAllDocsQuery(), filter, 1); docs = searcher.search(new MatchAllDocsQuery(), filter, 1);
@ -227,7 +225,6 @@ public class TestCachingWrapperFilter extends LuceneTestCase {
writer.deleteDocuments(new Term("id", "1")); writer.deleteDocuments(new Term("id", "1"));
reader = refreshReader(reader); reader = refreshReader(reader);
searcher.close();
searcher = newSearcher(reader, false); searcher = newSearcher(reader, false);
docs = searcher.search(new MatchAllDocsQuery(), filter, 1); docs = searcher.search(new MatchAllDocsQuery(), filter, 1);
@ -245,7 +242,6 @@ public class TestCachingWrapperFilter extends LuceneTestCase {
assertTrue(oldReader != null); assertTrue(oldReader != null);
assertTrue(oldReader2 != null); assertTrue(oldReader2 != null);
searcher.close();
reader.close(); reader.close();
writer.close(); writer.close();
dir.close(); dir.close();

View File

@ -125,7 +125,6 @@ public class TestConstantScoreQuery extends LuceneTestCase {
checkHits(searcher, bq, csq1.getBoost() + csq2.getBoost(), bucketScorerClass, null); checkHits(searcher, bq, csq1.getBoost() + csq2.getBoost(), bucketScorerClass, null);
checkHits(searcher, csqbq, csqbq.getBoost(), ConstantScoreQuery.ConstantScorer.class.getName(), bucketScorerClass); checkHits(searcher, csqbq, csqbq.getBoost(), ConstantScoreQuery.ConstantScorer.class.getName(), bucketScorerClass);
} finally { } finally {
if (searcher != null) searcher.close();
if (reader != null) reader.close(); if (reader != null) reader.close();
if (directory != null) directory.close(); if (directory != null) directory.close();
} }

View File

@ -97,7 +97,6 @@ public class TestDateFilter extends LuceneTestCase {
result = searcher.search(query2, df2, 1000).scoreDocs; result = searcher.search(query2, df2, 1000).scoreDocs;
assertEquals(0, result.length); assertEquals(0, result.length);
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }
@ -164,7 +163,6 @@ public class TestDateFilter extends LuceneTestCase {
result = searcher.search(query2, df2, 1000).scoreDocs; result = searcher.search(query2, df2, 1000).scoreDocs;
assertEquals(0, result.length); assertEquals(0, result.length);
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }

View File

@ -88,7 +88,6 @@ public class TestDateSort extends LuceneTestCase {
String text = document.get(TEXT_FIELD); String text = document.get(TEXT_FIELD);
actualOrder[i] = text; actualOrder[i] = text;
} }
searcher.close();
// Set up the expected order (i.e. Document 5, 4, 3, 2, 1). // Set up the expected order (i.e. Document 5, 4, 3, 2, 1).
String[] expectedOrder = new String[5]; String[] expectedOrder = new String[5];

View File

@ -163,7 +163,6 @@ public class TestDisjunctionMaxQuery extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
s.close();
r.close(); r.close();
index.close(); index.close();
super.tearDown(); super.tearDown();

View File

@ -121,7 +121,6 @@ public class TestDocIdSet extends LuceneTestCase {
}; };
Assert.assertEquals(0, searcher.search(new MatchAllDocsQuery(), f, 10).totalHits); Assert.assertEquals(0, searcher.search(new MatchAllDocsQuery(), f, 10).totalHits);
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
} }

View File

@ -133,9 +133,6 @@ public class TestDocValuesScoring extends LuceneTestCase {
assertEquals(boost.scoreDocs[0].score, noboost.scoreDocs[0].score, SCORE_EPSILON); assertEquals(boost.scoreDocs[0].score, noboost.scoreDocs[0].score, SCORE_EPSILON);
searcher1.close();
searcher2.close();
ir.close(); ir.close();
dir.close(); dir.close();
} }

View File

@ -61,7 +61,6 @@ public class TestElevationComparator extends LuceneTestCase {
runTest(searcher, true); runTest(searcher, true);
runTest(searcher, false); runTest(searcher, false);
searcher.close();
r.close(); r.close();
directory.close(); directory.close();
} }

View File

@ -61,7 +61,6 @@ public class TestExplanations extends LuceneTestCase {
@AfterClass @AfterClass
public static void afterClassTestExplanations() throws Exception { public static void afterClassTestExplanations() throws Exception {
searcher.close();
searcher = null; searcher = null;
reader.close(); reader.close();
reader = null; reader = null;

View File

@ -122,7 +122,6 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter {
result = search.search(q,FieldCacheRangeFilter.newStringRange("id",medIP,medIP,T,T), numDocs).scoreDocs; result = search.search(q,FieldCacheRangeFilter.newStringRange("id",medIP,medIP,T,T), numDocs).scoreDocs;
assertEquals("med,med,T,T", 1, result.length); assertEquals("med,med,T,T", 1, result.length);
search.close();
} }
@Test @Test
@ -185,7 +184,6 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter {
assertEquals("max,max,T,T", 1, result.length); assertEquals("max,max,T,T", 1, result.length);
result = search.search(q,FieldCacheRangeFilter.newStringRange("rand",maxRP,null,T,F), numDocs).scoreDocs; result = search.search(q,FieldCacheRangeFilter.newStringRange("rand",maxRP,null,T,F), numDocs).scoreDocs;
assertEquals("max,nul,T,T", 1, result.length); assertEquals("max,nul,T,T", 1, result.length);
search.close();
} }
// byte-ranges cannot be tested, because all ranges are too big for bytes, need an extra range for that // byte-ranges cannot be tested, because all ranges are too big for bytes, need an extra range for that
@ -278,7 +276,6 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter {
assertEquals("overflow special case", 0, result.length); assertEquals("overflow special case", 0, result.length);
result = search.search(q,FieldCacheRangeFilter.newShortRange("id",maxIdO,minIdO,T,T), numDocs).scoreDocs; result = search.search(q,FieldCacheRangeFilter.newShortRange("id",maxIdO,minIdO,T,T), numDocs).scoreDocs;
assertEquals("inverse range", 0, result.length); assertEquals("inverse range", 0, result.length);
search.close();
} }
@Test @Test
@ -370,7 +367,6 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter {
assertEquals("overflow special case", 0, result.length); assertEquals("overflow special case", 0, result.length);
result = search.search(q,FieldCacheRangeFilter.newIntRange("id",maxIdO,minIdO,T,T), numDocs).scoreDocs; result = search.search(q,FieldCacheRangeFilter.newIntRange("id",maxIdO,minIdO,T,T), numDocs).scoreDocs;
assertEquals("inverse range", 0, result.length); assertEquals("inverse range", 0, result.length);
search.close();
} }
@Test @Test
@ -462,7 +458,6 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter {
assertEquals("overflow special case", 0, result.length); assertEquals("overflow special case", 0, result.length);
result = search.search(q,FieldCacheRangeFilter.newLongRange("id",maxIdO,minIdO,T,T), numDocs).scoreDocs; result = search.search(q,FieldCacheRangeFilter.newLongRange("id",maxIdO,minIdO,T,T), numDocs).scoreDocs;
assertEquals("inverse range", 0, result.length); assertEquals("inverse range", 0, result.length);
search.close();
} }
// float and double tests are a bit minimalistic, but its complicated, because missing precision // float and double tests are a bit minimalistic, but its complicated, because missing precision
@ -494,7 +489,6 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter {
assertEquals("infinity special case", 0, result.length); assertEquals("infinity special case", 0, result.length);
result = search.search(q,FieldCacheRangeFilter.newFloatRange("id",null,Float.valueOf(Float.NEGATIVE_INFINITY),F,F), numDocs).scoreDocs; result = search.search(q,FieldCacheRangeFilter.newFloatRange("id",null,Float.valueOf(Float.NEGATIVE_INFINITY),F,F), numDocs).scoreDocs;
assertEquals("infinity special case", 0, result.length); assertEquals("infinity special case", 0, result.length);
search.close();
} }
@Test @Test
@ -524,7 +518,6 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter {
assertEquals("infinity special case", 0, result.length); assertEquals("infinity special case", 0, result.length);
result = search.search(q,FieldCacheRangeFilter.newDoubleRange("id",null, Double.valueOf(Double.NEGATIVE_INFINITY),F,F), numDocs).scoreDocs; result = search.search(q,FieldCacheRangeFilter.newDoubleRange("id",null, Double.valueOf(Double.NEGATIVE_INFINITY),F,F), numDocs).scoreDocs;
assertEquals("infinity special case", 0, result.length); assertEquals("infinity special case", 0, result.length);
search.close();
} }
// test using a sparse index (with deleted docs). // test using a sparse index (with deleted docs).
@ -565,7 +558,6 @@ public class TestFieldCacheRangeFilter extends BaseTestRangeFilter {
result = search.search(q,FieldCacheRangeFilter.newByteRange("id",Byte.valueOf((byte) -20),Byte.valueOf((byte) -10),T,T), 100).scoreDocs; result = search.search(q,FieldCacheRangeFilter.newByteRange("id",Byte.valueOf((byte) -20),Byte.valueOf((byte) -10),T,T), 100).scoreDocs;
assertEquals("find all", 11, result.length); assertEquals("find all", 11, result.length);
search.close();
reader.close(); reader.close();
dir.close(); dir.close();
} }

View File

@ -68,7 +68,6 @@ public class TestFieldCacheTermsFilter extends LuceneTestCase {
results = searcher.search(q, new FieldCacheTermsFilter(fieldName, terms.toArray(new String[0])), numDocs).scoreDocs; results = searcher.search(q, new FieldCacheTermsFilter(fieldName, terms.toArray(new String[0])), numDocs).scoreDocs;
assertEquals("Must match 2", 2, results.length); assertEquals("Must match 2", 2, results.length);
searcher.close();
reader.close(); reader.close();
rd.close(); rd.close();
} }

View File

@ -58,7 +58,6 @@ public class TestFieldValueFilter extends LuceneTestCase {
} }
reader.close(); reader.close();
searcher.close();
directory.close(); directory.close();
} }
@ -86,7 +85,6 @@ public class TestFieldValueFilter extends LuceneTestCase {
} }
reader.close(); reader.close();
searcher.close();
directory.close(); directory.close();
} }

View File

@ -104,7 +104,6 @@ public class TestFilteredQuery extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();

View File

@ -82,7 +82,6 @@ public class TestFilteredSearch extends LuceneTestCase {
IndexSearcher indexSearcher = new IndexSearcher(reader); IndexSearcher indexSearcher = new IndexSearcher(reader);
ScoreDoc[] hits = indexSearcher.search(booleanQuery, filter, 1000).scoreDocs; ScoreDoc[] hits = indexSearcher.search(booleanQuery, filter, 1000).scoreDocs;
assertEquals("Number of matched documents", 1, hits.length); assertEquals("Number of matched documents", 1, hits.length);
indexSearcher.close();
reader.close(); reader.close();
} }
catch (IOException e) { catch (IOException e) {

View File

@ -185,7 +185,6 @@ public class TestFuzzyQuery extends LuceneTestCase {
hits = searcher.search(query, null, 1000).scoreDocs; hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals(0, hits.length); assertEquals(0, hits.length);
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }
@ -275,7 +274,6 @@ public class TestFuzzyQuery extends LuceneTestCase {
// expecting exception // expecting exception
} }
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }
@ -312,7 +310,6 @@ public class TestFuzzyQuery extends LuceneTestCase {
FuzzyQuery fq = new FuzzyQuery(new Term("field", "z123456"), 1f, 0, 2); FuzzyQuery fq = new FuzzyQuery(new Term("field", "z123456"), 1f, 0, 2);
TopDocs docs = searcher.search(fq, 2); TopDocs docs = searcher.search(fq, 2);
assertEquals(5, docs.totalHits); // 5 docs, from the a and b's assertEquals(5, docs.totalHits); // 5 docs, from the a and b's
searcher.close();
mr.close(); mr.close();
ir1.close(); ir1.close();
ir2.close(); ir2.close();
@ -353,7 +350,6 @@ public class TestFuzzyQuery extends LuceneTestCase {
hits = searcher.search(query, null, 1000).scoreDocs; hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals(0, hits.length); assertEquals(0, hits.length);
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }
@ -378,7 +374,6 @@ public class TestFuzzyQuery extends LuceneTestCase {
assertEquals("Lucene", reader.document(hits[0].doc).get("field")); assertEquals("Lucene", reader.document(hits[0].doc).get("field"));
assertEquals("Lucene", reader.document(hits[1].doc).get("field")); assertEquals("Lucene", reader.document(hits[1].doc).get("field"));
assertEquals("Lucenne", reader.document(hits[2].doc).get("field")); assertEquals("Lucenne", reader.document(hits[2].doc).get("field"));
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }
@ -416,7 +411,6 @@ public class TestFuzzyQuery extends LuceneTestCase {
ScoreDoc[] hits = searcher.search(q, 10).scoreDocs; ScoreDoc[] hits = searcher.search(q, 10).scoreDocs;
assertEquals(1, hits.length); assertEquals(1, hits.length);
assertEquals("Giga byte", searcher.doc(hits[0].doc).get("field")); assertEquals("Giga byte", searcher.doc(hits[0].doc).get("field"));
searcher.close();
r.close(); r.close();
index.close(); index.close();
} }
@ -457,7 +451,6 @@ public class TestFuzzyQuery extends LuceneTestCase {
assertEquals("test", searcher.doc(hits[0].doc).get("field")); assertEquals("test", searcher.doc(hits[0].doc).get("field"));
assertEquals("foobar", searcher.doc(hits[1].doc).get("field")); assertEquals("foobar", searcher.doc(hits[1].doc).get("field"));
searcher.close();
reader.close(); reader.close();
index.close(); index.close();
} }

View File

@ -139,7 +139,6 @@ public class TestFuzzyQuery2 extends LuceneTestCase {
assertEquals(Float.parseFloat(scoreDoc[1]), docs.scoreDocs[i].score, epsilon); assertEquals(Float.parseFloat(scoreDoc[1]), docs.scoreDocs[i].score, epsilon);
} }
} }
searcher.close();
r.close(); r.close();
dir.close(); dir.close();
} }

View File

@ -70,7 +70,6 @@ public class TestMatchAllDocsQuery extends LuceneTestCase {
assertEquals(1, hits.length); assertEquals(1, hits.length);
iw.deleteDocuments(new Term("key", "one")); iw.deleteDocuments(new Term("key", "one"));
is.close();
ir.close(); ir.close();
ir = IndexReader.open(iw, true); ir = IndexReader.open(iw, true);
is = newSearcher(ir); is = newSearcher(ir);
@ -79,7 +78,6 @@ public class TestMatchAllDocsQuery extends LuceneTestCase {
assertEquals(2, hits.length); assertEquals(2, hits.length);
iw.close(); iw.close();
is.close();
ir.close(); ir.close();
dir.close(); dir.close();
} }

View File

@ -142,7 +142,6 @@ public class TestMultiPhraseQuery extends LuceneTestCase {
} }
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }
@ -162,7 +161,6 @@ public class TestMultiPhraseQuery extends LuceneTestCase {
q.add(new Term("body", "chocolate")); q.add(new Term("body", "chocolate"));
q.add(new Term[] {new Term("body", "pie"), new Term("body", "tart")}); q.add(new Term[] {new Term("body", "pie"), new Term("body", "tart")});
assertEquals(2, searcher.search(q, 1).totalHits); assertEquals(2, searcher.search(q, 1).totalHits);
searcher.close();
r.close(); r.close();
indexStore.close(); indexStore.close();
} }
@ -205,7 +203,6 @@ public class TestMultiPhraseQuery extends LuceneTestCase {
searcher.explain(q, 0); searcher.explain(q, 0);
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }
@ -233,7 +230,6 @@ public class TestMultiPhraseQuery extends LuceneTestCase {
ScoreDoc[] hits = searcher.search(q, null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(q, null, 1000).scoreDocs;
assertEquals("Wrong number of hits", 0, hits.length); assertEquals("Wrong number of hits", 0, hits.length);
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }
@ -256,7 +252,6 @@ public class TestMultiPhraseQuery extends LuceneTestCase {
searcher.explain(q, 0); searcher.explain(q, 0);
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }
@ -330,7 +325,6 @@ public class TestMultiPhraseQuery extends LuceneTestCase {
assertEquals(10f * 10f, weight.getValueForNormalization(), 0.001f); assertEquals(10f * 10f, weight.getValueForNormalization(), 0.001f);
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }

View File

@ -161,7 +161,6 @@ public class TestMultiTermConstantScore extends BaseTestRangeFilter {
result[i].score, SCORE_COMP_THRESH); result[i].score, SCORE_COMP_THRESH);
} }
search.close();
} }
@Test @Test
@ -247,7 +246,6 @@ public class TestMultiTermConstantScore extends BaseTestRangeFilter {
Assert.assertEquals(0, hits[0].doc); Assert.assertEquals(0, hits[0].doc);
Assert.assertEquals(1, hits[1].doc); Assert.assertEquals(1, hits[1].doc);
assertTrue(hits[0].score > hits[1].score); assertTrue(hits[0].score > hits[1].score);
search.close();
} }
@Test @Test
@ -278,8 +276,6 @@ public class TestMultiTermConstantScore extends BaseTestRangeFilter {
assertEquals("mismatch in docid for hit#" + i, expected[i].doc, assertEquals("mismatch in docid for hit#" + i, expected[i].doc,
actual[i].doc); actual[i].doc);
} }
search.close();
} }
@Test @Test
@ -412,8 +408,6 @@ public class TestMultiTermConstantScore extends BaseTestRangeFilter {
result = search.search(csrq("id", medIP, medIP, T, T, MultiTermQuery.CONSTANT_SCORE_AUTO_REWRITE_DEFAULT), null, numDocs).scoreDocs; result = search.search(csrq("id", medIP, medIP, T, T, MultiTermQuery.CONSTANT_SCORE_AUTO_REWRITE_DEFAULT), null, numDocs).scoreDocs;
assertEquals("med,med,T,T", 1, result.length); assertEquals("med,med,T,T", 1, result.length);
search.close();
} }
@Test @Test
@ -476,7 +470,5 @@ public class TestMultiTermConstantScore extends BaseTestRangeFilter {
assertEquals("max,max,T,T", 1, result.length); assertEquals("max,max,T,T", 1, result.length);
result = search.search(csrq("rand", maxRP, null, T, F), null, numDocs).scoreDocs; result = search.search(csrq("rand", maxRP, null, T, F), null, numDocs).scoreDocs;
assertEquals("max,nul,T,T", 1, result.length); assertEquals("max,nul,T,T", 1, result.length);
search.close();
} }
} }

View File

@ -73,7 +73,6 @@ public class TestMultiValuedNumericRangeQuery extends LuceneTestCase {
TopDocs nrTopDocs = searcher.search(tq, 1); TopDocs nrTopDocs = searcher.search(tq, 1);
assertEquals("Returned count for NumericRangeQuery and TermRangeQuery must be equal", trTopDocs.totalHits, nrTopDocs.totalHits ); assertEquals("Returned count for NumericRangeQuery and TermRangeQuery must be equal", trTopDocs.totalHits, nrTopDocs.totalHits );
} }
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }

View File

@ -51,7 +51,6 @@ public class TestNot extends LuceneTestCase {
ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals(0, hits.length); assertEquals(0, hits.length);
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
store.close(); store.close();
} }

View File

@ -96,7 +96,6 @@ public class TestNumericRangeQuery32 extends LuceneTestCase {
@AfterClass @AfterClass
public static void afterClass() throws Exception { public static void afterClass() throws Exception {
searcher.close();
searcher = null; searcher = null;
reader.close(); reader.close();
reader = null; reader = null;
@ -335,7 +334,6 @@ public class TestNumericRangeQuery32 extends LuceneTestCase {
topDocs = s.search(q, 10); topDocs = s.search(q, 10);
assertEquals("Score doc count", TestNumericUtils.FLOAT_NANs.length, topDocs.scoreDocs.length ); assertEquals("Score doc count", TestNumericUtils.FLOAT_NANs.length, topDocs.scoreDocs.length );
s.close();
r.close(); r.close();
dir.close(); dir.close();
} }

View File

@ -99,7 +99,6 @@ public class TestNumericRangeQuery64 extends LuceneTestCase {
@AfterClass @AfterClass
public static void afterClass() throws Exception { public static void afterClass() throws Exception {
searcher.close();
searcher = null; searcher = null;
reader.close(); reader.close();
reader = null; reader = null;
@ -354,7 +353,6 @@ public class TestNumericRangeQuery64 extends LuceneTestCase {
topDocs = s.search(q, 10); topDocs = s.search(q, 10);
assertEquals("Score doc count", TestNumericUtils.DOUBLE_NANs.length, topDocs.scoreDocs.length ); assertEquals("Score doc count", TestNumericUtils.DOUBLE_NANs.length, topDocs.scoreDocs.length );
s.close();
r.close(); r.close();
dir.close(); dir.close();
} }

View File

@ -93,7 +93,6 @@ public class TestPhrasePrefixQuery extends LuceneTestCase {
result = searcher.search(query2, null, 1000).scoreDocs; result = searcher.search(query2, null, 1000).scoreDocs;
assertEquals(0, result.length); assertEquals(0, result.length);
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }

View File

@ -98,7 +98,6 @@ public class TestPhraseQuery extends LuceneTestCase {
@AfterClass @AfterClass
public static void afterClass() throws Exception { public static void afterClass() throws Exception {
searcher.close();
searcher = null; searcher = null;
reader.close(); reader.close();
reader = null; reader = null;
@ -249,7 +248,6 @@ public class TestPhraseQuery extends LuceneTestCase {
QueryUtils.check(random, query,searcher); QueryUtils.check(random, query,searcher);
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }
@ -289,7 +287,6 @@ public class TestPhraseQuery extends LuceneTestCase {
QueryUtils.check(random, termQuery,searcher); QueryUtils.check(random, termQuery,searcher);
searcher.close();
reader.close(); reader.close();
writer = new RandomIndexWriter(random, directory, writer = new RandomIndexWriter(random, directory,
@ -336,7 +333,6 @@ public class TestPhraseQuery extends LuceneTestCase {
QueryUtils.check(random, booleanQuery,searcher); QueryUtils.check(random, booleanQuery,searcher);
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }
@ -380,7 +376,6 @@ public class TestPhraseQuery extends LuceneTestCase {
assertEquals(0.31, hits[2].score, 0.01); assertEquals(0.31, hits[2].score, 0.01);
assertEquals(2, hits[2].doc); assertEquals(2, hits[2].doc);
QueryUtils.check(random, query,searcher); QueryUtils.check(random, query,searcher);
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }
@ -695,7 +690,6 @@ public class TestPhraseQuery extends LuceneTestCase {
} }
reader.close(); reader.close();
s.close();
dir.close(); dir.close();
} }
} }

View File

@ -196,7 +196,6 @@ public class TestPositionIncrement extends LuceneTestCase {
hits = searcher.search(q, null, 1000).scoreDocs; hits = searcher.search(q, null, 1000).scoreDocs;
assertEquals(0, hits.length); assertEquals(0, hits.length);
searcher.close();
reader.close(); reader.close();
store.close(); store.close();
} }

View File

@ -89,7 +89,6 @@ public class TestPositiveScoresOnlyCollector extends LuceneTestCase {
for (int i = 0; i < sd.length; i++) { for (int i = 0; i < sd.length; i++) {
assertTrue("only positive scores should return: " + sd[i].score, sd[i].score > 0); assertTrue("only positive scores should return: " + sd[i].score, sd[i].score > 0);
} }
searcher.close();
ir.close(); ir.close();
directory.close(); directory.close();
} }

View File

@ -101,7 +101,6 @@ public class TestPrefixFilter extends LuceneTestCase {
assertEquals(0, hits.length); assertEquals(0, hits.length);
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }

View File

@ -76,7 +76,6 @@ public class TestPrefixInBooleanQuery extends LuceneTestCase {
@AfterClass @AfterClass
public static void afterClass() throws Exception { public static void afterClass() throws Exception {
searcher.close();
searcher = null; searcher = null;
reader.close(); reader.close();
reader = null; reader = null;

View File

@ -61,7 +61,6 @@ public class TestPrefixQuery extends LuceneTestCase {
hits = searcher.search(query, null, 1000).scoreDocs; hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals("everything", 3, hits.length); assertEquals("everything", 3, hits.length);
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
} }

View File

@ -75,7 +75,6 @@ public class TestPrefixRandom extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
reader.close(); reader.close();
searcher.close();
dir.close(); dir.close();
super.tearDown(); super.tearDown();
} }

View File

@ -81,7 +81,6 @@ public class TestQueryWrapperFilter extends LuceneTestCase {
assertEquals(0, hits.totalHits); assertEquals(0, hits.totalHits);
hits = searcher.search(new MatchAllDocsQuery(), new CachingWrapperFilter(qwf), 10); hits = searcher.search(new MatchAllDocsQuery(), new CachingWrapperFilter(qwf), 10);
assertEquals(0, hits.totalHits); assertEquals(0, hits.totalHits);
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
} }
@ -148,7 +147,6 @@ public class TestQueryWrapperFilter extends LuceneTestCase {
assertEquals(1, td.totalHits); assertEquals(1, td.totalHits);
} }
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
} }

View File

@ -59,7 +59,6 @@ public class TestRegexpQuery extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();

View File

@ -95,7 +95,6 @@ public class TestRegexpRandom extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
super.tearDown(); super.tearDown();

View File

@ -95,8 +95,6 @@ public class TestRegexpRandom2 extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
reader.close(); reader.close();
searcher1.close();
searcher2.close();
dir.close(); dir.close();
super.tearDown(); super.tearDown();
} }

View File

@ -119,7 +119,6 @@ public class TestScoreCachingWrappingScorer extends LuceneTestCase {
for (int i = 0; i < scores.length; i++) { for (int i = 0; i < scores.length; i++) {
assertEquals(scores[i], scc.mscores[i], 0f); assertEquals(scores[i], scc.mscores[i], 0f);
} }
searcher.close();
ir.close(); ir.close();
directory.close(); directory.close();
} }

View File

@ -323,7 +323,6 @@ public class TestScorerPerf extends LuceneTestCase {
sets=randBitSets(atLeast(1000), atLeast(10)); sets=randBitSets(atLeast(1000), atLeast(10));
doConjunctions(atLeast(10000), atLeast(5)); doConjunctions(atLeast(10000), atLeast(5));
doNestedConjunctions(atLeast(10000), atLeast(3), atLeast(3)); doNestedConjunctions(atLeast(10000), atLeast(3), atLeast(3));
s.close();
r.close(); r.close();
d.close(); d.close();
} }

View File

@ -54,7 +54,6 @@ public class TestSearchAfter extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
dir.close(); dir.close();
super.tearDown(); super.tearDown();

View File

@ -108,7 +108,6 @@ public class TestSearchWithThreads extends LuceneTestCase {
if (VERBOSE) System.out.println(NUM_SEARCH_THREADS + " threads did " + netSearch.get() + " searches"); if (VERBOSE) System.out.println(NUM_SEARCH_THREADS + " threads did " + netSearch.get() + " searches");
s.close();
r.close(); r.close();
dir.close(); dir.close();
} }

View File

@ -169,7 +169,6 @@ public class TestSimilarity extends LuceneTestCase {
} }
}); });
searcher.close();
reader.close(); reader.close();
store.close(); store.close();
} }

View File

@ -67,7 +67,6 @@ public class TestSimilarityProvider extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();

View File

@ -149,7 +149,6 @@ public class TestSloppyPhraseQuery extends LuceneTestCase {
//QueryUtils.check(query,searcher); //QueryUtils.check(query,searcher);
writer.close(); writer.close();
searcher.close();
reader.close(); reader.close();
ramDir.close(); ramDir.close();
@ -262,7 +261,6 @@ public class TestSloppyPhraseQuery extends LuceneTestCase {
assertEquals(3, is.search(pq, 4).totalHits); assertEquals(3, is.search(pq, 4).totalHits);
pq.setSlop(2); pq.setSlop(2);
assertEquals(4, is.search(pq, 4).totalHits); assertEquals(4, is.search(pq, 4).totalHits);
is.close();
ir.close(); ir.close();
dir.close(); dir.close();
} }
@ -286,7 +284,6 @@ public class TestSloppyPhraseQuery extends LuceneTestCase {
pq.add(new Term("lyrics", "drug"), 3); pq.add(new Term("lyrics", "drug"), 3);
pq.setSlop(1); pq.setSlop(1);
assertSaneScoring(pq, is); assertSaneScoring(pq, is);
is.close();
ir.close(); ir.close();
dir.close(); dir.close();
} }
@ -341,7 +338,6 @@ public class TestSloppyPhraseQuery extends LuceneTestCase {
pq.add(new Term("lyrics", "drug"), 3); pq.add(new Term("lyrics", "drug"), 3);
pq.setSlop(5); pq.setSlop(5);
assertSaneScoring(pq, is); assertSaneScoring(pq, is);
is.close();
ir.close(); ir.close();
dir.close(); dir.close();
} }

View File

@ -334,9 +334,6 @@ public class TestSort extends LuceneTestCase {
full.reader.close(); full.reader.close();
searchX.reader.close(); searchX.reader.close();
searchY.reader.close(); searchY.reader.close();
full.close();
searchX.close();
searchY.close();
for (Directory dir : dirs) for (Directory dir : dirs)
dir.close(); dir.close();
super.tearDown(); super.tearDown();
@ -559,7 +556,6 @@ public class TestSort extends LuceneTestCase {
} }
assertFalse("Found sort results out of order", fail); assertFalse("Found sort results out of order", fail);
searcher.getIndexReader().close(); searcher.getIndexReader().close();
searcher.close();
} }
/** /**
@ -816,7 +812,6 @@ public class TestSort extends LuceneTestCase {
new SortField ("string", SortField.Type.STRING), new SortField ("string", SortField.Type.STRING),
new SortField ("float", SortField.Type.FLOAT, true) ); new SortField ("float", SortField.Type.FLOAT, true) );
assertMatches (parallelSearcher, queryG, sort, "ZYXW"); assertMatches (parallelSearcher, queryG, sort, "ZYXW");
parallelSearcher.close();
exec.shutdown(); exec.shutdown();
exec.awaitTermination(1000, TimeUnit.MILLISECONDS); exec.awaitTermination(1000, TimeUnit.MILLISECONDS);
} }
@ -855,7 +850,6 @@ public class TestSort extends LuceneTestCase {
new IndexReader[] {searchX.getIndexReader(), new IndexReader[] {searchX.getIndexReader(),
searchY.getIndexReader()}), exec); searchY.getIndexReader()}), exec);
runMultiSorts(searcher, false); runMultiSorts(searcher, false);
searcher.close();
exec.shutdown(); exec.shutdown();
exec.awaitTermination(1000, TimeUnit.MILLISECONDS); exec.awaitTermination(1000, TimeUnit.MILLISECONDS);
} }
@ -1284,7 +1278,6 @@ public class TestSort extends LuceneTestCase {
// null sorts first // null sorts first
assertEquals(1, hits.scoreDocs[0].doc); assertEquals(1, hits.scoreDocs[0].doc);
assertEquals(0, hits.scoreDocs[1].doc); assertEquals(0, hits.scoreDocs[1].doc);
s.close();
r.close(); r.close();
dir.close(); dir.close();
} }
@ -1308,7 +1301,6 @@ public class TestSort extends LuceneTestCase {
IndexReader reader = IndexReader.open(indexStore); IndexReader reader = IndexReader.open(indexStore);
IndexSearcher searcher = new IndexSearcher(reader); IndexSearcher searcher = new IndexSearcher(reader);
searcher.search(new MatchAllDocsQuery(), null, 500, sort); searcher.search(new MatchAllDocsQuery(), null, 500, sort);
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }
@ -1329,7 +1321,6 @@ public class TestSort extends LuceneTestCase {
TotalHitCountCollector c = new TotalHitCountCollector(); TotalHitCountCollector c = new TotalHitCountCollector();
searcher.search(new MatchAllDocsQuery(), null, c); searcher.search(new MatchAllDocsQuery(), null, c);
assertEquals(5, c.getTotalHits()); assertEquals(5, c.getTotalHits());
searcher.close();
reader.close(); reader.close();
indexStore.close(); indexStore.close();
} }

View File

@ -61,7 +61,6 @@ public class TestSubScorerFreqs extends LuceneTestCase {
@AfterClass @AfterClass
public static void finish() throws Exception { public static void finish() throws Exception {
s.getIndexReader().close(); s.getIndexReader().close();
s.close();
s = null; s = null;
dir.close(); dir.close();
dir = null; dir = null;

View File

@ -134,8 +134,6 @@ public class TestTermRangeFilter extends BaseTestRangeFilter {
result = search.search(q, TermRangeFilter.newStringRange("id", medIP, medIP, T, T), result = search.search(q, TermRangeFilter.newStringRange("id", medIP, medIP, T, T),
numDocs).scoreDocs; numDocs).scoreDocs;
assertEquals("med,med,T,T", 1, result.length); assertEquals("med,med,T,T", 1, result.length);
search.close();
} }
@Test @Test
@ -212,7 +210,5 @@ public class TestTermRangeFilter extends BaseTestRangeFilter {
result = search.search(q, TermRangeFilter.newStringRange("rand", maxRP, null, T, F), result = search.search(q, TermRangeFilter.newStringRange("rand", maxRP, null, T, F),
numDocs).scoreDocs; numDocs).scoreDocs;
assertEquals("max,nul,T,T", 1, result.length); assertEquals("max,nul,T,T", 1, result.length);
search.close();
} }
} }

View File

@ -59,7 +59,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
IndexSearcher searcher = new IndexSearcher(reader); IndexSearcher searcher = new IndexSearcher(reader);
ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals("A,B,C,D, only B in range", 1, hits.length); assertEquals("A,B,C,D, only B in range", 1, hits.length);
searcher.close();
reader.close(); reader.close();
initializeIndex(new String[] {"A", "B", "D"}); initializeIndex(new String[] {"A", "B", "D"});
@ -67,7 +66,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
hits = searcher.search(query, null, 1000).scoreDocs; hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals("A,B,D, only B in range", 1, hits.length); assertEquals("A,B,D, only B in range", 1, hits.length);
searcher.close();
reader.close(); reader.close();
addDoc("C"); addDoc("C");
@ -75,7 +73,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
hits = searcher.search(query, null, 1000).scoreDocs; hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals("C added, still only B in range", 1, hits.length); assertEquals("C added, still only B in range", 1, hits.length);
searcher.close();
reader.close(); reader.close();
} }
@ -87,7 +84,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
IndexSearcher searcher = new IndexSearcher(reader); IndexSearcher searcher = new IndexSearcher(reader);
ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs; ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals("A,B,C,D - A,B,C in range", 3, hits.length); assertEquals("A,B,C,D - A,B,C in range", 3, hits.length);
searcher.close();
reader.close(); reader.close();
initializeIndex(new String[]{"A", "B", "D"}); initializeIndex(new String[]{"A", "B", "D"});
@ -95,7 +91,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
hits = searcher.search(query, null, 1000).scoreDocs; hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals("A,B,D - A and B in range", 2, hits.length); assertEquals("A,B,D - A and B in range", 2, hits.length);
searcher.close();
reader.close(); reader.close();
addDoc("C"); addDoc("C");
@ -103,7 +98,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
searcher = new IndexSearcher(reader); searcher = new IndexSearcher(reader);
hits = searcher.search(query, null, 1000).scoreDocs; hits = searcher.search(query, null, 1000).scoreDocs;
assertEquals("C added - A, B, C in range", 3, hits.length); assertEquals("C added - A, B, C in range", 3, hits.length);
searcher.close();
reader.close(); reader.close();
} }
@ -125,7 +119,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
query = TermRangeQuery.newStringRange("content", "B", null, true, false); query = TermRangeQuery.newStringRange("content", "B", null, true, false);
assertTrue(query.getTermsEnum(terms) instanceof TermRangeTermsEnum); assertTrue(query.getTermsEnum(terms) instanceof TermRangeTermsEnum);
assertEquals(3, searcher.search(query, null, 1000).scoreDocs.length); assertEquals(3, searcher.search(query, null, 1000).scoreDocs.length);
searcher.close();
reader.close(); reader.close();
} }
@ -146,7 +139,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
} finally { } finally {
BooleanQuery.setMaxClauseCount(savedClauseCount); BooleanQuery.setMaxClauseCount(savedClauseCount);
} }
searcher.close();
reader.close(); reader.close();
} }
@ -292,7 +284,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
// until Lucene-38 is fixed, use this assert: // until Lucene-38 is fixed, use this assert:
//assertEquals("A,B,<empty string>,C,D => A, B & <empty string> are in range", 2, hits.length()); //assertEquals("A,B,<empty string>,C,D => A, B & <empty string> are in range", 2, hits.length());
searcher.close();
reader.close(); reader.close();
initializeIndex(new String[] {"A", "B", "", "D"}, analyzer); initializeIndex(new String[] {"A", "B", "", "D"}, analyzer);
reader = IndexReader.open(dir); reader = IndexReader.open(dir);
@ -302,7 +293,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
assertEquals("A,B,<empty string>,D => A, B & <empty string> are in range", 3, numHits); assertEquals("A,B,<empty string>,D => A, B & <empty string> are in range", 3, numHits);
// until Lucene-38 is fixed, use this assert: // until Lucene-38 is fixed, use this assert:
//assertEquals("A,B,<empty string>,D => A, B & <empty string> are in range", 2, hits.length()); //assertEquals("A,B,<empty string>,D => A, B & <empty string> are in range", 2, hits.length());
searcher.close();
reader.close(); reader.close();
addDoc("C"); addDoc("C");
reader = IndexReader.open(dir); reader = IndexReader.open(dir);
@ -312,7 +302,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
assertEquals("C added, still A, B & <empty string> are in range", 3, numHits); assertEquals("C added, still A, B & <empty string> are in range", 3, numHits);
// until Lucene-38 is fixed, use this assert // until Lucene-38 is fixed, use this assert
//assertEquals("C added, still A, B & <empty string> are in range", 2, hits.length()); //assertEquals("C added, still A, B & <empty string> are in range", 2, hits.length());
searcher.close();
reader.close(); reader.close();
} }
@ -329,7 +318,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
assertEquals("A,B,<empty string>,C,D => A,B,<empty string>,C in range", 4, numHits); assertEquals("A,B,<empty string>,C,D => A,B,<empty string>,C in range", 4, numHits);
// until Lucene-38 is fixed, use this assert // until Lucene-38 is fixed, use this assert
//assertEquals("A,B,<empty string>,C,D => A,B,<empty string>,C in range", 3, hits.length()); //assertEquals("A,B,<empty string>,C,D => A,B,<empty string>,C in range", 3, hits.length());
searcher.close();
reader.close(); reader.close();
initializeIndex(new String[]{"A", "B", "", "D"}, analyzer); initializeIndex(new String[]{"A", "B", "", "D"}, analyzer);
reader = IndexReader.open(dir); reader = IndexReader.open(dir);
@ -339,7 +327,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
assertEquals("A,B,<empty string>,D - A, B and <empty string> in range", 3, numHits); assertEquals("A,B,<empty string>,D - A, B and <empty string> in range", 3, numHits);
// until Lucene-38 is fixed, use this assert // until Lucene-38 is fixed, use this assert
//assertEquals("A,B,<empty string>,D => A, B and <empty string> in range", 2, hits.length()); //assertEquals("A,B,<empty string>,D => A, B and <empty string> in range", 2, hits.length());
searcher.close();
reader.close(); reader.close();
addDoc("C"); addDoc("C");
reader = IndexReader.open(dir); reader = IndexReader.open(dir);
@ -349,7 +336,6 @@ public class TestTermRangeQuery extends LuceneTestCase {
assertEquals("C added => A,B,<empty string>,C in range", 4, numHits); assertEquals("C added => A,B,<empty string>,C in range", 4, numHits);
// until Lucene-38 is fixed, use this assert // until Lucene-38 is fixed, use this assert
//assertEquals("C added => A,B,<empty string>,C in range", 3, hits.length()); //assertEquals("C added => A,B,<empty string>,C in range", 3, hits.length());
searcher.close();
reader.close(); reader.close();
} }
} }

View File

@ -65,7 +65,6 @@ public class TestTermScorer extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
indexSearcher.close();
indexReader.close(); indexReader.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();

View File

@ -79,7 +79,6 @@ public class TestTermVectors extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
super.tearDown(); super.tearDown();

View File

@ -107,7 +107,6 @@ public class TestTimeLimitingCollector extends LuceneTestCase {
@Override @Override
public void tearDown() throws Exception { public void tearDown() throws Exception {
searcher.close();
reader.close(); reader.close();
directory.close(); directory.close();
counterThread.stopTimer(); counterThread.stopTimer();

Some files were not shown because too many files have changed in this diff Show More