mirror of https://github.com/apache/lucene.git
Fix missing variable assignment in testAllVersionHaveCfsAndNocfs() and other minor code cleanups (#12969)
This commit is contained in:
parent
0fc1e2c2f7
commit
5442748995
|
@ -182,7 +182,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
|
|
||||||
// These are only needed for the special upgrade test to verify
|
// These are only needed for the special upgrade test to verify
|
||||||
// that also single-segment indexes are correctly upgraded by IndexUpgrader.
|
// that also single-segment indexes are correctly upgraded by IndexUpgrader.
|
||||||
// You don't need them to be build for non-4.0 (the test is happy with just one
|
// You don't need them to be built for non-4.0 (the test is happy with just one
|
||||||
// "old" segment format, version is unimportant:
|
// "old" segment format, version is unimportant:
|
||||||
|
|
||||||
public void testCreateSingleSegmentCFS() throws IOException {
|
public void testCreateSingleSegmentCFS() throws IOException {
|
||||||
|
@ -263,7 +263,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
throw new AssertionError("failed to parse \"" + dateString + "\" as date");
|
throw new AssertionError("failed to parse \"" + dateString + "\" as date");
|
||||||
}
|
}
|
||||||
if (dateDVField == null) {
|
if (dateDVField == null) {
|
||||||
dateDVField = new NumericDocValuesField("dateDV", 0l);
|
dateDVField = new NumericDocValuesField("dateDV", 0L);
|
||||||
doc.add(dateDVField);
|
doc.add(dateDVField);
|
||||||
}
|
}
|
||||||
dateDVField.setLongValue(date.getTime());
|
dateDVField.setLongValue(date.getTime());
|
||||||
|
@ -727,6 +727,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
String prefix = prevFile.replace("-cfs", "");
|
String prefix = prevFile.replace("-cfs", "");
|
||||||
assertEquals("Missing -nocfs for backcompat index " + prefix, prefix + "-nocfs", file);
|
assertEquals("Missing -nocfs for backcompat index " + prefix, prefix + "-nocfs", file);
|
||||||
}
|
}
|
||||||
|
prevFile = file;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -746,7 +747,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
expectedVersions.add(v.toString() + "-cfs");
|
expectedVersions.add(v + "-cfs");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -765,7 +766,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
assertNotNull(lastPrevMajorVersion);
|
assertNotNull(lastPrevMajorVersion);
|
||||||
expectedVersions.remove(lastPrevMajorVersion.toString() + "-cfs");
|
expectedVersions.remove(lastPrevMajorVersion + "-cfs");
|
||||||
// END TRUNK ONLY BLOCK
|
// END TRUNK ONLY BLOCK
|
||||||
|
|
||||||
Collections.sort(expectedVersions);
|
Collections.sort(expectedVersions);
|
||||||
|
@ -814,17 +815,17 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
StringBuffer msg = new StringBuffer();
|
StringBuilder msg = new StringBuilder();
|
||||||
if (missingFiles.size() > 1) {
|
if (missingFiles.size() > 1) {
|
||||||
msg.append("Missing backcompat test files:\n");
|
msg.append("Missing backcompat test files:\n");
|
||||||
for (String missingFile : missingFiles) {
|
for (String missingFile : missingFiles) {
|
||||||
msg.append(" " + missingFile + "\n");
|
msg.append(" ").append(missingFile).append("\n");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (extraFiles.isEmpty() == false) {
|
if (extraFiles.isEmpty() == false) {
|
||||||
msg.append("Extra backcompat test files:\n");
|
msg.append("Extra backcompat test files:\n");
|
||||||
for (String extraFile : extraFiles) {
|
for (String extraFile : extraFiles) {
|
||||||
msg.append(" " + extraFile + "\n");
|
msg.append(" ").append(extraFile).append("\n");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
fail(msg.toString());
|
fail(msg.toString());
|
||||||
|
@ -924,7 +925,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
e.printStackTrace(System.out);
|
e.printStackTrace(System.out);
|
||||||
}
|
}
|
||||||
// Make sure exc message includes a path=
|
// Make sure exc message includes a path=
|
||||||
assertTrue("got exc message: " + e.getMessage(), e.getMessage().indexOf("path=\"") != -1);
|
assertTrue("got exc message: " + e.getMessage(), e.getMessage().contains("path=\""));
|
||||||
} finally {
|
} finally {
|
||||||
// we should fail to open IW, and so it should be null when we get here.
|
// we should fail to open IW, and so it should be null when we get here.
|
||||||
// However, if the test fails (i.e., IW did not fail on open), we need
|
// However, if the test fails (i.e., IW did not fail on open), we need
|
||||||
|
@ -1226,7 +1227,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
assertEquals(expectedRef, term);
|
assertEquals(expectedRef, term);
|
||||||
|
|
||||||
assertEquals(i, dvDouble.nextDoc());
|
assertEquals(i, dvDouble.nextDoc());
|
||||||
assertEquals((double) id, Double.longBitsToDouble(dvDouble.longValue()), 0D);
|
assertEquals(id, Double.longBitsToDouble(dvDouble.longValue()), 0D);
|
||||||
assertEquals(i, dvFloat.nextDoc());
|
assertEquals(i, dvFloat.nextDoc());
|
||||||
assertEquals((float) id, Float.intBitsToFloat((int) dvFloat.longValue()), 0F);
|
assertEquals((float) id, Float.intBitsToFloat((int) dvFloat.longValue()), 0F);
|
||||||
assertEquals(i, dvInt.nextDoc());
|
assertEquals(i, dvInt.nextDoc());
|
||||||
|
@ -1249,8 +1250,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
assertEquals(id, dvSortedNumeric.nextValue());
|
assertEquals(id, dvSortedNumeric.nextValue());
|
||||||
}
|
}
|
||||||
|
|
||||||
ScoreDoc[] hits =
|
ScoreDoc[] hits = searcher.search(new TermQuery(new Term("content", "aaa")), 1000).scoreDocs;
|
||||||
searcher.search(new TermQuery(new Term(new String("content"), "aaa")), 1000).scoreDocs;
|
|
||||||
|
|
||||||
// First document should be #0
|
// First document should be #0
|
||||||
Document d = storedFields.document(hits[0].doc);
|
Document d = storedFields.document(hits[0].doc);
|
||||||
|
@ -1258,19 +1258,18 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
|
|
||||||
doTestHits(hits, 34, searcher.getIndexReader());
|
doTestHits(hits, 34, searcher.getIndexReader());
|
||||||
|
|
||||||
hits = searcher.search(new TermQuery(new Term(new String("content5"), "aaa")), 1000).scoreDocs;
|
hits = searcher.search(new TermQuery(new Term("content5", "aaa")), 1000).scoreDocs;
|
||||||
|
|
||||||
doTestHits(hits, 34, searcher.getIndexReader());
|
doTestHits(hits, 34, searcher.getIndexReader());
|
||||||
|
|
||||||
hits = searcher.search(new TermQuery(new Term(new String("content6"), "aaa")), 1000).scoreDocs;
|
hits = searcher.search(new TermQuery(new Term("content6", "aaa")), 1000).scoreDocs;
|
||||||
|
|
||||||
doTestHits(hits, 34, searcher.getIndexReader());
|
doTestHits(hits, 34, searcher.getIndexReader());
|
||||||
|
|
||||||
hits = searcher.search(new TermQuery(new Term("utf8", "\u0000")), 1000).scoreDocs;
|
hits = searcher.search(new TermQuery(new Term("utf8", "\u0000")), 1000).scoreDocs;
|
||||||
assertEquals(34, hits.length);
|
assertEquals(34, hits.length);
|
||||||
hits =
|
hits =
|
||||||
searcher.search(
|
searcher.search(new TermQuery(new Term("utf8", "lu\uD834\uDD1Ece\uD834\uDD60ne")), 1000)
|
||||||
new TermQuery(new Term(new String("utf8"), "lu\uD834\uDD1Ece\uD834\uDD60ne")), 1000)
|
|
||||||
.scoreDocs;
|
.scoreDocs;
|
||||||
assertEquals(34, hits.length);
|
assertEquals(34, hits.length);
|
||||||
hits = searcher.search(new TermQuery(new Term("utf8", "ab\ud917\udc17cd")), 1000).scoreDocs;
|
hits = searcher.search(new TermQuery(new Term("utf8", "ab\ud917\udc17cd")), 1000).scoreDocs;
|
||||||
|
@ -1578,7 +1577,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
doc.add(new SortedDocValuesField("dvBytesSortedVar", ref));
|
doc.add(new SortedDocValuesField("dvBytesSortedVar", ref));
|
||||||
doc.add(new BinaryDocValuesField("dvBytesStraightFixed", ref));
|
doc.add(new BinaryDocValuesField("dvBytesStraightFixed", ref));
|
||||||
doc.add(new BinaryDocValuesField("dvBytesStraightVar", ref));
|
doc.add(new BinaryDocValuesField("dvBytesStraightVar", ref));
|
||||||
doc.add(new DoubleDocValuesField("dvDouble", (double) id));
|
doc.add(new DoubleDocValuesField("dvDouble", id));
|
||||||
doc.add(new FloatDocValuesField("dvFloat", (float) id));
|
doc.add(new FloatDocValuesField("dvFloat", (float) id));
|
||||||
doc.add(new NumericDocValuesField("dvInt", id));
|
doc.add(new NumericDocValuesField("dvInt", id));
|
||||||
doc.add(new NumericDocValuesField("dvLong", id));
|
doc.add(new NumericDocValuesField("dvLong", id));
|
||||||
|
@ -1593,8 +1592,8 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
doc.add(new FloatPoint("floatPoint2d", (float) id, (float) 2 * id));
|
doc.add(new FloatPoint("floatPoint2d", (float) id, (float) 2 * id));
|
||||||
doc.add(new LongPoint("longPoint1d", id));
|
doc.add(new LongPoint("longPoint1d", id));
|
||||||
doc.add(new LongPoint("longPoint2d", id, 2 * id));
|
doc.add(new LongPoint("longPoint2d", id, 2 * id));
|
||||||
doc.add(new DoublePoint("doublePoint1d", (double) id));
|
doc.add(new DoublePoint("doublePoint1d", id));
|
||||||
doc.add(new DoublePoint("doublePoint2d", (double) id, (double) 2 * id));
|
doc.add(new DoublePoint("doublePoint2d", id, (double) 2 * id));
|
||||||
doc.add(new BinaryPoint("binaryPoint1d", bytes));
|
doc.add(new BinaryPoint("binaryPoint1d", bytes));
|
||||||
doc.add(new BinaryPoint("binaryPoint2d", bytes, bytes));
|
doc.add(new BinaryPoint("binaryPoint2d", bytes, bytes));
|
||||||
|
|
||||||
|
@ -1803,7 +1802,7 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
public void testIndexUpgraderCommandLineArgs() throws Exception {
|
public void testIndexUpgraderCommandLineArgs() throws Exception {
|
||||||
|
|
||||||
PrintStream savedSystemOut = System.out;
|
PrintStream savedSystemOut = System.out;
|
||||||
System.setOut(new PrintStream(new ByteArrayOutputStream(), false, "UTF-8"));
|
System.setOut(new PrintStream(new ByteArrayOutputStream(), false, UTF_8));
|
||||||
try {
|
try {
|
||||||
for (Map.Entry<String, Directory> entry : oldIndexDirs.entrySet()) {
|
for (Map.Entry<String, Directory> entry : oldIndexDirs.entrySet()) {
|
||||||
String name = entry.getKey();
|
String name = entry.getKey();
|
||||||
|
@ -1845,10 +1844,8 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
IndexUpgrader.main(args.toArray(new String[0]));
|
IndexUpgrader.main(args.toArray(new String[0]));
|
||||||
|
|
||||||
Directory upgradedDir = newFSDirectory(dir);
|
Directory upgradedDir = newFSDirectory(dir);
|
||||||
try {
|
try (upgradedDir) {
|
||||||
checkAllSegmentsUpgraded(upgradedDir, indexCreatedVersion);
|
checkAllSegmentsUpgraded(upgradedDir, indexCreatedVersion);
|
||||||
} finally {
|
|
||||||
upgradedDir.close();
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
|
@ -1892,10 +1889,8 @@ public class TestBackwardsCompatibility extends LuceneTestCase {
|
||||||
IndexWriterConfig iwc = new IndexWriterConfig(null).setMergePolicy(mp);
|
IndexWriterConfig iwc = new IndexWriterConfig(null).setMergePolicy(mp);
|
||||||
IndexWriter w = new IndexWriter(dir, iwc);
|
IndexWriter w = new IndexWriter(dir, iwc);
|
||||||
w.addIndexes(ramDir);
|
w.addIndexes(ramDir);
|
||||||
try {
|
try (w) {
|
||||||
w.commit();
|
w.commit();
|
||||||
} finally {
|
|
||||||
w.close();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// determine count of segments in modified index
|
// determine count of segments in modified index
|
||||||
|
|
Loading…
Reference in New Issue