mirror of https://github.com/apache/lucene.git
LUCENE-9167: test speedup for slowest/pathological tests (round 3)
This commit is contained in:
parent
4d61e4aaab
commit
c53cc3edaf
|
@ -112,7 +112,7 @@ public class TestArabicAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
ArabicAnalyzer a = new ArabicAnalyzer();
|
||||
checkRandomData(random(), a, 200*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), a, 200 * RANDOM_MULTIPLIER);
|
||||
a.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -47,7 +47,7 @@ public class TestBengaliAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new BengaliAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -293,7 +293,7 @@ public class TestCJKAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomHugeStrings() throws Exception {
|
||||
Analyzer a = new CJKAnalyzer();
|
||||
checkRandomData(random(), a, 100*RANDOM_MULTIPLIER, 8192);
|
||||
checkRandomData(random(), a, 10 * RANDOM_MULTIPLIER, 8192);
|
||||
a.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -170,12 +170,12 @@ public class TestCJKBigramFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomUnibiStrings() throws Exception {
|
||||
checkRandomData(random(), unibiAnalyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), unibiAnalyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomUnibiHugeStrings() throws Exception {
|
||||
Random random = random();
|
||||
checkRandomData(random, unibiAnalyzer, 100*RANDOM_MULTIPLIER, 8192);
|
||||
checkRandomData(random, unibiAnalyzer, 10 * RANDOM_MULTIPLIER, 8192);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -74,7 +74,7 @@ public class TestCJKWidthFilter extends BaseTokenStreamTestCase {
|
|||
}
|
||||
|
||||
public void testRandomData() throws IOException {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -347,7 +347,7 @@ public class CommonGramsFilterTest extends BaseTokenStreamTestCase {
|
|||
}
|
||||
};
|
||||
|
||||
checkRandomData(random(), b, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), b, 200 * RANDOM_MULTIPLIER);
|
||||
b.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -389,7 +389,7 @@ public class TestCompoundWordTokenFilter extends BaseTokenStreamTestCase {
|
|||
return new TokenStreamComponents(tokenizer, filter);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), b, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), b, 200 * RANDOM_MULTIPLIER);
|
||||
b.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -180,7 +180,7 @@ public class TestDecimalDigitFilter extends BaseTokenStreamTestCase {
|
|||
* blast some random strings through the filter
|
||||
*/
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), tokenized, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), tokenized, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
/** returns a psuedo-random codepoint which is a Decimal Digit */
|
||||
|
|
|
@ -130,7 +130,7 @@ public class TestKeywordAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new KeywordAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -63,7 +63,7 @@ public class TestCzechAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new CzechAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestDanishAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new DanishAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -75,7 +75,7 @@ public class TestGermanLightStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -87,7 +87,7 @@ public class TestGermanMinimalStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -75,7 +75,7 @@ public class TestGermanNormalizationFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -82,7 +82,7 @@ public class TestGermanStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -64,7 +64,7 @@ public class TestEnglishMinimalStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -58,7 +58,7 @@ public class TestSpanishLightStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -55,6 +55,6 @@ public class TestPersianCharFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -75,7 +75,7 @@ public class TestFinnishLightStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -205,7 +205,7 @@ public class TestFrenchLightStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -106,7 +106,7 @@ public class TestFrenchMinimalStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestGalicianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new GalicianAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -59,7 +59,7 @@ public class TestHindiAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new HindiAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestHungarianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new HungarianAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -88,7 +88,7 @@ public class TestHunspellStemFilter extends BaseTokenStreamTestCase {
|
|||
return new TokenStreamComponents(tokenizer, new HunspellStemFilter(tokenizer, dictionary));
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestArmenianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new ArmenianAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestIndonesianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new IndonesianAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestItalianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new ItalianAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -58,7 +58,7 @@ public class TestItalianLightStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -48,6 +48,6 @@ public class TestLithuanianAnalyzer extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), new LithuanianAnalyzer(), 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), new LithuanianAnalyzer(), 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestLatvianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new LatvianAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -142,7 +142,7 @@ public class TestRemoveDuplicatesTokenFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
final int numIters = atLeast(10);
|
||||
final int numIters = atLeast(3);
|
||||
for (int i = 0; i < numIters; i++) {
|
||||
SynonymMap.Builder b = new SynonymMap.Builder(random().nextBoolean());
|
||||
final int numEntries = atLeast(10);
|
||||
|
|
|
@ -130,6 +130,6 @@ public class TestScandinavianFoldingFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomData() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -128,6 +128,6 @@ public class TestScandinavianNormalizationFilter extends BaseTokenStreamTestCase
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomData() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -440,7 +440,7 @@ public class TestWordDelimiterFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some enormous random strings through the analyzer */
|
||||
public void testRandomHugeStrings() throws Exception {
|
||||
int numIterations = atLeast(3);
|
||||
int numIterations = atLeast(1);
|
||||
for (int i = 0; i < numIterations; i++) {
|
||||
final int flags = random().nextInt(512);
|
||||
final CharArraySet protectedWords;
|
||||
|
|
|
@ -501,7 +501,7 @@ public class TestWordDelimiterGraphFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some enormous random strings through the analyzer */
|
||||
public void testRandomHugeStrings() throws Exception {
|
||||
int numIterations = atLeast(3);
|
||||
int numIterations = atLeast(1);
|
||||
for (int i = 0; i < numIterations; i++) {
|
||||
final int flags = random().nextInt(512);
|
||||
final CharArraySet protectedWords;
|
||||
|
|
|
@ -194,7 +194,7 @@ public class EdgeNGramTokenFilterTest extends BaseTokenStreamTestCase {
|
|||
new EdgeNGramTokenFilter(tokenizer, min, max, preserveOriginal));
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), a, 100*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), a, 10*RANDOM_MULTIPLIER);
|
||||
a.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -196,7 +196,7 @@ public class NGramTokenFilterTest extends BaseTokenStreamTestCase {
|
|||
new NGramTokenFilter(tokenizer, min, max, preserveOriginal));
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), a, 200*RANDOM_MULTIPLIER, 20);
|
||||
checkRandomData(random(), a, 10*RANDOM_MULTIPLIER, 20);
|
||||
a.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -113,7 +113,7 @@ public class NGramTokenizerTest extends BaseTokenStreamTestCase {
|
|||
return new TokenStreamComponents(tokenizer, tokenizer);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), a, 200*RANDOM_MULTIPLIER, 20);
|
||||
checkRandomData(random(), a, 200 * RANDOM_MULTIPLIER, 20);
|
||||
checkRandomData(random(), a, 10*RANDOM_MULTIPLIER, 1027);
|
||||
a.close();
|
||||
}
|
||||
|
|
|
@ -176,7 +176,7 @@ public class TestDutchAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new DutchAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestNorwegianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new NorwegianAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -93,7 +93,7 @@ public class TestNorwegianLightStemFilter extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Random random = random();
|
||||
checkRandomData(random, analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random, analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -93,7 +93,7 @@ public class TestNorwegianMinimalStemFilter extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Random random = random();
|
||||
checkRandomData(random, analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random, analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -226,7 +226,7 @@ public class TestPathHierarchyTokenizer extends BaseTokenStreamTestCase {
|
|||
}
|
||||
};
|
||||
// TODO: properly support positionLengthAttribute
|
||||
checkRandomData(random(), a, 1000*RANDOM_MULTIPLIER, 20, false, false);
|
||||
checkRandomData(random(), a, 200 * RANDOM_MULTIPLIER, 20, false, false);
|
||||
a.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -184,7 +184,7 @@ public class TestReversePathHierarchyTokenizer extends BaseTokenStreamTestCase {
|
|||
}
|
||||
};
|
||||
// TODO: properly support positionLengthAttribute
|
||||
checkRandomData(random(), a, 1000*RANDOM_MULTIPLIER, 20, false, false);
|
||||
checkRandomData(random(), a, 200 * RANDOM_MULTIPLIER, 20, false, false);
|
||||
a.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -99,7 +99,7 @@ public class TestPatternReplaceFilter extends BaseTokenStreamTestCase {
|
|||
return new TokenStreamComponents(tokenizer, filter);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), b, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), b, 200 * RANDOM_MULTIPLIER);
|
||||
b.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -142,7 +142,7 @@ public class TestPatternTokenizer extends BaseTokenStreamTestCase
|
|||
return new TokenStreamComponents(tokenizer);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), b, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), b, 200 * RANDOM_MULTIPLIER);
|
||||
b.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -267,7 +267,7 @@ public class TestSimplePatternSplitTokenizer extends BaseTokenStreamTestCase {
|
|||
return new TokenStreamComponents(tokenizer);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), b, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), b, 200 * RANDOM_MULTIPLIER);
|
||||
b.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -212,7 +212,7 @@ public class TestSimplePatternTokenizer extends BaseTokenStreamTestCase {
|
|||
return new TokenStreamComponents(tokenizer);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), b, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), b, 200 * RANDOM_MULTIPLIER);
|
||||
b.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestPortugueseAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new PortugueseAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -119,7 +119,7 @@ public class TestPortugueseLightStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -93,7 +93,7 @@ public class TestPortugueseMinimalStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -93,7 +93,7 @@ public class TestPortugueseStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestRomanianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new RomanianAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -59,7 +59,7 @@ public class TestRussianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new RussianAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -75,7 +75,7 @@ public class TestRussianLightStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -1125,7 +1125,7 @@ public class ShingleFilterTest extends BaseTokenStreamTestCase {
|
|||
return new TokenStreamComponents(tokenizer, new ShingleFilter(tokenizer));
|
||||
}
|
||||
};
|
||||
checkRandomData(random, a, 100*RANDOM_MULTIPLIER, 8192);
|
||||
checkRandomData(random, a, 3*RANDOM_MULTIPLIER, 8192);
|
||||
a.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -67,7 +67,7 @@ public class TestSerbianNormalizationFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -67,7 +67,7 @@ public class TestSerbianNormalizationRegularFilter extends BaseTokenStreamTestCa
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -334,14 +334,14 @@ public class TestClassicAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new ClassicAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
|
||||
/** blast some random large strings through the analyzer */
|
||||
public void testRandomHugeStrings() throws Exception {
|
||||
Analyzer analyzer = new ClassicAnalyzer();
|
||||
checkRandomData(random(), analyzer, 100*RANDOM_MULTIPLIER, 8192);
|
||||
checkRandomData(random(), analyzer, 10 * RANDOM_MULTIPLIER, 8192);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -627,7 +627,7 @@ public class TestUAX29URLEmailTokenizer extends BaseTokenStreamTestCase {
|
|||
/** blast some random large strings through the analyzer */
|
||||
public void testRandomHugeStrings() throws Exception {
|
||||
Random random = random();
|
||||
checkRandomData(random, a, 100*RANDOM_MULTIPLIER, 8192);
|
||||
checkRandomData(random, a, 3 * RANDOM_MULTIPLIER, 8192);
|
||||
}
|
||||
|
||||
public void testExampleURLs() throws Exception {
|
||||
|
|
|
@ -54,7 +54,7 @@ public class TestSwedishAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new SwedishAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -75,7 +75,7 @@ public class TestSwedishLightStemFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -1001,7 +1001,7 @@ public class TestSynonymGraphFilter extends BaseTokenStreamTestCase {
|
|||
/** simple random test like testRandom2, but for larger docs
|
||||
*/
|
||||
public void testRandomHuge() throws Exception {
|
||||
final int numIters = atLeast(3);
|
||||
final int numIters = atLeast(1);
|
||||
for (int i = 0; i < numIters; i++) {
|
||||
SynonymMap.Builder b = new SynonymMap.Builder(random().nextBoolean());
|
||||
final int numEntries = atLeast(10);
|
||||
|
@ -1406,7 +1406,7 @@ public class TestSynonymGraphFilter extends BaseTokenStreamTestCase {
|
|||
a = getAnalyzer(b, true);
|
||||
}
|
||||
|
||||
int iters = atLeast(10);
|
||||
int iters = atLeast(1);
|
||||
for(int iter=0;iter<iters;iter++) {
|
||||
|
||||
String doc = toTokenString(randomBinaryChars(50, 100, bias, 'a'));
|
||||
|
|
|
@ -532,7 +532,7 @@ public class TestSynonymMapFilter extends BaseTokenStreamTestCase {
|
|||
}
|
||||
};
|
||||
|
||||
checkRandomData(random, analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random, analyzer, 200*RANDOM_MULTIPLIER);
|
||||
}
|
||||
}
|
||||
*/
|
||||
|
|
|
@ -98,14 +98,14 @@ public class TestThaiAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new ThaiAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
|
||||
/** blast some random large strings through the analyzer */
|
||||
public void testRandomHugeStrings() throws Exception {
|
||||
Analyzer analyzer = new ThaiAnalyzer();
|
||||
checkRandomData(random(), analyzer, 100*RANDOM_MULTIPLIER, 8192);
|
||||
checkRandomData(random(), analyzer, 3*RANDOM_MULTIPLIER, 8192);
|
||||
analyzer.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -56,7 +56,7 @@ public class TestTurkishAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new TurkishAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.util.TestUtil;
|
|||
public class TestRollingCharBuffer extends LuceneTestCase {
|
||||
|
||||
public void test() throws Exception {
|
||||
final int ITERS = atLeast(1000);
|
||||
final int ITERS = atLeast(100);
|
||||
|
||||
RollingCharBuffer buffer = new RollingCharBuffer();
|
||||
|
||||
|
|
|
@ -138,8 +138,8 @@ public class TestSegmentingTokenizerBase extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
checkRandomData(random(), sentence, 10000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), sentenceAndWord, 10000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), sentence, 200*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), sentenceAndWord, 200*RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
// some tokenizers for testing
|
||||
|
|
|
@ -192,7 +192,7 @@ public class WikipediaTokenizerTest extends BaseTokenStreamTestCase {
|
|||
}
|
||||
};
|
||||
// TODO: properly support positionLengthAttribute
|
||||
checkRandomData(random(), a, 1000*RANDOM_MULTIPLIER, 20, false, false);
|
||||
checkRandomData(random(), a, 200 * RANDOM_MULTIPLIER, 20, false, false);
|
||||
a.close();
|
||||
}
|
||||
|
||||
|
@ -208,7 +208,7 @@ public class WikipediaTokenizerTest extends BaseTokenStreamTestCase {
|
|||
}
|
||||
};
|
||||
// TODO: properly support positionLengthAttribute
|
||||
checkRandomData(random, a, 100*RANDOM_MULTIPLIER, 8192, false, false);
|
||||
checkRandomData(random, a, 10 * RANDOM_MULTIPLIER, 8192, false, false);
|
||||
a.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -86,7 +86,7 @@ public class TestCollationDocValuesField extends LuceneTestCase {
|
|||
doc.add(field);
|
||||
doc.add(collationField);
|
||||
|
||||
int numDocs = atLeast(500);
|
||||
int numDocs = atLeast(100);
|
||||
for (int i = 0; i < numDocs; i++) {
|
||||
String value = TestUtil.randomSimpleString(random());
|
||||
field.setStringValue(value);
|
||||
|
@ -98,7 +98,7 @@ public class TestCollationDocValuesField extends LuceneTestCase {
|
|||
iw.close();
|
||||
IndexSearcher is = newSearcher(ir);
|
||||
|
||||
int numChecks = atLeast(100);
|
||||
int numChecks = atLeast(20);
|
||||
|
||||
try {
|
||||
for (int i = 0; i < numChecks; i++) {
|
||||
|
|
|
@ -339,7 +339,7 @@ public class TestICUTokenizer extends BaseTokenStreamTestCase {
|
|||
/** blast some random large strings through the analyzer */
|
||||
public void testRandomHugeStrings() throws Exception {
|
||||
Random random = random();
|
||||
checkRandomData(random, a, 100*RANDOM_MULTIPLIER, 8192);
|
||||
checkRandomData(random, a, 10 * RANDOM_MULTIPLIER, 8192);
|
||||
}
|
||||
|
||||
public void testTokenAttributes() throws Exception {
|
||||
|
|
|
@ -115,6 +115,6 @@ public class TestICUTokenizerCJK extends BaseTokenStreamTestCase {
|
|||
/** blast some random large strings through the analyzer */
|
||||
public void testRandomHugeStrings() throws Exception {
|
||||
Random random = random();
|
||||
checkRandomData(random, a, 100*RANDOM_MULTIPLIER, 8192);
|
||||
checkRandomData(random, a, 10 * RANDOM_MULTIPLIER, 8192);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -80,6 +80,12 @@ public class TestExtendedMode extends BaseTokenStreamTestCase {
|
|||
|
||||
/** blast some random large strings through the analyzer */
|
||||
public void testRandomHugeStrings() throws Exception {
|
||||
Random random = random();
|
||||
checkRandomData(random, analyzer, RANDOM_MULTIPLIER, 4096);
|
||||
}
|
||||
|
||||
@Nightly
|
||||
public void testRandomHugeStringsAtNight() throws Exception {
|
||||
Random random = random();
|
||||
checkRandomData(random, analyzer, 3*RANDOM_MULTIPLIER, 8192);
|
||||
}
|
||||
|
|
|
@ -94,7 +94,7 @@ public class TestJapaneseKatakanaStemFilter extends BaseTokenStreamTestCase {
|
|||
}
|
||||
|
||||
public void testRandomData() throws IOException {
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -103,8 +103,8 @@ public class TestJapaneseReadingFormFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
public void testRandomData() throws IOException {
|
||||
Random random = random();
|
||||
checkRandomData(random, katakanaAnalyzer, 200*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random, romajiAnalyzer, 200*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random, katakanaAnalyzer, 200 * RANDOM_MULTIPLIER);
|
||||
checkRandomData(random, romajiAnalyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -369,7 +369,8 @@ public class
|
|||
|
||||
|
||||
public void testLargeDocReliability() throws Exception {
|
||||
for (int i = 0; i < 10; i++) {
|
||||
int numIters = atLeast(1);
|
||||
for (int i = 0; i < numIters; i++) {
|
||||
String s = TestUtil.randomUnicodeString(random(), 10000);
|
||||
try (TokenStream ts = analyzer.tokenStream("foo", s)) {
|
||||
ts.reset();
|
||||
|
|
|
@ -73,7 +73,7 @@ public class TestUkrainianAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new UkrainianMorfologikAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -58,7 +58,7 @@ public class TestKoreanReadingFormFilter extends BaseTokenStreamTestCase {
|
|||
|
||||
public void testRandomData() throws IOException {
|
||||
Random random = random();
|
||||
checkRandomData(random, analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random, analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
}
|
||||
|
||||
public void testEmptyTerm() throws IOException {
|
||||
|
|
|
@ -400,7 +400,7 @@ public class TestKoreanTokenizer extends BaseTokenStreamTestCase {
|
|||
return new TokenStreamComponents(tokenizer, graph);
|
||||
}
|
||||
};
|
||||
checkRandomData(random, analyzer, 3*RANDOM_MULTIPLIER, 8192);
|
||||
checkRandomData(random, analyzer, RANDOM_MULTIPLIER, 4096);
|
||||
analyzer.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -105,7 +105,7 @@ public class TestPhoneticFilter extends BaseTokenStreamTestCase {
|
|||
}
|
||||
};
|
||||
|
||||
checkRandomData(random(), b, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), b, 200 * RANDOM_MULTIPLIER);
|
||||
b.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -53,7 +53,7 @@ public class TestPolishAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new PolishAnalyzer();
|
||||
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -55,11 +55,13 @@ import org.apache.lucene.index.TermsEnum;
|
|||
import org.apache.lucene.search.DocIdSetIterator;
|
||||
import org.apache.lucene.store.Directory;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.LuceneTestCase;
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
|
||||
/**
|
||||
* Test very simply that perf tasks - simple algorithms - are doing what they should.
|
||||
*/
|
||||
@LuceneTestCase.SuppressCodecs("SimpleText")
|
||||
public class TestPerfTasksLogic extends BenchmarkTestCase {
|
||||
|
||||
@Override
|
||||
|
|
|
@ -42,6 +42,7 @@ import org.junit.Test;
|
|||
/**
|
||||
* Testcase for {@link org.apache.lucene.classification.utils.DatasetSplitter}
|
||||
*/
|
||||
@LuceneTestCase.SuppressCodecs("SimpleText")
|
||||
public class DataSplitterTest extends LuceneTestCase {
|
||||
|
||||
private LeafReader originalIndex;
|
||||
|
|
|
@ -439,7 +439,7 @@ public class TestStandardAnalyzer extends BaseTokenStreamTestCase {
|
|||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer analyzer = new StandardAnalyzer();
|
||||
checkRandomData(random(), analyzer, 200*RANDOM_MULTIPLIER);
|
||||
checkRandomData(random(), analyzer, 200 * RANDOM_MULTIPLIER);
|
||||
analyzer.close();
|
||||
}
|
||||
|
||||
|
|
|
@ -51,6 +51,7 @@ public class TestCharTermAttributeImpl extends LuceneTestCase {
|
|||
});
|
||||
}
|
||||
|
||||
@Slow
|
||||
public void testGrow() {
|
||||
CharTermAttributeImpl t = new CharTermAttributeImpl();
|
||||
StringBuilder buf = new StringBuilder("ab");
|
||||
|
|
|
@ -108,7 +108,7 @@ public class TestLucene60PointsFormat extends BasePointsFormatTestCase {
|
|||
byte[] pointValue = new byte[3];
|
||||
byte[] uniquePointValue = new byte[3];
|
||||
random().nextBytes(uniquePointValue);
|
||||
final int numDocs = atLeast(10000); // make sure we have several leaves
|
||||
final int numDocs = TEST_NIGHTLY ? atLeast(10000) : atLeast(500); // at night, make sure we have several leaves
|
||||
final boolean multiValues = random().nextBoolean();
|
||||
for (int i = 0; i < numDocs; ++i) {
|
||||
Document doc = new Document();
|
||||
|
@ -215,7 +215,7 @@ public class TestLucene60PointsFormat extends BasePointsFormatTestCase {
|
|||
uniquePointValue[1] = new byte[3];
|
||||
random().nextBytes(uniquePointValue[0]);
|
||||
random().nextBytes(uniquePointValue[1]);
|
||||
final int numDocs = atLeast(10000); // make sure we have several leaves
|
||||
final int numDocs = TEST_NIGHTLY? atLeast(10000) : atLeast(1000); // in nightly, make sure we have several leaves
|
||||
final boolean multiValues = random().nextBoolean();
|
||||
for (int i = 0; i < numDocs; ++i) {
|
||||
Document doc = new Document();
|
||||
|
|
|
@ -319,7 +319,8 @@ public class TestIndexedDISI extends LuceneTestCase {
|
|||
|
||||
public void testFewMissingDocs() throws IOException {
|
||||
try (Directory dir = newDirectory()) {
|
||||
for (int iter = 0; iter < 100; ++iter) {
|
||||
int numIters = atLeast(10);
|
||||
for (int iter = 0; iter < numIters; ++iter) {
|
||||
int maxDoc = TestUtil.nextInt(random(), 1, 100000);
|
||||
FixedBitSet set = new FixedBitSet(maxDoc);
|
||||
set.set(0, maxDoc);
|
||||
|
@ -412,7 +413,8 @@ public class TestIndexedDISI extends LuceneTestCase {
|
|||
|
||||
public void testRandom() throws IOException {
|
||||
try (Directory dir = newDirectory()) {
|
||||
for (int i = 0; i < 10; ++i) {
|
||||
int numIters = atLeast(3);
|
||||
for (int i = 0; i < numIters; ++i) {
|
||||
doTestRandom(dir);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -67,6 +67,7 @@ import org.junit.Test;
|
|||
//TODO: would be better in this test to pull termsenums and instanceof or something?
|
||||
// this way we can verify PFPF is doing the right thing.
|
||||
// for now we do termqueries.
|
||||
@LuceneTestCase.SuppressCodecs("SimpleText")
|
||||
public class TestPerFieldPostingsFormat2 extends LuceneTestCase {
|
||||
|
||||
private IndexWriter newWriter(Directory dir, IndexWriterConfig conf)
|
||||
|
|
|
@ -70,7 +70,7 @@ public abstract class BaseShapeTestCase extends LuceneTestCase {
|
|||
|
||||
// A particularly tricky adversary for BKD tree:
|
||||
public void testSameShapeManyTimes() throws Exception {
|
||||
int numShapes = TEST_NIGHTLY ? atLeast(50) : atLeast(10);
|
||||
int numShapes = TEST_NIGHTLY ? atLeast(50) : atLeast(3);
|
||||
|
||||
// Every doc has 2 points:
|
||||
Object theShape = nextShape();
|
||||
|
|
|
@ -338,7 +338,7 @@ public class TestLatLonPointDistanceFeatureQuery extends LuceneTestCase {
|
|||
LatLonDocValuesField docValue = new LatLonDocValuesField("foo", 0., 0.);
|
||||
doc.add(docValue);
|
||||
|
||||
int numDocs = atLeast(10000);
|
||||
int numDocs = atLeast(1000);
|
||||
for (int i = 0; i < numDocs; ++i) {
|
||||
double lat = random().nextDouble() * 180 - 90;
|
||||
double lon = random().nextDouble() * 360 - 180;
|
||||
|
@ -350,7 +350,8 @@ public class TestLatLonPointDistanceFeatureQuery extends LuceneTestCase {
|
|||
IndexReader reader = DirectoryReader.open(w);
|
||||
IndexSearcher searcher = newSearcher(reader);
|
||||
|
||||
for (int iter = 0; iter < 10; ++iter) {
|
||||
int numIters = atLeast(3);
|
||||
for (int iter = 0; iter < numIters; ++iter) {
|
||||
double lat = random().nextDouble() * 180 - 90;
|
||||
double lon = random().nextDouble() * 360 - 180;
|
||||
double pivotDistance = random().nextDouble() * random().nextDouble() * Math.PI * GeoUtils.EARTH_MEAN_RADIUS_METERS;
|
||||
|
|
|
@ -25,8 +25,10 @@ import org.apache.lucene.geo.Tessellator;
|
|||
import org.apache.lucene.geo.XYPolygon;
|
||||
import org.apache.lucene.geo.XYRectangle;
|
||||
import org.apache.lucene.geo.XYRectangle2D;
|
||||
import org.apache.lucene.util.LuceneTestCase;
|
||||
|
||||
/** random cartesian bounding box, line, and polygon query tests for random indexed arrays of cartesian {@link XYPolygon} types */
|
||||
@LuceneTestCase.SuppressCodecs("SimpleText")
|
||||
public class TestXYMultiPolygonShapeQueries extends BaseXYShapeTestCase {
|
||||
@Override
|
||||
protected ShapeType getShapeType() {
|
||||
|
|
|
@ -50,6 +50,7 @@ import org.junit.Assume;
|
|||
|
||||
import static org.apache.lucene.search.DocIdSetIterator.NO_MORE_DOCS;
|
||||
|
||||
@LuceneTestCase.SuppressCodecs("SimpleText")
|
||||
public class TestDirectoryReader extends LuceneTestCase {
|
||||
|
||||
public void testDocument() throws IOException {
|
||||
|
|
|
@ -168,6 +168,8 @@ public class TestIndexWriterCommit extends LuceneTestCase {
|
|||
* file. We check this by using MockDirectoryWrapper to
|
||||
* measure max temp disk space used.
|
||||
*/
|
||||
// TODO: can this write less docs/indexes?
|
||||
@Nightly
|
||||
public void testCommitOnCloseDiskUsage() throws IOException {
|
||||
// MemoryCodec, since it uses FST, is not necessarily
|
||||
// "additive", ie if you add up N small FSTs, then merge
|
||||
|
|
|
@ -980,7 +980,8 @@ public class TestIndexWriterDelete extends LuceneTestCase {
|
|||
dir.close();
|
||||
}
|
||||
|
||||
@Slow
|
||||
// TODO: this test can hit pathological cases (IW settings?) where it runs for far too long
|
||||
@Nightly
|
||||
public void testIndexingThenDeleting() throws Exception {
|
||||
// TODO: move this test to its own class and just @SuppressCodecs?
|
||||
// TODO: is it enough to just use newFSDirectory?
|
||||
|
|
|
@ -362,6 +362,8 @@ public class TestIndexWriterMaxDocs extends LuceneTestCase {
|
|||
/**
|
||||
* LUCENE-6299: Test if addindexes(Dir[]) prevents exceeding max docs.
|
||||
*/
|
||||
// TODO: can we use the setter to lower the amount of docs to be written here?
|
||||
@Nightly
|
||||
public void testAddTooManyIndexesDir() throws Exception {
|
||||
// we cheat and add the same one over again... IW wants a write lock on each
|
||||
Directory dir = newDirectory(random(), NoLockFactory.INSTANCE);
|
||||
|
|
|
@ -128,6 +128,7 @@ public class TestIndexWriterMergePolicy extends LuceneTestCase {
|
|||
}
|
||||
|
||||
// Test the case where both mergeFactor and maxBufferedDocs change
|
||||
@Nightly
|
||||
public void testMaxBufferedDocsChange() throws IOException {
|
||||
Directory dir = newDirectory();
|
||||
|
||||
|
|
|
@ -42,6 +42,7 @@ import org.apache.lucene.util.StringHelper;
|
|||
import org.apache.lucene.util.TestUtil;
|
||||
import org.apache.lucene.util.Version;
|
||||
|
||||
@LuceneTestCase.SuppressCodecs("SimpleText")
|
||||
public class TestIndexWriterThreadsToSegments extends LuceneTestCase {
|
||||
|
||||
// LUCENE-5644: for first segment, two threads each indexed one doc (likely concurrently), but for second segment, each thread indexed the
|
||||
|
|
|
@ -50,7 +50,7 @@ import org.apache.lucene.util.LuceneTestCase.Slow;
|
|||
/**
|
||||
* MultiThreaded IndexWriter tests
|
||||
*/
|
||||
@Slow
|
||||
@Slow @LuceneTestCase.SuppressCodecs("SimpleText")
|
||||
public class TestIndexWriterWithThreads extends LuceneTestCase {
|
||||
|
||||
// Used by test cases below
|
||||
|
@ -141,7 +141,7 @@ public class TestIndexWriterWithThreads extends LuceneTestCase {
|
|||
public void testImmediateDiskFullWithThreads() throws Exception {
|
||||
|
||||
int NUM_THREADS = 3;
|
||||
final int numIterations = TEST_NIGHTLY ? 10 : 3;
|
||||
final int numIterations = TEST_NIGHTLY ? 10 : 1;
|
||||
for (int iter=0;iter<numIterations;iter++) {
|
||||
if (VERBOSE) {
|
||||
System.out.println("\nTEST: iter=" + iter);
|
||||
|
|
|
@ -30,6 +30,7 @@ import org.apache.lucene.store.Directory;
|
|||
import org.apache.lucene.util.LuceneTestCase;
|
||||
|
||||
/** Test that creates way, way, way too many fields */
|
||||
@LuceneTestCase.SuppressCodecs("SimpleText")
|
||||
public class TestManyFields extends LuceneTestCase {
|
||||
private static final FieldType storedTextType = new FieldType(TextField.TYPE_NOT_STORED);
|
||||
|
||||
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue