LUCENE-3994: tone down analysis tests (remove n^2) so they dont take hours in nightly builds

git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@1327162 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Robert Muir 2012-04-17 16:23:14 +00:00
parent 7ad9f7aad5
commit 19c450884b
7 changed files with 12 additions and 11 deletions

View File

@ -221,7 +221,8 @@ public class TestMappingCharFilter extends BaseTokenStreamTestCase {
//@Ignore("wrong finalOffset: https://issues.apache.org/jira/browse/LUCENE-3971")
public void testRandomMaps() throws Exception {
for (int i = 0; i < 100; i++) {
int numIterations = atLeast(10);
for (int i = 0; i < numIterations; i++) {
final NormalizeCharMap map = randomMap();
Analyzer analyzer = new Analyzer() {
@Override
@ -235,7 +236,7 @@ public class TestMappingCharFilter extends BaseTokenStreamTestCase {
return new MappingCharFilter(map, CharReader.get(reader));
}
};
int numRounds = RANDOM_MULTIPLIER * 100;
int numRounds = 100;
checkRandomData(random(), analyzer, numRounds);
}
}

View File

@ -845,7 +845,7 @@ public class TestRandomChains extends BaseTokenStreamTestCase {
}
public void testRandomChains() throws Throwable {
int numIterations = atLeast(20);
int numIterations = atLeast(10);
Random random = random();
for (int i = 0; i < numIterations; i++) {
MockRandomAnalyzer a = new MockRandomAnalyzer(random.nextLong());
@ -853,7 +853,7 @@ public class TestRandomChains extends BaseTokenStreamTestCase {
System.out.println("Creating random analyzer:" + a);
}
try {
checkRandomData(random, a, 1000, 20, false,
checkRandomData(random, a, 200, 20, false,
false /* We already validate our own offsets... */);
} catch (Throwable e) {
System.err.println("Exception from random analyzer: " + a);

View File

@ -163,7 +163,7 @@ public class TestRemoveDuplicatesTokenFilter extends BaseTokenStreamTestCase {
}
};
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
checkRandomData(random(), analyzer, 200);
}
}

View File

@ -351,7 +351,7 @@ public class TestWordDelimiterFilter extends BaseTokenStreamTestCase {
return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(tokenizer, flags, protectedWords));
}
};
checkRandomData(random(), a, 10000*RANDOM_MULTIPLIER, 20, false, false);
checkRandomData(random(), a, 200, 20, false, false);
}
}

View File

@ -445,7 +445,7 @@ public class TestSynonymMapFilter extends BaseTokenStreamTestCase {
}
};
checkRandomData(random(), analyzer, 1000*RANDOM_MULTIPLIER);
checkRandomData(random(), analyzer, 200);
}
}
@ -503,7 +503,7 @@ public class TestSynonymMapFilter extends BaseTokenStreamTestCase {
}
};
checkRandomData(random, analyzer, 1000*RANDOM_MULTIPLIER);
checkRandomData(random, analyzer, 200);
}
}
@ -553,7 +553,7 @@ public class TestSynonymMapFilter extends BaseTokenStreamTestCase {
}
};
checkRandomData(random, analyzer, 100*RANDOM_MULTIPLIER, 8192);
checkRandomData(random, analyzer, 100, 8192);
}
}

View File

@ -430,7 +430,7 @@ public abstract class BaseTokenStreamTestCase extends LuceneTestCase {
checkRandomData(new Random(seed), a, iterations, maxWordLength, useCharFilter, simple, offsetsAreCorrect);
// now test with multiple threads: note we do the EXACT same thing we did before in each thread,
// so this should only really fail from another thread if its an actual thread problem
int numThreads = _TestUtil.nextInt(random, 4, 8);
int numThreads = _TestUtil.nextInt(random, 2, 4);
AnalysisThread threads[] = new AnalysisThread[numThreads];
for (int i = 0; i < threads.length; i++) {
threads[i] = new AnalysisThread(seed, a, iterations, maxWordLength, useCharFilter, simple, offsetsAreCorrect);

View File

@ -952,7 +952,7 @@ public abstract class LuceneTestCase extends Assert {
* is active and {@link #RANDOM_MULTIPLIER}, but also with some random fudge.
*/
public static int atLeast(Random random, int i) {
int min = (TEST_NIGHTLY ? 3*i : i) * RANDOM_MULTIPLIER;
int min = (TEST_NIGHTLY ? 2*i : i) * RANDOM_MULTIPLIER;
int max = min+(min/2);
return _TestUtil.nextInt(random, min, max);
}