mirror of https://github.com/apache/lucene.git
LUCENE-3378: move VocabularyAssert to test-framework
git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@1158730 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
f06edac7c5
commit
8a0578dfe2
|
@ -1,4 +1,4 @@
|
||||||
package org.apache.lucene.analysis.util;
|
package org.apache.lucene.analysis;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
* Licensed to the Apache Software Foundation (ASF) under one or more
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link GermanLightStemFilter}
|
* Simple tests for {@link GermanLightStemFilter}
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link GermanMinimalStemFilter}
|
* Simple tests for {@link GermanMinimalStemFilter}
|
||||||
|
|
|
@ -27,7 +27,7 @@ import org.apache.lucene.analysis.core.KeywordTokenizer;
|
||||||
import org.apache.lucene.analysis.core.LowerCaseFilter;
|
import org.apache.lucene.analysis.core.LowerCaseFilter;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test the German stemmer. The stemming algorithm is known to work less
|
* Test the German stemmer. The stemming algorithm is known to work less
|
||||||
|
|
|
@ -17,7 +17,7 @@ package org.apache.lucene.analysis.en;
|
||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.assertVocabulary;
|
import static org.apache.lucene.analysis.VocabularyAssert.assertVocabulary;
|
||||||
|
|
||||||
import java.io.Reader;
|
import java.io.Reader;
|
||||||
|
|
||||||
|
|
|
@ -30,7 +30,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.TokenStream;
|
import org.apache.lucene.analysis.TokenStream;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test the PorterStemFilter with Martin Porter's test data.
|
* Test the PorterStemFilter with Martin Porter's test data.
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link SpanishLightStemFilter}
|
* Simple tests for {@link SpanishLightStemFilter}
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link FinnishLightStemFilter}
|
* Simple tests for {@link FinnishLightStemFilter}
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link FrenchLightStemFilter}
|
* Simple tests for {@link FrenchLightStemFilter}
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link FrenchMinimalStemFilter}
|
* Simple tests for {@link FrenchMinimalStemFilter}
|
||||||
|
|
|
@ -17,7 +17,7 @@ package org.apache.lucene.analysis.gl;
|
||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.assertVocabulary;
|
import static org.apache.lucene.analysis.VocabularyAssert.assertVocabulary;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.io.Reader;
|
import java.io.Reader;
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link HungarianLightStemFilter}
|
* Simple tests for {@link HungarianLightStemFilter}
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link ItalianLightStemFilter}
|
* Simple tests for {@link ItalianLightStemFilter}
|
||||||
|
|
|
@ -28,7 +28,7 @@ import org.apache.lucene.analysis.core.LowerCaseFilter;
|
||||||
import org.apache.lucene.analysis.standard.StandardTokenizer;
|
import org.apache.lucene.analysis.standard.StandardTokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link PortugueseLightStemFilter}
|
* Simple tests for {@link PortugueseLightStemFilter}
|
||||||
|
|
|
@ -28,7 +28,7 @@ import org.apache.lucene.analysis.core.LowerCaseFilter;
|
||||||
import org.apache.lucene.analysis.standard.StandardTokenizer;
|
import org.apache.lucene.analysis.standard.StandardTokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link PortugueseMinimalStemFilter}
|
* Simple tests for {@link PortugueseMinimalStemFilter}
|
||||||
|
|
|
@ -17,7 +17,7 @@ package org.apache.lucene.analysis.pt;
|
||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.assertVocabulary;
|
import static org.apache.lucene.analysis.VocabularyAssert.assertVocabulary;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.io.Reader;
|
import java.io.Reader;
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link RussianLightStemFilter}
|
* Simple tests for {@link RussianLightStemFilter}
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.core.KeywordTokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
import org.apache.lucene.util.LuceneTestCase;
|
import org.apache.lucene.util.LuceneTestCase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Test the snowball filters against the snowball data tests
|
* Test the snowball filters against the snowball data tests
|
||||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.analysis.MockTokenizer;
|
||||||
import org.apache.lucene.analysis.Tokenizer;
|
import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
import org.apache.lucene.analysis.util.ReusableAnalyzerBase;
|
||||||
|
|
||||||
import static org.apache.lucene.analysis.util.VocabularyAssert.*;
|
import static org.apache.lucene.analysis.VocabularyAssert.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple tests for {@link SwedishLightStemFilter}
|
* Simple tests for {@link SwedishLightStemFilter}
|
||||||
|
|
Loading…
Reference in New Issue