mirror of https://github.com/apache/lucene.git
Remove HyphenationCompoundWordTokenFilter and DictionaryCompoundWordTokenFilter from the list of offset offenders.
git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@1486913 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
d7f397056b
commit
5de9c6af0b
|
@ -59,7 +59,6 @@ import org.apache.lucene.analysis.charfilter.NormalizeCharMap;
|
|||
import org.apache.lucene.analysis.cjk.CJKBigramFilter;
|
||||
import org.apache.lucene.analysis.commongrams.CommonGramsFilter;
|
||||
import org.apache.lucene.analysis.commongrams.CommonGramsQueryFilter;
|
||||
import org.apache.lucene.analysis.compound.DictionaryCompoundWordTokenFilter;
|
||||
import org.apache.lucene.analysis.compound.HyphenationCompoundWordTokenFilter;
|
||||
import org.apache.lucene.analysis.compound.TestCompoundWordTokenFilter;
|
||||
import org.apache.lucene.analysis.compound.hyphenation.HyphenationTree;
|
||||
|
@ -165,8 +164,6 @@ public class TestRandomChains extends BaseTokenStreamTestCase {
|
|||
for (Class<?> c : Arrays.<Class<?>>asList(
|
||||
ReversePathHierarchyTokenizer.class,
|
||||
PathHierarchyTokenizer.class,
|
||||
HyphenationCompoundWordTokenFilter.class,
|
||||
DictionaryCompoundWordTokenFilter.class,
|
||||
// TODO: it seems to mess up offsets!?
|
||||
WikipediaTokenizer.class,
|
||||
// TODO: doesn't handle graph inputs
|
||||
|
|
Loading…
Reference in New Issue