mirror of https://github.com/apache/lucene.git
SOLR-2059: Add types attribute to WordDelimiterFilterFactory
git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@990451 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
eb899ea918
commit
33cc5a041e
|
@ -21,8 +21,9 @@ import static org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter.*;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* A BreakIterator-like API for iterating over subwords in text, according to WordDelimiterFilter rules.
|
* A BreakIterator-like API for iterating over subwords in text, according to WordDelimiterFilter rules.
|
||||||
|
* @lucene.internal
|
||||||
*/
|
*/
|
||||||
final class WordDelimiterIterator {
|
public final class WordDelimiterIterator {
|
||||||
|
|
||||||
/** Indicates the end of iteration */
|
/** Indicates the end of iteration */
|
||||||
public static final int DONE = -1;
|
public static final int DONE = -1;
|
||||||
|
@ -270,6 +271,16 @@ final class WordDelimiterIterator {
|
||||||
if (ch < charTypeTable.length) {
|
if (ch < charTypeTable.length) {
|
||||||
return charTypeTable[ch];
|
return charTypeTable[ch];
|
||||||
}
|
}
|
||||||
|
return getType(ch);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Computes the type of the given character
|
||||||
|
*
|
||||||
|
* @param ch Character whose type is to be determined
|
||||||
|
* @return Type of the character
|
||||||
|
*/
|
||||||
|
public static byte getType(int ch) {
|
||||||
switch (Character.getType(ch)) {
|
switch (Character.getType(ch)) {
|
||||||
case Character.UPPERCASE_LETTER: return UPPER;
|
case Character.UPPERCASE_LETTER: return UPPER;
|
||||||
case Character.LOWERCASE_LETTER: return LOWER;
|
case Character.LOWERCASE_LETTER: return LOWER;
|
||||||
|
|
|
@ -246,6 +246,10 @@ New Features
|
||||||
* SOLR-1316: Create autosuggest component.
|
* SOLR-1316: Create autosuggest component.
|
||||||
(Ankul Garg, Jason Rutherglen, Shalin Shekhar Mangar, gsingers, Robert Muir, ab)
|
(Ankul Garg, Jason Rutherglen, Shalin Shekhar Mangar, gsingers, Robert Muir, ab)
|
||||||
|
|
||||||
|
* SOLR-2059: Add "types" attribute to WordDelimiterFilterFactory, which
|
||||||
|
allows you to customize how WordDelimiterFilter tokenizes text with
|
||||||
|
a configuration file. (Peter Karich, rmuir)
|
||||||
|
|
||||||
|
|
||||||
Optimizations
|
Optimizations
|
||||||
----------------------
|
----------------------
|
||||||
|
|
|
@ -16,14 +16,23 @@
|
||||||
*/
|
*/
|
||||||
|
|
||||||
package org.apache.solr.analysis;
|
package org.apache.solr.analysis;
|
||||||
|
|
||||||
import org.apache.lucene.analysis.TokenStream;
|
import org.apache.lucene.analysis.TokenStream;
|
||||||
import org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter;
|
import org.apache.lucene.analysis.miscellaneous.WordDelimiterFilter;
|
||||||
|
import org.apache.lucene.analysis.miscellaneous.WordDelimiterIterator;
|
||||||
import org.apache.lucene.analysis.util.CharArraySet;
|
import org.apache.lucene.analysis.util.CharArraySet;
|
||||||
|
|
||||||
import org.apache.solr.util.plugin.ResourceLoaderAware;
|
import org.apache.solr.util.plugin.ResourceLoaderAware;
|
||||||
import org.apache.solr.common.ResourceLoader;
|
import org.apache.solr.common.ResourceLoader;
|
||||||
|
import org.apache.solr.common.util.StrUtils;
|
||||||
|
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
|
import java.util.SortedMap;
|
||||||
|
import java.util.TreeMap;
|
||||||
|
import java.util.regex.Matcher;
|
||||||
|
import java.util.regex.Pattern;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
|
|
||||||
|
|
||||||
|
@ -32,6 +41,7 @@ import java.io.IOException;
|
||||||
*/
|
*/
|
||||||
public class WordDelimiterFilterFactory extends BaseTokenFilterFactory implements ResourceLoaderAware {
|
public class WordDelimiterFilterFactory extends BaseTokenFilterFactory implements ResourceLoaderAware {
|
||||||
public static final String PROTECTED_TOKENS = "protected";
|
public static final String PROTECTED_TOKENS = "protected";
|
||||||
|
public static final String TYPES = "types";
|
||||||
|
|
||||||
public void inform(ResourceLoader loader) {
|
public void inform(ResourceLoader loader) {
|
||||||
String wordFiles = args.get(PROTECTED_TOKENS);
|
String wordFiles = args.get(PROTECTED_TOKENS);
|
||||||
|
@ -42,6 +52,20 @@ public class WordDelimiterFilterFactory extends BaseTokenFilterFactory implement
|
||||||
throw new RuntimeException(e);
|
throw new RuntimeException(e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
String types = args.get(TYPES);
|
||||||
|
if (types != null) {
|
||||||
|
try {
|
||||||
|
List<String> files = StrUtils.splitFileNames( types );
|
||||||
|
List<String> wlist = new ArrayList<String>();
|
||||||
|
for( String file : files ){
|
||||||
|
List<String> lines = loader.getLines( file.trim() );
|
||||||
|
wlist.addAll( lines );
|
||||||
|
}
|
||||||
|
typeTable = parseTypes(wlist);
|
||||||
|
} catch (IOException e) {
|
||||||
|
throw new RuntimeException(e);
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private CharArraySet protectedWords = null;
|
private CharArraySet protectedWords = null;
|
||||||
|
@ -55,6 +79,7 @@ public class WordDelimiterFilterFactory extends BaseTokenFilterFactory implement
|
||||||
int splitOnNumerics=0;
|
int splitOnNumerics=0;
|
||||||
int preserveOriginal=0;
|
int preserveOriginal=0;
|
||||||
int stemEnglishPossessive=0;
|
int stemEnglishPossessive=0;
|
||||||
|
byte[] typeTable = null;
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void init(Map<String, String> args) {
|
public void init(Map<String, String> args) {
|
||||||
|
@ -71,10 +96,87 @@ public class WordDelimiterFilterFactory extends BaseTokenFilterFactory implement
|
||||||
}
|
}
|
||||||
|
|
||||||
public WordDelimiterFilter create(TokenStream input) {
|
public WordDelimiterFilter create(TokenStream input) {
|
||||||
return new WordDelimiterFilter(input,
|
return new WordDelimiterFilter(input, typeTable == null ? WordDelimiterIterator.DEFAULT_WORD_DELIM_TABLE : typeTable,
|
||||||
generateWordParts, generateNumberParts,
|
generateWordParts, generateNumberParts,
|
||||||
catenateWords, catenateNumbers, catenateAll,
|
catenateWords, catenateNumbers, catenateAll,
|
||||||
splitOnCaseChange, preserveOriginal,
|
splitOnCaseChange, preserveOriginal,
|
||||||
splitOnNumerics, stemEnglishPossessive, protectedWords);
|
splitOnNumerics, stemEnglishPossessive, protectedWords);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// source => type
|
||||||
|
private static Pattern typePattern = Pattern.compile( "(.*)\\s*=>\\s*(.*)\\s*$" );
|
||||||
|
|
||||||
|
/** parses a list of MappingCharFilter style rules into a custom byte[] type table */
|
||||||
|
private byte[] parseTypes(List<String> rules) {
|
||||||
|
SortedMap<Character,Byte> typeMap = new TreeMap<Character,Byte>();
|
||||||
|
for( String rule : rules ){
|
||||||
|
Matcher m = typePattern.matcher(rule);
|
||||||
|
if( !m.find() )
|
||||||
|
throw new RuntimeException("Invalid Mapping Rule : [" + rule + "]");
|
||||||
|
String lhs = parseString(m.group(1).trim());
|
||||||
|
Byte rhs = parseType(m.group(2).trim());
|
||||||
|
if (lhs.length() != 1)
|
||||||
|
throw new RuntimeException("Invalid Mapping Rule : [" + rule + "]. Only a single character is allowed.");
|
||||||
|
if (rhs == null)
|
||||||
|
throw new RuntimeException("Invalid Mapping Rule : [" + rule + "]. Illegal type.");
|
||||||
|
typeMap.put(lhs.charAt(0), rhs);
|
||||||
|
}
|
||||||
|
|
||||||
|
// ensure the table is always at least as big as DEFAULT_WORD_DELIM_TABLE for performance
|
||||||
|
byte types[] = new byte[Math.max(typeMap.lastKey()+1, WordDelimiterIterator.DEFAULT_WORD_DELIM_TABLE.length)];
|
||||||
|
for (int i = 0; i < types.length; i++)
|
||||||
|
types[i] = WordDelimiterIterator.getType(i);
|
||||||
|
for (Map.Entry<Character,Byte> mapping : typeMap.entrySet())
|
||||||
|
types[mapping.getKey()] = mapping.getValue();
|
||||||
|
return types;
|
||||||
|
}
|
||||||
|
|
||||||
|
private Byte parseType(String s) {
|
||||||
|
if (s.equals("LOWER"))
|
||||||
|
return WordDelimiterFilter.LOWER;
|
||||||
|
else if (s.equals("UPPER"))
|
||||||
|
return WordDelimiterFilter.UPPER;
|
||||||
|
else if (s.equals("ALPHA"))
|
||||||
|
return WordDelimiterFilter.ALPHA;
|
||||||
|
else if (s.equals("DIGIT"))
|
||||||
|
return WordDelimiterFilter.DIGIT;
|
||||||
|
else if (s.equals("ALPHANUM"))
|
||||||
|
return WordDelimiterFilter.ALPHANUM;
|
||||||
|
else if (s.equals("SUBWORD_DELIM"))
|
||||||
|
return WordDelimiterFilter.SUBWORD_DELIM;
|
||||||
|
else
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
char[] out = new char[256];
|
||||||
|
|
||||||
|
private String parseString(String s){
|
||||||
|
int readPos = 0;
|
||||||
|
int len = s.length();
|
||||||
|
int writePos = 0;
|
||||||
|
while( readPos < len ){
|
||||||
|
char c = s.charAt( readPos++ );
|
||||||
|
if( c == '\\' ){
|
||||||
|
if( readPos >= len )
|
||||||
|
throw new RuntimeException( "Invalid escaped char in [" + s + "]" );
|
||||||
|
c = s.charAt( readPos++ );
|
||||||
|
switch( c ) {
|
||||||
|
case '\\' : c = '\\'; break;
|
||||||
|
case 'n' : c = '\n'; break;
|
||||||
|
case 't' : c = '\t'; break;
|
||||||
|
case 'r' : c = '\r'; break;
|
||||||
|
case 'b' : c = '\b'; break;
|
||||||
|
case 'f' : c = '\f'; break;
|
||||||
|
case 'u' :
|
||||||
|
if( readPos + 3 >= len )
|
||||||
|
throw new RuntimeException( "Invalid escaped char in [" + s + "]" );
|
||||||
|
c = (char)Integer.parseInt( s.substring( readPos, readPos + 4 ), 16 );
|
||||||
|
readPos += 4;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
out[writePos++] = c;
|
||||||
|
}
|
||||||
|
return new String( out, 0, writePos );
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -17,7 +17,15 @@
|
||||||
|
|
||||||
package org.apache.solr.analysis;
|
package org.apache.solr.analysis;
|
||||||
|
|
||||||
|
import java.io.StringReader;
|
||||||
|
import java.util.HashMap;
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
|
import org.apache.lucene.analysis.TokenStream;
|
||||||
|
import org.apache.lucene.analysis.core.WhitespaceTokenizer;
|
||||||
import org.apache.solr.SolrTestCaseJ4;
|
import org.apache.solr.SolrTestCaseJ4;
|
||||||
|
import org.apache.solr.common.ResourceLoader;
|
||||||
|
import org.apache.solr.core.SolrResourceLoader;
|
||||||
import org.junit.BeforeClass;
|
import org.junit.BeforeClass;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
|
||||||
|
@ -183,4 +191,51 @@ public class TestWordDelimiterFilterFactory extends SolrTestCaseJ4 {
|
||||||
);
|
);
|
||||||
clearIndex();
|
clearIndex();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testCustomTypes() throws Exception {
|
||||||
|
String testText = "I borrowed $5,400.00 at 25% interest-rate";
|
||||||
|
WordDelimiterFilterFactory factoryDefault = new WordDelimiterFilterFactory();
|
||||||
|
ResourceLoader loader = new SolrResourceLoader(null, null);
|
||||||
|
Map<String,String> args = new HashMap<String,String>();
|
||||||
|
args.put("generateWordParts", "1");
|
||||||
|
args.put("generateNumberParts", "1");
|
||||||
|
args.put("catenateWords", "1");
|
||||||
|
args.put("catenateNumbers", "1");
|
||||||
|
args.put("catenateAll", "0");
|
||||||
|
args.put("splitOnCaseChange", "1");
|
||||||
|
|
||||||
|
/* default behavior */
|
||||||
|
factoryDefault.init(args);
|
||||||
|
factoryDefault.inform(loader);
|
||||||
|
|
||||||
|
TokenStream ts = factoryDefault.create(
|
||||||
|
new WhitespaceTokenizer(BaseTokenTestCase.DEFAULT_VERSION, new StringReader(testText)));
|
||||||
|
BaseTokenTestCase.assertTokenStreamContents(ts,
|
||||||
|
new String[] { "I", "borrowed", "5", "400", "00", "540000", "at", "25", "interest", "rate", "interestrate" });
|
||||||
|
|
||||||
|
ts = factoryDefault.create(
|
||||||
|
new WhitespaceTokenizer(BaseTokenTestCase.DEFAULT_VERSION, new StringReader("foo\u200Dbar")));
|
||||||
|
BaseTokenTestCase.assertTokenStreamContents(ts,
|
||||||
|
new String[] { "foo", "bar", "foobar" });
|
||||||
|
|
||||||
|
|
||||||
|
/* custom behavior */
|
||||||
|
WordDelimiterFilterFactory factoryCustom = new WordDelimiterFilterFactory();
|
||||||
|
// use a custom type mapping
|
||||||
|
args.put("types", "wdftypes.txt");
|
||||||
|
factoryCustom.init(args);
|
||||||
|
factoryCustom.inform(loader);
|
||||||
|
|
||||||
|
ts = factoryCustom.create(
|
||||||
|
new WhitespaceTokenizer(BaseTokenTestCase.DEFAULT_VERSION, new StringReader(testText)));
|
||||||
|
BaseTokenTestCase.assertTokenStreamContents(ts,
|
||||||
|
new String[] { "I", "borrowed", "$5,400.00", "at", "25%", "interest", "rate", "interestrate" });
|
||||||
|
|
||||||
|
/* test custom behavior with a char > 0x7F, because we had to make a larger byte[] */
|
||||||
|
ts = factoryCustom.create(
|
||||||
|
new WhitespaceTokenizer(BaseTokenTestCase.DEFAULT_VERSION, new StringReader("foo\u200Dbar")));
|
||||||
|
BaseTokenTestCase.assertTokenStreamContents(ts,
|
||||||
|
new String[] { "foo\u200Dbar" });
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,32 @@
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
# contributor license agreements. See the NOTICE file distributed with
|
||||||
|
# this work for additional information regarding copyright ownership.
|
||||||
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
||||||
|
# (the "License"); you may not use this file except in compliance with
|
||||||
|
# the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
#
|
||||||
|
# A customized type mapping for WordDelimiterFilterFactory
|
||||||
|
# the allowable types are: LOWER, UPPER, ALPHA, DIGIT, ALPHANUM, SUBWORD_DELIM
|
||||||
|
#
|
||||||
|
# the default for any character without a mapping is always computed from
|
||||||
|
# Unicode character properties
|
||||||
|
|
||||||
|
# Map the $, %, '.', and ',' characters to DIGIT
|
||||||
|
# This might be useful for financial data.
|
||||||
|
$ => DIGIT
|
||||||
|
% => DIGIT
|
||||||
|
. => DIGIT
|
||||||
|
\u002C => DIGIT
|
||||||
|
|
||||||
|
# in some cases you might not want to split on ZWJ
|
||||||
|
# this also tests the case where we need a bigger byte[]
|
||||||
|
# see http://en.wikipedia.org/wiki/Zero-width_joiner
|
||||||
|
\u200D => ALPHANUM
|
Loading…
Reference in New Issue