Adrien Grand e021451d6c LUCENE-5042: Fix the n-gram tokenizers and filters.
This commit fixes n-gram tokenizers and filters so that they handle
supplementary characters correctly and adds the ability to pre-tokenize the
stream in tokenizers.


git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@1492185 13f79535-47bb-0310-9956-ffa450edef68
2013-06-12 13:17:49 +00:00
2013-06-10 12:02:09 +00:00
2010-12-12 15:36:08 +00:00

Apache Lucene/Solr

lucene/ is a search engine library
solr/ is a search engine server that uses lucene

To compile the sources run 'ant compile'
To run all the tests run 'ant test'
To setup your ide run 'ant idea' or 'ant eclipse'
For Maven info, see dev-tools/maven/README.maven.

For more information on how to contribute see:
http://wiki.apache.org/lucene-java/HowToContribute
http://wiki.apache.org/solr/HowToContribute
Description
Apache Lucene open-source search software
Readme 846 MiB
Languages
Java 97.7%
HTML 1%
Python 0.9%
Lex 0.3%