mirror of https://github.com/apache/lucene.git
LUCENE-1378 - Removed the remaining 199 @author references
git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@695514 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
8d02d7859f
commit
0195fcd03d
|
@ -37,7 +37,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
|
||||||
* will not be indexed at all) and an external list of exclusions (word that will
|
* will not be indexed at all) and an external list of exclusions (word that will
|
||||||
* not be stemmed, but indexed).
|
* not be stemmed, but indexed).
|
||||||
*
|
*
|
||||||
* @author João Kramer
|
|
||||||
*/
|
*/
|
||||||
public final class BrazilianAnalyzer extends Analyzer {
|
public final class BrazilianAnalyzer extends Analyzer {
|
||||||
|
|
||||||
|
|
|
@ -28,7 +28,6 @@ import java.util.Set;
|
||||||
/**
|
/**
|
||||||
* Based on GermanStemFilter
|
* Based on GermanStemFilter
|
||||||
*
|
*
|
||||||
* @author João Kramer
|
|
||||||
*/
|
*/
|
||||||
public final class BrazilianStemFilter extends TokenFilter {
|
public final class BrazilianStemFilter extends TokenFilter {
|
||||||
|
|
||||||
|
|
|
@ -28,7 +28,6 @@ import java.util.Set;
|
||||||
/**
|
/**
|
||||||
* Filters CJKTokenizer with StopFilter.
|
* Filters CJKTokenizer with StopFilter.
|
||||||
*
|
*
|
||||||
* @author Che, Dong
|
|
||||||
*/
|
*/
|
||||||
public class CJKAnalyzer extends Analyzer {
|
public class CJKAnalyzer extends Analyzer {
|
||||||
//~ Static fields/initializers ---------------------------------------------
|
//~ Static fields/initializers ---------------------------------------------
|
||||||
|
|
|
@ -34,7 +34,6 @@ import java.io.Reader;
|
||||||
* please search <a
|
* please search <a
|
||||||
* href="http://www.google.com/search?q=word+chinese+segment">google</a>
|
* href="http://www.google.com/search?q=word+chinese+segment">google</a>
|
||||||
*
|
*
|
||||||
* @author Che, Dong
|
|
||||||
*/
|
*/
|
||||||
public final class CJKTokenizer extends Tokenizer {
|
public final class CJKTokenizer extends Tokenizer {
|
||||||
//~ Static fields/initializers ---------------------------------------------
|
//~ Static fields/initializers ---------------------------------------------
|
||||||
|
|
|
@ -28,7 +28,6 @@ import org.apache.lucene.analysis.TokenStream;
|
||||||
* build from a ChineseTokenizer, filtered with ChineseFilter.
|
* build from a ChineseTokenizer, filtered with ChineseFilter.
|
||||||
* Copyright: Copyright (c) 2001
|
* Copyright: Copyright (c) 2001
|
||||||
* Company:
|
* Company:
|
||||||
* @author Yiyi Sun
|
|
||||||
* @version 1.0
|
* @version 1.0
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -37,7 +37,6 @@ import org.apache.lucene.analysis.TokenStream;
|
||||||
*
|
*
|
||||||
* Copyright: Copyright (c) 2001
|
* Copyright: Copyright (c) 2001
|
||||||
* Company:
|
* Company:
|
||||||
* @author Yiyi Sun
|
|
||||||
* @version 1.0
|
* @version 1.0
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -47,7 +47,6 @@ import org.apache.lucene.analysis.Tokenizer;
|
||||||
* C4C2, C1C2C3 ... the ChineseTokenizer works, but the
|
* C4C2, C1C2C3 ... the ChineseTokenizer works, but the
|
||||||
* CJKTokenizer will not work.
|
* CJKTokenizer will not work.
|
||||||
*
|
*
|
||||||
* @author Yiyi Sun
|
|
||||||
* @version 1.0
|
* @version 1.0
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -35,7 +35,6 @@ import java.util.Set;
|
||||||
* A default set of stopwords is used unless an alternative list is specified, the
|
* A default set of stopwords is used unless an alternative list is specified, the
|
||||||
* exclusion list is empty by default.
|
* exclusion list is empty by default.
|
||||||
*
|
*
|
||||||
* @author Lukas Zapletal [lzap@root.cz]
|
|
||||||
*/
|
*/
|
||||||
public final class CzechAnalyzer extends Analyzer {
|
public final class CzechAnalyzer extends Analyzer {
|
||||||
|
|
||||||
|
|
|
@ -32,7 +32,6 @@ import java.util.Set;
|
||||||
* that will not be indexed at all).
|
* that will not be indexed at all).
|
||||||
* A default set of stopwords is used unless an alternative list is specified.
|
* A default set of stopwords is used unless an alternative list is specified.
|
||||||
*
|
*
|
||||||
* @author Panagiotis Astithas, past@ebs.gr
|
|
||||||
*/
|
*/
|
||||||
public final class GreekAnalyzer extends Analyzer
|
public final class GreekAnalyzer extends Analyzer
|
||||||
{
|
{
|
||||||
|
|
|
@ -23,7 +23,6 @@ package org.apache.lucene.analysis.el;
|
||||||
* including accented ones. One should be able to add other encoding schemes (see RFC 1947) by adding
|
* including accented ones. One should be able to add other encoding schemes (see RFC 1947) by adding
|
||||||
* the definition of a new charset as well as the required logic in the toLowerCase() method.
|
* the definition of a new charset as well as the required logic in the toLowerCase() method.
|
||||||
*
|
*
|
||||||
* @author Panagiotis Astithas, past@ebs.gr
|
|
||||||
*/
|
*/
|
||||||
public class GreekCharsets
|
public class GreekCharsets
|
||||||
{
|
{
|
||||||
|
|
|
@ -23,7 +23,6 @@ import org.apache.lucene.analysis.TokenStream;
|
||||||
/**
|
/**
|
||||||
* Normalizes token text to lower case, analyzing given ("greek") charset.
|
* Normalizes token text to lower case, analyzing given ("greek") charset.
|
||||||
*
|
*
|
||||||
* @author Panagiotis Astithas, past@ebs.gr
|
|
||||||
*/
|
*/
|
||||||
public final class GreekLowerCaseFilter extends TokenFilter
|
public final class GreekLowerCaseFilter extends TokenFilter
|
||||||
{
|
{
|
||||||
|
|
|
@ -33,7 +33,6 @@ import org.apache.lucene.analysis.TokenFilter;
|
||||||
* Note that StandardTokenizer sees " ' " as a space, and cuts it out.
|
* Note that StandardTokenizer sees " ' " as a space, and cuts it out.
|
||||||
*
|
*
|
||||||
* @see <a href="http://fr.wikipedia.org/wiki/%C3%89lision">Elision in Wikipedia</a>
|
* @see <a href="http://fr.wikipedia.org/wiki/%C3%89lision">Elision in Wikipedia</a>
|
||||||
* @author <a href="mailto:mlecarme@openwide.fr">Mathieu Lecarme</a>
|
|
||||||
*/
|
*/
|
||||||
public class ElisionFilter extends TokenFilter {
|
public class ElisionFilter extends TokenFilter {
|
||||||
private Set articles = null;
|
private Set articles = null;
|
||||||
|
|
|
@ -31,7 +31,6 @@ import java.util.Set;
|
||||||
* not be stemmed at all. The used stemmer can be changed at runtime after the
|
* not be stemmed at all. The used stemmer can be changed at runtime after the
|
||||||
* filter object is created (as long as it is a FrenchStemmer).
|
* filter object is created (as long as it is a FrenchStemmer).
|
||||||
*
|
*
|
||||||
* @author Patrick Talbot (based on Gerhard Schwarz work for German)
|
|
||||||
*/
|
*/
|
||||||
public final class FrenchStemFilter extends TokenFilter {
|
public final class FrenchStemFilter extends TokenFilter {
|
||||||
|
|
||||||
|
|
|
@ -23,7 +23,6 @@ package org.apache.lucene.analysis.fr;
|
||||||
* refer to http://snowball.sourceforge.net/french/stemmer.html<br>
|
* refer to http://snowball.sourceforge.net/french/stemmer.html<br>
|
||||||
* (French stemming algorithm) for details
|
* (French stemming algorithm) for details
|
||||||
*
|
*
|
||||||
* @author Patrick Talbot
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class FrenchStemmer {
|
public class FrenchStemmer {
|
||||||
|
|
|
@ -26,7 +26,6 @@ import java.util.LinkedList;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tokenizes the input into n-grams of the given size(s).
|
* Tokenizes the input into n-grams of the given size(s).
|
||||||
* @author Otis Gospodnetic
|
|
||||||
*/
|
*/
|
||||||
public class NGramTokenFilter extends TokenFilter {
|
public class NGramTokenFilter extends TokenFilter {
|
||||||
public static final int DEFAULT_MIN_NGRAM_SIZE = 1;
|
public static final int DEFAULT_MIN_NGRAM_SIZE = 1;
|
||||||
|
|
|
@ -25,7 +25,6 @@ import java.io.Reader;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tokenizes the input into n-grams of the given size(s).
|
* Tokenizes the input into n-grams of the given size(s).
|
||||||
* @author Otis Gospodnetic
|
|
||||||
*/
|
*/
|
||||||
public class NGramTokenizer extends Tokenizer {
|
public class NGramTokenizer extends Tokenizer {
|
||||||
public static final int DEFAULT_MIN_NGRAM_SIZE = 1;
|
public static final int DEFAULT_MIN_NGRAM_SIZE = 1;
|
||||||
|
|
|
@ -39,7 +39,6 @@ import java.util.Map;
|
||||||
* A default set of stopwords is used unless an alternative list is specified, the
|
* A default set of stopwords is used unless an alternative list is specified, the
|
||||||
* exclusion list is empty by default.
|
* exclusion list is empty by default.
|
||||||
*
|
*
|
||||||
* @author Edwin de Jonge
|
|
||||||
*/
|
*/
|
||||||
public class DutchAnalyzer extends Analyzer {
|
public class DutchAnalyzer extends Analyzer {
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -32,7 +32,6 @@ import java.util.Map;
|
||||||
* not be stemmed at all. The stemmer used can be changed at runtime after the
|
* not be stemmed at all. The stemmer used can be changed at runtime after the
|
||||||
* filter object is created (as long as it is a DutchStemmer).
|
* filter object is created (as long as it is a DutchStemmer).
|
||||||
*
|
*
|
||||||
* @author Edwin de Jonge
|
|
||||||
*/
|
*/
|
||||||
public final class DutchStemFilter extends TokenFilter {
|
public final class DutchStemFilter extends TokenFilter {
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -25,7 +25,6 @@ import java.util.Map;
|
||||||
* the <a href="http://snowball.tartarus.org/algorithms/dutch/stemmer.html">dutch stemming</a>
|
* the <a href="http://snowball.tartarus.org/algorithms/dutch/stemmer.html">dutch stemming</a>
|
||||||
* algorithm in Martin Porter's snowball project.
|
* algorithm in Martin Porter's snowball project.
|
||||||
*
|
*
|
||||||
* @author Edwin de Jonge (ejne at cbs.nl)
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class DutchStemmer {
|
public class DutchStemmer {
|
||||||
|
|
|
@ -29,7 +29,6 @@ import java.util.HashMap;
|
||||||
* should contain only one word. If the file is not found or on any error, an
|
* should contain only one word. If the file is not found or on any error, an
|
||||||
* empty table is returned.
|
* empty table is returned.
|
||||||
*
|
*
|
||||||
* @author Gerhard Schwarz
|
|
||||||
* @deprecated use {@link org.apache.lucene.analysis.WordlistLoader} instead
|
* @deprecated use {@link org.apache.lucene.analysis.WordlistLoader} instead
|
||||||
*/
|
*/
|
||||||
public class WordlistLoader {
|
public class WordlistLoader {
|
||||||
|
|
|
@ -26,7 +26,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Analyzer for Thai language. It uses java.text.BreakIterator to break words.
|
* Analyzer for Thai language. It uses java.text.BreakIterator to break words.
|
||||||
* @author Samphan Raruenrom <samphan@osdev.co.th> for To-Be-One Technology Co., Ltd.
|
|
||||||
* @version 0.2
|
* @version 0.2
|
||||||
*/
|
*/
|
||||||
public class ThaiAnalyzer extends Analyzer {
|
public class ThaiAnalyzer extends Analyzer {
|
||||||
|
|
|
@ -27,7 +27,6 @@ import java.text.BreakIterator;
|
||||||
/**
|
/**
|
||||||
* TokenFilter that use java.text.BreakIterator to break each
|
* TokenFilter that use java.text.BreakIterator to break each
|
||||||
* Token that is Thai into separate Token(s) for each Thai word.
|
* Token that is Thai into separate Token(s) for each Thai word.
|
||||||
* @author Samphan Raruenrom <samphan@osdev.co.th> for To-Be-One Technology Co., Ltd.
|
|
||||||
* @version 0.2
|
* @version 0.2
|
||||||
*/
|
*/
|
||||||
public class ThaiWordFilter extends TokenFilter {
|
public class ThaiWordFilter extends TokenFilter {
|
||||||
|
|
|
@ -24,9 +24,7 @@ import junit.framework.TestCase;
|
||||||
|
|
||||||
import org.apache.lucene.analysis.Token;
|
import org.apache.lucene.analysis.Token;
|
||||||
|
|
||||||
/**
|
|
||||||
* @author rayt
|
|
||||||
*/
|
|
||||||
public class TestChineseTokenizer extends TestCase
|
public class TestChineseTokenizer extends TestCase
|
||||||
{
|
{
|
||||||
public void testOtherLetterOffset() throws IOException
|
public void testOtherLetterOffset() throws IOException
|
||||||
|
|
|
@ -34,7 +34,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
|
||||||
* than perfect, as it doesn't use any word lists with exceptions. We
|
* than perfect, as it doesn't use any word lists with exceptions. We
|
||||||
* also check some of the cases where the algorithm is wrong.
|
* also check some of the cases where the algorithm is wrong.
|
||||||
*
|
*
|
||||||
* @author Daniel Naber
|
|
||||||
*/
|
*/
|
||||||
public class TestGermanStemFilter extends TestCase {
|
public class TestGermanStemFilter extends TestCase {
|
||||||
|
|
||||||
|
|
|
@ -28,7 +28,6 @@ import junit.framework.TestCase;
|
||||||
/**
|
/**
|
||||||
* A unit test class for verifying the correct operation of the GreekAnalyzer.
|
* A unit test class for verifying the correct operation of the GreekAnalyzer.
|
||||||
*
|
*
|
||||||
* @author past
|
|
||||||
*/
|
*/
|
||||||
public class GreekAnalyzerTest extends TestCase {
|
public class GreekAnalyzerTest extends TestCase {
|
||||||
|
|
||||||
|
|
|
@ -32,7 +32,6 @@ import org.apache.lucene.analysis.Tokenizer;
|
||||||
import org.apache.lucene.analysis.standard.StandardTokenizer;
|
import org.apache.lucene.analysis.standard.StandardTokenizer;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @author Mathieu Lecarme<mlecarme@openwide.fr>
|
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
public class TestElision extends TestCase {
|
public class TestElision extends TestCase {
|
||||||
|
|
|
@ -65,7 +65,6 @@ import org.apache.lucene.analysis.TokenStream;
|
||||||
/**
|
/**
|
||||||
* Test case for FrenchAnalyzer.
|
* Test case for FrenchAnalyzer.
|
||||||
*
|
*
|
||||||
* @author Jean-François Halleux
|
|
||||||
* @version $version$
|
* @version $version$
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
|
|
@ -27,7 +27,6 @@ import junit.framework.TestCase;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tests {@link EdgeNGramTokenFilter} for correctness.
|
* Tests {@link EdgeNGramTokenFilter} for correctness.
|
||||||
* @author Otis Gospodnetic
|
|
||||||
*/
|
*/
|
||||||
public class EdgeNGramTokenFilterTest extends TestCase {
|
public class EdgeNGramTokenFilterTest extends TestCase {
|
||||||
private TokenStream input;
|
private TokenStream input;
|
||||||
|
|
|
@ -25,7 +25,6 @@ import junit.framework.TestCase;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tests {@link EdgeNGramTokenizer} for correctness.
|
* Tests {@link EdgeNGramTokenizer} for correctness.
|
||||||
* @author Otis Gospodnetic
|
|
||||||
*/
|
*/
|
||||||
public class EdgeNGramTokenizerTest extends TestCase {
|
public class EdgeNGramTokenizerTest extends TestCase {
|
||||||
private StringReader input;
|
private StringReader input;
|
||||||
|
|
|
@ -28,7 +28,6 @@ import junit.framework.TestCase;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tests {@link NGramTokenFilter} for correctness.
|
* Tests {@link NGramTokenFilter} for correctness.
|
||||||
* @author Otis Gospodnetic
|
|
||||||
*/
|
*/
|
||||||
public class NGramTokenFilterTest extends TestCase {
|
public class NGramTokenFilterTest extends TestCase {
|
||||||
private TokenStream input;
|
private TokenStream input;
|
||||||
|
|
|
@ -26,7 +26,6 @@ import junit.framework.TestCase;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tests {@link NGramTokenizer} for correctness.
|
* Tests {@link NGramTokenizer} for correctness.
|
||||||
* @author Otis Gospodnetic
|
|
||||||
*/
|
*/
|
||||||
public class NGramTokenizerTest extends TestCase {
|
public class NGramTokenizerTest extends TestCase {
|
||||||
private StringReader input;
|
private StringReader input;
|
||||||
|
|
|
@ -26,7 +26,6 @@ import org.apache.lucene.analysis.TokenStream;
|
||||||
/**
|
/**
|
||||||
* Test case for ThaiAnalyzer, modified from TestFrenchAnalyzer
|
* Test case for ThaiAnalyzer, modified from TestFrenchAnalyzer
|
||||||
*
|
*
|
||||||
* @author Samphan Raruenrom <samphan@osdev.co.th>
|
|
||||||
* @version 0.1
|
* @version 0.1
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
|
|
@ -24,7 +24,6 @@ import java.io.File;
|
||||||
/**
|
/**
|
||||||
* Allows a class to act as a Lucene document handler
|
* Allows a class to act as a Lucene document handler
|
||||||
*
|
*
|
||||||
*@author Erik Hatcher
|
|
||||||
*@since October 27, 2001
|
*@since October 27, 2001
|
||||||
*/
|
*/
|
||||||
public interface DocumentHandler {
|
public interface DocumentHandler {
|
||||||
|
|
|
@ -20,8 +20,7 @@ package org.apache.lucene.ant;
|
||||||
import java.io.PrintStream;
|
import java.io.PrintStream;
|
||||||
import java.io.PrintWriter;
|
import java.io.PrintWriter;
|
||||||
|
|
||||||
/**
|
|
||||||
*/
|
|
||||||
public class DocumentHandlerException extends Exception {
|
public class DocumentHandlerException extends Exception {
|
||||||
private Throwable cause;
|
private Throwable cause;
|
||||||
|
|
||||||
|
|
|
@ -26,7 +26,6 @@ import java.io.File;
|
||||||
* based on a files extension. Currently only .html and .txt
|
* based on a files extension. Currently only .html and .txt
|
||||||
* files are handled, other extensions ignored.
|
* files are handled, other extensions ignored.
|
||||||
*
|
*
|
||||||
*@author Erik Hatcher
|
|
||||||
*@since October 28, 2001
|
*@since October 28, 2001
|
||||||
*@todo Implement dynamic document type lookup
|
*@todo Implement dynamic document type lookup
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -39,7 +39,6 @@ import java.io.StringWriter;
|
||||||
* It does this by using JTidy package. It can take input input
|
* It does this by using JTidy package. It can take input input
|
||||||
* from {@link java.io.File} or {@link java.io.InputStream}.
|
* from {@link java.io.File} or {@link java.io.InputStream}.
|
||||||
*
|
*
|
||||||
*@author Erik Hatcher
|
|
||||||
*/
|
*/
|
||||||
public class HtmlDocument {
|
public class HtmlDocument {
|
||||||
private Element rawDoc;
|
private Element rawDoc;
|
||||||
|
|
|
@ -52,7 +52,6 @@ import java.text.ParseException;
|
||||||
/**
|
/**
|
||||||
* Ant task to index files with Lucene
|
* Ant task to index files with Lucene
|
||||||
*
|
*
|
||||||
*@author Erik Hatcher
|
|
||||||
*/
|
*/
|
||||||
public class IndexTask extends Task {
|
public class IndexTask extends Task {
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -29,7 +29,6 @@ import java.io.StringWriter;
|
||||||
/**
|
/**
|
||||||
* A utility for making Lucene Documents from a File.
|
* A utility for making Lucene Documents from a File.
|
||||||
*
|
*
|
||||||
*@author Erik Hatcher
|
|
||||||
*@since December 6, 2001
|
*@since December 6, 2001
|
||||||
*@todo Fix JavaDoc comments here
|
*@todo Fix JavaDoc comments here
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -37,7 +37,6 @@ import org.apache.tools.ant.types.FileSet;
|
||||||
/**
|
/**
|
||||||
* Test cases for index task
|
* Test cases for index task
|
||||||
*
|
*
|
||||||
*@author Erik Hatcher
|
|
||||||
*/
|
*/
|
||||||
public class IndexTaskTest extends TestCase {
|
public class IndexTaskTest extends TestCase {
|
||||||
private final static String docHandler =
|
private final static String docHandler =
|
||||||
|
|
|
@ -18,7 +18,6 @@ package org.apache.lucene.benchmark.stats;
|
||||||
/**
|
/**
|
||||||
* This class holds a set of memory usage values.
|
* This class holds a set of memory usage values.
|
||||||
*
|
*
|
||||||
* @author Andrzej Bialecki <ab@getopt.org>
|
|
||||||
*/
|
*/
|
||||||
public class MemUsage {
|
public class MemUsage {
|
||||||
public long maxFree, minFree, avgFree;
|
public long maxFree, minFree, avgFree;
|
||||||
|
|
|
@ -23,7 +23,6 @@ import org.apache.lucene.benchmark.Constants;
|
||||||
/**
|
/**
|
||||||
* This class holds parameters for a query benchmark.
|
* This class holds parameters for a query benchmark.
|
||||||
*
|
*
|
||||||
* @author Andrzej Bialecki <ab@getopt.org>
|
|
||||||
*/
|
*/
|
||||||
public class QueryData {
|
public class QueryData {
|
||||||
/** Benchmark id */
|
/** Benchmark id */
|
||||||
|
|
|
@ -36,7 +36,6 @@ import org.apache.lucene.store.Directory;
|
||||||
* This class holds together all parameters related to a test. Single test is
|
* This class holds together all parameters related to a test. Single test is
|
||||||
* performed several times, and all results are averaged.
|
* performed several times, and all results are averaged.
|
||||||
*
|
*
|
||||||
* @author Andrzej Bialecki <ab@getopt.org>
|
|
||||||
*/
|
*/
|
||||||
public class TestData
|
public class TestData
|
||||||
{
|
{
|
||||||
|
|
|
@ -26,7 +26,6 @@ import java.util.Iterator;
|
||||||
* values may contribute to different measurements, so this class provides also
|
* values may contribute to different measurements, so this class provides also
|
||||||
* some useful methods to separate them.
|
* some useful methods to separate them.
|
||||||
*
|
*
|
||||||
* @author Andrzej Bialecki <ab@getopt.org>
|
|
||||||
*/
|
*/
|
||||||
public class TestRunData {
|
public class TestRunData {
|
||||||
private String id;
|
private String id;
|
||||||
|
|
|
@ -19,7 +19,6 @@ package org.apache.lucene.benchmark.stats;
|
||||||
/**
|
/**
|
||||||
* This class holds a data point measuring speed of processing.
|
* This class holds a data point measuring speed of processing.
|
||||||
*
|
*
|
||||||
* @author Andrzej Bialecki <ab@getopt.org>
|
|
||||||
*/
|
*/
|
||||||
public class TimeData {
|
public class TimeData {
|
||||||
/** Name of the data point - usually one of a data series with the same name */
|
/** Name of the data point - usually one of a data series with the same name */
|
||||||
|
|
|
@ -25,7 +25,6 @@ import com.sleepycat.je.DatabaseException;
|
||||||
/**
|
/**
|
||||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||||
*
|
*
|
||||||
* @author Aaron Donovan
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class Block extends Object {
|
public class Block extends Object {
|
||||||
|
|
|
@ -34,7 +34,6 @@ import com.sleepycat.je.Transaction;
|
||||||
/**
|
/**
|
||||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||||
*
|
*
|
||||||
* @author Aaron Donovan
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class File extends Object {
|
public class File extends Object {
|
||||||
|
|
|
@ -47,7 +47,6 @@ import com.sleepycat.je.Transaction;
|
||||||
* {@link com.sleepycat.je.Database Db} database handles, one for storing file
|
* {@link com.sleepycat.je.Database Db} database handles, one for storing file
|
||||||
* records and another for storing file data blocks.
|
* records and another for storing file data blocks.
|
||||||
*
|
*
|
||||||
* @author Aaron Donovan
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class JEDirectory extends Directory {
|
public class JEDirectory extends Directory {
|
||||||
|
|
|
@ -23,7 +23,6 @@ import org.apache.lucene.store.IndexInput;
|
||||||
/**
|
/**
|
||||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||||
*
|
*
|
||||||
* @author Aaron Donovan
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class JEIndexInput extends IndexInput {
|
public class JEIndexInput extends IndexInput {
|
||||||
|
|
|
@ -24,7 +24,6 @@ import org.apache.lucene.store.IndexOutput;
|
||||||
/**
|
/**
|
||||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||||
*
|
*
|
||||||
* @author Aaron Donovan
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class JEIndexOutput extends IndexOutput {
|
public class JEIndexOutput extends IndexOutput {
|
||||||
|
|
|
@ -22,7 +22,6 @@ import org.apache.lucene.store.Lock;
|
||||||
/**
|
/**
|
||||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||||
*
|
*
|
||||||
* @author Aaron Donovan
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class JELock extends Lock {
|
public class JELock extends Lock {
|
||||||
|
|
|
@ -45,7 +45,6 @@ import com.sleepycat.je.Transaction;
|
||||||
*
|
*
|
||||||
* Adapted from Andi Vajda's org.apache.lucene.db.DbStoreTest.
|
* Adapted from Andi Vajda's org.apache.lucene.db.DbStoreTest.
|
||||||
*
|
*
|
||||||
* @author Aaron Donovan
|
|
||||||
*/
|
*/
|
||||||
public class JEStoreTest extends TestCase {
|
public class JEStoreTest extends TestCase {
|
||||||
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");
|
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");
|
||||||
|
|
|
@ -30,7 +30,6 @@ import com.sleepycat.db.internal.DbTxn;
|
||||||
* that were the entrypoint objects in the pre-4.3 Java API and that wrap the
|
* that were the entrypoint objects in the pre-4.3 Java API and that wrap the
|
||||||
* actual Berkeley DB C objects via SWIG.
|
* actual Berkeley DB C objects via SWIG.
|
||||||
*
|
*
|
||||||
* @author Andi Vajda
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class DbHandleExtractor {
|
public class DbHandleExtractor {
|
||||||
|
|
|
@ -24,9 +24,7 @@ import com.sleepycat.db.internal.Db;
|
||||||
import com.sleepycat.db.internal.DbTxn;
|
import com.sleepycat.db.internal.DbTxn;
|
||||||
import com.sleepycat.db.DatabaseException;
|
import com.sleepycat.db.DatabaseException;
|
||||||
|
|
||||||
/**
|
|
||||||
* @author Andi Vajda
|
|
||||||
*/
|
|
||||||
|
|
||||||
public class Block extends Object {
|
public class Block extends Object {
|
||||||
protected DatabaseEntry key, data;
|
protected DatabaseEntry key, data;
|
||||||
|
|
|
@ -50,7 +50,6 @@ import com.sleepycat.db.DbHandleExtractor;
|
||||||
* {@link com.sleepycat.db.internal.Db Db} database handles, one for storing file
|
* {@link com.sleepycat.db.internal.Db Db} database handles, one for storing file
|
||||||
* records and another for storing file data blocks.
|
* records and another for storing file data blocks.
|
||||||
*
|
*
|
||||||
* @author Andi Vajda
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class DbDirectory extends Directory {
|
public class DbDirectory extends Directory {
|
||||||
|
|
|
@ -21,9 +21,7 @@ import java.io.IOException;
|
||||||
import org.apache.lucene.store.IndexInput;
|
import org.apache.lucene.store.IndexInput;
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @author Andi Vajda
|
|
||||||
*/
|
|
||||||
|
|
||||||
public class DbIndexInput extends IndexInput {
|
public class DbIndexInput extends IndexInput {
|
||||||
|
|
||||||
|
|
|
@ -21,9 +21,7 @@ import java.io.IOException;
|
||||||
import org.apache.lucene.store.IndexOutput;
|
import org.apache.lucene.store.IndexOutput;
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @author Andi Vajda
|
|
||||||
*/
|
|
||||||
|
|
||||||
public class DbIndexOutput extends IndexOutput {
|
public class DbIndexOutput extends IndexOutput {
|
||||||
|
|
||||||
|
|
|
@ -24,7 +24,6 @@ import org.apache.lucene.store.Lock;
|
||||||
* trivial as {@link DbDirectory} operations are managed by the Berkeley DB
|
* trivial as {@link DbDirectory} operations are managed by the Berkeley DB
|
||||||
* locking system.
|
* locking system.
|
||||||
*
|
*
|
||||||
* @author Andi Vajda
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class DbLock extends Lock {
|
public class DbLock extends Lock {
|
||||||
|
|
|
@ -31,9 +31,7 @@ import com.sleepycat.db.internal.Db;
|
||||||
import com.sleepycat.db.internal.DbTxn;
|
import com.sleepycat.db.internal.DbTxn;
|
||||||
import com.sleepycat.db.DatabaseException;
|
import com.sleepycat.db.DatabaseException;
|
||||||
|
|
||||||
/**
|
|
||||||
* @author Andi Vajda
|
|
||||||
*/
|
|
||||||
|
|
||||||
public class File extends Object {
|
public class File extends Object {
|
||||||
|
|
||||||
|
|
|
@ -42,7 +42,6 @@ import org.apache.lucene.store.IndexOutput;
|
||||||
* Tests {@link DbDirectory}.
|
* Tests {@link DbDirectory}.
|
||||||
*
|
*
|
||||||
* Adapted from org.apache.lucene.StoreTest with larger files and random bytes.
|
* Adapted from org.apache.lucene.StoreTest with larger files and random bytes.
|
||||||
* @author Andi Vajda
|
|
||||||
*/
|
*/
|
||||||
public class DbStoreTest extends TestCase {
|
public class DbStoreTest extends TestCase {
|
||||||
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");
|
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");
|
||||||
|
|
|
@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple {@link Encoder} implementation that does not modify the output
|
* Simple {@link Encoder} implementation that does not modify the output
|
||||||
* @author Nicko Cadell
|
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
public class DefaultEncoder implements Encoder
|
public class DefaultEncoder implements Encoder
|
||||||
|
|
|
@ -19,7 +19,6 @@ package org.apache.lucene.search.highlight;
|
||||||
/**
|
/**
|
||||||
* Encodes original text. The Encoder works with the Formatter to generate the output.
|
* Encodes original text. The Encoder works with the Formatter to generate the output.
|
||||||
*
|
*
|
||||||
* @author Nicko Cadell
|
|
||||||
*/
|
*/
|
||||||
public interface Encoder
|
public interface Encoder
|
||||||
{
|
{
|
||||||
|
|
|
@ -22,7 +22,6 @@ import org.apache.lucene.analysis.Token;
|
||||||
* Implements the policy for breaking text into multiple fragments for consideration
|
* Implements the policy for breaking text into multiple fragments for consideration
|
||||||
* by the {@link Highlighter} class. A sophisticated implementation may do this on the basis
|
* by the {@link Highlighter} class. A sophisticated implementation may do this on the basis
|
||||||
* of detecting end of sentences in the text.
|
* of detecting end of sentences in the text.
|
||||||
* @author mark@searcharea.co.uk
|
|
||||||
*/
|
*/
|
||||||
public interface Fragmenter
|
public interface Fragmenter
|
||||||
{
|
{
|
||||||
|
|
|
@ -20,7 +20,6 @@ package org.apache.lucene.search.highlight;
|
||||||
* Formats text with different color intensity depending on the score of the
|
* Formats text with different color intensity depending on the score of the
|
||||||
* term.
|
* term.
|
||||||
*
|
*
|
||||||
* @author maharwood
|
|
||||||
*/
|
*/
|
||||||
public class GradientFormatter implements Formatter
|
public class GradientFormatter implements Formatter
|
||||||
{
|
{
|
||||||
|
|
|
@ -30,7 +30,6 @@ import org.apache.lucene.util.PriorityQueue;
|
||||||
* Class used to markup highlighted terms found in the best sections of a
|
* Class used to markup highlighted terms found in the best sections of a
|
||||||
* text, using configurable {@link Fragmenter}, {@link Scorer}, {@link Formatter},
|
* text, using configurable {@link Fragmenter}, {@link Scorer}, {@link Formatter},
|
||||||
* {@link Encoder} and tokenizers.
|
* {@link Encoder} and tokenizers.
|
||||||
* @author mark@searcharea.co.uk
|
|
||||||
*/
|
*/
|
||||||
public class Highlighter
|
public class Highlighter
|
||||||
{
|
{
|
||||||
|
@ -503,8 +502,7 @@ public class Highlighter
|
||||||
this.maxDocCharsToAnalyze = maxDocCharsToAnalyze;
|
this.maxDocCharsToAnalyze = maxDocCharsToAnalyze;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
*/
|
|
||||||
public Fragmenter getTextFragmenter()
|
public Fragmenter getTextFragmenter()
|
||||||
{
|
{
|
||||||
return textFragmenter;
|
return textFragmenter;
|
||||||
|
|
|
@ -27,7 +27,6 @@ import org.apache.lucene.search.Query;
|
||||||
* {@link Scorer} implementation which scores text fragments by the number of unique query terms found.
|
* {@link Scorer} implementation which scores text fragments by the number of unique query terms found.
|
||||||
* This class uses the {@link QueryTermExtractor} class to process determine the query terms and
|
* This class uses the {@link QueryTermExtractor} class to process determine the query terms and
|
||||||
* their boosts to be used.
|
* their boosts to be used.
|
||||||
* @author mark@searcharea.co.uk
|
|
||||||
*/
|
*/
|
||||||
//TODO: provide option to boost score of fragments near beginning of document
|
//TODO: provide option to boost score of fragments near beginning of document
|
||||||
// based on fragment.getFragNum()
|
// based on fragment.getFragNum()
|
||||||
|
|
|
@ -20,7 +20,6 @@ import org.apache.lucene.analysis.Token;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Adds to the score for a fragment based on its tokens
|
* Adds to the score for a fragment based on its tokens
|
||||||
* @author mark@searcharea.co.uk
|
|
||||||
*/
|
*/
|
||||||
public interface Scorer
|
public interface Scorer
|
||||||
{
|
{
|
||||||
|
|
|
@ -21,7 +21,6 @@ import org.apache.lucene.analysis.Token;
|
||||||
/**
|
/**
|
||||||
* {@link Fragmenter} implementation which breaks text up into same-size
|
* {@link Fragmenter} implementation which breaks text up into same-size
|
||||||
* fragments with no concerns over spotting sentence boundaries.
|
* fragments with no concerns over spotting sentence boundaries.
|
||||||
* @author mark@searcharea.co.uk
|
|
||||||
*/
|
*/
|
||||||
public class SimpleFragmenter implements Fragmenter
|
public class SimpleFragmenter implements Fragmenter
|
||||||
{
|
{
|
||||||
|
|
|
@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple {@link Encoder} implementation to escape text for HTML output
|
* Simple {@link Encoder} implementation to escape text for HTML output
|
||||||
* @author Nicko Cadell
|
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
public class SimpleHTMLEncoder implements Encoder
|
public class SimpleHTMLEncoder implements Encoder
|
||||||
|
|
|
@ -18,7 +18,6 @@ package org.apache.lucene.search.highlight;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Simple {@link Formatter} implementation to highlight terms with a pre and post tag
|
* Simple {@link Formatter} implementation to highlight terms with a pre and post tag
|
||||||
* @author MAHarwood
|
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
public class SimpleHTMLFormatter implements Formatter
|
public class SimpleHTMLFormatter implements Formatter
|
||||||
|
|
|
@ -20,7 +20,6 @@ package org.apache.lucene.search.highlight;
|
||||||
/**
|
/**
|
||||||
* Low-level class used to record information about a section of a document
|
* Low-level class used to record information about a section of a document
|
||||||
* with a score.
|
* with a score.
|
||||||
* @author MAHarwood
|
|
||||||
*
|
*
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -20,7 +20,6 @@ import org.apache.lucene.analysis.Token;
|
||||||
/**
|
/**
|
||||||
* One, or several overlapping tokens, along with the score(s) and the
|
* One, or several overlapping tokens, along with the score(s) and the
|
||||||
* scope of the original text
|
* scope of the original text
|
||||||
* @author MAHarwood
|
|
||||||
*/
|
*/
|
||||||
public class TokenGroup
|
public class TokenGroup
|
||||||
{
|
{
|
||||||
|
|
|
@ -39,7 +39,6 @@ import org.apache.lucene.index.TermVectorOffsetInfo;
|
||||||
* Hides implementation issues associated with obtaining a TokenStream for use with
|
* Hides implementation issues associated with obtaining a TokenStream for use with
|
||||||
* the higlighter - can obtain from TermFreqVectors with offsets and (optionally) positions or
|
* the higlighter - can obtain from TermFreqVectors with offsets and (optionally) positions or
|
||||||
* from Analyzer class reparsing the stored content.
|
* from Analyzer class reparsing the stored content.
|
||||||
* @author maharwood
|
|
||||||
*/
|
*/
|
||||||
public class TokenSources
|
public class TokenSources
|
||||||
{
|
{
|
||||||
|
|
|
@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
|
||||||
*/
|
*/
|
||||||
|
|
||||||
/** Lightweight class to hold term and a weight value used for scoring this term
|
/** Lightweight class to hold term and a weight value used for scoring this term
|
||||||
* @author Mark Harwood
|
|
||||||
*/
|
*/
|
||||||
public class WeightedTerm
|
public class WeightedTerm
|
||||||
{
|
{
|
||||||
|
|
|
@ -77,7 +77,6 @@ import org.w3c.dom.NodeList;
|
||||||
/**
|
/**
|
||||||
* JUnit Test for Highlighter class.
|
* JUnit Test for Highlighter class.
|
||||||
*
|
*
|
||||||
* @author mark@searcharea.co.uk
|
|
||||||
*/
|
*/
|
||||||
public class HighlighterTest extends TestCase implements Formatter {
|
public class HighlighterTest extends TestCase implements Formatter {
|
||||||
private IndexReader reader;
|
private IndexReader reader;
|
||||||
|
@ -1448,7 +1447,6 @@ class SynonymAnalyzer extends Analyzer {
|
||||||
/**
|
/**
|
||||||
* Expands a token stream with synonyms (TODO - make the synonyms analyzed by choice of analyzer)
|
* Expands a token stream with synonyms (TODO - make the synonyms analyzed by choice of analyzer)
|
||||||
*
|
*
|
||||||
* @author MAHarwood
|
|
||||||
*/
|
*/
|
||||||
class SynonymTokenizer extends TokenStream {
|
class SynonymTokenizer extends TokenStream {
|
||||||
private TokenStream realStream;
|
private TokenStream realStream;
|
||||||
|
|
|
@ -39,7 +39,6 @@ import org.apache.lucene.analysis.TokenStream;
|
||||||
* Various fulltext analysis utilities avoiding redundant code in several
|
* Various fulltext analysis utilities avoiding redundant code in several
|
||||||
* classes.
|
* classes.
|
||||||
*
|
*
|
||||||
* @author whoschek.AT.lbl.DOT.gov
|
|
||||||
*/
|
*/
|
||||||
public class AnalyzerUtil {
|
public class AnalyzerUtil {
|
||||||
|
|
||||||
|
|
|
@ -167,7 +167,6 @@ import org.apache.lucene.search.Similarity;
|
||||||
* href="http://java.sun.com/developer/technicalArticles/Programming/HPROF.html">
|
* href="http://java.sun.com/developer/technicalArticles/Programming/HPROF.html">
|
||||||
* hprof tracing </a>).
|
* hprof tracing </a>).
|
||||||
*
|
*
|
||||||
* @author whoschek.AT.lbl.DOT.gov
|
|
||||||
*/
|
*/
|
||||||
public class MemoryIndex implements Serializable {
|
public class MemoryIndex implements Serializable {
|
||||||
|
|
||||||
|
|
|
@ -61,7 +61,6 @@ import org.apache.lucene.analysis.TokenStream;
|
||||||
* "English"));
|
* "English"));
|
||||||
* </pre>
|
* </pre>
|
||||||
*
|
*
|
||||||
* @author whoschek.AT.lbl.DOT.gov
|
|
||||||
*/
|
*/
|
||||||
public class PatternAnalyzer extends Analyzer {
|
public class PatternAnalyzer extends Analyzer {
|
||||||
|
|
||||||
|
|
|
@ -68,7 +68,6 @@ import java.util.TreeSet;
|
||||||
* xxxx:[]
|
* xxxx:[]
|
||||||
* </pre>
|
* </pre>
|
||||||
*
|
*
|
||||||
* @author whoschek.AT.lbl.DOT.gov
|
|
||||||
* @see <a target="_blank"
|
* @see <a target="_blank"
|
||||||
* href="http://www.cogsci.princeton.edu/~wn/man/prologdb.5WN.html">prologdb
|
* href="http://www.cogsci.princeton.edu/~wn/man/prologdb.5WN.html">prologdb
|
||||||
* man page </a>
|
* man page </a>
|
||||||
|
|
|
@ -28,7 +28,6 @@ import org.apache.lucene.analysis.TokenStream;
|
||||||
* underlying child stream; the child stream must deliver lowercase tokens
|
* underlying child stream; the child stream must deliver lowercase tokens
|
||||||
* for synonyms to be found.
|
* for synonyms to be found.
|
||||||
*
|
*
|
||||||
* @author whoschek.AT.lbl.DOT.gov
|
|
||||||
*/
|
*/
|
||||||
public class SynonymTokenFilter extends TokenFilter {
|
public class SynonymTokenFilter extends TokenFilter {
|
||||||
|
|
||||||
|
|
|
@ -196,7 +196,6 @@ hello^2.0
|
||||||
the^3
|
the^3
|
||||||
</pre>
|
</pre>
|
||||||
|
|
||||||
@author whoschek.AT.lbl.DOT.gov
|
|
||||||
*/
|
*/
|
||||||
public class MemoryIndexTest extends TestCase {
|
public class MemoryIndexTest extends TestCase {
|
||||||
|
|
||||||
|
|
|
@ -58,7 +58,6 @@ Lucene features: CharTokenizer.MAX_WORD_LEN = 255.
|
||||||
Thus the PatternAnalyzer produces correct output, whereas the WhitespaceAnalyzer
|
Thus the PatternAnalyzer produces correct output, whereas the WhitespaceAnalyzer
|
||||||
silently truncates text, and so the comparison results in assertEquals() don't match up.
|
silently truncates text, and so the comparison results in assertEquals() don't match up.
|
||||||
|
|
||||||
@author whoschek.AT.lbl.DOT.gov
|
|
||||||
*/
|
*/
|
||||||
public class PatternAnalyzerTest extends TestCase {
|
public class PatternAnalyzerTest extends TestCase {
|
||||||
|
|
||||||
|
|
|
@ -34,8 +34,6 @@ import org.apache.lucene.store.FSDirectory;
|
||||||
* NOTE: This will overwrite any length normalization or field/document boosts.
|
* NOTE: This will overwrite any length normalization or field/document boosts.
|
||||||
* </p>
|
* </p>
|
||||||
*
|
*
|
||||||
* @author Chris Hostetter
|
|
||||||
* @author Otis Gospodnetic
|
|
||||||
*/
|
*/
|
||||||
public class FieldNormModifier {
|
public class FieldNormModifier {
|
||||||
|
|
||||||
|
|
|
@ -79,7 +79,6 @@ import org.apache.lucene.util.SortedVIntList;
|
||||||
* /least computationally-intensive filters first.
|
* /least computationally-intensive filters first.
|
||||||
* </p>
|
* </p>
|
||||||
*
|
*
|
||||||
* @author <a href="mailto:kelvint@apache.org">Kelvin Tan</a>
|
|
||||||
*/
|
*/
|
||||||
public class ChainedFilter extends Filter
|
public class ChainedFilter extends Filter
|
||||||
{
|
{
|
||||||
|
|
|
@ -27,7 +27,6 @@ import java.io.IOException;
|
||||||
/**
|
/**
|
||||||
* Merges indices specified on the command line into the index
|
* Merges indices specified on the command line into the index
|
||||||
* specified as the first command line argument.
|
* specified as the first command line argument.
|
||||||
* @author Erik Hatcher
|
|
||||||
* @version $Id$
|
* @version $Id$
|
||||||
*/
|
*/
|
||||||
public class IndexMergeTool {
|
public class IndexMergeTool {
|
||||||
|
|
|
@ -39,7 +39,6 @@ import org.apache.lucene.search.Query;
|
||||||
* become <code>h?user</code> when using this parser and thus no match would be found (i.e.
|
* become <code>h?user</code> when using this parser and thus no match would be found (i.e.
|
||||||
* using this parser will be no improvement over QueryParser in such cases).
|
* using this parser will be no improvement over QueryParser in such cases).
|
||||||
*
|
*
|
||||||
* @author Ronnie Kolehmainen (ronnie.kolehmainen at ub.uu.se)
|
|
||||||
* @version $Revision$, $Date$
|
* @version $Revision$, $Date$
|
||||||
*/
|
*/
|
||||||
public class AnalyzingQueryParser extends org.apache.lucene.queryParser.QueryParser {
|
public class AnalyzingQueryParser extends org.apache.lucene.queryParser.QueryParser {
|
||||||
|
|
|
@ -70,9 +70,6 @@ import org.apache.lucene.util.Parameter;
|
||||||
* documentation</a>.
|
* documentation</a>.
|
||||||
* </p>
|
* </p>
|
||||||
*
|
*
|
||||||
* @author Brian Goetz
|
|
||||||
* @author Peter Halacsy
|
|
||||||
* @author Tatu Saloranta
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class PrecedenceQueryParser implements PrecedenceQueryParserConstants {
|
public class PrecedenceQueryParser implements PrecedenceQueryParserConstants {
|
||||||
|
|
|
@ -30,7 +30,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
|
||||||
import org.apache.lucene.queryParser.ParseException;
|
import org.apache.lucene.queryParser.ParseException;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @author Ronnie Kolehmainen (ronnie.kolehmainen at ub.uu.se)
|
|
||||||
* @version $Revision$, $Date$
|
* @version $Revision$, $Date$
|
||||||
*/
|
*/
|
||||||
public class TestAnalyzingQueryParser extends TestCase {
|
public class TestAnalyzingQueryParser extends TestCase {
|
||||||
|
|
|
@ -36,7 +36,6 @@ import org.apache.lucene.util.SortedVIntList;
|
||||||
* SHOULD Filters are OR'd together
|
* SHOULD Filters are OR'd together
|
||||||
* The resulting Filter is NOT'd with the NOT Filters
|
* The resulting Filter is NOT'd with the NOT Filters
|
||||||
* The resulting Filter is AND'd with the MUST Filters
|
* The resulting Filter is AND'd with the MUST Filters
|
||||||
* @author BPDThebault
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class BooleanFilter extends Filter
|
public class BooleanFilter extends Filter
|
||||||
|
|
|
@ -24,7 +24,6 @@ import org.apache.lucene.search.BooleanClause.Occur;
|
||||||
* is used when composed with another filter.
|
* is used when composed with another filter.
|
||||||
* (Follows the boolean logic in BooleanClause for composition
|
* (Follows the boolean logic in BooleanClause for composition
|
||||||
* of queries.)
|
* of queries.)
|
||||||
* @author BPDThebault
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
public class FilterClause implements java.io.Serializable
|
public class FilterClause implements java.io.Serializable
|
||||||
|
|
|
@ -46,7 +46,6 @@ import org.apache.lucene.util.PriorityQueue;
|
||||||
* terms eg misspellings. Instead, all variants use the same IDF ranking (the one for the source query
|
* terms eg misspellings. Instead, all variants use the same IDF ranking (the one for the source query
|
||||||
* term) and this is factored into the variant's boost. If the source query term does not exist in the
|
* term) and this is factored into the variant's boost. If the source query term does not exist in the
|
||||||
* index the average IDF of the variants is used.
|
* index the average IDF of the variants is used.
|
||||||
* @author maharwood
|
|
||||||
*/
|
*/
|
||||||
public class FuzzyLikeThisQuery extends Query
|
public class FuzzyLikeThisQuery extends Query
|
||||||
{
|
{
|
||||||
|
|
|
@ -35,7 +35,6 @@ import org.apache.lucene.util.OpenBitSet;
|
||||||
* a choice of "category" labels picked by the end user. As a filter, this is much faster than the
|
* a choice of "category" labels picked by the end user. As a filter, this is much faster than the
|
||||||
* equivalent query (a BooleanQuery with many "should" TermQueries)
|
* equivalent query (a BooleanQuery with many "should" TermQueries)
|
||||||
*
|
*
|
||||||
* @author maharwood
|
|
||||||
*/
|
*/
|
||||||
public class TermsFilter extends Filter
|
public class TermsFilter extends Filter
|
||||||
{
|
{
|
||||||
|
|
|
@ -144,9 +144,6 @@ import java.util.ArrayList;
|
||||||
* - optimise: when no termvector support available - used maxNumTermsParsed to limit amount of tokenization
|
* - optimise: when no termvector support available - used maxNumTermsParsed to limit amount of tokenization
|
||||||
* </pre>
|
* </pre>
|
||||||
*
|
*
|
||||||
* @author David Spencer
|
|
||||||
* @author Bruce Ritchie
|
|
||||||
* @author Mark Harwood
|
|
||||||
*/
|
*/
|
||||||
public final class MoreLikeThis {
|
public final class MoreLikeThis {
|
||||||
|
|
||||||
|
|
|
@ -35,7 +35,6 @@ import org.apache.lucene.search.similar.MoreLikeThis;
|
||||||
* A simple wrapper for MoreLikeThis for use in scenarios where a Query object is required eg
|
* A simple wrapper for MoreLikeThis for use in scenarios where a Query object is required eg
|
||||||
* in custom QueryParser extensions. At query.rewrite() time the reader is used to construct the
|
* in custom QueryParser extensions. At query.rewrite() time the reader is used to construct the
|
||||||
* actual MoreLikeThis object and obtain the real Query object.
|
* actual MoreLikeThis object and obtain the real Query object.
|
||||||
* @author maharwood
|
|
||||||
*/
|
*/
|
||||||
public class MoreLikeThisQuery extends Query
|
public class MoreLikeThisQuery extends Query
|
||||||
{
|
{
|
||||||
|
|
|
@ -35,7 +35,6 @@ import org.apache.lucene.store.RAMDirectory;
|
||||||
* It first creates a simple index and then a couple of instances of LuceneDictionary
|
* It first creates a simple index and then a couple of instances of LuceneDictionary
|
||||||
* on different fields and checks if all the right text comes back.
|
* on different fields and checks if all the right text comes back.
|
||||||
*
|
*
|
||||||
* @author Christian Mallwitz
|
|
||||||
*/
|
*/
|
||||||
public class TestLuceneDictionary extends TestCase {
|
public class TestLuceneDictionary extends TestCase {
|
||||||
|
|
||||||
|
|
|
@ -27,7 +27,6 @@ import org.apache.lucene.store.RAMDirectory;
|
||||||
/**
|
/**
|
||||||
* Test case for PlainTextDictionary
|
* Test case for PlainTextDictionary
|
||||||
*
|
*
|
||||||
* @author Daniel Naber
|
|
||||||
*/
|
*/
|
||||||
public class TestPlainTextDictionary extends TestCase {
|
public class TestPlainTextDictionary extends TestCase {
|
||||||
|
|
||||||
|
|
|
@ -36,7 +36,6 @@ import java.util.ArrayList;
|
||||||
/**
|
/**
|
||||||
* See table searcher explanation.
|
* See table searcher explanation.
|
||||||
*
|
*
|
||||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
|
||||||
*/
|
*/
|
||||||
public class ListSearcher extends AbstractListModel {
|
public class ListSearcher extends AbstractListModel {
|
||||||
private ListModel listModel;
|
private ListModel listModel;
|
||||||
|
|
|
@ -62,7 +62,6 @@ import java.util.ArrayList;
|
||||||
* at any time by searching with an empty string. Additionally, you can
|
* at any time by searching with an empty string. Additionally, you can
|
||||||
* add a button calling the clearSearch() method.
|
* add a button calling the clearSearch() method.
|
||||||
*
|
*
|
||||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
|
||||||
*/
|
*/
|
||||||
public class TableSearcher extends AbstractTableModel {
|
public class TableSearcher extends AbstractTableModel {
|
||||||
|
|
||||||
|
|
|
@ -22,9 +22,7 @@ import java.util.List;
|
||||||
|
|
||||||
import javax.swing.AbstractListModel;
|
import javax.swing.AbstractListModel;
|
||||||
|
|
||||||
/**
|
|
||||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
|
||||||
*/
|
|
||||||
public class BaseListModel extends AbstractListModel {
|
public class BaseListModel extends AbstractListModel {
|
||||||
private List data = new ArrayList();
|
private List data = new ArrayList();
|
||||||
|
|
||||||
|
|
|
@ -22,9 +22,7 @@ import java.util.List;
|
||||||
|
|
||||||
import javax.swing.table.AbstractTableModel;
|
import javax.swing.table.AbstractTableModel;
|
||||||
|
|
||||||
/**
|
|
||||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
|
||||||
*/
|
|
||||||
public class BaseTableModel extends AbstractTableModel {
|
public class BaseTableModel extends AbstractTableModel {
|
||||||
private List columnNames = new ArrayList();
|
private List columnNames = new ArrayList();
|
||||||
private List rows = new ArrayList();
|
private List rows = new ArrayList();
|
||||||
|
|
|
@ -22,9 +22,7 @@ import java.util.ArrayList;
|
||||||
import java.util.Collection;
|
import java.util.Collection;
|
||||||
import java.util.Iterator;
|
import java.util.Iterator;
|
||||||
|
|
||||||
/**
|
|
||||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
|
||||||
*/
|
|
||||||
public class DataStore {
|
public class DataStore {
|
||||||
|
|
||||||
private static final String ITALIAN_CATEGORY = "Italian";
|
private static final String ITALIAN_CATEGORY = "Italian";
|
||||||
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue