mirror of https://github.com/apache/lucene.git
LUCENE-1378 - Removed the remaining 199 @author references
git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@695514 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
8d02d7859f
commit
0195fcd03d
|
@ -37,7 +37,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
|
|||
* will not be indexed at all) and an external list of exclusions (word that will
|
||||
* not be stemmed, but indexed).
|
||||
*
|
||||
* @author João Kramer
|
||||
*/
|
||||
public final class BrazilianAnalyzer extends Analyzer {
|
||||
|
||||
|
|
|
@ -28,7 +28,6 @@ import java.util.Set;
|
|||
/**
|
||||
* Based on GermanStemFilter
|
||||
*
|
||||
* @author João Kramer
|
||||
*/
|
||||
public final class BrazilianStemFilter extends TokenFilter {
|
||||
|
||||
|
|
|
@ -28,7 +28,6 @@ import java.util.Set;
|
|||
/**
|
||||
* Filters CJKTokenizer with StopFilter.
|
||||
*
|
||||
* @author Che, Dong
|
||||
*/
|
||||
public class CJKAnalyzer extends Analyzer {
|
||||
//~ Static fields/initializers ---------------------------------------------
|
||||
|
|
|
@ -34,7 +34,6 @@ import java.io.Reader;
|
|||
* please search <a
|
||||
* href="http://www.google.com/search?q=word+chinese+segment">google</a>
|
||||
*
|
||||
* @author Che, Dong
|
||||
*/
|
||||
public final class CJKTokenizer extends Tokenizer {
|
||||
//~ Static fields/initializers ---------------------------------------------
|
||||
|
|
|
@ -28,7 +28,6 @@ import org.apache.lucene.analysis.TokenStream;
|
|||
* build from a ChineseTokenizer, filtered with ChineseFilter.
|
||||
* Copyright: Copyright (c) 2001
|
||||
* Company:
|
||||
* @author Yiyi Sun
|
||||
* @version 1.0
|
||||
*
|
||||
*/
|
||||
|
|
|
@ -37,7 +37,6 @@ import org.apache.lucene.analysis.TokenStream;
|
|||
*
|
||||
* Copyright: Copyright (c) 2001
|
||||
* Company:
|
||||
* @author Yiyi Sun
|
||||
* @version 1.0
|
||||
*
|
||||
*/
|
||||
|
|
|
@ -47,7 +47,6 @@ import org.apache.lucene.analysis.Tokenizer;
|
|||
* C4C2, C1C2C3 ... the ChineseTokenizer works, but the
|
||||
* CJKTokenizer will not work.
|
||||
*
|
||||
* @author Yiyi Sun
|
||||
* @version 1.0
|
||||
*
|
||||
*/
|
||||
|
|
|
@ -35,7 +35,6 @@ import java.util.Set;
|
|||
* A default set of stopwords is used unless an alternative list is specified, the
|
||||
* exclusion list is empty by default.
|
||||
*
|
||||
* @author Lukas Zapletal [lzap@root.cz]
|
||||
*/
|
||||
public final class CzechAnalyzer extends Analyzer {
|
||||
|
||||
|
|
|
@ -32,7 +32,6 @@ import java.util.Set;
|
|||
* that will not be indexed at all).
|
||||
* A default set of stopwords is used unless an alternative list is specified.
|
||||
*
|
||||
* @author Panagiotis Astithas, past@ebs.gr
|
||||
*/
|
||||
public final class GreekAnalyzer extends Analyzer
|
||||
{
|
||||
|
|
|
@ -23,7 +23,6 @@ package org.apache.lucene.analysis.el;
|
|||
* including accented ones. One should be able to add other encoding schemes (see RFC 1947) by adding
|
||||
* the definition of a new charset as well as the required logic in the toLowerCase() method.
|
||||
*
|
||||
* @author Panagiotis Astithas, past@ebs.gr
|
||||
*/
|
||||
public class GreekCharsets
|
||||
{
|
||||
|
|
|
@ -23,7 +23,6 @@ import org.apache.lucene.analysis.TokenStream;
|
|||
/**
|
||||
* Normalizes token text to lower case, analyzing given ("greek") charset.
|
||||
*
|
||||
* @author Panagiotis Astithas, past@ebs.gr
|
||||
*/
|
||||
public final class GreekLowerCaseFilter extends TokenFilter
|
||||
{
|
||||
|
|
|
@ -32,8 +32,7 @@ import org.apache.lucene.analysis.TokenFilter;
|
|||
* <p>
|
||||
* Note that StandardTokenizer sees " ' " as a space, and cuts it out.
|
||||
*
|
||||
* @see <a href="http://fr.wikipedia.org/wiki/%C3%89lision">Elision in Wikipedia</a>
|
||||
* @author <a href="mailto:mlecarme@openwide.fr">Mathieu Lecarme</a>
|
||||
* @see <a href="http://fr.wikipedia.org/wiki/%C3%89lision">Elision in Wikipedia</a>
|
||||
*/
|
||||
public class ElisionFilter extends TokenFilter {
|
||||
private Set articles = null;
|
||||
|
|
|
@ -31,7 +31,6 @@ import java.util.Set;
|
|||
* not be stemmed at all. The used stemmer can be changed at runtime after the
|
||||
* filter object is created (as long as it is a FrenchStemmer).
|
||||
*
|
||||
* @author Patrick Talbot (based on Gerhard Schwarz work for German)
|
||||
*/
|
||||
public final class FrenchStemFilter extends TokenFilter {
|
||||
|
||||
|
|
|
@ -23,7 +23,6 @@ package org.apache.lucene.analysis.fr;
|
|||
* refer to http://snowball.sourceforge.net/french/stemmer.html<br>
|
||||
* (French stemming algorithm) for details
|
||||
*
|
||||
* @author Patrick Talbot
|
||||
*/
|
||||
|
||||
public class FrenchStemmer {
|
||||
|
|
|
@ -26,7 +26,6 @@ import java.util.LinkedList;
|
|||
|
||||
/**
|
||||
* Tokenizes the input into n-grams of the given size(s).
|
||||
* @author Otis Gospodnetic
|
||||
*/
|
||||
public class NGramTokenFilter extends TokenFilter {
|
||||
public static final int DEFAULT_MIN_NGRAM_SIZE = 1;
|
||||
|
|
|
@ -25,7 +25,6 @@ import java.io.Reader;
|
|||
|
||||
/**
|
||||
* Tokenizes the input into n-grams of the given size(s).
|
||||
* @author Otis Gospodnetic
|
||||
*/
|
||||
public class NGramTokenizer extends Tokenizer {
|
||||
public static final int DEFAULT_MIN_NGRAM_SIZE = 1;
|
||||
|
|
|
@ -38,8 +38,7 @@ import java.util.Map;
|
|||
* the algorithm (dictionary stemming).
|
||||
* A default set of stopwords is used unless an alternative list is specified, the
|
||||
* exclusion list is empty by default.
|
||||
*
|
||||
* @author Edwin de Jonge
|
||||
*
|
||||
*/
|
||||
public class DutchAnalyzer extends Analyzer {
|
||||
/**
|
||||
|
|
|
@ -31,8 +31,7 @@ import java.util.Map;
|
|||
* A filter that stems Dutch words. It supports a table of words that should
|
||||
* not be stemmed at all. The stemmer used can be changed at runtime after the
|
||||
* filter object is created (as long as it is a DutchStemmer).
|
||||
*
|
||||
* @author Edwin de Jonge
|
||||
*
|
||||
*/
|
||||
public final class DutchStemFilter extends TokenFilter {
|
||||
/**
|
||||
|
|
|
@ -24,8 +24,7 @@ import java.util.Map;
|
|||
* A stemmer for Dutch words. The algorithm is an implementation of
|
||||
* the <a href="http://snowball.tartarus.org/algorithms/dutch/stemmer.html">dutch stemming</a>
|
||||
* algorithm in Martin Porter's snowball project.
|
||||
*
|
||||
* @author Edwin de Jonge (ejne at cbs.nl)
|
||||
*
|
||||
*/
|
||||
|
||||
public class DutchStemmer {
|
||||
|
|
|
@ -28,8 +28,7 @@ import java.util.HashMap;
|
|||
* Loads a text file and adds every line as an entry to a Hashtable. Every line
|
||||
* should contain only one word. If the file is not found or on any error, an
|
||||
* empty table is returned.
|
||||
*
|
||||
* @author Gerhard Schwarz
|
||||
*
|
||||
* @deprecated use {@link org.apache.lucene.analysis.WordlistLoader} instead
|
||||
*/
|
||||
public class WordlistLoader {
|
||||
|
|
|
@ -26,7 +26,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
|
|||
|
||||
/**
|
||||
* Analyzer for Thai language. It uses java.text.BreakIterator to break words.
|
||||
* @author Samphan Raruenrom <samphan@osdev.co.th> for To-Be-One Technology Co., Ltd.
|
||||
* @version 0.2
|
||||
*/
|
||||
public class ThaiAnalyzer extends Analyzer {
|
||||
|
|
|
@ -27,7 +27,6 @@ import java.text.BreakIterator;
|
|||
/**
|
||||
* TokenFilter that use java.text.BreakIterator to break each
|
||||
* Token that is Thai into separate Token(s) for each Thai word.
|
||||
* @author Samphan Raruenrom <samphan@osdev.co.th> for To-Be-One Technology Co., Ltd.
|
||||
* @version 0.2
|
||||
*/
|
||||
public class ThaiWordFilter extends TokenFilter {
|
||||
|
|
|
@ -24,9 +24,7 @@ import junit.framework.TestCase;
|
|||
|
||||
import org.apache.lucene.analysis.Token;
|
||||
|
||||
/**
|
||||
* @author rayt
|
||||
*/
|
||||
|
||||
public class TestChineseTokenizer extends TestCase
|
||||
{
|
||||
public void testOtherLetterOffset() throws IOException
|
||||
|
|
|
@ -33,8 +33,7 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
|
|||
* Test the German stemmer. The stemming algorithm is known to work less
|
||||
* than perfect, as it doesn't use any word lists with exceptions. We
|
||||
* also check some of the cases where the algorithm is wrong.
|
||||
*
|
||||
* @author Daniel Naber
|
||||
*
|
||||
*/
|
||||
public class TestGermanStemFilter extends TestCase {
|
||||
|
||||
|
|
|
@ -28,7 +28,6 @@ import junit.framework.TestCase;
|
|||
/**
|
||||
* A unit test class for verifying the correct operation of the GreekAnalyzer.
|
||||
*
|
||||
* @author past
|
||||
*/
|
||||
public class GreekAnalyzerTest extends TestCase {
|
||||
|
||||
|
|
|
@ -32,7 +32,6 @@ import org.apache.lucene.analysis.Tokenizer;
|
|||
import org.apache.lucene.analysis.standard.StandardTokenizer;
|
||||
|
||||
/**
|
||||
* @author Mathieu Lecarme<mlecarme@openwide.fr>
|
||||
*
|
||||
*/
|
||||
public class TestElision extends TestCase {
|
||||
|
|
|
@ -65,7 +65,6 @@ import org.apache.lucene.analysis.TokenStream;
|
|||
/**
|
||||
* Test case for FrenchAnalyzer.
|
||||
*
|
||||
* @author Jean-François Halleux
|
||||
* @version $version$
|
||||
*/
|
||||
|
||||
|
|
|
@ -27,7 +27,6 @@ import junit.framework.TestCase;
|
|||
|
||||
/**
|
||||
* Tests {@link EdgeNGramTokenFilter} for correctness.
|
||||
* @author Otis Gospodnetic
|
||||
*/
|
||||
public class EdgeNGramTokenFilterTest extends TestCase {
|
||||
private TokenStream input;
|
||||
|
|
|
@ -25,7 +25,6 @@ import junit.framework.TestCase;
|
|||
|
||||
/**
|
||||
* Tests {@link EdgeNGramTokenizer} for correctness.
|
||||
* @author Otis Gospodnetic
|
||||
*/
|
||||
public class EdgeNGramTokenizerTest extends TestCase {
|
||||
private StringReader input;
|
||||
|
|
|
@ -28,7 +28,6 @@ import junit.framework.TestCase;
|
|||
|
||||
/**
|
||||
* Tests {@link NGramTokenFilter} for correctness.
|
||||
* @author Otis Gospodnetic
|
||||
*/
|
||||
public class NGramTokenFilterTest extends TestCase {
|
||||
private TokenStream input;
|
||||
|
|
|
@ -26,7 +26,6 @@ import junit.framework.TestCase;
|
|||
|
||||
/**
|
||||
* Tests {@link NGramTokenizer} for correctness.
|
||||
* @author Otis Gospodnetic
|
||||
*/
|
||||
public class NGramTokenizerTest extends TestCase {
|
||||
private StringReader input;
|
||||
|
|
|
@ -26,7 +26,6 @@ import org.apache.lucene.analysis.TokenStream;
|
|||
/**
|
||||
* Test case for ThaiAnalyzer, modified from TestFrenchAnalyzer
|
||||
*
|
||||
* @author Samphan Raruenrom <samphan@osdev.co.th>
|
||||
* @version 0.1
|
||||
*/
|
||||
|
||||
|
|
|
@ -24,7 +24,6 @@ import java.io.File;
|
|||
/**
|
||||
* Allows a class to act as a Lucene document handler
|
||||
*
|
||||
*@author Erik Hatcher
|
||||
*@since October 27, 2001
|
||||
*/
|
||||
public interface DocumentHandler {
|
||||
|
|
|
@ -20,8 +20,7 @@ package org.apache.lucene.ant;
|
|||
import java.io.PrintStream;
|
||||
import java.io.PrintWriter;
|
||||
|
||||
/**
|
||||
*/
|
||||
|
||||
public class DocumentHandlerException extends Exception {
|
||||
private Throwable cause;
|
||||
|
||||
|
|
|
@ -26,7 +26,6 @@ import java.io.File;
|
|||
* based on a files extension. Currently only .html and .txt
|
||||
* files are handled, other extensions ignored.
|
||||
*
|
||||
*@author Erik Hatcher
|
||||
*@since October 28, 2001
|
||||
*@todo Implement dynamic document type lookup
|
||||
*/
|
||||
|
|
|
@ -39,7 +39,6 @@ import java.io.StringWriter;
|
|||
* It does this by using JTidy package. It can take input input
|
||||
* from {@link java.io.File} or {@link java.io.InputStream}.
|
||||
*
|
||||
*@author Erik Hatcher
|
||||
*/
|
||||
public class HtmlDocument {
|
||||
private Element rawDoc;
|
||||
|
|
|
@ -52,7 +52,6 @@ import java.text.ParseException;
|
|||
/**
|
||||
* Ant task to index files with Lucene
|
||||
*
|
||||
*@author Erik Hatcher
|
||||
*/
|
||||
public class IndexTask extends Task {
|
||||
/**
|
||||
|
|
|
@ -29,7 +29,6 @@ import java.io.StringWriter;
|
|||
/**
|
||||
* A utility for making Lucene Documents from a File.
|
||||
*
|
||||
*@author Erik Hatcher
|
||||
*@since December 6, 2001
|
||||
*@todo Fix JavaDoc comments here
|
||||
*/
|
||||
|
|
|
@ -37,7 +37,6 @@ import org.apache.tools.ant.types.FileSet;
|
|||
/**
|
||||
* Test cases for index task
|
||||
*
|
||||
*@author Erik Hatcher
|
||||
*/
|
||||
public class IndexTaskTest extends TestCase {
|
||||
private final static String docHandler =
|
||||
|
|
|
@ -18,7 +18,6 @@ package org.apache.lucene.benchmark.stats;
|
|||
/**
|
||||
* This class holds a set of memory usage values.
|
||||
*
|
||||
* @author Andrzej Bialecki <ab@getopt.org>
|
||||
*/
|
||||
public class MemUsage {
|
||||
public long maxFree, minFree, avgFree;
|
||||
|
|
|
@ -23,7 +23,6 @@ import org.apache.lucene.benchmark.Constants;
|
|||
/**
|
||||
* This class holds parameters for a query benchmark.
|
||||
*
|
||||
* @author Andrzej Bialecki <ab@getopt.org>
|
||||
*/
|
||||
public class QueryData {
|
||||
/** Benchmark id */
|
||||
|
|
|
@ -36,7 +36,6 @@ import org.apache.lucene.store.Directory;
|
|||
* This class holds together all parameters related to a test. Single test is
|
||||
* performed several times, and all results are averaged.
|
||||
*
|
||||
* @author Andrzej Bialecki <ab@getopt.org>
|
||||
*/
|
||||
public class TestData
|
||||
{
|
||||
|
|
|
@ -26,7 +26,6 @@ import java.util.Iterator;
|
|||
* values may contribute to different measurements, so this class provides also
|
||||
* some useful methods to separate them.
|
||||
*
|
||||
* @author Andrzej Bialecki <ab@getopt.org>
|
||||
*/
|
||||
public class TestRunData {
|
||||
private String id;
|
||||
|
|
|
@ -19,7 +19,6 @@ package org.apache.lucene.benchmark.stats;
|
|||
/**
|
||||
* This class holds a data point measuring speed of processing.
|
||||
*
|
||||
* @author Andrzej Bialecki <ab@getopt.org>
|
||||
*/
|
||||
public class TimeData {
|
||||
/** Name of the data point - usually one of a data series with the same name */
|
||||
|
|
|
@ -24,8 +24,7 @@ import com.sleepycat.je.DatabaseException;
|
|||
|
||||
/**
|
||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||
*
|
||||
* @author Aaron Donovan
|
||||
*
|
||||
*/
|
||||
|
||||
public class Block extends Object {
|
||||
|
|
|
@ -33,8 +33,7 @@ import com.sleepycat.je.Transaction;
|
|||
|
||||
/**
|
||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||
*
|
||||
* @author Aaron Donovan
|
||||
*
|
||||
*/
|
||||
|
||||
public class File extends Object {
|
||||
|
|
|
@ -46,8 +46,7 @@ import com.sleepycat.je.Transaction;
|
|||
* {@link org.apache.lucene.store.Directory Directory}. It uses two
|
||||
* {@link com.sleepycat.je.Database Db} database handles, one for storing file
|
||||
* records and another for storing file data blocks.
|
||||
*
|
||||
* @author Aaron Donovan
|
||||
*
|
||||
*/
|
||||
|
||||
public class JEDirectory extends Directory {
|
||||
|
|
|
@ -22,8 +22,7 @@ import org.apache.lucene.store.IndexInput;
|
|||
|
||||
/**
|
||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||
*
|
||||
* @author Aaron Donovan
|
||||
*
|
||||
*/
|
||||
|
||||
public class JEIndexInput extends IndexInput {
|
||||
|
|
|
@ -23,8 +23,7 @@ import org.apache.lucene.store.IndexOutput;
|
|||
|
||||
/**
|
||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||
*
|
||||
* @author Aaron Donovan
|
||||
*
|
||||
*/
|
||||
|
||||
public class JEIndexOutput extends IndexOutput {
|
||||
|
|
|
@ -21,8 +21,7 @@ import org.apache.lucene.store.Lock;
|
|||
|
||||
/**
|
||||
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
|
||||
*
|
||||
* @author Aaron Donovan
|
||||
*
|
||||
*/
|
||||
|
||||
public class JELock extends Lock {
|
||||
|
|
|
@ -44,8 +44,7 @@ import com.sleepycat.je.Transaction;
|
|||
* Tests {@link JEDirectory}.
|
||||
*
|
||||
* Adapted from Andi Vajda's org.apache.lucene.db.DbStoreTest.
|
||||
*
|
||||
* @author Aaron Donovan
|
||||
*
|
||||
*/
|
||||
public class JEStoreTest extends TestCase {
|
||||
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");
|
||||
|
|
|
@ -30,7 +30,6 @@ import com.sleepycat.db.internal.DbTxn;
|
|||
* that were the entrypoint objects in the pre-4.3 Java API and that wrap the
|
||||
* actual Berkeley DB C objects via SWIG.
|
||||
*
|
||||
* @author Andi Vajda
|
||||
*/
|
||||
|
||||
public class DbHandleExtractor {
|
||||
|
|
|
@ -24,9 +24,7 @@ import com.sleepycat.db.internal.Db;
|
|||
import com.sleepycat.db.internal.DbTxn;
|
||||
import com.sleepycat.db.DatabaseException;
|
||||
|
||||
/**
|
||||
* @author Andi Vajda
|
||||
*/
|
||||
|
||||
|
||||
public class Block extends Object {
|
||||
protected DatabaseEntry key, data;
|
||||
|
|
|
@ -50,7 +50,6 @@ import com.sleepycat.db.DbHandleExtractor;
|
|||
* {@link com.sleepycat.db.internal.Db Db} database handles, one for storing file
|
||||
* records and another for storing file data blocks.
|
||||
*
|
||||
* @author Andi Vajda
|
||||
*/
|
||||
|
||||
public class DbDirectory extends Directory {
|
||||
|
|
|
@ -21,9 +21,7 @@ import java.io.IOException;
|
|||
import org.apache.lucene.store.IndexInput;
|
||||
|
||||
|
||||
/**
|
||||
* @author Andi Vajda
|
||||
*/
|
||||
|
||||
|
||||
public class DbIndexInput extends IndexInput {
|
||||
|
||||
|
|
|
@ -21,9 +21,7 @@ import java.io.IOException;
|
|||
import org.apache.lucene.store.IndexOutput;
|
||||
|
||||
|
||||
/**
|
||||
* @author Andi Vajda
|
||||
*/
|
||||
|
||||
|
||||
public class DbIndexOutput extends IndexOutput {
|
||||
|
||||
|
|
|
@ -24,7 +24,6 @@ import org.apache.lucene.store.Lock;
|
|||
* trivial as {@link DbDirectory} operations are managed by the Berkeley DB
|
||||
* locking system.
|
||||
*
|
||||
* @author Andi Vajda
|
||||
*/
|
||||
|
||||
public class DbLock extends Lock {
|
||||
|
|
|
@ -31,9 +31,7 @@ import com.sleepycat.db.internal.Db;
|
|||
import com.sleepycat.db.internal.DbTxn;
|
||||
import com.sleepycat.db.DatabaseException;
|
||||
|
||||
/**
|
||||
* @author Andi Vajda
|
||||
*/
|
||||
|
||||
|
||||
public class File extends Object {
|
||||
|
||||
|
|
|
@ -42,7 +42,6 @@ import org.apache.lucene.store.IndexOutput;
|
|||
* Tests {@link DbDirectory}.
|
||||
*
|
||||
* Adapted from org.apache.lucene.StoreTest with larger files and random bytes.
|
||||
* @author Andi Vajda
|
||||
*/
|
||||
public class DbStoreTest extends TestCase {
|
||||
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");
|
||||
|
|
|
@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
|
|||
|
||||
/**
|
||||
* Simple {@link Encoder} implementation that does not modify the output
|
||||
* @author Nicko Cadell
|
||||
*
|
||||
*/
|
||||
public class DefaultEncoder implements Encoder
|
||||
|
|
|
@ -19,7 +19,6 @@ package org.apache.lucene.search.highlight;
|
|||
/**
|
||||
* Encodes original text. The Encoder works with the Formatter to generate the output.
|
||||
*
|
||||
* @author Nicko Cadell
|
||||
*/
|
||||
public interface Encoder
|
||||
{
|
||||
|
|
|
@ -21,8 +21,7 @@ import org.apache.lucene.analysis.Token;
|
|||
/**
|
||||
* Implements the policy for breaking text into multiple fragments for consideration
|
||||
* by the {@link Highlighter} class. A sophisticated implementation may do this on the basis
|
||||
* of detecting end of sentences in the text.
|
||||
* @author mark@searcharea.co.uk
|
||||
* of detecting end of sentences in the text.
|
||||
*/
|
||||
public interface Fragmenter
|
||||
{
|
||||
|
|
|
@ -19,8 +19,7 @@ package org.apache.lucene.search.highlight;
|
|||
/**
|
||||
* Formats text with different color intensity depending on the score of the
|
||||
* term.
|
||||
*
|
||||
* @author maharwood
|
||||
*
|
||||
*/
|
||||
public class GradientFormatter implements Formatter
|
||||
{
|
||||
|
|
|
@ -30,7 +30,6 @@ import org.apache.lucene.util.PriorityQueue;
|
|||
* Class used to markup highlighted terms found in the best sections of a
|
||||
* text, using configurable {@link Fragmenter}, {@link Scorer}, {@link Formatter},
|
||||
* {@link Encoder} and tokenizers.
|
||||
* @author mark@searcharea.co.uk
|
||||
*/
|
||||
public class Highlighter
|
||||
{
|
||||
|
@ -503,8 +502,7 @@ public class Highlighter
|
|||
this.maxDocCharsToAnalyze = maxDocCharsToAnalyze;
|
||||
}
|
||||
|
||||
/**
|
||||
*/
|
||||
|
||||
public Fragmenter getTextFragmenter()
|
||||
{
|
||||
return textFragmenter;
|
||||
|
|
|
@ -26,8 +26,7 @@ import org.apache.lucene.search.Query;
|
|||
/**
|
||||
* {@link Scorer} implementation which scores text fragments by the number of unique query terms found.
|
||||
* This class uses the {@link QueryTermExtractor} class to process determine the query terms and
|
||||
* their boosts to be used.
|
||||
* @author mark@searcharea.co.uk
|
||||
* their boosts to be used.
|
||||
*/
|
||||
//TODO: provide option to boost score of fragments near beginning of document
|
||||
// based on fragment.getFragNum()
|
||||
|
|
|
@ -20,7 +20,6 @@ import org.apache.lucene.analysis.Token;
|
|||
|
||||
/**
|
||||
* Adds to the score for a fragment based on its tokens
|
||||
* @author mark@searcharea.co.uk
|
||||
*/
|
||||
public interface Scorer
|
||||
{
|
||||
|
|
|
@ -21,7 +21,6 @@ import org.apache.lucene.analysis.Token;
|
|||
/**
|
||||
* {@link Fragmenter} implementation which breaks text up into same-size
|
||||
* fragments with no concerns over spotting sentence boundaries.
|
||||
* @author mark@searcharea.co.uk
|
||||
*/
|
||||
public class SimpleFragmenter implements Fragmenter
|
||||
{
|
||||
|
|
|
@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
|
|||
|
||||
/**
|
||||
* Simple {@link Encoder} implementation to escape text for HTML output
|
||||
* @author Nicko Cadell
|
||||
*
|
||||
*/
|
||||
public class SimpleHTMLEncoder implements Encoder
|
||||
|
|
|
@ -18,7 +18,6 @@ package org.apache.lucene.search.highlight;
|
|||
|
||||
/**
|
||||
* Simple {@link Formatter} implementation to highlight terms with a pre and post tag
|
||||
* @author MAHarwood
|
||||
*
|
||||
*/
|
||||
public class SimpleHTMLFormatter implements Formatter
|
||||
|
|
|
@ -20,7 +20,6 @@ package org.apache.lucene.search.highlight;
|
|||
/**
|
||||
* Low-level class used to record information about a section of a document
|
||||
* with a score.
|
||||
* @author MAHarwood
|
||||
*
|
||||
*
|
||||
*/
|
||||
|
|
|
@ -20,7 +20,6 @@ import org.apache.lucene.analysis.Token;
|
|||
/**
|
||||
* One, or several overlapping tokens, along with the score(s) and the
|
||||
* scope of the original text
|
||||
* @author MAHarwood
|
||||
*/
|
||||
public class TokenGroup
|
||||
{
|
||||
|
|
|
@ -38,8 +38,7 @@ import org.apache.lucene.index.TermVectorOffsetInfo;
|
|||
/**
|
||||
* Hides implementation issues associated with obtaining a TokenStream for use with
|
||||
* the higlighter - can obtain from TermFreqVectors with offsets and (optionally) positions or
|
||||
* from Analyzer class reparsing the stored content.
|
||||
* @author maharwood
|
||||
* from Analyzer class reparsing the stored content.
|
||||
*/
|
||||
public class TokenSources
|
||||
{
|
||||
|
|
|
@ -16,8 +16,7 @@ package org.apache.lucene.search.highlight;
|
|||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/** Lightweight class to hold term and a weight value used for scoring this term
|
||||
* @author Mark Harwood
|
||||
/** Lightweight class to hold term and a weight value used for scoring this term
|
||||
*/
|
||||
public class WeightedTerm
|
||||
{
|
||||
|
|
|
@ -76,8 +76,7 @@ import org.w3c.dom.NodeList;
|
|||
|
||||
/**
|
||||
* JUnit Test for Highlighter class.
|
||||
*
|
||||
* @author mark@searcharea.co.uk
|
||||
*
|
||||
*/
|
||||
public class HighlighterTest extends TestCase implements Formatter {
|
||||
private IndexReader reader;
|
||||
|
@ -1447,8 +1446,7 @@ class SynonymAnalyzer extends Analyzer {
|
|||
|
||||
/**
|
||||
* Expands a token stream with synonyms (TODO - make the synonyms analyzed by choice of analyzer)
|
||||
*
|
||||
* @author MAHarwood
|
||||
*
|
||||
*/
|
||||
class SynonymTokenizer extends TokenStream {
|
||||
private TokenStream realStream;
|
||||
|
|
|
@ -38,8 +38,7 @@ import org.apache.lucene.analysis.TokenStream;
|
|||
/**
|
||||
* Various fulltext analysis utilities avoiding redundant code in several
|
||||
* classes.
|
||||
*
|
||||
* @author whoschek.AT.lbl.DOT.gov
|
||||
*
|
||||
*/
|
||||
public class AnalyzerUtil {
|
||||
|
||||
|
|
|
@ -166,8 +166,7 @@ import org.apache.lucene.search.Similarity;
|
|||
* target="_blank"
|
||||
* href="http://java.sun.com/developer/technicalArticles/Programming/HPROF.html">
|
||||
* hprof tracing </a>).
|
||||
*
|
||||
* @author whoschek.AT.lbl.DOT.gov
|
||||
*
|
||||
*/
|
||||
public class MemoryIndex implements Serializable {
|
||||
|
||||
|
|
|
@ -60,8 +60,7 @@ import org.apache.lucene.analysis.TokenStream;
|
|||
* pat.tokenStream("content", "James is running round in the woods"),
|
||||
* "English"));
|
||||
* </pre>
|
||||
*
|
||||
* @author whoschek.AT.lbl.DOT.gov
|
||||
*
|
||||
*/
|
||||
public class PatternAnalyzer extends Analyzer {
|
||||
|
||||
|
|
|
@ -67,8 +67,7 @@ import java.util.TreeSet;
|
|||
* wolfish:[edacious, esurient, rapacious, ravening, ravenous, voracious, wolflike]
|
||||
* xxxx:[]
|
||||
* </pre>
|
||||
*
|
||||
* @author whoschek.AT.lbl.DOT.gov
|
||||
*
|
||||
* @see <a target="_blank"
|
||||
* href="http://www.cogsci.princeton.edu/~wn/man/prologdb.5WN.html">prologdb
|
||||
* man page </a>
|
||||
|
|
|
@ -27,8 +27,7 @@ import org.apache.lucene.analysis.TokenStream;
|
|||
* Injects additional tokens for synonyms of token terms fetched from the
|
||||
* underlying child stream; the child stream must deliver lowercase tokens
|
||||
* for synonyms to be found.
|
||||
*
|
||||
* @author whoschek.AT.lbl.DOT.gov
|
||||
*
|
||||
*/
|
||||
public class SynonymTokenFilter extends TokenFilter {
|
||||
|
||||
|
|
|
@ -196,7 +196,6 @@ hello^2.0
|
|||
the^3
|
||||
</pre>
|
||||
|
||||
@author whoschek.AT.lbl.DOT.gov
|
||||
*/
|
||||
public class MemoryIndexTest extends TestCase {
|
||||
|
||||
|
|
|
@ -58,7 +58,6 @@ Lucene features: CharTokenizer.MAX_WORD_LEN = 255.
|
|||
Thus the PatternAnalyzer produces correct output, whereas the WhitespaceAnalyzer
|
||||
silently truncates text, and so the comparison results in assertEquals() don't match up.
|
||||
|
||||
@author whoschek.AT.lbl.DOT.gov
|
||||
*/
|
||||
public class PatternAnalyzerTest extends TestCase {
|
||||
|
||||
|
|
|
@ -34,8 +34,6 @@ import org.apache.lucene.store.FSDirectory;
|
|||
* NOTE: This will overwrite any length normalization or field/document boosts.
|
||||
* </p>
|
||||
*
|
||||
* @author Chris Hostetter
|
||||
* @author Otis Gospodnetic
|
||||
*/
|
||||
public class FieldNormModifier {
|
||||
|
||||
|
|
|
@ -79,7 +79,6 @@ import org.apache.lucene.util.SortedVIntList;
|
|||
* /least computationally-intensive filters first.
|
||||
* </p>
|
||||
*
|
||||
* @author <a href="mailto:kelvint@apache.org">Kelvin Tan</a>
|
||||
*/
|
||||
public class ChainedFilter extends Filter
|
||||
{
|
||||
|
|
|
@ -27,7 +27,6 @@ import java.io.IOException;
|
|||
/**
|
||||
* Merges indices specified on the command line into the index
|
||||
* specified as the first command line argument.
|
||||
* @author Erik Hatcher
|
||||
* @version $Id$
|
||||
*/
|
||||
public class IndexMergeTool {
|
||||
|
|
|
@ -39,7 +39,6 @@ import org.apache.lucene.search.Query;
|
|||
* become <code>h?user</code> when using this parser and thus no match would be found (i.e.
|
||||
* using this parser will be no improvement over QueryParser in such cases).
|
||||
*
|
||||
* @author Ronnie Kolehmainen (ronnie.kolehmainen at ub.uu.se)
|
||||
* @version $Revision$, $Date$
|
||||
*/
|
||||
public class AnalyzingQueryParser extends org.apache.lucene.queryParser.QueryParser {
|
||||
|
|
|
@ -70,9 +70,6 @@ import org.apache.lucene.util.Parameter;
|
|||
* documentation</a>.
|
||||
* </p>
|
||||
*
|
||||
* @author Brian Goetz
|
||||
* @author Peter Halacsy
|
||||
* @author Tatu Saloranta
|
||||
*/
|
||||
|
||||
public class PrecedenceQueryParser implements PrecedenceQueryParserConstants {
|
||||
|
|
|
@ -30,7 +30,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
|
|||
import org.apache.lucene.queryParser.ParseException;
|
||||
|
||||
/**
|
||||
* @author Ronnie Kolehmainen (ronnie.kolehmainen at ub.uu.se)
|
||||
* @version $Revision$, $Date$
|
||||
*/
|
||||
public class TestAnalyzingQueryParser extends TestCase {
|
||||
|
|
|
@ -36,7 +36,6 @@ import org.apache.lucene.util.SortedVIntList;
|
|||
* SHOULD Filters are OR'd together
|
||||
* The resulting Filter is NOT'd with the NOT Filters
|
||||
* The resulting Filter is AND'd with the MUST Filters
|
||||
* @author BPDThebault
|
||||
*/
|
||||
|
||||
public class BooleanFilter extends Filter
|
||||
|
|
|
@ -24,7 +24,6 @@ import org.apache.lucene.search.BooleanClause.Occur;
|
|||
* is used when composed with another filter.
|
||||
* (Follows the boolean logic in BooleanClause for composition
|
||||
* of queries.)
|
||||
* @author BPDThebault
|
||||
*/
|
||||
|
||||
public class FilterClause implements java.io.Serializable
|
||||
|
|
|
@ -45,8 +45,7 @@ import org.apache.lucene.util.PriorityQueue;
|
|||
* TermQuery is used for variants and does not use that variant term's IDF because this would favour rarer
|
||||
* terms eg misspellings. Instead, all variants use the same IDF ranking (the one for the source query
|
||||
* term) and this is factored into the variant's boost. If the source query term does not exist in the
|
||||
* index the average IDF of the variants is used.
|
||||
* @author maharwood
|
||||
* index the average IDF of the variants is used.
|
||||
*/
|
||||
public class FuzzyLikeThisQuery extends Query
|
||||
{
|
||||
|
|
|
@ -34,8 +34,7 @@ import org.apache.lucene.util.OpenBitSet;
|
|||
* a sequence. An example might be a collection of primary keys from a database query result or perhaps
|
||||
* a choice of "category" labels picked by the end user. As a filter, this is much faster than the
|
||||
* equivalent query (a BooleanQuery with many "should" TermQueries)
|
||||
*
|
||||
* @author maharwood
|
||||
*
|
||||
*/
|
||||
public class TermsFilter extends Filter
|
||||
{
|
||||
|
|
|
@ -143,10 +143,7 @@ import java.util.ArrayList;
|
|||
* - refactor: moved common code into isNoiseWord()
|
||||
* - optimise: when no termvector support available - used maxNumTermsParsed to limit amount of tokenization
|
||||
* </pre>
|
||||
*
|
||||
* @author David Spencer
|
||||
* @author Bruce Ritchie
|
||||
* @author Mark Harwood
|
||||
*
|
||||
*/
|
||||
public final class MoreLikeThis {
|
||||
|
||||
|
|
|
@ -35,7 +35,6 @@ import org.apache.lucene.search.similar.MoreLikeThis;
|
|||
* A simple wrapper for MoreLikeThis for use in scenarios where a Query object is required eg
|
||||
* in custom QueryParser extensions. At query.rewrite() time the reader is used to construct the
|
||||
* actual MoreLikeThis object and obtain the real Query object.
|
||||
* @author maharwood
|
||||
*/
|
||||
public class MoreLikeThisQuery extends Query
|
||||
{
|
||||
|
|
|
@ -35,7 +35,6 @@ import org.apache.lucene.store.RAMDirectory;
|
|||
* It first creates a simple index and then a couple of instances of LuceneDictionary
|
||||
* on different fields and checks if all the right text comes back.
|
||||
*
|
||||
* @author Christian Mallwitz
|
||||
*/
|
||||
public class TestLuceneDictionary extends TestCase {
|
||||
|
||||
|
|
|
@ -27,7 +27,6 @@ import org.apache.lucene.store.RAMDirectory;
|
|||
/**
|
||||
* Test case for PlainTextDictionary
|
||||
*
|
||||
* @author Daniel Naber
|
||||
*/
|
||||
public class TestPlainTextDictionary extends TestCase {
|
||||
|
||||
|
|
|
@ -36,7 +36,6 @@ import java.util.ArrayList;
|
|||
/**
|
||||
* See table searcher explanation.
|
||||
*
|
||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
||||
*/
|
||||
public class ListSearcher extends AbstractListModel {
|
||||
private ListModel listModel;
|
||||
|
|
|
@ -62,7 +62,6 @@ import java.util.ArrayList;
|
|||
* at any time by searching with an empty string. Additionally, you can
|
||||
* add a button calling the clearSearch() method.
|
||||
*
|
||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
||||
*/
|
||||
public class TableSearcher extends AbstractTableModel {
|
||||
|
||||
|
|
|
@ -22,9 +22,7 @@ import java.util.List;
|
|||
|
||||
import javax.swing.AbstractListModel;
|
||||
|
||||
/**
|
||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
||||
*/
|
||||
|
||||
public class BaseListModel extends AbstractListModel {
|
||||
private List data = new ArrayList();
|
||||
|
||||
|
|
|
@ -22,9 +22,7 @@ import java.util.List;
|
|||
|
||||
import javax.swing.table.AbstractTableModel;
|
||||
|
||||
/**
|
||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
||||
*/
|
||||
|
||||
public class BaseTableModel extends AbstractTableModel {
|
||||
private List columnNames = new ArrayList();
|
||||
private List rows = new ArrayList();
|
||||
|
|
|
@ -22,9 +22,7 @@ import java.util.ArrayList;
|
|||
import java.util.Collection;
|
||||
import java.util.Iterator;
|
||||
|
||||
/**
|
||||
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
|
||||
*/
|
||||
|
||||
public class DataStore {
|
||||
|
||||
private static final String ITALIAN_CATEGORY = "Italian";
|
||||
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue