LUCENE-1378 - Removed the remaining 199 @author references

git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@695514 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Otis Gospodnetic 2008-09-15 15:42:11 +00:00
parent 8d02d7859f
commit 0195fcd03d
193 changed files with 86 additions and 325 deletions

View File

@ -37,7 +37,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
* will not be indexed at all) and an external list of exclusions (word that will * will not be indexed at all) and an external list of exclusions (word that will
* not be stemmed, but indexed). * not be stemmed, but indexed).
* *
* @author João Kramer
*/ */
public final class BrazilianAnalyzer extends Analyzer { public final class BrazilianAnalyzer extends Analyzer {

View File

@ -28,7 +28,6 @@ import java.util.Set;
/** /**
* Based on GermanStemFilter * Based on GermanStemFilter
* *
* @author João Kramer
*/ */
public final class BrazilianStemFilter extends TokenFilter { public final class BrazilianStemFilter extends TokenFilter {

View File

@ -28,7 +28,6 @@ import java.util.Set;
/** /**
* Filters CJKTokenizer with StopFilter. * Filters CJKTokenizer with StopFilter.
* *
* @author Che, Dong
*/ */
public class CJKAnalyzer extends Analyzer { public class CJKAnalyzer extends Analyzer {
//~ Static fields/initializers --------------------------------------------- //~ Static fields/initializers ---------------------------------------------

View File

@ -34,7 +34,6 @@ import java.io.Reader;
* please search <a * please search <a
* href="http://www.google.com/search?q=word+chinese+segment">google</a> * href="http://www.google.com/search?q=word+chinese+segment">google</a>
* *
* @author Che, Dong
*/ */
public final class CJKTokenizer extends Tokenizer { public final class CJKTokenizer extends Tokenizer {
//~ Static fields/initializers --------------------------------------------- //~ Static fields/initializers ---------------------------------------------

View File

@ -28,7 +28,6 @@ import org.apache.lucene.analysis.TokenStream;
* build from a ChineseTokenizer, filtered with ChineseFilter. * build from a ChineseTokenizer, filtered with ChineseFilter.
* Copyright: Copyright (c) 2001 * Copyright: Copyright (c) 2001
* Company: * Company:
* @author Yiyi Sun
* @version 1.0 * @version 1.0
* *
*/ */

View File

@ -37,7 +37,6 @@ import org.apache.lucene.analysis.TokenStream;
* *
* Copyright: Copyright (c) 2001 * Copyright: Copyright (c) 2001
* Company: * Company:
* @author Yiyi Sun
* @version 1.0 * @version 1.0
* *
*/ */

View File

@ -47,7 +47,6 @@ import org.apache.lucene.analysis.Tokenizer;
* C4C2, C1C2C3 ... the ChineseTokenizer works, but the * C4C2, C1C2C3 ... the ChineseTokenizer works, but the
* CJKTokenizer will not work. * CJKTokenizer will not work.
* *
* @author Yiyi Sun
* @version 1.0 * @version 1.0
* *
*/ */

View File

@ -35,7 +35,6 @@ import java.util.Set;
* A default set of stopwords is used unless an alternative list is specified, the * A default set of stopwords is used unless an alternative list is specified, the
* exclusion list is empty by default. * exclusion list is empty by default.
* *
* @author Lukas Zapletal [lzap@root.cz]
*/ */
public final class CzechAnalyzer extends Analyzer { public final class CzechAnalyzer extends Analyzer {

View File

@ -32,7 +32,6 @@ import java.util.Set;
* that will not be indexed at all). * that will not be indexed at all).
* A default set of stopwords is used unless an alternative list is specified. * A default set of stopwords is used unless an alternative list is specified.
* *
* @author Panagiotis Astithas, past@ebs.gr
*/ */
public final class GreekAnalyzer extends Analyzer public final class GreekAnalyzer extends Analyzer
{ {

View File

@ -23,7 +23,6 @@ package org.apache.lucene.analysis.el;
* including accented ones. One should be able to add other encoding schemes (see RFC 1947) by adding * including accented ones. One should be able to add other encoding schemes (see RFC 1947) by adding
* the definition of a new charset as well as the required logic in the toLowerCase() method. * the definition of a new charset as well as the required logic in the toLowerCase() method.
* *
* @author Panagiotis Astithas, past@ebs.gr
*/ */
public class GreekCharsets public class GreekCharsets
{ {

View File

@ -23,7 +23,6 @@ import org.apache.lucene.analysis.TokenStream;
/** /**
* Normalizes token text to lower case, analyzing given ("greek") charset. * Normalizes token text to lower case, analyzing given ("greek") charset.
* *
* @author Panagiotis Astithas, past@ebs.gr
*/ */
public final class GreekLowerCaseFilter extends TokenFilter public final class GreekLowerCaseFilter extends TokenFilter
{ {

View File

@ -33,7 +33,6 @@ import org.apache.lucene.analysis.TokenFilter;
* Note that StandardTokenizer sees " ' " as a space, and cuts it out. * Note that StandardTokenizer sees " ' " as a space, and cuts it out.
* *
* @see <a href="http://fr.wikipedia.org/wiki/%C3%89lision">Elision in Wikipedia</a> * @see <a href="http://fr.wikipedia.org/wiki/%C3%89lision">Elision in Wikipedia</a>
* @author <a href="mailto:mlecarme@openwide.fr">Mathieu Lecarme</a>
*/ */
public class ElisionFilter extends TokenFilter { public class ElisionFilter extends TokenFilter {
private Set articles = null; private Set articles = null;

View File

@ -31,7 +31,6 @@ import java.util.Set;
* not be stemmed at all. The used stemmer can be changed at runtime after the * not be stemmed at all. The used stemmer can be changed at runtime after the
* filter object is created (as long as it is a FrenchStemmer). * filter object is created (as long as it is a FrenchStemmer).
* *
* @author Patrick Talbot (based on Gerhard Schwarz work for German)
*/ */
public final class FrenchStemFilter extends TokenFilter { public final class FrenchStemFilter extends TokenFilter {

View File

@ -23,7 +23,6 @@ package org.apache.lucene.analysis.fr;
* refer to http://snowball.sourceforge.net/french/stemmer.html<br> * refer to http://snowball.sourceforge.net/french/stemmer.html<br>
* (French stemming algorithm) for details * (French stemming algorithm) for details
* *
* @author Patrick Talbot
*/ */
public class FrenchStemmer { public class FrenchStemmer {

View File

@ -26,7 +26,6 @@ import java.util.LinkedList;
/** /**
* Tokenizes the input into n-grams of the given size(s). * Tokenizes the input into n-grams of the given size(s).
* @author Otis Gospodnetic
*/ */
public class NGramTokenFilter extends TokenFilter { public class NGramTokenFilter extends TokenFilter {
public static final int DEFAULT_MIN_NGRAM_SIZE = 1; public static final int DEFAULT_MIN_NGRAM_SIZE = 1;

View File

@ -25,7 +25,6 @@ import java.io.Reader;
/** /**
* Tokenizes the input into n-grams of the given size(s). * Tokenizes the input into n-grams of the given size(s).
* @author Otis Gospodnetic
*/ */
public class NGramTokenizer extends Tokenizer { public class NGramTokenizer extends Tokenizer {
public static final int DEFAULT_MIN_NGRAM_SIZE = 1; public static final int DEFAULT_MIN_NGRAM_SIZE = 1;

View File

@ -39,7 +39,6 @@ import java.util.Map;
* A default set of stopwords is used unless an alternative list is specified, the * A default set of stopwords is used unless an alternative list is specified, the
* exclusion list is empty by default. * exclusion list is empty by default.
* *
* @author Edwin de Jonge
*/ */
public class DutchAnalyzer extends Analyzer { public class DutchAnalyzer extends Analyzer {
/** /**

View File

@ -32,7 +32,6 @@ import java.util.Map;
* not be stemmed at all. The stemmer used can be changed at runtime after the * not be stemmed at all. The stemmer used can be changed at runtime after the
* filter object is created (as long as it is a DutchStemmer). * filter object is created (as long as it is a DutchStemmer).
* *
* @author Edwin de Jonge
*/ */
public final class DutchStemFilter extends TokenFilter { public final class DutchStemFilter extends TokenFilter {
/** /**

View File

@ -25,7 +25,6 @@ import java.util.Map;
* the <a href="http://snowball.tartarus.org/algorithms/dutch/stemmer.html">dutch stemming</a> * the <a href="http://snowball.tartarus.org/algorithms/dutch/stemmer.html">dutch stemming</a>
* algorithm in Martin Porter's snowball project. * algorithm in Martin Porter's snowball project.
* *
* @author Edwin de Jonge (ejne at cbs.nl)
*/ */
public class DutchStemmer { public class DutchStemmer {

View File

@ -29,7 +29,6 @@ import java.util.HashMap;
* should contain only one word. If the file is not found or on any error, an * should contain only one word. If the file is not found or on any error, an
* empty table is returned. * empty table is returned.
* *
* @author Gerhard Schwarz
* @deprecated use {@link org.apache.lucene.analysis.WordlistLoader} instead * @deprecated use {@link org.apache.lucene.analysis.WordlistLoader} instead
*/ */
public class WordlistLoader { public class WordlistLoader {

View File

@ -26,7 +26,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
/** /**
* Analyzer for Thai language. It uses java.text.BreakIterator to break words. * Analyzer for Thai language. It uses java.text.BreakIterator to break words.
* @author Samphan Raruenrom <samphan@osdev.co.th> for To-Be-One Technology Co., Ltd.
* @version 0.2 * @version 0.2
*/ */
public class ThaiAnalyzer extends Analyzer { public class ThaiAnalyzer extends Analyzer {

View File

@ -27,7 +27,6 @@ import java.text.BreakIterator;
/** /**
* TokenFilter that use java.text.BreakIterator to break each * TokenFilter that use java.text.BreakIterator to break each
* Token that is Thai into separate Token(s) for each Thai word. * Token that is Thai into separate Token(s) for each Thai word.
* @author Samphan Raruenrom <samphan@osdev.co.th> for To-Be-One Technology Co., Ltd.
* @version 0.2 * @version 0.2
*/ */
public class ThaiWordFilter extends TokenFilter { public class ThaiWordFilter extends TokenFilter {

View File

@ -24,9 +24,7 @@ import junit.framework.TestCase;
import org.apache.lucene.analysis.Token; import org.apache.lucene.analysis.Token;
/**
* @author rayt
*/
public class TestChineseTokenizer extends TestCase public class TestChineseTokenizer extends TestCase
{ {
public void testOtherLetterOffset() throws IOException public void testOtherLetterOffset() throws IOException

View File

@ -34,7 +34,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
* than perfect, as it doesn't use any word lists with exceptions. We * than perfect, as it doesn't use any word lists with exceptions. We
* also check some of the cases where the algorithm is wrong. * also check some of the cases where the algorithm is wrong.
* *
* @author Daniel Naber
*/ */
public class TestGermanStemFilter extends TestCase { public class TestGermanStemFilter extends TestCase {

View File

@ -28,7 +28,6 @@ import junit.framework.TestCase;
/** /**
* A unit test class for verifying the correct operation of the GreekAnalyzer. * A unit test class for verifying the correct operation of the GreekAnalyzer.
* *
* @author past
*/ */
public class GreekAnalyzerTest extends TestCase { public class GreekAnalyzerTest extends TestCase {

View File

@ -32,7 +32,6 @@ import org.apache.lucene.analysis.Tokenizer;
import org.apache.lucene.analysis.standard.StandardTokenizer; import org.apache.lucene.analysis.standard.StandardTokenizer;
/** /**
* @author Mathieu Lecarme<mlecarme@openwide.fr>
* *
*/ */
public class TestElision extends TestCase { public class TestElision extends TestCase {

View File

@ -65,7 +65,6 @@ import org.apache.lucene.analysis.TokenStream;
/** /**
* Test case for FrenchAnalyzer. * Test case for FrenchAnalyzer.
* *
* @author Jean-François Halleux
* @version $version$ * @version $version$
*/ */

View File

@ -27,7 +27,6 @@ import junit.framework.TestCase;
/** /**
* Tests {@link EdgeNGramTokenFilter} for correctness. * Tests {@link EdgeNGramTokenFilter} for correctness.
* @author Otis Gospodnetic
*/ */
public class EdgeNGramTokenFilterTest extends TestCase { public class EdgeNGramTokenFilterTest extends TestCase {
private TokenStream input; private TokenStream input;

View File

@ -25,7 +25,6 @@ import junit.framework.TestCase;
/** /**
* Tests {@link EdgeNGramTokenizer} for correctness. * Tests {@link EdgeNGramTokenizer} for correctness.
* @author Otis Gospodnetic
*/ */
public class EdgeNGramTokenizerTest extends TestCase { public class EdgeNGramTokenizerTest extends TestCase {
private StringReader input; private StringReader input;

View File

@ -28,7 +28,6 @@ import junit.framework.TestCase;
/** /**
* Tests {@link NGramTokenFilter} for correctness. * Tests {@link NGramTokenFilter} for correctness.
* @author Otis Gospodnetic
*/ */
public class NGramTokenFilterTest extends TestCase { public class NGramTokenFilterTest extends TestCase {
private TokenStream input; private TokenStream input;

View File

@ -26,7 +26,6 @@ import junit.framework.TestCase;
/** /**
* Tests {@link NGramTokenizer} for correctness. * Tests {@link NGramTokenizer} for correctness.
* @author Otis Gospodnetic
*/ */
public class NGramTokenizerTest extends TestCase { public class NGramTokenizerTest extends TestCase {
private StringReader input; private StringReader input;

View File

@ -26,7 +26,6 @@ import org.apache.lucene.analysis.TokenStream;
/** /**
* Test case for ThaiAnalyzer, modified from TestFrenchAnalyzer * Test case for ThaiAnalyzer, modified from TestFrenchAnalyzer
* *
* @author Samphan Raruenrom <samphan@osdev.co.th>
* @version 0.1 * @version 0.1
*/ */

View File

@ -24,7 +24,6 @@ import java.io.File;
/** /**
* Allows a class to act as a Lucene document handler * Allows a class to act as a Lucene document handler
* *
*@author Erik Hatcher
*@since October 27, 2001 *@since October 27, 2001
*/ */
public interface DocumentHandler { public interface DocumentHandler {

View File

@ -20,8 +20,7 @@ package org.apache.lucene.ant;
import java.io.PrintStream; import java.io.PrintStream;
import java.io.PrintWriter; import java.io.PrintWriter;
/**
*/
public class DocumentHandlerException extends Exception { public class DocumentHandlerException extends Exception {
private Throwable cause; private Throwable cause;

View File

@ -26,7 +26,6 @@ import java.io.File;
* based on a files extension. Currently only .html and .txt * based on a files extension. Currently only .html and .txt
* files are handled, other extensions ignored. * files are handled, other extensions ignored.
* *
*@author Erik Hatcher
*@since October 28, 2001 *@since October 28, 2001
*@todo Implement dynamic document type lookup *@todo Implement dynamic document type lookup
*/ */

View File

@ -39,7 +39,6 @@ import java.io.StringWriter;
* It does this by using JTidy package. It can take input input * It does this by using JTidy package. It can take input input
* from {@link java.io.File} or {@link java.io.InputStream}. * from {@link java.io.File} or {@link java.io.InputStream}.
* *
*@author Erik Hatcher
*/ */
public class HtmlDocument { public class HtmlDocument {
private Element rawDoc; private Element rawDoc;

View File

@ -52,7 +52,6 @@ import java.text.ParseException;
/** /**
* Ant task to index files with Lucene * Ant task to index files with Lucene
* *
*@author Erik Hatcher
*/ */
public class IndexTask extends Task { public class IndexTask extends Task {
/** /**

View File

@ -29,7 +29,6 @@ import java.io.StringWriter;
/** /**
* A utility for making Lucene Documents from a File. * A utility for making Lucene Documents from a File.
* *
*@author Erik Hatcher
*@since December 6, 2001 *@since December 6, 2001
*@todo Fix JavaDoc comments here *@todo Fix JavaDoc comments here
*/ */

View File

@ -37,7 +37,6 @@ import org.apache.tools.ant.types.FileSet;
/** /**
* Test cases for index task * Test cases for index task
* *
*@author Erik Hatcher
*/ */
public class IndexTaskTest extends TestCase { public class IndexTaskTest extends TestCase {
private final static String docHandler = private final static String docHandler =

View File

@ -18,7 +18,6 @@ package org.apache.lucene.benchmark.stats;
/** /**
* This class holds a set of memory usage values. * This class holds a set of memory usage values.
* *
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/ */
public class MemUsage { public class MemUsage {
public long maxFree, minFree, avgFree; public long maxFree, minFree, avgFree;

View File

@ -23,7 +23,6 @@ import org.apache.lucene.benchmark.Constants;
/** /**
* This class holds parameters for a query benchmark. * This class holds parameters for a query benchmark.
* *
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/ */
public class QueryData { public class QueryData {
/** Benchmark id */ /** Benchmark id */

View File

@ -36,7 +36,6 @@ import org.apache.lucene.store.Directory;
* This class holds together all parameters related to a test. Single test is * This class holds together all parameters related to a test. Single test is
* performed several times, and all results are averaged. * performed several times, and all results are averaged.
* *
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/ */
public class TestData public class TestData
{ {

View File

@ -26,7 +26,6 @@ import java.util.Iterator;
* values may contribute to different measurements, so this class provides also * values may contribute to different measurements, so this class provides also
* some useful methods to separate them. * some useful methods to separate them.
* *
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/ */
public class TestRunData { public class TestRunData {
private String id; private String id;

View File

@ -19,7 +19,6 @@ package org.apache.lucene.benchmark.stats;
/** /**
* This class holds a data point measuring speed of processing. * This class holds a data point measuring speed of processing.
* *
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/ */
public class TimeData { public class TimeData {
/** Name of the data point - usually one of a data series with the same name */ /** Name of the data point - usually one of a data series with the same name */

View File

@ -25,7 +25,6 @@ import com.sleepycat.je.DatabaseException;
/** /**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database * Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
* *
* @author Aaron Donovan
*/ */
public class Block extends Object { public class Block extends Object {

View File

@ -34,7 +34,6 @@ import com.sleepycat.je.Transaction;
/** /**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database * Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
* *
* @author Aaron Donovan
*/ */
public class File extends Object { public class File extends Object {

View File

@ -47,7 +47,6 @@ import com.sleepycat.je.Transaction;
* {@link com.sleepycat.je.Database Db} database handles, one for storing file * {@link com.sleepycat.je.Database Db} database handles, one for storing file
* records and another for storing file data blocks. * records and another for storing file data blocks.
* *
* @author Aaron Donovan
*/ */
public class JEDirectory extends Directory { public class JEDirectory extends Directory {

View File

@ -23,7 +23,6 @@ import org.apache.lucene.store.IndexInput;
/** /**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database * Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
* *
* @author Aaron Donovan
*/ */
public class JEIndexInput extends IndexInput { public class JEIndexInput extends IndexInput {

View File

@ -24,7 +24,6 @@ import org.apache.lucene.store.IndexOutput;
/** /**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database * Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
* *
* @author Aaron Donovan
*/ */
public class JEIndexOutput extends IndexOutput { public class JEIndexOutput extends IndexOutput {

View File

@ -22,7 +22,6 @@ import org.apache.lucene.store.Lock;
/** /**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database * Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
* *
* @author Aaron Donovan
*/ */
public class JELock extends Lock { public class JELock extends Lock {

View File

@ -45,7 +45,6 @@ import com.sleepycat.je.Transaction;
* *
* Adapted from Andi Vajda's org.apache.lucene.db.DbStoreTest. * Adapted from Andi Vajda's org.apache.lucene.db.DbStoreTest.
* *
* @author Aaron Donovan
*/ */
public class JEStoreTest extends TestCase { public class JEStoreTest extends TestCase {
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index"); protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");

View File

@ -30,7 +30,6 @@ import com.sleepycat.db.internal.DbTxn;
* that were the entrypoint objects in the pre-4.3 Java API and that wrap the * that were the entrypoint objects in the pre-4.3 Java API and that wrap the
* actual Berkeley DB C objects via SWIG. * actual Berkeley DB C objects via SWIG.
* *
* @author Andi Vajda
*/ */
public class DbHandleExtractor { public class DbHandleExtractor {

View File

@ -24,9 +24,7 @@ import com.sleepycat.db.internal.Db;
import com.sleepycat.db.internal.DbTxn; import com.sleepycat.db.internal.DbTxn;
import com.sleepycat.db.DatabaseException; import com.sleepycat.db.DatabaseException;
/**
* @author Andi Vajda
*/
public class Block extends Object { public class Block extends Object {
protected DatabaseEntry key, data; protected DatabaseEntry key, data;

View File

@ -50,7 +50,6 @@ import com.sleepycat.db.DbHandleExtractor;
* {@link com.sleepycat.db.internal.Db Db} database handles, one for storing file * {@link com.sleepycat.db.internal.Db Db} database handles, one for storing file
* records and another for storing file data blocks. * records and another for storing file data blocks.
* *
* @author Andi Vajda
*/ */
public class DbDirectory extends Directory { public class DbDirectory extends Directory {

View File

@ -21,9 +21,7 @@ import java.io.IOException;
import org.apache.lucene.store.IndexInput; import org.apache.lucene.store.IndexInput;
/**
* @author Andi Vajda
*/
public class DbIndexInput extends IndexInput { public class DbIndexInput extends IndexInput {

View File

@ -21,9 +21,7 @@ import java.io.IOException;
import org.apache.lucene.store.IndexOutput; import org.apache.lucene.store.IndexOutput;
/**
* @author Andi Vajda
*/
public class DbIndexOutput extends IndexOutput { public class DbIndexOutput extends IndexOutput {

View File

@ -24,7 +24,6 @@ import org.apache.lucene.store.Lock;
* trivial as {@link DbDirectory} operations are managed by the Berkeley DB * trivial as {@link DbDirectory} operations are managed by the Berkeley DB
* locking system. * locking system.
* *
* @author Andi Vajda
*/ */
public class DbLock extends Lock { public class DbLock extends Lock {

View File

@ -31,9 +31,7 @@ import com.sleepycat.db.internal.Db;
import com.sleepycat.db.internal.DbTxn; import com.sleepycat.db.internal.DbTxn;
import com.sleepycat.db.DatabaseException; import com.sleepycat.db.DatabaseException;
/**
* @author Andi Vajda
*/
public class File extends Object { public class File extends Object {

View File

@ -42,7 +42,6 @@ import org.apache.lucene.store.IndexOutput;
* Tests {@link DbDirectory}. * Tests {@link DbDirectory}.
* *
* Adapted from org.apache.lucene.StoreTest with larger files and random bytes. * Adapted from org.apache.lucene.StoreTest with larger files and random bytes.
* @author Andi Vajda
*/ */
public class DbStoreTest extends TestCase { public class DbStoreTest extends TestCase {
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index"); protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");

View File

@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
/** /**
* Simple {@link Encoder} implementation that does not modify the output * Simple {@link Encoder} implementation that does not modify the output
* @author Nicko Cadell
* *
*/ */
public class DefaultEncoder implements Encoder public class DefaultEncoder implements Encoder

View File

@ -19,7 +19,6 @@ package org.apache.lucene.search.highlight;
/** /**
* Encodes original text. The Encoder works with the Formatter to generate the output. * Encodes original text. The Encoder works with the Formatter to generate the output.
* *
* @author Nicko Cadell
*/ */
public interface Encoder public interface Encoder
{ {

View File

@ -22,7 +22,6 @@ import org.apache.lucene.analysis.Token;
* Implements the policy for breaking text into multiple fragments for consideration * Implements the policy for breaking text into multiple fragments for consideration
* by the {@link Highlighter} class. A sophisticated implementation may do this on the basis * by the {@link Highlighter} class. A sophisticated implementation may do this on the basis
* of detecting end of sentences in the text. * of detecting end of sentences in the text.
* @author mark@searcharea.co.uk
*/ */
public interface Fragmenter public interface Fragmenter
{ {

View File

@ -20,7 +20,6 @@ package org.apache.lucene.search.highlight;
* Formats text with different color intensity depending on the score of the * Formats text with different color intensity depending on the score of the
* term. * term.
* *
* @author maharwood
*/ */
public class GradientFormatter implements Formatter public class GradientFormatter implements Formatter
{ {

View File

@ -30,7 +30,6 @@ import org.apache.lucene.util.PriorityQueue;
* Class used to markup highlighted terms found in the best sections of a * Class used to markup highlighted terms found in the best sections of a
* text, using configurable {@link Fragmenter}, {@link Scorer}, {@link Formatter}, * text, using configurable {@link Fragmenter}, {@link Scorer}, {@link Formatter},
* {@link Encoder} and tokenizers. * {@link Encoder} and tokenizers.
* @author mark@searcharea.co.uk
*/ */
public class Highlighter public class Highlighter
{ {
@ -503,8 +502,7 @@ public class Highlighter
this.maxDocCharsToAnalyze = maxDocCharsToAnalyze; this.maxDocCharsToAnalyze = maxDocCharsToAnalyze;
} }
/**
*/
public Fragmenter getTextFragmenter() public Fragmenter getTextFragmenter()
{ {
return textFragmenter; return textFragmenter;

View File

@ -27,7 +27,6 @@ import org.apache.lucene.search.Query;
* {@link Scorer} implementation which scores text fragments by the number of unique query terms found. * {@link Scorer} implementation which scores text fragments by the number of unique query terms found.
* This class uses the {@link QueryTermExtractor} class to process determine the query terms and * This class uses the {@link QueryTermExtractor} class to process determine the query terms and
* their boosts to be used. * their boosts to be used.
* @author mark@searcharea.co.uk
*/ */
//TODO: provide option to boost score of fragments near beginning of document //TODO: provide option to boost score of fragments near beginning of document
// based on fragment.getFragNum() // based on fragment.getFragNum()

View File

@ -20,7 +20,6 @@ import org.apache.lucene.analysis.Token;
/** /**
* Adds to the score for a fragment based on its tokens * Adds to the score for a fragment based on its tokens
* @author mark@searcharea.co.uk
*/ */
public interface Scorer public interface Scorer
{ {

View File

@ -21,7 +21,6 @@ import org.apache.lucene.analysis.Token;
/** /**
* {@link Fragmenter} implementation which breaks text up into same-size * {@link Fragmenter} implementation which breaks text up into same-size
* fragments with no concerns over spotting sentence boundaries. * fragments with no concerns over spotting sentence boundaries.
* @author mark@searcharea.co.uk
*/ */
public class SimpleFragmenter implements Fragmenter public class SimpleFragmenter implements Fragmenter
{ {

View File

@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
/** /**
* Simple {@link Encoder} implementation to escape text for HTML output * Simple {@link Encoder} implementation to escape text for HTML output
* @author Nicko Cadell
* *
*/ */
public class SimpleHTMLEncoder implements Encoder public class SimpleHTMLEncoder implements Encoder

View File

@ -18,7 +18,6 @@ package org.apache.lucene.search.highlight;
/** /**
* Simple {@link Formatter} implementation to highlight terms with a pre and post tag * Simple {@link Formatter} implementation to highlight terms with a pre and post tag
* @author MAHarwood
* *
*/ */
public class SimpleHTMLFormatter implements Formatter public class SimpleHTMLFormatter implements Formatter

View File

@ -20,7 +20,6 @@ package org.apache.lucene.search.highlight;
/** /**
* Low-level class used to record information about a section of a document * Low-level class used to record information about a section of a document
* with a score. * with a score.
* @author MAHarwood
* *
* *
*/ */

View File

@ -20,7 +20,6 @@ import org.apache.lucene.analysis.Token;
/** /**
* One, or several overlapping tokens, along with the score(s) and the * One, or several overlapping tokens, along with the score(s) and the
* scope of the original text * scope of the original text
* @author MAHarwood
*/ */
public class TokenGroup public class TokenGroup
{ {

View File

@ -39,7 +39,6 @@ import org.apache.lucene.index.TermVectorOffsetInfo;
* Hides implementation issues associated with obtaining a TokenStream for use with * Hides implementation issues associated with obtaining a TokenStream for use with
* the higlighter - can obtain from TermFreqVectors with offsets and (optionally) positions or * the higlighter - can obtain from TermFreqVectors with offsets and (optionally) positions or
* from Analyzer class reparsing the stored content. * from Analyzer class reparsing the stored content.
* @author maharwood
*/ */
public class TokenSources public class TokenSources
{ {

View File

@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
*/ */
/** Lightweight class to hold term and a weight value used for scoring this term /** Lightweight class to hold term and a weight value used for scoring this term
* @author Mark Harwood
*/ */
public class WeightedTerm public class WeightedTerm
{ {

View File

@ -77,7 +77,6 @@ import org.w3c.dom.NodeList;
/** /**
* JUnit Test for Highlighter class. * JUnit Test for Highlighter class.
* *
* @author mark@searcharea.co.uk
*/ */
public class HighlighterTest extends TestCase implements Formatter { public class HighlighterTest extends TestCase implements Formatter {
private IndexReader reader; private IndexReader reader;
@ -1448,7 +1447,6 @@ class SynonymAnalyzer extends Analyzer {
/** /**
* Expands a token stream with synonyms (TODO - make the synonyms analyzed by choice of analyzer) * Expands a token stream with synonyms (TODO - make the synonyms analyzed by choice of analyzer)
* *
* @author MAHarwood
*/ */
class SynonymTokenizer extends TokenStream { class SynonymTokenizer extends TokenStream {
private TokenStream realStream; private TokenStream realStream;

View File

@ -39,7 +39,6 @@ import org.apache.lucene.analysis.TokenStream;
* Various fulltext analysis utilities avoiding redundant code in several * Various fulltext analysis utilities avoiding redundant code in several
* classes. * classes.
* *
* @author whoschek.AT.lbl.DOT.gov
*/ */
public class AnalyzerUtil { public class AnalyzerUtil {

View File

@ -167,7 +167,6 @@ import org.apache.lucene.search.Similarity;
* href="http://java.sun.com/developer/technicalArticles/Programming/HPROF.html"> * href="http://java.sun.com/developer/technicalArticles/Programming/HPROF.html">
* hprof tracing </a>). * hprof tracing </a>).
* *
* @author whoschek.AT.lbl.DOT.gov
*/ */
public class MemoryIndex implements Serializable { public class MemoryIndex implements Serializable {

View File

@ -61,7 +61,6 @@ import org.apache.lucene.analysis.TokenStream;
* "English")); * "English"));
* </pre> * </pre>
* *
* @author whoschek.AT.lbl.DOT.gov
*/ */
public class PatternAnalyzer extends Analyzer { public class PatternAnalyzer extends Analyzer {

View File

@ -68,7 +68,6 @@ import java.util.TreeSet;
* xxxx:[] * xxxx:[]
* </pre> * </pre>
* *
* @author whoschek.AT.lbl.DOT.gov
* @see <a target="_blank" * @see <a target="_blank"
* href="http://www.cogsci.princeton.edu/~wn/man/prologdb.5WN.html">prologdb * href="http://www.cogsci.princeton.edu/~wn/man/prologdb.5WN.html">prologdb
* man page </a> * man page </a>

View File

@ -28,7 +28,6 @@ import org.apache.lucene.analysis.TokenStream;
* underlying child stream; the child stream must deliver lowercase tokens * underlying child stream; the child stream must deliver lowercase tokens
* for synonyms to be found. * for synonyms to be found.
* *
* @author whoschek.AT.lbl.DOT.gov
*/ */
public class SynonymTokenFilter extends TokenFilter { public class SynonymTokenFilter extends TokenFilter {

View File

@ -196,7 +196,6 @@ hello^2.0
the^3 the^3
</pre> </pre>
@author whoschek.AT.lbl.DOT.gov
*/ */
public class MemoryIndexTest extends TestCase { public class MemoryIndexTest extends TestCase {

View File

@ -58,7 +58,6 @@ Lucene features: CharTokenizer.MAX_WORD_LEN = 255.
Thus the PatternAnalyzer produces correct output, whereas the WhitespaceAnalyzer Thus the PatternAnalyzer produces correct output, whereas the WhitespaceAnalyzer
silently truncates text, and so the comparison results in assertEquals() don't match up. silently truncates text, and so the comparison results in assertEquals() don't match up.
@author whoschek.AT.lbl.DOT.gov
*/ */
public class PatternAnalyzerTest extends TestCase { public class PatternAnalyzerTest extends TestCase {

View File

@ -34,8 +34,6 @@ import org.apache.lucene.store.FSDirectory;
* NOTE: This will overwrite any length normalization or field/document boosts. * NOTE: This will overwrite any length normalization or field/document boosts.
* </p> * </p>
* *
* @author Chris Hostetter
* @author Otis Gospodnetic
*/ */
public class FieldNormModifier { public class FieldNormModifier {

View File

@ -79,7 +79,6 @@ import org.apache.lucene.util.SortedVIntList;
* /least computationally-intensive filters first. * /least computationally-intensive filters first.
* </p> * </p>
* *
* @author <a href="mailto:kelvint@apache.org">Kelvin Tan</a>
*/ */
public class ChainedFilter extends Filter public class ChainedFilter extends Filter
{ {

View File

@ -27,7 +27,6 @@ import java.io.IOException;
/** /**
* Merges indices specified on the command line into the index * Merges indices specified on the command line into the index
* specified as the first command line argument. * specified as the first command line argument.
* @author Erik Hatcher
* @version $Id$ * @version $Id$
*/ */
public class IndexMergeTool { public class IndexMergeTool {

View File

@ -39,7 +39,6 @@ import org.apache.lucene.search.Query;
* become <code>h?user</code> when using this parser and thus no match would be found (i.e. * become <code>h?user</code> when using this parser and thus no match would be found (i.e.
* using this parser will be no improvement over QueryParser in such cases). * using this parser will be no improvement over QueryParser in such cases).
* *
* @author Ronnie Kolehmainen (ronnie.kolehmainen at ub.uu.se)
* @version $Revision$, $Date$ * @version $Revision$, $Date$
*/ */
public class AnalyzingQueryParser extends org.apache.lucene.queryParser.QueryParser { public class AnalyzingQueryParser extends org.apache.lucene.queryParser.QueryParser {

View File

@ -70,9 +70,6 @@ import org.apache.lucene.util.Parameter;
* documentation</a>. * documentation</a>.
* </p> * </p>
* *
* @author Brian Goetz
* @author Peter Halacsy
* @author Tatu Saloranta
*/ */
public class PrecedenceQueryParser implements PrecedenceQueryParserConstants { public class PrecedenceQueryParser implements PrecedenceQueryParserConstants {

View File

@ -30,7 +30,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
import org.apache.lucene.queryParser.ParseException; import org.apache.lucene.queryParser.ParseException;
/** /**
* @author Ronnie Kolehmainen (ronnie.kolehmainen at ub.uu.se)
* @version $Revision$, $Date$ * @version $Revision$, $Date$
*/ */
public class TestAnalyzingQueryParser extends TestCase { public class TestAnalyzingQueryParser extends TestCase {

View File

@ -36,7 +36,6 @@ import org.apache.lucene.util.SortedVIntList;
* SHOULD Filters are OR'd together * SHOULD Filters are OR'd together
* The resulting Filter is NOT'd with the NOT Filters * The resulting Filter is NOT'd with the NOT Filters
* The resulting Filter is AND'd with the MUST Filters * The resulting Filter is AND'd with the MUST Filters
* @author BPDThebault
*/ */
public class BooleanFilter extends Filter public class BooleanFilter extends Filter

View File

@ -24,7 +24,6 @@ import org.apache.lucene.search.BooleanClause.Occur;
* is used when composed with another filter. * is used when composed with another filter.
* (Follows the boolean logic in BooleanClause for composition * (Follows the boolean logic in BooleanClause for composition
* of queries.) * of queries.)
* @author BPDThebault
*/ */
public class FilterClause implements java.io.Serializable public class FilterClause implements java.io.Serializable

View File

@ -46,7 +46,6 @@ import org.apache.lucene.util.PriorityQueue;
* terms eg misspellings. Instead, all variants use the same IDF ranking (the one for the source query * terms eg misspellings. Instead, all variants use the same IDF ranking (the one for the source query
* term) and this is factored into the variant's boost. If the source query term does not exist in the * term) and this is factored into the variant's boost. If the source query term does not exist in the
* index the average IDF of the variants is used. * index the average IDF of the variants is used.
* @author maharwood
*/ */
public class FuzzyLikeThisQuery extends Query public class FuzzyLikeThisQuery extends Query
{ {

View File

@ -35,7 +35,6 @@ import org.apache.lucene.util.OpenBitSet;
* a choice of "category" labels picked by the end user. As a filter, this is much faster than the * a choice of "category" labels picked by the end user. As a filter, this is much faster than the
* equivalent query (a BooleanQuery with many "should" TermQueries) * equivalent query (a BooleanQuery with many "should" TermQueries)
* *
* @author maharwood
*/ */
public class TermsFilter extends Filter public class TermsFilter extends Filter
{ {

View File

@ -144,9 +144,6 @@ import java.util.ArrayList;
* - optimise: when no termvector support available - used maxNumTermsParsed to limit amount of tokenization * - optimise: when no termvector support available - used maxNumTermsParsed to limit amount of tokenization
* </pre> * </pre>
* *
* @author David Spencer
* @author Bruce Ritchie
* @author Mark Harwood
*/ */
public final class MoreLikeThis { public final class MoreLikeThis {

View File

@ -35,7 +35,6 @@ import org.apache.lucene.search.similar.MoreLikeThis;
* A simple wrapper for MoreLikeThis for use in scenarios where a Query object is required eg * A simple wrapper for MoreLikeThis for use in scenarios where a Query object is required eg
* in custom QueryParser extensions. At query.rewrite() time the reader is used to construct the * in custom QueryParser extensions. At query.rewrite() time the reader is used to construct the
* actual MoreLikeThis object and obtain the real Query object. * actual MoreLikeThis object and obtain the real Query object.
* @author maharwood
*/ */
public class MoreLikeThisQuery extends Query public class MoreLikeThisQuery extends Query
{ {

View File

@ -35,7 +35,6 @@ import org.apache.lucene.store.RAMDirectory;
* It first creates a simple index and then a couple of instances of LuceneDictionary * It first creates a simple index and then a couple of instances of LuceneDictionary
* on different fields and checks if all the right text comes back. * on different fields and checks if all the right text comes back.
* *
* @author Christian Mallwitz
*/ */
public class TestLuceneDictionary extends TestCase { public class TestLuceneDictionary extends TestCase {

View File

@ -27,7 +27,6 @@ import org.apache.lucene.store.RAMDirectory;
/** /**
* Test case for PlainTextDictionary * Test case for PlainTextDictionary
* *
* @author Daniel Naber
*/ */
public class TestPlainTextDictionary extends TestCase { public class TestPlainTextDictionary extends TestCase {

View File

@ -36,7 +36,6 @@ import java.util.ArrayList;
/** /**
* See table searcher explanation. * See table searcher explanation.
* *
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/ */
public class ListSearcher extends AbstractListModel { public class ListSearcher extends AbstractListModel {
private ListModel listModel; private ListModel listModel;

View File

@ -62,7 +62,6 @@ import java.util.ArrayList;
* at any time by searching with an empty string. Additionally, you can * at any time by searching with an empty string. Additionally, you can
* add a button calling the clearSearch() method. * add a button calling the clearSearch() method.
* *
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/ */
public class TableSearcher extends AbstractTableModel { public class TableSearcher extends AbstractTableModel {

View File

@ -22,9 +22,7 @@ import java.util.List;
import javax.swing.AbstractListModel; import javax.swing.AbstractListModel;
/**
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/
public class BaseListModel extends AbstractListModel { public class BaseListModel extends AbstractListModel {
private List data = new ArrayList(); private List data = new ArrayList();

View File

@ -22,9 +22,7 @@ import java.util.List;
import javax.swing.table.AbstractTableModel; import javax.swing.table.AbstractTableModel;
/**
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/
public class BaseTableModel extends AbstractTableModel { public class BaseTableModel extends AbstractTableModel {
private List columnNames = new ArrayList(); private List columnNames = new ArrayList();
private List rows = new ArrayList(); private List rows = new ArrayList();

View File

@ -22,9 +22,7 @@ import java.util.ArrayList;
import java.util.Collection; import java.util.Collection;
import java.util.Iterator; import java.util.Iterator;
/**
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/
public class DataStore { public class DataStore {
private static final String ITALIAN_CATEGORY = "Italian"; private static final String ITALIAN_CATEGORY = "Italian";

Some files were not shown because too many files have changed in this diff Show More