LUCENE-1378 - Removed the remaining 199 @author references

git-svn-id: https://svn.apache.org/repos/asf/lucene/java/trunk@695514 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Otis Gospodnetic 2008-09-15 15:42:11 +00:00
parent 8d02d7859f
commit 0195fcd03d
193 changed files with 86 additions and 325 deletions

View File

@ -37,7 +37,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
* will not be indexed at all) and an external list of exclusions (word that will
* not be stemmed, but indexed).
*
* @author João Kramer
*/
public final class BrazilianAnalyzer extends Analyzer {

View File

@ -28,7 +28,6 @@ import java.util.Set;
/**
* Based on GermanStemFilter
*
* @author João Kramer
*/
public final class BrazilianStemFilter extends TokenFilter {

View File

@ -28,7 +28,6 @@ import java.util.Set;
/**
* Filters CJKTokenizer with StopFilter.
*
* @author Che, Dong
*/
public class CJKAnalyzer extends Analyzer {
//~ Static fields/initializers ---------------------------------------------

View File

@ -34,7 +34,6 @@ import java.io.Reader;
* please search <a
* href="http://www.google.com/search?q=word+chinese+segment">google</a>
*
* @author Che, Dong
*/
public final class CJKTokenizer extends Tokenizer {
//~ Static fields/initializers ---------------------------------------------

View File

@ -28,7 +28,6 @@ import org.apache.lucene.analysis.TokenStream;
* build from a ChineseTokenizer, filtered with ChineseFilter.
* Copyright: Copyright (c) 2001
* Company:
* @author Yiyi Sun
* @version 1.0
*
*/

View File

@ -37,7 +37,6 @@ import org.apache.lucene.analysis.TokenStream;
*
* Copyright: Copyright (c) 2001
* Company:
* @author Yiyi Sun
* @version 1.0
*
*/

View File

@ -47,7 +47,6 @@ import org.apache.lucene.analysis.Tokenizer;
* C4C2, C1C2C3 ... the ChineseTokenizer works, but the
* CJKTokenizer will not work.
*
* @author Yiyi Sun
* @version 1.0
*
*/

View File

@ -35,7 +35,6 @@ import java.util.Set;
* A default set of stopwords is used unless an alternative list is specified, the
* exclusion list is empty by default.
*
* @author Lukas Zapletal [lzap@root.cz]
*/
public final class CzechAnalyzer extends Analyzer {

View File

@ -32,7 +32,6 @@ import java.util.Set;
* that will not be indexed at all).
* A default set of stopwords is used unless an alternative list is specified.
*
* @author Panagiotis Astithas, past@ebs.gr
*/
public final class GreekAnalyzer extends Analyzer
{

View File

@ -23,7 +23,6 @@ package org.apache.lucene.analysis.el;
* including accented ones. One should be able to add other encoding schemes (see RFC 1947) by adding
* the definition of a new charset as well as the required logic in the toLowerCase() method.
*
* @author Panagiotis Astithas, past@ebs.gr
*/
public class GreekCharsets
{

View File

@ -23,7 +23,6 @@ import org.apache.lucene.analysis.TokenStream;
/**
* Normalizes token text to lower case, analyzing given ("greek") charset.
*
* @author Panagiotis Astithas, past@ebs.gr
*/
public final class GreekLowerCaseFilter extends TokenFilter
{

View File

@ -32,8 +32,7 @@ import org.apache.lucene.analysis.TokenFilter;
* <p>
* Note that StandardTokenizer sees " ' " as a space, and cuts it out.
*
* @see <a href="http://fr.wikipedia.org/wiki/%C3%89lision">Elision in Wikipedia</a>
* @author <a href="mailto:mlecarme@openwide.fr">Mathieu Lecarme</a>
* @see <a href="http://fr.wikipedia.org/wiki/%C3%89lision">Elision in Wikipedia</a>
*/
public class ElisionFilter extends TokenFilter {
private Set articles = null;

View File

@ -31,7 +31,6 @@ import java.util.Set;
* not be stemmed at all. The used stemmer can be changed at runtime after the
* filter object is created (as long as it is a FrenchStemmer).
*
* @author Patrick Talbot (based on Gerhard Schwarz work for German)
*/
public final class FrenchStemFilter extends TokenFilter {

View File

@ -23,7 +23,6 @@ package org.apache.lucene.analysis.fr;
* refer to http://snowball.sourceforge.net/french/stemmer.html<br>
* (French stemming algorithm) for details
*
* @author Patrick Talbot
*/
public class FrenchStemmer {

View File

@ -26,7 +26,6 @@ import java.util.LinkedList;
/**
* Tokenizes the input into n-grams of the given size(s).
* @author Otis Gospodnetic
*/
public class NGramTokenFilter extends TokenFilter {
public static final int DEFAULT_MIN_NGRAM_SIZE = 1;

View File

@ -25,7 +25,6 @@ import java.io.Reader;
/**
* Tokenizes the input into n-grams of the given size(s).
* @author Otis Gospodnetic
*/
public class NGramTokenizer extends Tokenizer {
public static final int DEFAULT_MIN_NGRAM_SIZE = 1;

View File

@ -38,8 +38,7 @@ import java.util.Map;
* the algorithm (dictionary stemming).
* A default set of stopwords is used unless an alternative list is specified, the
* exclusion list is empty by default.
*
* @author Edwin de Jonge
*
*/
public class DutchAnalyzer extends Analyzer {
/**

View File

@ -31,8 +31,7 @@ import java.util.Map;
* A filter that stems Dutch words. It supports a table of words that should
* not be stemmed at all. The stemmer used can be changed at runtime after the
* filter object is created (as long as it is a DutchStemmer).
*
* @author Edwin de Jonge
*
*/
public final class DutchStemFilter extends TokenFilter {
/**

View File

@ -24,8 +24,7 @@ import java.util.Map;
* A stemmer for Dutch words. The algorithm is an implementation of
* the <a href="http://snowball.tartarus.org/algorithms/dutch/stemmer.html">dutch stemming</a>
* algorithm in Martin Porter's snowball project.
*
* @author Edwin de Jonge (ejne at cbs.nl)
*
*/
public class DutchStemmer {

View File

@ -28,8 +28,7 @@ import java.util.HashMap;
* Loads a text file and adds every line as an entry to a Hashtable. Every line
* should contain only one word. If the file is not found or on any error, an
* empty table is returned.
*
* @author Gerhard Schwarz
*
* @deprecated use {@link org.apache.lucene.analysis.WordlistLoader} instead
*/
public class WordlistLoader {

View File

@ -26,7 +26,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
/**
* Analyzer for Thai language. It uses java.text.BreakIterator to break words.
* @author Samphan Raruenrom <samphan@osdev.co.th> for To-Be-One Technology Co., Ltd.
* @version 0.2
*/
public class ThaiAnalyzer extends Analyzer {

View File

@ -27,7 +27,6 @@ import java.text.BreakIterator;
/**
* TokenFilter that use java.text.BreakIterator to break each
* Token that is Thai into separate Token(s) for each Thai word.
* @author Samphan Raruenrom <samphan@osdev.co.th> for To-Be-One Technology Co., Ltd.
* @version 0.2
*/
public class ThaiWordFilter extends TokenFilter {

View File

@ -24,9 +24,7 @@ import junit.framework.TestCase;
import org.apache.lucene.analysis.Token;
/**
* @author rayt
*/
public class TestChineseTokenizer extends TestCase
{
public void testOtherLetterOffset() throws IOException

View File

@ -33,8 +33,7 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
* Test the German stemmer. The stemming algorithm is known to work less
* than perfect, as it doesn't use any word lists with exceptions. We
* also check some of the cases where the algorithm is wrong.
*
* @author Daniel Naber
*
*/
public class TestGermanStemFilter extends TestCase {

View File

@ -28,7 +28,6 @@ import junit.framework.TestCase;
/**
* A unit test class for verifying the correct operation of the GreekAnalyzer.
*
* @author past
*/
public class GreekAnalyzerTest extends TestCase {

View File

@ -32,7 +32,6 @@ import org.apache.lucene.analysis.Tokenizer;
import org.apache.lucene.analysis.standard.StandardTokenizer;
/**
* @author Mathieu Lecarme<mlecarme@openwide.fr>
*
*/
public class TestElision extends TestCase {

View File

@ -65,7 +65,6 @@ import org.apache.lucene.analysis.TokenStream;
/**
* Test case for FrenchAnalyzer.
*
* @author Jean-François Halleux
* @version $version$
*/

View File

@ -27,7 +27,6 @@ import junit.framework.TestCase;
/**
* Tests {@link EdgeNGramTokenFilter} for correctness.
* @author Otis Gospodnetic
*/
public class EdgeNGramTokenFilterTest extends TestCase {
private TokenStream input;

View File

@ -25,7 +25,6 @@ import junit.framework.TestCase;
/**
* Tests {@link EdgeNGramTokenizer} for correctness.
* @author Otis Gospodnetic
*/
public class EdgeNGramTokenizerTest extends TestCase {
private StringReader input;

View File

@ -28,7 +28,6 @@ import junit.framework.TestCase;
/**
* Tests {@link NGramTokenFilter} for correctness.
* @author Otis Gospodnetic
*/
public class NGramTokenFilterTest extends TestCase {
private TokenStream input;

View File

@ -26,7 +26,6 @@ import junit.framework.TestCase;
/**
* Tests {@link NGramTokenizer} for correctness.
* @author Otis Gospodnetic
*/
public class NGramTokenizerTest extends TestCase {
private StringReader input;

View File

@ -26,7 +26,6 @@ import org.apache.lucene.analysis.TokenStream;
/**
* Test case for ThaiAnalyzer, modified from TestFrenchAnalyzer
*
* @author Samphan Raruenrom <samphan@osdev.co.th>
* @version 0.1
*/

View File

@ -24,7 +24,6 @@ import java.io.File;
/**
* Allows a class to act as a Lucene document handler
*
*@author Erik Hatcher
*@since October 27, 2001
*/
public interface DocumentHandler {

View File

@ -20,8 +20,7 @@ package org.apache.lucene.ant;
import java.io.PrintStream;
import java.io.PrintWriter;
/**
*/
public class DocumentHandlerException extends Exception {
private Throwable cause;

View File

@ -26,7 +26,6 @@ import java.io.File;
* based on a files extension. Currently only .html and .txt
* files are handled, other extensions ignored.
*
*@author Erik Hatcher
*@since October 28, 2001
*@todo Implement dynamic document type lookup
*/

View File

@ -39,7 +39,6 @@ import java.io.StringWriter;
* It does this by using JTidy package. It can take input input
* from {@link java.io.File} or {@link java.io.InputStream}.
*
*@author Erik Hatcher
*/
public class HtmlDocument {
private Element rawDoc;

View File

@ -52,7 +52,6 @@ import java.text.ParseException;
/**
* Ant task to index files with Lucene
*
*@author Erik Hatcher
*/
public class IndexTask extends Task {
/**

View File

@ -29,7 +29,6 @@ import java.io.StringWriter;
/**
* A utility for making Lucene Documents from a File.
*
*@author Erik Hatcher
*@since December 6, 2001
*@todo Fix JavaDoc comments here
*/

View File

@ -37,7 +37,6 @@ import org.apache.tools.ant.types.FileSet;
/**
* Test cases for index task
*
*@author Erik Hatcher
*/
public class IndexTaskTest extends TestCase {
private final static String docHandler =

View File

@ -18,7 +18,6 @@ package org.apache.lucene.benchmark.stats;
/**
* This class holds a set of memory usage values.
*
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/
public class MemUsage {
public long maxFree, minFree, avgFree;

View File

@ -23,7 +23,6 @@ import org.apache.lucene.benchmark.Constants;
/**
* This class holds parameters for a query benchmark.
*
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/
public class QueryData {
/** Benchmark id */

View File

@ -36,7 +36,6 @@ import org.apache.lucene.store.Directory;
* This class holds together all parameters related to a test. Single test is
* performed several times, and all results are averaged.
*
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/
public class TestData
{

View File

@ -26,7 +26,6 @@ import java.util.Iterator;
* values may contribute to different measurements, so this class provides also
* some useful methods to separate them.
*
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/
public class TestRunData {
private String id;

View File

@ -19,7 +19,6 @@ package org.apache.lucene.benchmark.stats;
/**
* This class holds a data point measuring speed of processing.
*
* @author Andrzej Bialecki &lt;ab@getopt.org&gt;
*/
public class TimeData {
/** Name of the data point - usually one of a data series with the same name */

View File

@ -24,8 +24,7 @@ import com.sleepycat.je.DatabaseException;
/**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
*
* @author Aaron Donovan
*
*/
public class Block extends Object {

View File

@ -33,8 +33,7 @@ import com.sleepycat.je.Transaction;
/**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
*
* @author Aaron Donovan
*
*/
public class File extends Object {

View File

@ -46,8 +46,7 @@ import com.sleepycat.je.Transaction;
* {@link org.apache.lucene.store.Directory Directory}. It uses two
* {@link com.sleepycat.je.Database Db} database handles, one for storing file
* records and another for storing file data blocks.
*
* @author Aaron Donovan
*
*/
public class JEDirectory extends Directory {

View File

@ -22,8 +22,7 @@ import org.apache.lucene.store.IndexInput;
/**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
*
* @author Aaron Donovan
*
*/
public class JEIndexInput extends IndexInput {

View File

@ -23,8 +23,7 @@ import org.apache.lucene.store.IndexOutput;
/**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
*
* @author Aaron Donovan
*
*/
public class JEIndexOutput extends IndexOutput {

View File

@ -21,8 +21,7 @@ import org.apache.lucene.store.Lock;
/**
* Port of Andi Vajda's DbDirectory to Java Edition of Berkeley Database
*
* @author Aaron Donovan
*
*/
public class JELock extends Lock {

View File

@ -44,8 +44,7 @@ import com.sleepycat.je.Transaction;
* Tests {@link JEDirectory}.
*
* Adapted from Andi Vajda's org.apache.lucene.db.DbStoreTest.
*
* @author Aaron Donovan
*
*/
public class JEStoreTest extends TestCase {
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");

View File

@ -30,7 +30,6 @@ import com.sleepycat.db.internal.DbTxn;
* that were the entrypoint objects in the pre-4.3 Java API and that wrap the
* actual Berkeley DB C objects via SWIG.
*
* @author Andi Vajda
*/
public class DbHandleExtractor {

View File

@ -24,9 +24,7 @@ import com.sleepycat.db.internal.Db;
import com.sleepycat.db.internal.DbTxn;
import com.sleepycat.db.DatabaseException;
/**
* @author Andi Vajda
*/
public class Block extends Object {
protected DatabaseEntry key, data;

View File

@ -50,7 +50,6 @@ import com.sleepycat.db.DbHandleExtractor;
* {@link com.sleepycat.db.internal.Db Db} database handles, one for storing file
* records and another for storing file data blocks.
*
* @author Andi Vajda
*/
public class DbDirectory extends Directory {

View File

@ -21,9 +21,7 @@ import java.io.IOException;
import org.apache.lucene.store.IndexInput;
/**
* @author Andi Vajda
*/
public class DbIndexInput extends IndexInput {

View File

@ -21,9 +21,7 @@ import java.io.IOException;
import org.apache.lucene.store.IndexOutput;
/**
* @author Andi Vajda
*/
public class DbIndexOutput extends IndexOutput {

View File

@ -24,7 +24,6 @@ import org.apache.lucene.store.Lock;
* trivial as {@link DbDirectory} operations are managed by the Berkeley DB
* locking system.
*
* @author Andi Vajda
*/
public class DbLock extends Lock {

View File

@ -31,9 +31,7 @@ import com.sleepycat.db.internal.Db;
import com.sleepycat.db.internal.DbTxn;
import com.sleepycat.db.DatabaseException;
/**
* @author Andi Vajda
*/
public class File extends Object {

View File

@ -42,7 +42,6 @@ import org.apache.lucene.store.IndexOutput;
* Tests {@link DbDirectory}.
*
* Adapted from org.apache.lucene.StoreTest with larger files and random bytes.
* @author Andi Vajda
*/
public class DbStoreTest extends TestCase {
protected File dbHome = new File(System.getProperty("java.io.tmpdir"),"index");

View File

@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
/**
* Simple {@link Encoder} implementation that does not modify the output
* @author Nicko Cadell
*
*/
public class DefaultEncoder implements Encoder

View File

@ -19,7 +19,6 @@ package org.apache.lucene.search.highlight;
/**
* Encodes original text. The Encoder works with the Formatter to generate the output.
*
* @author Nicko Cadell
*/
public interface Encoder
{

View File

@ -21,8 +21,7 @@ import org.apache.lucene.analysis.Token;
/**
* Implements the policy for breaking text into multiple fragments for consideration
* by the {@link Highlighter} class. A sophisticated implementation may do this on the basis
* of detecting end of sentences in the text.
* @author mark@searcharea.co.uk
* of detecting end of sentences in the text.
*/
public interface Fragmenter
{

View File

@ -19,8 +19,7 @@ package org.apache.lucene.search.highlight;
/**
* Formats text with different color intensity depending on the score of the
* term.
*
* @author maharwood
*
*/
public class GradientFormatter implements Formatter
{

View File

@ -30,7 +30,6 @@ import org.apache.lucene.util.PriorityQueue;
* Class used to markup highlighted terms found in the best sections of a
* text, using configurable {@link Fragmenter}, {@link Scorer}, {@link Formatter},
* {@link Encoder} and tokenizers.
* @author mark@searcharea.co.uk
*/
public class Highlighter
{
@ -503,8 +502,7 @@ public class Highlighter
this.maxDocCharsToAnalyze = maxDocCharsToAnalyze;
}
/**
*/
public Fragmenter getTextFragmenter()
{
return textFragmenter;

View File

@ -26,8 +26,7 @@ import org.apache.lucene.search.Query;
/**
* {@link Scorer} implementation which scores text fragments by the number of unique query terms found.
* This class uses the {@link QueryTermExtractor} class to process determine the query terms and
* their boosts to be used.
* @author mark@searcharea.co.uk
* their boosts to be used.
*/
//TODO: provide option to boost score of fragments near beginning of document
// based on fragment.getFragNum()

View File

@ -20,7 +20,6 @@ import org.apache.lucene.analysis.Token;
/**
* Adds to the score for a fragment based on its tokens
* @author mark@searcharea.co.uk
*/
public interface Scorer
{

View File

@ -21,7 +21,6 @@ import org.apache.lucene.analysis.Token;
/**
* {@link Fragmenter} implementation which breaks text up into same-size
* fragments with no concerns over spotting sentence boundaries.
* @author mark@searcharea.co.uk
*/
public class SimpleFragmenter implements Fragmenter
{

View File

@ -17,7 +17,6 @@ package org.apache.lucene.search.highlight;
/**
* Simple {@link Encoder} implementation to escape text for HTML output
* @author Nicko Cadell
*
*/
public class SimpleHTMLEncoder implements Encoder

View File

@ -18,7 +18,6 @@ package org.apache.lucene.search.highlight;
/**
* Simple {@link Formatter} implementation to highlight terms with a pre and post tag
* @author MAHarwood
*
*/
public class SimpleHTMLFormatter implements Formatter

View File

@ -20,7 +20,6 @@ package org.apache.lucene.search.highlight;
/**
* Low-level class used to record information about a section of a document
* with a score.
* @author MAHarwood
*
*
*/

View File

@ -20,7 +20,6 @@ import org.apache.lucene.analysis.Token;
/**
* One, or several overlapping tokens, along with the score(s) and the
* scope of the original text
* @author MAHarwood
*/
public class TokenGroup
{

View File

@ -38,8 +38,7 @@ import org.apache.lucene.index.TermVectorOffsetInfo;
/**
* Hides implementation issues associated with obtaining a TokenStream for use with
* the higlighter - can obtain from TermFreqVectors with offsets and (optionally) positions or
* from Analyzer class reparsing the stored content.
* @author maharwood
* from Analyzer class reparsing the stored content.
*/
public class TokenSources
{

View File

@ -16,8 +16,7 @@ package org.apache.lucene.search.highlight;
* limitations under the License.
*/
/** Lightweight class to hold term and a weight value used for scoring this term
* @author Mark Harwood
/** Lightweight class to hold term and a weight value used for scoring this term
*/
public class WeightedTerm
{

View File

@ -76,8 +76,7 @@ import org.w3c.dom.NodeList;
/**
* JUnit Test for Highlighter class.
*
* @author mark@searcharea.co.uk
*
*/
public class HighlighterTest extends TestCase implements Formatter {
private IndexReader reader;
@ -1447,8 +1446,7 @@ class SynonymAnalyzer extends Analyzer {
/**
* Expands a token stream with synonyms (TODO - make the synonyms analyzed by choice of analyzer)
*
* @author MAHarwood
*
*/
class SynonymTokenizer extends TokenStream {
private TokenStream realStream;

View File

@ -38,8 +38,7 @@ import org.apache.lucene.analysis.TokenStream;
/**
* Various fulltext analysis utilities avoiding redundant code in several
* classes.
*
* @author whoschek.AT.lbl.DOT.gov
*
*/
public class AnalyzerUtil {

View File

@ -166,8 +166,7 @@ import org.apache.lucene.search.Similarity;
* target="_blank"
* href="http://java.sun.com/developer/technicalArticles/Programming/HPROF.html">
* hprof tracing </a>).
*
* @author whoschek.AT.lbl.DOT.gov
*
*/
public class MemoryIndex implements Serializable {

View File

@ -60,8 +60,7 @@ import org.apache.lucene.analysis.TokenStream;
* pat.tokenStream("content", "James is running round in the woods"),
* "English"));
* </pre>
*
* @author whoschek.AT.lbl.DOT.gov
*
*/
public class PatternAnalyzer extends Analyzer {

View File

@ -67,8 +67,7 @@ import java.util.TreeSet;
* wolfish:[edacious, esurient, rapacious, ravening, ravenous, voracious, wolflike]
* xxxx:[]
* </pre>
*
* @author whoschek.AT.lbl.DOT.gov
*
* @see <a target="_blank"
* href="http://www.cogsci.princeton.edu/~wn/man/prologdb.5WN.html">prologdb
* man page </a>

View File

@ -27,8 +27,7 @@ import org.apache.lucene.analysis.TokenStream;
* Injects additional tokens for synonyms of token terms fetched from the
* underlying child stream; the child stream must deliver lowercase tokens
* for synonyms to be found.
*
* @author whoschek.AT.lbl.DOT.gov
*
*/
public class SynonymTokenFilter extends TokenFilter {

View File

@ -196,7 +196,6 @@ hello^2.0
the^3
</pre>
@author whoschek.AT.lbl.DOT.gov
*/
public class MemoryIndexTest extends TestCase {

View File

@ -58,7 +58,6 @@ Lucene features: CharTokenizer.MAX_WORD_LEN = 255.
Thus the PatternAnalyzer produces correct output, whereas the WhitespaceAnalyzer
silently truncates text, and so the comparison results in assertEquals() don't match up.
@author whoschek.AT.lbl.DOT.gov
*/
public class PatternAnalyzerTest extends TestCase {

View File

@ -34,8 +34,6 @@ import org.apache.lucene.store.FSDirectory;
* NOTE: This will overwrite any length normalization or field/document boosts.
* </p>
*
* @author Chris Hostetter
* @author Otis Gospodnetic
*/
public class FieldNormModifier {

View File

@ -79,7 +79,6 @@ import org.apache.lucene.util.SortedVIntList;
* /least computationally-intensive filters first.
* </p>
*
* @author <a href="mailto:kelvint@apache.org">Kelvin Tan</a>
*/
public class ChainedFilter extends Filter
{

View File

@ -27,7 +27,6 @@ import java.io.IOException;
/**
* Merges indices specified on the command line into the index
* specified as the first command line argument.
* @author Erik Hatcher
* @version $Id$
*/
public class IndexMergeTool {

View File

@ -39,7 +39,6 @@ import org.apache.lucene.search.Query;
* become <code>h?user</code> when using this parser and thus no match would be found (i.e.
* using this parser will be no improvement over QueryParser in such cases).
*
* @author Ronnie Kolehmainen (ronnie.kolehmainen at ub.uu.se)
* @version $Revision$, $Date$
*/
public class AnalyzingQueryParser extends org.apache.lucene.queryParser.QueryParser {

View File

@ -70,9 +70,6 @@ import org.apache.lucene.util.Parameter;
* documentation</a>.
* </p>
*
* @author Brian Goetz
* @author Peter Halacsy
* @author Tatu Saloranta
*/
public class PrecedenceQueryParser implements PrecedenceQueryParserConstants {

View File

@ -30,7 +30,6 @@ import org.apache.lucene.analysis.standard.StandardTokenizer;
import org.apache.lucene.queryParser.ParseException;
/**
* @author Ronnie Kolehmainen (ronnie.kolehmainen at ub.uu.se)
* @version $Revision$, $Date$
*/
public class TestAnalyzingQueryParser extends TestCase {

View File

@ -36,7 +36,6 @@ import org.apache.lucene.util.SortedVIntList;
* SHOULD Filters are OR'd together
* The resulting Filter is NOT'd with the NOT Filters
* The resulting Filter is AND'd with the MUST Filters
* @author BPDThebault
*/
public class BooleanFilter extends Filter

View File

@ -24,7 +24,6 @@ import org.apache.lucene.search.BooleanClause.Occur;
* is used when composed with another filter.
* (Follows the boolean logic in BooleanClause for composition
* of queries.)
* @author BPDThebault
*/
public class FilterClause implements java.io.Serializable

View File

@ -45,8 +45,7 @@ import org.apache.lucene.util.PriorityQueue;
* TermQuery is used for variants and does not use that variant term's IDF because this would favour rarer
* terms eg misspellings. Instead, all variants use the same IDF ranking (the one for the source query
* term) and this is factored into the variant's boost. If the source query term does not exist in the
* index the average IDF of the variants is used.
* @author maharwood
* index the average IDF of the variants is used.
*/
public class FuzzyLikeThisQuery extends Query
{

View File

@ -34,8 +34,7 @@ import org.apache.lucene.util.OpenBitSet;
* a sequence. An example might be a collection of primary keys from a database query result or perhaps
* a choice of "category" labels picked by the end user. As a filter, this is much faster than the
* equivalent query (a BooleanQuery with many "should" TermQueries)
*
* @author maharwood
*
*/
public class TermsFilter extends Filter
{

View File

@ -143,10 +143,7 @@ import java.util.ArrayList;
* - refactor: moved common code into isNoiseWord()
* - optimise: when no termvector support available - used maxNumTermsParsed to limit amount of tokenization
* </pre>
*
* @author David Spencer
* @author Bruce Ritchie
* @author Mark Harwood
*
*/
public final class MoreLikeThis {

View File

@ -35,7 +35,6 @@ import org.apache.lucene.search.similar.MoreLikeThis;
* A simple wrapper for MoreLikeThis for use in scenarios where a Query object is required eg
* in custom QueryParser extensions. At query.rewrite() time the reader is used to construct the
* actual MoreLikeThis object and obtain the real Query object.
* @author maharwood
*/
public class MoreLikeThisQuery extends Query
{

View File

@ -35,7 +35,6 @@ import org.apache.lucene.store.RAMDirectory;
* It first creates a simple index and then a couple of instances of LuceneDictionary
* on different fields and checks if all the right text comes back.
*
* @author Christian Mallwitz
*/
public class TestLuceneDictionary extends TestCase {

View File

@ -27,7 +27,6 @@ import org.apache.lucene.store.RAMDirectory;
/**
* Test case for PlainTextDictionary
*
* @author Daniel Naber
*/
public class TestPlainTextDictionary extends TestCase {

View File

@ -36,7 +36,6 @@ import java.util.ArrayList;
/**
* See table searcher explanation.
*
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/
public class ListSearcher extends AbstractListModel {
private ListModel listModel;

View File

@ -62,7 +62,6 @@ import java.util.ArrayList;
* at any time by searching with an empty string. Additionally, you can
* add a button calling the clearSearch() method.
*
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/
public class TableSearcher extends AbstractTableModel {

View File

@ -22,9 +22,7 @@ import java.util.List;
import javax.swing.AbstractListModel;
/**
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/
public class BaseListModel extends AbstractListModel {
private List data = new ArrayList();

View File

@ -22,9 +22,7 @@ import java.util.List;
import javax.swing.table.AbstractTableModel;
/**
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/
public class BaseTableModel extends AbstractTableModel {
private List columnNames = new ArrayList();
private List rows = new ArrayList();

View File

@ -22,9 +22,7 @@ import java.util.ArrayList;
import java.util.Collection;
import java.util.Iterator;
/**
* @author Jonathan Simon - jonathan_s_simon@yahoo.com
*/
public class DataStore {
private static final String ITALIAN_CATEGORY = "Italian";

Some files were not shown because too many files have changed in this diff Show More