Revert "[Rename] server/src/main/java/org/apache (#162)"

This reverts commit c50e8c83a2476ca4ca2f1dd05fa5a608bc0e9ef6
which went should have merged to the rename branch instead of
the main branch.

Signed-off-by: Peter Nied <petern@amazon.com>
This commit is contained in:
Nicholas Knize 2021-03-02 15:28:07 -06:00 committed by Peter Nied
parent 4c709c21f4
commit abf5850485
11 changed files with 25 additions and 25 deletions

View File

@ -24,7 +24,7 @@ import org.apache.lucene.analysis.TokenFilter;
import org.apache.lucene.analysis.TokenStream;
import org.apache.lucene.analysis.tokenattributes.TermToBytesRefAttribute;
import org.apache.lucene.util.BytesRef;
import org.opensearch.common.hash.MurmurHash3;
import org.elasticsearch.common.hash.MurmurHash3;
import java.io.IOException;
import java.util.ArrayList;
@ -34,14 +34,14 @@ import java.util.ArrayList;
* have a minimum length - 6 is a good heuristic as it avoids filtering common
* idioms/phrases but detects longer sections that are typical of cut+paste
* copies of text.
*
*
* <p>
* Internally each token is hashed/moduloed into a single byte (so 256 possible
* values for each token) and then recorded in a trie of seen byte sequences
* using a {@link DuplicateByteSequenceSpotter}. This trie is passed into the
* TokenFilter constructor so a single object can be reused across multiple
* documents.
*
*
* <p>
* The emitDuplicates setting controls if duplicate tokens are filtered from
* results or are output (the {@link DuplicateSequenceAttribute} attribute can
@ -57,7 +57,7 @@ public class DeDuplicatingTokenFilter extends FilteringTokenFilter {
}
/**
*
*
* @param in
* The input token stream
* @param byteStreamDuplicateSpotter
@ -110,9 +110,9 @@ public class DeDuplicatingTokenFilter extends FilteringTokenFilter {
}
public void loadAllTokens() throws IOException {
// TODO consider changing this implementation to emit tokens as-we-go
// rather than buffering all. However this array is perhaps not the
// bulk of memory usage (in practice the dupSequenceSpotter requires
// TODO consider changing this implementation to emit tokens as-we-go
// rather than buffering all. However this array is perhaps not the
// bulk of memory usage (in practice the dupSequenceSpotter requires
// ~5x the original content size in its internal tree ).
allTokens = new ArrayList<State>(256);
@ -198,4 +198,4 @@ public class DeDuplicatingTokenFilter extends FilteringTokenFilter {
}
}
}
}

View File

@ -19,7 +19,7 @@
package org.apache.lucene.index;
import org.opensearch.common.lucene.Lucene;
import org.elasticsearch.common.lucene.Lucene;
import java.io.IOException;
import java.util.ArrayList;

View File

@ -32,7 +32,7 @@ import org.apache.lucene.search.TwoPhaseIterator;
import org.apache.lucene.search.Weight;
import org.apache.lucene.store.ByteArrayDataInput;
import org.apache.lucene.util.BytesRef;
import org.opensearch.index.mapper.RangeType;
import org.elasticsearch.index.mapper.RangeType;
import java.io.IOException;
import java.util.Objects;

View File

@ -20,7 +20,7 @@
package org.apache.lucene.queries;
import org.apache.lucene.search.BooleanClause.Occur;
import org.opensearch.common.lucene.search.Queries;
import org.elasticsearch.common.lucene.search.Queries;
/**
* Extended version of {@link CommonTermsQuery} that allows to pass in a

View File

@ -33,7 +33,7 @@ import org.apache.lucene.search.Scorer;
import org.apache.lucene.search.Sort;
import org.apache.lucene.search.SortField;
import org.apache.lucene.search.Weight;
import org.opensearch.common.lucene.Lucene;
import org.elasticsearch.common.lucene.Lucene;
import java.io.IOException;
import java.util.Arrays;

View File

@ -26,7 +26,7 @@ import org.apache.lucene.search.SortField;
import org.apache.lucene.search.TopFieldDocs;
import org.apache.lucene.search.TotalHits;
import org.apache.lucene.util.PriorityQueue;
import org.opensearch.common.util.CollectionUtils;
import org.elasticsearch.common.util.CollectionUtils;
import java.util.ArrayList;
import java.util.HashSet;

View File

@ -29,9 +29,9 @@ import org.apache.lucene.index.SortedNumericDocValues;
import org.apache.lucene.index.SortedSetDocValues;
import org.apache.lucene.search.Scorable;
import org.apache.lucene.util.BytesRef;
import org.opensearch.index.fielddata.AbstractNumericDocValues;
import org.opensearch.index.fielddata.AbstractSortedDocValues;
import org.opensearch.index.mapper.MappedFieldType;
import org.elasticsearch.index.fielddata.AbstractNumericDocValues;
import org.elasticsearch.index.fielddata.AbstractSortedDocValues;
import org.elasticsearch.index.mapper.MappedFieldType;
import java.io.IOException;
import java.util.Collection;

View File

@ -25,7 +25,7 @@ import org.apache.lucene.search.ScoreMode;
import org.apache.lucene.search.Sort;
import org.apache.lucene.search.SortField;
import org.apache.lucene.search.TotalHits;
import org.opensearch.index.mapper.MappedFieldType;
import org.elasticsearch.index.mapper.MappedFieldType;
import java.io.IOException;
import java.util.Collection;

View File

@ -20,7 +20,7 @@
package org.apache.lucene.search.uhighlight;
import org.apache.lucene.search.highlight.Encoder;
import org.opensearch.search.fetch.subphase.highlight.HighlightUtils;
import org.elasticsearch.search.fetch.subphase.highlight.HighlightUtils;
/**
* Custom passage formatter that allows us to:

View File

@ -31,10 +31,10 @@ import org.apache.lucene.search.spans.SpanOrQuery;
import org.apache.lucene.search.spans.SpanQuery;
import org.apache.lucene.search.spans.SpanTermQuery;
import org.apache.lucene.util.BytesRef;
import org.opensearch.common.CheckedSupplier;
import org.opensearch.common.Nullable;
import org.opensearch.common.lucene.search.MultiPhrasePrefixQuery;
import org.opensearch.index.IndexSettings;
import org.elasticsearch.common.CheckedSupplier;
import org.elasticsearch.common.Nullable;
import org.elasticsearch.common.lucene.search.MultiPhrasePrefixQuery;
import org.elasticsearch.index.IndexSettings;
import java.io.IOException;
import java.text.BreakIterator;

View File

@ -30,9 +30,9 @@ import org.apache.lucene.search.Query;
import org.apache.lucene.search.SynonymQuery;
import org.apache.lucene.search.TermQuery;
import org.apache.lucene.search.spans.SpanTermQuery;
import org.opensearch.common.lucene.search.MultiPhrasePrefixQuery;
import org.opensearch.common.lucene.search.function.FunctionScoreQuery;
import org.opensearch.index.search.ESToParentBlockJoinQuery;
import org.elasticsearch.common.lucene.search.MultiPhrasePrefixQuery;
import org.elasticsearch.common.lucene.search.function.FunctionScoreQuery;
import org.elasticsearch.index.search.ESToParentBlockJoinQuery;
import java.io.IOException;
import java.util.Collection;