Revert "[Rename] server/src/main/java/org/apache (#162)"
This reverts commit c50e8c83a2476ca4ca2f1dd05fa5a608bc0e9ef6 which went should have merged to the rename branch instead of the main branch. Signed-off-by: Peter Nied <petern@amazon.com>
This commit is contained in:
parent
4c709c21f4
commit
abf5850485
|
@ -24,7 +24,7 @@ import org.apache.lucene.analysis.TokenFilter;
|
|||
import org.apache.lucene.analysis.TokenStream;
|
||||
import org.apache.lucene.analysis.tokenattributes.TermToBytesRefAttribute;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.opensearch.common.hash.MurmurHash3;
|
||||
import org.elasticsearch.common.hash.MurmurHash3;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
|
@ -34,14 +34,14 @@ import java.util.ArrayList;
|
|||
* have a minimum length - 6 is a good heuristic as it avoids filtering common
|
||||
* idioms/phrases but detects longer sections that are typical of cut+paste
|
||||
* copies of text.
|
||||
*
|
||||
*
|
||||
* <p>
|
||||
* Internally each token is hashed/moduloed into a single byte (so 256 possible
|
||||
* values for each token) and then recorded in a trie of seen byte sequences
|
||||
* using a {@link DuplicateByteSequenceSpotter}. This trie is passed into the
|
||||
* TokenFilter constructor so a single object can be reused across multiple
|
||||
* documents.
|
||||
*
|
||||
*
|
||||
* <p>
|
||||
* The emitDuplicates setting controls if duplicate tokens are filtered from
|
||||
* results or are output (the {@link DuplicateSequenceAttribute} attribute can
|
||||
|
@ -57,7 +57,7 @@ public class DeDuplicatingTokenFilter extends FilteringTokenFilter {
|
|||
}
|
||||
|
||||
/**
|
||||
*
|
||||
*
|
||||
* @param in
|
||||
* The input token stream
|
||||
* @param byteStreamDuplicateSpotter
|
||||
|
@ -110,9 +110,9 @@ public class DeDuplicatingTokenFilter extends FilteringTokenFilter {
|
|||
}
|
||||
|
||||
public void loadAllTokens() throws IOException {
|
||||
// TODO consider changing this implementation to emit tokens as-we-go
|
||||
// rather than buffering all. However this array is perhaps not the
|
||||
// bulk of memory usage (in practice the dupSequenceSpotter requires
|
||||
// TODO consider changing this implementation to emit tokens as-we-go
|
||||
// rather than buffering all. However this array is perhaps not the
|
||||
// bulk of memory usage (in practice the dupSequenceSpotter requires
|
||||
// ~5x the original content size in its internal tree ).
|
||||
allTokens = new ArrayList<State>(256);
|
||||
|
||||
|
@ -198,4 +198,4 @@ public class DeDuplicatingTokenFilter extends FilteringTokenFilter {
|
|||
}
|
||||
|
||||
}
|
||||
}
|
||||
}
|
|
@ -19,7 +19,7 @@
|
|||
|
||||
package org.apache.lucene.index;
|
||||
|
||||
import org.opensearch.common.lucene.Lucene;
|
||||
import org.elasticsearch.common.lucene.Lucene;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
|
|
|
@ -32,7 +32,7 @@ import org.apache.lucene.search.TwoPhaseIterator;
|
|||
import org.apache.lucene.search.Weight;
|
||||
import org.apache.lucene.store.ByteArrayDataInput;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.opensearch.index.mapper.RangeType;
|
||||
import org.elasticsearch.index.mapper.RangeType;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Objects;
|
||||
|
|
|
@ -20,7 +20,7 @@
|
|||
package org.apache.lucene.queries;
|
||||
|
||||
import org.apache.lucene.search.BooleanClause.Occur;
|
||||
import org.opensearch.common.lucene.search.Queries;
|
||||
import org.elasticsearch.common.lucene.search.Queries;
|
||||
|
||||
/**
|
||||
* Extended version of {@link CommonTermsQuery} that allows to pass in a
|
||||
|
|
|
@ -33,7 +33,7 @@ import org.apache.lucene.search.Scorer;
|
|||
import org.apache.lucene.search.Sort;
|
||||
import org.apache.lucene.search.SortField;
|
||||
import org.apache.lucene.search.Weight;
|
||||
import org.opensearch.common.lucene.Lucene;
|
||||
import org.elasticsearch.common.lucene.Lucene;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Arrays;
|
||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.search.SortField;
|
|||
import org.apache.lucene.search.TopFieldDocs;
|
||||
import org.apache.lucene.search.TotalHits;
|
||||
import org.apache.lucene.util.PriorityQueue;
|
||||
import org.opensearch.common.util.CollectionUtils;
|
||||
import org.elasticsearch.common.util.CollectionUtils;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.HashSet;
|
||||
|
|
|
@ -29,9 +29,9 @@ import org.apache.lucene.index.SortedNumericDocValues;
|
|||
import org.apache.lucene.index.SortedSetDocValues;
|
||||
import org.apache.lucene.search.Scorable;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.opensearch.index.fielddata.AbstractNumericDocValues;
|
||||
import org.opensearch.index.fielddata.AbstractSortedDocValues;
|
||||
import org.opensearch.index.mapper.MappedFieldType;
|
||||
import org.elasticsearch.index.fielddata.AbstractNumericDocValues;
|
||||
import org.elasticsearch.index.fielddata.AbstractSortedDocValues;
|
||||
import org.elasticsearch.index.mapper.MappedFieldType;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Collection;
|
||||
|
|
|
@ -25,7 +25,7 @@ import org.apache.lucene.search.ScoreMode;
|
|||
import org.apache.lucene.search.Sort;
|
||||
import org.apache.lucene.search.SortField;
|
||||
import org.apache.lucene.search.TotalHits;
|
||||
import org.opensearch.index.mapper.MappedFieldType;
|
||||
import org.elasticsearch.index.mapper.MappedFieldType;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Collection;
|
||||
|
|
|
@ -20,7 +20,7 @@
|
|||
package org.apache.lucene.search.uhighlight;
|
||||
|
||||
import org.apache.lucene.search.highlight.Encoder;
|
||||
import org.opensearch.search.fetch.subphase.highlight.HighlightUtils;
|
||||
import org.elasticsearch.search.fetch.subphase.highlight.HighlightUtils;
|
||||
|
||||
/**
|
||||
* Custom passage formatter that allows us to:
|
||||
|
|
|
@ -31,10 +31,10 @@ import org.apache.lucene.search.spans.SpanOrQuery;
|
|||
import org.apache.lucene.search.spans.SpanQuery;
|
||||
import org.apache.lucene.search.spans.SpanTermQuery;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.opensearch.common.CheckedSupplier;
|
||||
import org.opensearch.common.Nullable;
|
||||
import org.opensearch.common.lucene.search.MultiPhrasePrefixQuery;
|
||||
import org.opensearch.index.IndexSettings;
|
||||
import org.elasticsearch.common.CheckedSupplier;
|
||||
import org.elasticsearch.common.Nullable;
|
||||
import org.elasticsearch.common.lucene.search.MultiPhrasePrefixQuery;
|
||||
import org.elasticsearch.index.IndexSettings;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.text.BreakIterator;
|
||||
|
|
|
@ -30,9 +30,9 @@ import org.apache.lucene.search.Query;
|
|||
import org.apache.lucene.search.SynonymQuery;
|
||||
import org.apache.lucene.search.TermQuery;
|
||||
import org.apache.lucene.search.spans.SpanTermQuery;
|
||||
import org.opensearch.common.lucene.search.MultiPhrasePrefixQuery;
|
||||
import org.opensearch.common.lucene.search.function.FunctionScoreQuery;
|
||||
import org.opensearch.index.search.ESToParentBlockJoinQuery;
|
||||
import org.elasticsearch.common.lucene.search.MultiPhrasePrefixQuery;
|
||||
import org.elasticsearch.common.lucene.search.function.FunctionScoreQuery;
|
||||
import org.elasticsearch.index.search.ESToParentBlockJoinQuery;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Collection;
|
||||
|
|
Loading…
Reference in New Issue