parent
bdccab7c7a
commit
31c026f25c
|
@ -1,5 +1,5 @@
|
|||
elasticsearch = 7.10.0
|
||||
lucene = 8.6.2
|
||||
lucene = 8.7.0-snapshot-61ea26a
|
||||
|
||||
bundled_jdk_vendor = adoptopenjdk
|
||||
bundled_jdk = 14.0.1+7
|
||||
|
|
|
@ -1,8 +1,8 @@
|
|||
|
||||
include::{docs-root}/shared/versions/stack/{source_branch}.asciidoc[]
|
||||
|
||||
:lucene_version: 8.6.2
|
||||
:lucene_version_path: 8_6_2
|
||||
:lucene_version: 8.7.0
|
||||
:lucene_version_path: 8.7.0
|
||||
:jdk: 1.8.0_131
|
||||
:jdk_major: 8
|
||||
:build_flavor: default
|
||||
|
|
|
@ -1 +0,0 @@
|
|||
2f8451c08cd6d19ecacea5f35db2f94014a3a423
|
|
@ -0,0 +1 @@
|
|||
d952f8254b64ff63b1846233371511073736e77d
|
|
@ -1 +0,0 @@
|
|||
75b90e16b9f0773a5fde05715f07f5a0b8a66658
|
|
@ -0,0 +1 @@
|
|||
f69abfe7fcdb8a9c1a929f5e7292924bbd38557a
|
|
@ -1 +0,0 @@
|
|||
665b1e0480983dd2e2996996a104cf05dca783c9
|
|
@ -0,0 +1 @@
|
|||
005c312e96d1c9d47fd3bf0980c810c30c497500
|
|
@ -1 +0,0 @@
|
|||
f3f001555ee4719adf4f8ce4d2bfe3de4ebd1f35
|
|
@ -0,0 +1 @@
|
|||
b0fb357b731fa483f747e2107c2e01b75160af72
|
|
@ -1 +0,0 @@
|
|||
a3676e5a7fb6ed20d0d641f893e3e7b1bda824ea
|
|
@ -0,0 +1 @@
|
|||
689b5f0e3371d0709765c9ba7fc21042cf41754d
|
|
@ -1 +0,0 @@
|
|||
6213af5b0dc92962cf9adc05a14f04682b9bbf15
|
|
@ -0,0 +1 @@
|
|||
5b0a0c7a88076ab8ee21d0b3781fae8d8353da84
|
|
@ -1 +0,0 @@
|
|||
0cb86dd6a32c867802fe5604dbab6e0fd84aa9fc
|
|
@ -0,0 +1 @@
|
|||
8cbea549bf71ca41bfd6ed985d188ee74d6e9098
|
|
@ -1 +0,0 @@
|
|||
3e44b8d571e18a67abb9e3bc941cd1a7cc4203c8
|
|
@ -0,0 +1 @@
|
|||
05cccc203ace5a4af42c77fb32b235c94f8ceb5a
|
|
@ -1 +0,0 @@
|
|||
c4c58571911464138b7ad8881a4319cb0d0ce504
|
|
@ -0,0 +1 @@
|
|||
2191c2c337bbc483c1bce0d7b35f0e00bece50ab
|
|
@ -1 +0,0 @@
|
|||
3719b09652d0c9de87de529060d9c2ce527f929c
|
|
@ -0,0 +1 @@
|
|||
d9d677d9b01528b984c722dd30fb1ceca9f601b7
|
|
@ -1 +0,0 @@
|
|||
de97930c2f5242124e52579eb3d23722e13f0f9d
|
|
@ -0,0 +1 @@
|
|||
331c1607af5fea2251e8c3cd6a0bce04f3ec2bc6
|
|
@ -1 +0,0 @@
|
|||
d6fd5936f99ab50622e75ba6dacff97090cb62e1
|
|
@ -0,0 +1 @@
|
|||
d5fa5c12ccb43629d79591c0079b5648a200e3cb
|
|
@ -1 +0,0 @@
|
|||
275ce757580877efb9370452c282a1de4d10fd3e
|
|
@ -0,0 +1 @@
|
|||
3e24abaa4ad85106b22dbdbd61756bda369c6d04
|
|
@ -1 +0,0 @@
|
|||
8f7e21a00095e0ebed741e2ee425da8019b47b6b
|
|
@ -0,0 +1 @@
|
|||
13b6769b7a956bc4d1b4e162a45aea9a8064948c
|
|
@ -1 +0,0 @@
|
|||
75412c308ff3d9f2bf45248538b9806ce624cc3d
|
|
@ -0,0 +1 @@
|
|||
8f17e777d3782bbec1036770e40823a867a72586
|
|
@ -1 +0,0 @@
|
|||
1072a1d3b3c6ffa40116a88c267bfa8523ccc166
|
|
@ -0,0 +1 @@
|
|||
3ddfc214dc0ea4b39532f2a3800181d1142afa02
|
|
@ -1 +0,0 @@
|
|||
f1edd4d4645d7e8ba07de3302a272123f061cbb5
|
|
@ -0,0 +1 @@
|
|||
b1b5e237cdf3a37f78dfecacdd13373427c6c2d6
|
|
@ -1 +0,0 @@
|
|||
27c206ac7cd7ba7b3edb176c781718c99efb14e2
|
|
@ -0,0 +1 @@
|
|||
115b85c09a0c0d6bd4213d100fbe2ac2494ff5e1
|
|
@ -1 +0,0 @@
|
|||
82d432972fecbeb3713329eca9240fd2bb322bd5
|
|
@ -0,0 +1 @@
|
|||
88c87545785f724c7a9babcf58818ae68c287423
|
|
@ -1 +0,0 @@
|
|||
9b145bd48c34e4a478f1e090a8424a3974ee23f1
|
|
@ -0,0 +1 @@
|
|||
e43bf5d798b5b12c2f4f815df0c1b238d02d1795
|
|
@ -1 +0,0 @@
|
|||
e9b93e2f6c5cdb74332b6b9aada4a190f20d45c3
|
|
@ -0,0 +1 @@
|
|||
a957497d9bb5c0c842703bbaa452125f462de309
|
|
@ -1 +0,0 @@
|
|||
1ffe015e1f765ec0e8aaccd3b82d7578584aebbf
|
|
@ -0,0 +1 @@
|
|||
c8978f546b65eefe4647f3bddcfffe2c574d0c84
|
File diff suppressed because it is too large
Load Diff
|
@ -1,143 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.lucene.search;
|
||||
|
||||
|
||||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.AutomatonProvider;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
|
||||
/**
|
||||
* Copy of Lucene's RegExpQuery class coming in 8.7 with case
|
||||
* insensitive search option
|
||||
* @deprecated
|
||||
*/
|
||||
@Deprecated
|
||||
public class RegexpQuery87 extends AutomatonQuery {
|
||||
/**
|
||||
* A provider that provides no named automata
|
||||
*/
|
||||
private static AutomatonProvider defaultProvider = new AutomatonProvider() {
|
||||
@Override
|
||||
public Automaton getAutomaton(String name) {
|
||||
return null;
|
||||
}
|
||||
};
|
||||
|
||||
/**
|
||||
* Constructs a query for terms matching <code>term</code>.
|
||||
* <p>
|
||||
* By default, all regular expression features are enabled.
|
||||
* </p>
|
||||
*
|
||||
* @param term regular expression.
|
||||
*/
|
||||
public RegexpQuery87(Term term) {
|
||||
this(term, RegExp87.ALL);
|
||||
}
|
||||
|
||||
/**
|
||||
* Constructs a query for terms matching <code>term</code>.
|
||||
*
|
||||
* @param term regular expression.
|
||||
* @param flags optional RegExp features from {@link RegExp87}
|
||||
*/
|
||||
public RegexpQuery87(Term term, int flags) {
|
||||
this(term, flags, defaultProvider,
|
||||
Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
}
|
||||
|
||||
/**
|
||||
* Constructs a query for terms matching <code>term</code>.
|
||||
*
|
||||
* @param term regular expression.
|
||||
* @param flags optional RegExp syntax features from {@link RegExp87}
|
||||
* @param maxDeterminizedStates maximum number of states that compiling the
|
||||
* automaton for the regexp can result in. Set higher to allow more complex
|
||||
* queries and lower to prevent memory exhaustion.
|
||||
*/
|
||||
public RegexpQuery87(Term term, int flags, int maxDeterminizedStates) {
|
||||
this(term, flags, defaultProvider, maxDeterminizedStates);
|
||||
}
|
||||
|
||||
/**
|
||||
* Constructs a query for terms matching <code>term</code>.
|
||||
*
|
||||
* @param term regular expression.
|
||||
* @param syntax_flags optional RegExp syntax features from {@link RegExp87}
|
||||
* automaton for the regexp can result in. Set higher to allow more complex
|
||||
* queries and lower to prevent memory exhaustion.
|
||||
* @param match_flags boolean 'or' of match behavior options such as case insensitivity
|
||||
* @param maxDeterminizedStates maximum number of states that compiling the
|
||||
*/
|
||||
public RegexpQuery87(Term term, int syntax_flags, int match_flags, int maxDeterminizedStates) {
|
||||
this(term, syntax_flags, match_flags, defaultProvider, maxDeterminizedStates);
|
||||
}
|
||||
|
||||
/**
|
||||
* Constructs a query for terms matching <code>term</code>.
|
||||
*
|
||||
* @param term regular expression.
|
||||
* @param syntax_flags optional RegExp features from {@link RegExp87}
|
||||
* @param provider custom AutomatonProvider for named automata
|
||||
* @param maxDeterminizedStates maximum number of states that compiling the
|
||||
* automaton for the regexp can result in. Set higher to allow more complex
|
||||
* queries and lower to prevent memory exhaustion.
|
||||
*/
|
||||
public RegexpQuery87(Term term, int syntax_flags, AutomatonProvider provider,
|
||||
int maxDeterminizedStates) {
|
||||
this(term, syntax_flags, 0, provider, maxDeterminizedStates);
|
||||
}
|
||||
|
||||
/**
|
||||
* Constructs a query for terms matching <code>term</code>.
|
||||
*
|
||||
* @param term regular expression.
|
||||
* @param syntax_flags optional RegExp features from {@link RegExp87}
|
||||
* @param match_flags boolean 'or' of match behavior options such as case insensitivity
|
||||
* @param provider custom AutomatonProvider for named automata
|
||||
* @param maxDeterminizedStates maximum number of states that compiling the
|
||||
* automaton for the regexp can result in. Set higher to allow more complex
|
||||
* queries and lower to prevent memory exhaustion.
|
||||
*/
|
||||
public RegexpQuery87(Term term, int syntax_flags, int match_flags, AutomatonProvider provider,
|
||||
int maxDeterminizedStates) {
|
||||
super(term,
|
||||
new RegExp87(term.text(), syntax_flags, match_flags).toAutomaton(
|
||||
provider, maxDeterminizedStates), maxDeterminizedStates);
|
||||
}
|
||||
|
||||
/** Returns the regexp of this query wrapped in a Term. */
|
||||
public Term getRegexp() {
|
||||
return term;
|
||||
}
|
||||
|
||||
/** Prints a user-readable version of this query. */
|
||||
@Override
|
||||
public String toString(String field) {
|
||||
StringBuilder buffer = new StringBuilder();
|
||||
if (!term.field().equals(field)) {
|
||||
buffer.append(term.field());
|
||||
buffer.append(":");
|
||||
}
|
||||
buffer.append('/');
|
||||
buffer.append(term.text());
|
||||
buffer.append('/');
|
||||
return buffer.toString();
|
||||
}
|
||||
}
|
|
@ -145,7 +145,7 @@ public class Version implements Comparable<Version>, ToXContentFragment {
|
|||
public static final Version V_7_9_0 = new Version(7090099, org.apache.lucene.util.Version.LUCENE_8_6_0);
|
||||
public static final Version V_7_9_1 = new Version(7090199, org.apache.lucene.util.Version.LUCENE_8_6_2);
|
||||
public static final Version V_7_9_2 = new Version(7090299, org.apache.lucene.util.Version.LUCENE_8_6_2);
|
||||
public static final Version V_7_10_0 = new Version(7100099, org.apache.lucene.util.Version.LUCENE_8_6_2);
|
||||
public static final Version V_7_10_0 = new Version(7100099, org.apache.lucene.util.Version.LUCENE_8_7_0);
|
||||
public static final Version CURRENT = V_7_10_0;
|
||||
|
||||
private static final ImmutableOpenIntMap<Version> idToVersion;
|
||||
|
|
|
@ -104,7 +104,7 @@ import java.util.List;
|
|||
import java.util.Map;
|
||||
|
||||
public class Lucene {
|
||||
public static final String LATEST_CODEC = "Lucene86";
|
||||
public static final String LATEST_CODEC = "Lucene87";
|
||||
|
||||
public static final String SOFT_DELETES_FIELD = "__soft_deletes";
|
||||
|
||||
|
|
|
@ -21,8 +21,8 @@ package org.elasticsearch.index.codec;
|
|||
|
||||
import org.apache.logging.log4j.Logger;
|
||||
import org.apache.lucene.codecs.Codec;
|
||||
import org.apache.lucene.codecs.lucene50.Lucene50StoredFieldsFormat.Mode;
|
||||
import org.apache.lucene.codecs.lucene86.Lucene86Codec;
|
||||
import org.apache.lucene.codecs.lucene87.Lucene87Codec;
|
||||
import org.apache.lucene.codecs.lucene87.Lucene87StoredFieldsFormat.Mode;
|
||||
import org.elasticsearch.common.Nullable;
|
||||
import org.elasticsearch.common.collect.MapBuilder;
|
||||
import org.elasticsearch.index.mapper.MapperService;
|
||||
|
@ -47,8 +47,8 @@ public class CodecService {
|
|||
public CodecService(@Nullable MapperService mapperService, Logger logger) {
|
||||
final MapBuilder<String, Codec> codecs = MapBuilder.<String, Codec>newMapBuilder();
|
||||
if (mapperService == null) {
|
||||
codecs.put(DEFAULT_CODEC, new Lucene86Codec());
|
||||
codecs.put(BEST_COMPRESSION_CODEC, new Lucene86Codec(Mode.BEST_COMPRESSION));
|
||||
codecs.put(DEFAULT_CODEC, new Lucene87Codec());
|
||||
codecs.put(BEST_COMPRESSION_CODEC, new Lucene87Codec(Mode.BEST_COMPRESSION));
|
||||
} else {
|
||||
codecs.put(DEFAULT_CODEC,
|
||||
new PerFieldMappingPostingFormatCodec(Mode.BEST_SPEED, mapperService, logger));
|
||||
|
|
|
@ -22,8 +22,8 @@ package org.elasticsearch.index.codec;
|
|||
import org.apache.logging.log4j.Logger;
|
||||
import org.apache.lucene.codecs.Codec;
|
||||
import org.apache.lucene.codecs.PostingsFormat;
|
||||
import org.apache.lucene.codecs.lucene50.Lucene50StoredFieldsFormat;
|
||||
import org.apache.lucene.codecs.lucene86.Lucene86Codec;
|
||||
import org.apache.lucene.codecs.lucene87.Lucene87StoredFieldsFormat;
|
||||
import org.apache.lucene.codecs.lucene87.Lucene87Codec;
|
||||
import org.elasticsearch.common.lucene.Lucene;
|
||||
import org.elasticsearch.index.mapper.CompletionFieldMapper;
|
||||
import org.elasticsearch.index.mapper.MappedFieldType;
|
||||
|
@ -37,7 +37,7 @@ import org.elasticsearch.index.mapper.MapperService;
|
|||
* per index in real time via the mapping API. If no specific postings format is
|
||||
* configured for a specific field the default postings format is used.
|
||||
*/
|
||||
public class PerFieldMappingPostingFormatCodec extends Lucene86Codec {
|
||||
public class PerFieldMappingPostingFormatCodec extends Lucene87Codec {
|
||||
private final Logger logger;
|
||||
private final MapperService mapperService;
|
||||
|
||||
|
@ -46,7 +46,7 @@ public class PerFieldMappingPostingFormatCodec extends Lucene86Codec {
|
|||
"PerFieldMappingPostingFormatCodec must subclass the latest " + "lucene codec: " + Lucene.LATEST_CODEC;
|
||||
}
|
||||
|
||||
public PerFieldMappingPostingFormatCodec(Lucene50StoredFieldsFormat.Mode compressionMode, MapperService mapperService, Logger logger) {
|
||||
public PerFieldMappingPostingFormatCodec(Lucene87StoredFieldsFormat.Mode compressionMode, MapperService mapperService, Logger logger) {
|
||||
super(compressionMode);
|
||||
this.mapperService = mapperService;
|
||||
this.logger = logger;
|
||||
|
|
|
@ -25,7 +25,7 @@ import org.apache.lucene.search.FuzzyQuery;
|
|||
import org.apache.lucene.search.MultiTermQuery;
|
||||
import org.apache.lucene.search.PrefixQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegexpQuery87;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.apache.lucene.search.TermRangeQuery;
|
||||
import org.apache.lucene.search.WildcardQuery;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
|
@ -141,7 +141,7 @@ public abstract class StringFieldType extends TermBasedFieldType {
|
|||
ALLOW_EXPENSIVE_QUERIES.getKey() + "' is set to false.");
|
||||
}
|
||||
failIfNotIndexed();
|
||||
RegexpQuery87 query = new RegexpQuery87(new Term(name(), indexedValueForSearch(value)), syntaxFlags,
|
||||
RegexpQuery query = new RegexpQuery(new Term(name(), indexedValueForSearch(value)), syntaxFlags,
|
||||
matchFlags, maxDeterminizedStates);
|
||||
if (method != null) {
|
||||
query.setRewriteMethod(method);
|
||||
|
|
|
@ -18,7 +18,7 @@
|
|||
*/
|
||||
package org.elasticsearch.index.query;
|
||||
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.elasticsearch.common.Strings;
|
||||
|
||||
import java.util.Locale;
|
||||
|
@ -43,37 +43,37 @@ public enum RegexpFlag {
|
|||
/**
|
||||
* Enables intersection of the form: {@code <expression> & <expression>}
|
||||
*/
|
||||
INTERSECTION(RegExp87.INTERSECTION),
|
||||
INTERSECTION(RegExp.INTERSECTION),
|
||||
|
||||
/**
|
||||
* Enables complement expression of the form: {@code ~<expression>}
|
||||
*/
|
||||
COMPLEMENT(RegExp87.COMPLEMENT),
|
||||
COMPLEMENT(RegExp.COMPLEMENT),
|
||||
|
||||
/**
|
||||
* Enables empty language expression: {@code #}
|
||||
*/
|
||||
EMPTY(RegExp87.EMPTY),
|
||||
EMPTY(RegExp.EMPTY),
|
||||
|
||||
/**
|
||||
* Enables any string expression: {@code @}
|
||||
*/
|
||||
ANYSTRING(RegExp87.ANYSTRING),
|
||||
ANYSTRING(RegExp.ANYSTRING),
|
||||
|
||||
/**
|
||||
* Enables numerical interval expression: {@code <n-m>}
|
||||
*/
|
||||
INTERVAL(RegExp87.INTERVAL),
|
||||
INTERVAL(RegExp.INTERVAL),
|
||||
|
||||
/**
|
||||
* Disables all available option flags
|
||||
*/
|
||||
NONE(RegExp87.NONE),
|
||||
NONE(RegExp.NONE),
|
||||
|
||||
/**
|
||||
* Enables all available option flags
|
||||
*/
|
||||
ALL(RegExp87.ALL);
|
||||
ALL(RegExp.ALL);
|
||||
|
||||
|
||||
final int value;
|
||||
|
@ -110,9 +110,9 @@ public enum RegexpFlag {
|
|||
*/
|
||||
public static int resolveValue(String flags) {
|
||||
if (flags == null || flags.isEmpty()) {
|
||||
return RegExp87.ALL;
|
||||
return RegExp.ALL;
|
||||
}
|
||||
int magic = RegExp87.NONE;
|
||||
int magic = RegExp.NONE;
|
||||
for (String s : Strings.delimitedListToStringArray(flags, "|")) {
|
||||
if (s.isEmpty()) {
|
||||
continue;
|
||||
|
|
|
@ -22,9 +22,9 @@ package org.elasticsearch.index.query;
|
|||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.search.MultiTermQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.search.RegexpQuery87;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.common.ParseField;
|
||||
import org.elasticsearch.common.ParsingException;
|
||||
|
@ -152,19 +152,19 @@ public class RegexpQueryBuilder extends AbstractQueryBuilder<RegexpQueryBuilder>
|
|||
public int flags() {
|
||||
return this.syntaxFlagsValue;
|
||||
}
|
||||
|
||||
|
||||
public RegexpQueryBuilder caseInsensitive(boolean caseInsensitive) {
|
||||
if (caseInsensitive == false) {
|
||||
throw new IllegalArgumentException("The case insensitive setting cannot be set to false.");
|
||||
}
|
||||
this.caseInsensitive = caseInsensitive;
|
||||
return this;
|
||||
}
|
||||
}
|
||||
|
||||
public boolean caseInsensitive() {
|
||||
return this.caseInsensitive;
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Sets the regexp maxDeterminizedStates.
|
||||
*/
|
||||
|
@ -288,17 +288,17 @@ public class RegexpQueryBuilder extends AbstractQueryBuilder<RegexpQueryBuilder>
|
|||
}
|
||||
MultiTermQuery.RewriteMethod method = QueryParsers.parseRewriteMethod(rewrite, null, LoggingDeprecationHandler.INSTANCE);
|
||||
|
||||
int matchFlagsValue = caseInsensitive ? RegExp87.ASCII_CASE_INSENSITIVE : 0;
|
||||
int matchFlagsValue = caseInsensitive ? RegExp.ASCII_CASE_INSENSITIVE : 0;
|
||||
Query query = null;
|
||||
// For BWC we mask irrelevant bits (RegExp changed ALL from 0xffff to 0xff)
|
||||
int sanitisedSyntaxFlag = syntaxFlagsValue & RegExp87.ALL;
|
||||
int sanitisedSyntaxFlag = syntaxFlagsValue & RegExp.ALL;
|
||||
|
||||
MappedFieldType fieldType = context.fieldMapper(fieldName);
|
||||
if (fieldType != null) {
|
||||
query = fieldType.regexpQuery(value, sanitisedSyntaxFlag, matchFlagsValue, maxDeterminizedStates, method, context);
|
||||
}
|
||||
if (query == null) {
|
||||
RegexpQuery87 regexpQuery = new RegexpQuery87(new Term(fieldName, BytesRefs.toBytesRef(value)), sanitisedSyntaxFlag,
|
||||
RegexpQuery regexpQuery = new RegexpQuery(new Term(fieldName, BytesRefs.toBytesRef(value)), sanitisedSyntaxFlag,
|
||||
matchFlagsValue, maxDeterminizedStates);
|
||||
if (method != null) {
|
||||
regexpQuery.setRewriteMethod(method);
|
||||
|
|
|
@ -37,13 +37,13 @@ import org.apache.lucene.search.MultiPhraseQuery;
|
|||
import org.apache.lucene.search.MultiTermQuery;
|
||||
import org.apache.lucene.search.PhraseQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.search.SynonymQuery;
|
||||
import org.apache.lucene.search.WildcardQuery;
|
||||
import org.apache.lucene.search.spans.SpanNearQuery;
|
||||
import org.apache.lucene.search.spans.SpanOrQuery;
|
||||
import org.apache.lucene.search.spans.SpanQuery;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.elasticsearch.common.lucene.search.Queries;
|
||||
import org.elasticsearch.common.regex.Regex;
|
||||
import org.elasticsearch.common.unit.Fuzziness;
|
||||
|
@ -675,13 +675,13 @@ public class QueryStringQueryParser extends XQueryParser {
|
|||
MappedFieldType currentFieldType = queryBuilder.context.fieldMapper(field);
|
||||
if (currentFieldType == null) {
|
||||
return newUnmappedFieldQuery(field);
|
||||
}
|
||||
if (forceAnalyzer != null &&
|
||||
}
|
||||
if (forceAnalyzer != null &&
|
||||
(analyzeWildcard || currentFieldType.getTextSearchInfo().isTokenized())) {
|
||||
setAnalyzer(forceAnalyzer);
|
||||
return super.getWildcardQuery(currentFieldType.name(), termStr);
|
||||
}
|
||||
|
||||
|
||||
return currentFieldType.wildcardQuery(termStr, getMultiTermRewriteMethod(), context);
|
||||
} catch (RuntimeException e) {
|
||||
if (lenient) {
|
||||
|
@ -730,8 +730,8 @@ public class QueryStringQueryParser extends XQueryParser {
|
|||
if (forceAnalyzer != null) {
|
||||
setAnalyzer(forceAnalyzer);
|
||||
return super.getRegexpQuery(field, termStr);
|
||||
}
|
||||
return currentFieldType.regexpQuery(termStr, RegExp87.ALL, 0, getMaxDeterminizedStates(),
|
||||
}
|
||||
return currentFieldType.regexpQuery(termStr, RegExp.ALL, 0, getMaxDeterminizedStates(),
|
||||
getMultiTermRewriteMethod(), context);
|
||||
} catch (RuntimeException e) {
|
||||
if (lenient) {
|
||||
|
|
|
@ -19,15 +19,14 @@
|
|||
|
||||
package org.elasticsearch.search.suggest.completion;
|
||||
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.elasticsearch.ElasticsearchParseException;
|
||||
import org.elasticsearch.common.ParseField;
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||
import org.elasticsearch.common.io.stream.Writeable;
|
||||
import org.elasticsearch.common.xcontent.ObjectParser;
|
||||
import org.elasticsearch.common.xcontent.ToXContent.Params;
|
||||
import org.elasticsearch.common.xcontent.ToXContentFragment;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
|
@ -143,7 +142,7 @@ public class RegexOptions implements ToXContentFragment, Writeable {
|
|||
* Options for regular expression queries
|
||||
*/
|
||||
public static class Builder {
|
||||
private int flagsValue = RegExp87.ALL;
|
||||
private int flagsValue = RegExp.ALL;
|
||||
private int maxDeterminizedStates = Operations.DEFAULT_MAX_DETERMINIZED_STATES;
|
||||
|
||||
public Builder() {
|
||||
|
|
|
@ -21,9 +21,9 @@ package org.elasticsearch.index.codec;
|
|||
|
||||
import org.apache.logging.log4j.LogManager;
|
||||
import org.apache.lucene.codecs.Codec;
|
||||
import org.apache.lucene.codecs.lucene50.Lucene50StoredFieldsFormat;
|
||||
import org.apache.lucene.codecs.lucene50.Lucene50StoredFieldsFormat.Mode;
|
||||
import org.apache.lucene.codecs.lucene86.Lucene86Codec;
|
||||
import org.apache.lucene.codecs.lucene87.Lucene87Codec;
|
||||
import org.apache.lucene.codecs.lucene87.Lucene87StoredFieldsFormat;
|
||||
import org.apache.lucene.codecs.lucene87.Lucene87StoredFieldsFormat.Mode;
|
||||
import org.apache.lucene.document.Document;
|
||||
import org.apache.lucene.index.DirectoryReader;
|
||||
import org.apache.lucene.index.IndexWriter;
|
||||
|
@ -53,8 +53,8 @@ public class CodecTests extends ESTestCase {
|
|||
public void testResolveDefaultCodecs() throws Exception {
|
||||
CodecService codecService = createCodecService();
|
||||
assertThat(codecService.codec("default"), instanceOf(PerFieldMappingPostingFormatCodec.class));
|
||||
assertThat(codecService.codec("default"), instanceOf(Lucene86Codec.class));
|
||||
assertThat(codecService.codec("Lucene86"), instanceOf(Lucene86Codec.class));
|
||||
assertThat(codecService.codec("default"), instanceOf(Lucene87Codec.class));
|
||||
assertThat(codecService.codec("Lucene87"), instanceOf(Lucene87Codec.class));
|
||||
}
|
||||
|
||||
public void testDefault() throws Exception {
|
||||
|
@ -78,7 +78,7 @@ public class CodecTests extends ESTestCase {
|
|||
iw.close();
|
||||
DirectoryReader ir = DirectoryReader.open(dir);
|
||||
SegmentReader sr = (SegmentReader) ir.leaves().get(0).reader();
|
||||
String v = sr.getSegmentInfo().info.getAttribute(Lucene50StoredFieldsFormat.MODE_KEY);
|
||||
String v = sr.getSegmentInfo().info.getAttribute(Lucene87StoredFieldsFormat.MODE_KEY);
|
||||
assertNotNull(v);
|
||||
assertEquals(expected, Mode.valueOf(v));
|
||||
ir.close();
|
||||
|
|
|
@ -19,7 +19,7 @@
|
|||
package org.elasticsearch.index.engine;
|
||||
|
||||
import org.apache.lucene.codecs.PostingsFormat;
|
||||
import org.apache.lucene.codecs.lucene86.Lucene86Codec;
|
||||
import org.apache.lucene.codecs.lucene87.Lucene87Codec;
|
||||
import org.apache.lucene.document.Document;
|
||||
import org.apache.lucene.index.DirectoryReader;
|
||||
import org.apache.lucene.index.IndexWriter;
|
||||
|
@ -57,7 +57,7 @@ public class CompletionStatsCacheTests extends ESTestCase {
|
|||
public void testCompletionStatsCache() throws IOException, InterruptedException {
|
||||
final IndexWriterConfig indexWriterConfig = newIndexWriterConfig();
|
||||
final PostingsFormat postingsFormat = new Completion84PostingsFormat();
|
||||
indexWriterConfig.setCodec(new Lucene86Codec() {
|
||||
indexWriterConfig.setCodec(new Lucene87Codec() {
|
||||
@Override
|
||||
public PostingsFormat getPostingsFormatForField(String field) {
|
||||
return postingsFormat; // all fields are suggest fields
|
||||
|
|
|
@ -22,7 +22,6 @@ import org.apache.lucene.analysis.standard.StandardAnalyzer;
|
|||
import org.apache.lucene.document.SortedSetDocValuesField;
|
||||
import org.apache.lucene.index.IndexableField;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.search.suggest.document.CompletionAnalyzer;
|
||||
import org.apache.lucene.search.suggest.document.ContextSuggestField;
|
||||
import org.apache.lucene.search.suggest.document.FuzzyCompletionQuery;
|
||||
|
@ -32,6 +31,7 @@ import org.apache.lucene.search.suggest.document.SuggestField;
|
|||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.CharsRefBuilder;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.cluster.metadata.IndexMetadata;
|
||||
import org.elasticsearch.common.Strings;
|
||||
|
@ -889,7 +889,7 @@ public class CompletionFieldMapperTests extends ESSingleNodeTestCase {
|
|||
Mapper fieldMapper = defaultMapper.mappers().getMapper("completion");
|
||||
CompletionFieldMapper completionFieldMapper = (CompletionFieldMapper) fieldMapper;
|
||||
Query prefixQuery = completionFieldMapper.fieldType()
|
||||
.regexpQuery(new BytesRef("co"), RegExp87.ALL, Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
.regexpQuery(new BytesRef("co"), RegExp.ALL, Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
assertThat(prefixQuery, instanceOf(RegexCompletionQuery.class));
|
||||
}
|
||||
|
||||
|
|
|
@ -22,7 +22,7 @@ package org.elasticsearch.index.mapper;
|
|||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.search.PrefixQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegexpQuery87;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.apache.lucene.search.WildcardQuery;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.elasticsearch.ElasticsearchException;
|
||||
|
@ -44,7 +44,7 @@ public class IgnoredFieldTypeTests extends FieldTypeTestCase {
|
|||
public void testRegexpQuery() {
|
||||
MappedFieldType ft = IgnoredFieldMapper.IgnoredFieldType.INSTANCE;
|
||||
|
||||
Query expected = new RegexpQuery87(new Term("_ignored", new BytesRef("foo?")));
|
||||
Query expected = new RegexpQuery(new Term("_ignored", new BytesRef("foo?")));
|
||||
assertEquals(expected, ft.regexpQuery("foo?", 0, 0, 10, null, MOCK_QSC));
|
||||
|
||||
ElasticsearchException ee = expectThrows(ElasticsearchException.class,
|
||||
|
|
|
@ -29,7 +29,7 @@ import org.apache.lucene.index.Term;
|
|||
import org.apache.lucene.search.DocValuesFieldExistsQuery;
|
||||
import org.apache.lucene.search.FuzzyQuery;
|
||||
import org.apache.lucene.search.NormsFieldExistsQuery;
|
||||
import org.apache.lucene.search.RegexpQuery87;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.apache.lucene.search.TermInSetQuery;
|
||||
import org.apache.lucene.search.TermQuery;
|
||||
import org.apache.lucene.search.TermRangeQuery;
|
||||
|
@ -128,7 +128,7 @@ public class KeywordFieldTypeTests extends FieldTypeTestCase {
|
|||
|
||||
public void testRegexpQuery() {
|
||||
MappedFieldType ft = new KeywordFieldType("field");
|
||||
assertEquals(new RegexpQuery87(new Term("field","foo.*")),
|
||||
assertEquals(new RegexpQuery(new Term("field","foo.*")),
|
||||
ft.regexpQuery("foo.*", 0, 0, 10, null, MOCK_QSC));
|
||||
|
||||
MappedFieldType unsearchable = new KeywordFieldType("field", false, true, Collections.emptyMap());
|
||||
|
|
|
@ -21,7 +21,7 @@ package org.elasticsearch.index.mapper;
|
|||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.search.PrefixQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegexpQuery87;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.apache.lucene.search.WildcardQuery;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.elasticsearch.ElasticsearchException;
|
||||
|
@ -43,7 +43,7 @@ public class RoutingFieldTypeTests extends FieldTypeTestCase {
|
|||
public void testRegexpQuery() {
|
||||
MappedFieldType ft = RoutingFieldMapper.RoutingFieldType.INSTANCE;
|
||||
|
||||
Query expected = new RegexpQuery87(new Term("_routing", new BytesRef("foo?")));
|
||||
Query expected = new RegexpQuery(new Term("_routing", new BytesRef("foo?")));
|
||||
assertEquals(expected, ft.regexpQuery("foo?", 0, 0, 10, null, MOCK_QSC));
|
||||
|
||||
ElasticsearchException ee = expectThrows(ElasticsearchException.class,
|
||||
|
|
|
@ -26,7 +26,7 @@ import org.apache.lucene.search.ConstantScoreQuery;
|
|||
import org.apache.lucene.search.FuzzyQuery;
|
||||
import org.apache.lucene.search.PrefixQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegexpQuery87;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.apache.lucene.search.TermInSetQuery;
|
||||
import org.apache.lucene.search.TermQuery;
|
||||
import org.apache.lucene.search.TermRangeQuery;
|
||||
|
@ -86,7 +86,7 @@ public class TextFieldTypeTests extends FieldTypeTestCase {
|
|||
|
||||
public void testRegexpQuery() {
|
||||
MappedFieldType ft = new TextFieldType("field");
|
||||
assertEquals(new RegexpQuery87(new Term("field","foo.*")),
|
||||
assertEquals(new RegexpQuery(new Term("field","foo.*")),
|
||||
ft.regexpQuery("foo.*", 0, 0, 10, null, MOCK_QSC));
|
||||
|
||||
MappedFieldType unsearchable = new TextFieldType("field", false, Collections.emptyMap());
|
||||
|
|
|
@ -40,7 +40,7 @@ import org.apache.lucene.search.NormsFieldExistsQuery;
|
|||
import org.apache.lucene.search.PhraseQuery;
|
||||
import org.apache.lucene.search.PrefixQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegexpQuery87;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.apache.lucene.search.SynonymQuery;
|
||||
import org.apache.lucene.search.TermQuery;
|
||||
import org.apache.lucene.search.TermRangeQuery;
|
||||
|
@ -733,8 +733,8 @@ public class QueryStringQueryBuilderTests extends AbstractQueryTestCase<QueryStr
|
|||
Query query = queryStringQuery("/foo*bar/").defaultField(TEXT_FIELD_NAME)
|
||||
.maxDeterminizedStates(5000)
|
||||
.toQuery(createShardContext());
|
||||
assertThat(query, instanceOf(RegexpQuery87.class));
|
||||
RegexpQuery87 regexpQuery = (RegexpQuery87) query;
|
||||
assertThat(query, instanceOf(RegexpQuery.class));
|
||||
RegexpQuery regexpQuery = (RegexpQuery) query;
|
||||
assertTrue(regexpQuery.toString().contains("/foo*bar/"));
|
||||
}
|
||||
|
||||
|
|
|
@ -20,9 +20,7 @@
|
|||
package org.elasticsearch.index.query;
|
||||
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.search.RegexpQuery87;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.elasticsearch.common.ParsingException;
|
||||
import org.elasticsearch.test.AbstractQueryTestCase;
|
||||
|
||||
|
@ -50,7 +48,7 @@ public class RegexpQueryBuilderTests extends AbstractQueryTestCase<RegexpQueryBu
|
|||
}
|
||||
if (randomBoolean()) {
|
||||
query.caseInsensitive(true);
|
||||
}
|
||||
}
|
||||
if (randomBoolean()) {
|
||||
query.maxDeterminizedStates(randomInt(50000));
|
||||
}
|
||||
|
@ -82,8 +80,8 @@ public class RegexpQueryBuilderTests extends AbstractQueryTestCase<RegexpQueryBu
|
|||
|
||||
@Override
|
||||
protected void doAssertLuceneQuery(RegexpQueryBuilder queryBuilder, Query query, QueryShardContext context) throws IOException {
|
||||
assertThat(query, instanceOf(RegexpQuery87.class));
|
||||
RegexpQuery87 regexpQuery = (RegexpQuery87) query;
|
||||
assertThat(query, instanceOf(RegexpQuery.class));
|
||||
RegexpQuery regexpQuery = (RegexpQuery) query;
|
||||
|
||||
String expectedFieldName = expectedFieldName( queryBuilder.fieldName());
|
||||
assertThat(regexpQuery.getField(), equalTo(expectedFieldName));
|
||||
|
@ -152,8 +150,8 @@ public class RegexpQueryBuilderTests extends AbstractQueryTestCase<RegexpQueryBu
|
|||
"}";
|
||||
e = expectThrows(ParsingException.class, () -> parseQuery(shortJson));
|
||||
assertEquals("[regexp] query doesn't support multiple fields, found [user1] and [user2]", e.getMessage());
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
public void testParseFailsWithCaseSensitive() throws IOException {
|
||||
String json =
|
||||
"{\n" +
|
||||
|
@ -166,12 +164,5 @@ public class RegexpQueryBuilderTests extends AbstractQueryTestCase<RegexpQueryBu
|
|||
"}";
|
||||
ParsingException e = expectThrows(ParsingException.class, () -> parseQuery(json));
|
||||
assertEquals("[regexp] query does not support [case_insensitive] = false", e.getMessage());
|
||||
}
|
||||
|
||||
public void testDeadCode() {
|
||||
assertTrue(RegExp87.class + " should be replaced with 8.7's "+RegExp.class,
|
||||
org.apache.lucene.util.Version.LATEST.major == 8 && org.apache.lucene.util.Version.LATEST.minor < 7);
|
||||
}
|
||||
|
||||
|
||||
}
|
||||
|
|
|
@ -61,14 +61,13 @@ import java.util.concurrent.atomic.AtomicInteger;
|
|||
import java.util.function.Supplier;
|
||||
|
||||
import static org.apache.lucene.codecs.compressing.CompressingStoredFieldsWriter.FIELDS_EXTENSION;
|
||||
import static org.apache.lucene.codecs.compressing.CompressingStoredFieldsWriter.INDEX_EXTENSION_PREFIX;
|
||||
import static org.apache.lucene.codecs.compressing.FieldsIndexWriter.FIELDS_INDEX_EXTENSION_SUFFIX;
|
||||
import static org.apache.lucene.codecs.compressing.FieldsIndexWriter.FIELDS_META_EXTENSION_SUFFIX;
|
||||
import static org.apache.lucene.codecs.compressing.CompressingStoredFieldsWriter.INDEX_EXTENSION;
|
||||
import static org.apache.lucene.codecs.compressing.CompressingStoredFieldsWriter.META_EXTENSION;
|
||||
|
||||
public class SourceOnlySnapshot {
|
||||
|
||||
private static final String FIELDS_INDEX_EXTENSION = INDEX_EXTENSION_PREFIX + FIELDS_INDEX_EXTENSION_SUFFIX;
|
||||
private static final String FIELDS_META_EXTENSION = INDEX_EXTENSION_PREFIX + FIELDS_META_EXTENSION_SUFFIX;
|
||||
private static final String FIELDS_INDEX_EXTENSION = INDEX_EXTENSION;
|
||||
private static final String FIELDS_META_EXTENSION = META_EXTENSION;
|
||||
private final LinkedFilesDirectory targetDirectory;
|
||||
private final Supplier<Query> deleteByQuerySupplier;
|
||||
|
||||
|
|
|
@ -13,12 +13,12 @@ import org.apache.lucene.search.MatchAllDocsQuery;
|
|||
import org.apache.lucene.search.MatchNoDocsQuery;
|
||||
import org.apache.lucene.search.MultiTermQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.UnicodeUtil;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.CharacterRunAutomaton;
|
||||
import org.apache.lucene.util.automaton.LevenshteinAutomata;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.elasticsearch.common.geo.ShapeRelation;
|
||||
import org.elasticsearch.common.lucene.BytesRefs;
|
||||
import org.elasticsearch.common.regex.Regex;
|
||||
|
@ -216,7 +216,7 @@ public class ConstantKeywordFieldMapper extends FieldMapper {
|
|||
return new MatchNoDocsQuery();
|
||||
}
|
||||
|
||||
final Automaton automaton = new RegExp87(value, syntaxFlags, matchFlags).toAutomaton(maxDeterminizedStates);
|
||||
final Automaton automaton = new RegExp(value, syntaxFlags, matchFlags).toAutomaton(maxDeterminizedStates);
|
||||
final CharacterRunAutomaton runAutomaton = new CharacterRunAutomaton(automaton);
|
||||
if (runAutomaton.run(this.value)) {
|
||||
return new MatchAllDocsQuery();
|
||||
|
|
|
@ -8,7 +8,7 @@ package org.elasticsearch.xpack.constantkeyword.mapper;
|
|||
|
||||
import org.apache.lucene.search.MatchAllDocsQuery;
|
||||
import org.apache.lucene.search.MatchNoDocsQuery;
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.elasticsearch.common.unit.Fuzziness;
|
||||
import org.elasticsearch.index.mapper.FieldTypeTestCase;
|
||||
import org.elasticsearch.xpack.constantkeyword.mapper.ConstantKeywordFieldMapper.ConstantKeywordFieldType;
|
||||
|
@ -86,9 +86,9 @@ public class ConstantKeywordFieldTypeTests extends FieldTypeTestCase {
|
|||
|
||||
public void testRegexpQuery() {
|
||||
ConstantKeywordFieldType none = new ConstantKeywordFieldType("f", null);
|
||||
assertEquals(new MatchNoDocsQuery(), none.regexpQuery("f..o", RegExp87.ALL, 0, 10, null, null));
|
||||
assertEquals(new MatchNoDocsQuery(), none.regexpQuery("f..o", RegExp.ALL, 0, 10, null, null));
|
||||
ConstantKeywordFieldType ft = new ConstantKeywordFieldType("f", "foo");
|
||||
assertEquals(new MatchAllDocsQuery(), ft.regexpQuery("f.o", RegExp87.ALL, 0, 10, null, null));
|
||||
assertEquals(new MatchNoDocsQuery(), ft.regexpQuery("f..o", RegExp87.ALL, 0, 10, null, null));
|
||||
assertEquals(new MatchAllDocsQuery(), ft.regexpQuery("f.o", RegExp.ALL, 0, 10, null, null));
|
||||
assertEquals(new MatchNoDocsQuery(), ft.regexpQuery("f..o", RegExp.ALL, 0, 10, null, null));
|
||||
}
|
||||
}
|
||||
|
|
|
@ -10,7 +10,7 @@ import org.apache.lucene.index.Term;
|
|||
import org.apache.lucene.search.DocValuesFieldExistsQuery;
|
||||
import org.apache.lucene.search.FuzzyQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegexpQuery87;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.apache.lucene.search.TermQuery;
|
||||
import org.apache.lucene.search.TermRangeQuery;
|
||||
import org.apache.lucene.search.WildcardQuery;
|
||||
|
@ -96,7 +96,7 @@ public class RootFlatObjectFieldTypeTests extends FieldTypeTestCase {
|
|||
public void testRegexpQuery() {
|
||||
RootFlatObjectFieldType ft = createDefaultFieldType();
|
||||
|
||||
Query expected = new RegexpQuery87(new Term("field", "val.*"));
|
||||
Query expected = new RegexpQuery(new Term("field", "val.*"));
|
||||
Query actual = ft.regexpQuery("val.*", 0, 0, 10, null, MOCK_QSC);
|
||||
assertEquals(expected, actual);
|
||||
|
||||
|
|
|
@ -1 +0,0 @@
|
|||
de97930c2f5242124e52579eb3d23722e13f0f9d
|
|
@ -0,0 +1 @@
|
|||
331c1607af5fea2251e8c3cd6a0bce04f3ec2bc6
|
|
@ -29,13 +29,13 @@ import org.apache.lucene.search.MultiTermQuery;
|
|||
import org.apache.lucene.search.MultiTermQuery.RewriteMethod;
|
||||
import org.apache.lucene.search.PrefixQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.search.RegExp87.Kind;
|
||||
import org.apache.lucene.search.TermQuery;
|
||||
import org.apache.lucene.search.TermRangeQuery;
|
||||
import org.apache.lucene.search.WildcardQuery;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.apache.lucene.util.automaton.RegExp.Kind;
|
||||
import org.elasticsearch.ElasticsearchParseException;
|
||||
import org.elasticsearch.common.geo.ShapeRelation;
|
||||
import org.elasticsearch.common.lucene.BytesRefs;
|
||||
|
@ -301,7 +301,7 @@ public class WildcardFieldMapper extends FieldMapper {
|
|||
return new MatchNoDocsQuery();
|
||||
}
|
||||
|
||||
RegExp87 ngramRegex = new RegExp87(addLineEndChars(toLowerCase(value)), syntaxFlags, matchFlags);
|
||||
RegExp ngramRegex = new RegExp(addLineEndChars(toLowerCase(value)), syntaxFlags, matchFlags);
|
||||
|
||||
Query approxBooleanQuery = toApproximationQuery(ngramRegex);
|
||||
Query approxNgramQuery = rewriteBoolToNgramQuery(approxBooleanQuery);
|
||||
|
@ -312,7 +312,7 @@ public class WildcardFieldMapper extends FieldMapper {
|
|||
return existsQuery(context);
|
||||
}
|
||||
Supplier<Automaton> deferredAutomatonSupplier = ()-> {
|
||||
RegExp87 regex = new RegExp87(value, syntaxFlags, matchFlags);
|
||||
RegExp regex = new RegExp(value, syntaxFlags, matchFlags);
|
||||
return regex.toAutomaton(maxDeterminizedStates);
|
||||
};
|
||||
|
||||
|
@ -341,7 +341,7 @@ public class WildcardFieldMapper extends FieldMapper {
|
|||
// * If an expression resolves to a RegExpQuery eg ?? then only the verification
|
||||
// query is run.
|
||||
// * Anything else is a concrete query that should be run on the ngram index.
|
||||
public static Query toApproximationQuery(RegExp87 r) throws IllegalArgumentException {
|
||||
public static Query toApproximationQuery(RegExp r) throws IllegalArgumentException {
|
||||
Query result = null;
|
||||
switch (r.kind) {
|
||||
case REGEXP_UNION:
|
||||
|
@ -402,7 +402,7 @@ public class WildcardFieldMapper extends FieldMapper {
|
|||
return result;
|
||||
}
|
||||
|
||||
private static Query createConcatenationQuery(RegExp87 r) {
|
||||
private static Query createConcatenationQuery(RegExp r) {
|
||||
// Create ANDs of expressions plus collapse consecutive TermQuerys into single longer ones
|
||||
ArrayList<Query> queries = new ArrayList<>();
|
||||
findLeaves(r.exp1, Kind.REGEXP_CONCATENATION, queries);
|
||||
|
@ -433,7 +433,7 @@ public class WildcardFieldMapper extends FieldMapper {
|
|||
|
||||
}
|
||||
|
||||
private static Query createUnionQuery(RegExp87 r) {
|
||||
private static Query createUnionQuery(RegExp r) {
|
||||
// Create an OR of clauses
|
||||
ArrayList<Query> queries = new ArrayList<>();
|
||||
findLeaves(r.exp1, Kind.REGEXP_UNION, queries);
|
||||
|
@ -460,7 +460,7 @@ public class WildcardFieldMapper extends FieldMapper {
|
|||
return new MatchAllButRequireVerificationQuery();
|
||||
}
|
||||
|
||||
private static void findLeaves(RegExp87 exp, Kind kind, List<Query> queries) {
|
||||
private static void findLeaves(RegExp exp, Kind kind, List<Query> queries) {
|
||||
if (exp.kind == kind) {
|
||||
findLeaves(exp.exp1, kind, queries);
|
||||
findLeaves( exp.exp2, kind, queries);
|
||||
|
|
|
@ -26,7 +26,6 @@ import org.apache.lucene.search.IndexSearcher;
|
|||
import org.apache.lucene.search.MatchAllDocsQuery;
|
||||
import org.apache.lucene.search.MatchNoDocsQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegExp87;
|
||||
import org.apache.lucene.search.ScoreDoc;
|
||||
import org.apache.lucene.search.Sort;
|
||||
import org.apache.lucene.search.SortField;
|
||||
|
@ -38,6 +37,7 @@ import org.apache.lucene.store.Directory;
|
|||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.ByteRunAutomaton;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.cluster.metadata.IndexMetadata;
|
||||
import org.elasticsearch.common.collect.List;
|
||||
|
@ -170,7 +170,7 @@ public class WildcardFieldMapperTests extends ESTestCase {
|
|||
assertThat(wildcardFieldTopDocs.totalHits.value, equalTo(0L));
|
||||
|
||||
// Test regexp query
|
||||
wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(queryString, RegExp87.ALL, 0, 20000, null, MOCK_QSC);
|
||||
wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(queryString, RegExp.ALL, 0, 20000, null, MOCK_QSC);
|
||||
wildcardFieldTopDocs = searcher.search(wildcardFieldQuery, 10, Sort.INDEXORDER);
|
||||
assertThat(wildcardFieldTopDocs.totalHits.value, equalTo(0L));
|
||||
|
||||
|
@ -227,8 +227,8 @@ public class WildcardFieldMapperTests extends ESTestCase {
|
|||
break;
|
||||
case 1:
|
||||
pattern = getRandomRegexPattern(values);
|
||||
wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(pattern, RegExp87.ALL, 0, 20000, null, MOCK_QSC);
|
||||
keywordFieldQuery = keywordFieldType.fieldType().regexpQuery(pattern, RegExp87.ALL, 0,20000, null, MOCK_QSC);
|
||||
wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(pattern, RegExp.ALL, 0, 20000, null, MOCK_QSC);
|
||||
keywordFieldQuery = keywordFieldType.fieldType().regexpQuery(pattern, RegExp.ALL, 0,20000, null, MOCK_QSC);
|
||||
break;
|
||||
case 2:
|
||||
pattern = randomABString(5);
|
||||
|
@ -381,12 +381,12 @@ public class WildcardFieldMapperTests extends ESTestCase {
|
|||
// All these expressions should rewrite to a match all with no verification step required at all
|
||||
String superfastRegexes[]= { ".*", "...*..", "(foo|bar|.*)", "@"};
|
||||
for (String regex : superfastRegexes) {
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp87.ALL, 0, 20000, null, MOCK_QSC);
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp.ALL, 0, 20000, null, MOCK_QSC);
|
||||
assertTrue(wildcardFieldQuery instanceof DocValuesFieldExistsQuery);
|
||||
}
|
||||
String matchNoDocsRegexes[]= { ""};
|
||||
for (String regex : matchNoDocsRegexes) {
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp87.ALL, 0, 20000, null, MOCK_QSC);
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp.ALL, 0, 20000, null, MOCK_QSC);
|
||||
assertTrue(wildcardFieldQuery instanceof MatchNoDocsQuery);
|
||||
}
|
||||
|
||||
|
@ -406,7 +406,7 @@ public class WildcardFieldMapperTests extends ESTestCase {
|
|||
for (String[] test : acceleratedTests) {
|
||||
String regex = test[0];
|
||||
String expectedAccelerationQueryString = test[1].replaceAll("_", ""+WildcardFieldMapper.TOKEN_START_OR_END_CHAR);
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp87.ALL, 0, 20000, null, MOCK_QSC);
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp.ALL, 0, 20000, null, MOCK_QSC);
|
||||
testExpectedAccelerationQuery(regex, wildcardFieldQuery, expectedAccelerationQueryString);
|
||||
}
|
||||
|
||||
|
@ -414,7 +414,7 @@ public class WildcardFieldMapperTests extends ESTestCase {
|
|||
// TODO we can possibly improve on some of these
|
||||
String matchAllButVerifyTests[]= { "..", "(a)?","(a|b){0,3}", "((foo)?|(foo|bar)?)", "@&~(abc.+)", "aaa.+&.+bbb"};
|
||||
for (String regex : matchAllButVerifyTests) {
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp87.ALL, 0, 20000, null, MOCK_QSC);
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp.ALL, 0, 20000, null, MOCK_QSC);
|
||||
assertTrue(regex +" was not a pure verify query " +formatQuery(wildcardFieldQuery),
|
||||
wildcardFieldQuery instanceof AutomatonQueryOnBinaryDv);
|
||||
}
|
||||
|
@ -430,7 +430,7 @@ public class WildcardFieldMapperTests extends ESTestCase {
|
|||
for (String[] test : suboptimalTests) {
|
||||
String regex = test[0];
|
||||
String expectedAccelerationQueryString = test[1].replaceAll("_", ""+WildcardFieldMapper.TOKEN_START_OR_END_CHAR);
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp87.ALL, 0, 20000, null, MOCK_QSC);
|
||||
Query wildcardFieldQuery = wildcardFieldType.fieldType().regexpQuery(regex, RegExp.ALL, 0, 20000, null, MOCK_QSC);
|
||||
|
||||
testExpectedAccelerationQuery(regex, wildcardFieldQuery, expectedAccelerationQueryString);
|
||||
}
|
||||
|
@ -769,12 +769,12 @@ public class WildcardFieldMapperTests extends ESTestCase {
|
|||
}
|
||||
|
||||
//Assert our randomly generated regex actually matches the provided raw input.
|
||||
RegExp87 regex = new RegExp87(result.toString());
|
||||
RegExp regex = new RegExp(result.toString());
|
||||
Automaton automaton = regex.toAutomaton();
|
||||
ByteRunAutomaton bytesMatcher = new ByteRunAutomaton(automaton);
|
||||
BytesRef br = new BytesRef(randomValue);
|
||||
assertTrue("[" + result.toString() + "]should match [" + randomValue + "]" + substitutionPoint + "-" + substitutionLength + "/"
|
||||
+ randomValue.length(), bytesMatcher.run(br.bytes, br.offset, br.length));
|
||||
+ randomValue.length(), bytesMatcher.run(br.bytes, br.offset, br.length));
|
||||
return result.toString();
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue