Remove some unused code

This commit is contained in:
Tanguy Leroux 2016-07-01 13:05:41 +02:00
parent 8c40b2b54e
commit 0a293fad29
102 changed files with 148 additions and 317 deletions

View File

@ -26,7 +26,7 @@ import org.elasticsearch.common.unit.ByteSizeValue;
*/ */
public class StoreRateLimiting { public class StoreRateLimiting {
public static interface Provider { public interface Provider {
StoreRateLimiting rateLimiting(); StoreRateLimiting rateLimiting();
} }

View File

@ -38,7 +38,7 @@ public class BulkShardRequest extends ReplicatedWriteRequest<BulkShardRequest> {
public BulkShardRequest() { public BulkShardRequest() {
} }
BulkShardRequest(BulkRequest bulkRequest, ShardId shardId, RefreshPolicy refreshPolicy, BulkItemRequest[] items) { BulkShardRequest(ShardId shardId, RefreshPolicy refreshPolicy, BulkItemRequest[] items) {
super(shardId); super(shardId);
this.items = items; this.items = items;
setRefreshPolicy(refreshPolicy); setRefreshPolicy(refreshPolicy);

View File

@ -344,7 +344,7 @@ public class TransportBulkAction extends HandledTransportAction<BulkRequest, Bul
for (Map.Entry<ShardId, List<BulkItemRequest>> entry : requestsByShard.entrySet()) { for (Map.Entry<ShardId, List<BulkItemRequest>> entry : requestsByShard.entrySet()) {
final ShardId shardId = entry.getKey(); final ShardId shardId = entry.getKey();
final List<BulkItemRequest> requests = entry.getValue(); final List<BulkItemRequest> requests = entry.getValue();
BulkShardRequest bulkShardRequest = new BulkShardRequest(bulkRequest, shardId, bulkRequest.getRefreshPolicy(), BulkShardRequest bulkShardRequest = new BulkShardRequest(shardId, bulkRequest.getRefreshPolicy(),
requests.toArray(new BulkItemRequest[requests.size()])); requests.toArray(new BulkItemRequest[requests.size()]));
bulkShardRequest.consistencyLevel(bulkRequest.consistencyLevel()); bulkShardRequest.consistencyLevel(bulkRequest.consistencyLevel());
bulkShardRequest.timeout(bulkRequest.timeout()); bulkShardRequest.timeout(bulkRequest.timeout());

View File

@ -131,7 +131,7 @@ public class TransportExplainAction extends TransportSingleShardAction<ExplainRe
// Advantage is that we're not opening a second searcher to retrieve the _source. Also // Advantage is that we're not opening a second searcher to retrieve the _source. Also
// because we are working in the same searcher in engineGetResult we can be sure that a // because we are working in the same searcher in engineGetResult we can be sure that a
// doc isn't deleted between the initial get and this call. // doc isn't deleted between the initial get and this call.
GetResult getResult = indexShard.getService().get(result, request.id(), request.type(), request.fields(), request.fetchSourceContext(), false); GetResult getResult = indexShard.getService().get(result, request.id(), request.type(), request.fields(), request.fetchSourceContext());
return new ExplainResponse(shardId.getIndexName(), request.type(), request.id(), true, explanation, getResult); return new ExplainResponse(shardId.getIndexName(), request.type(), request.id(), true, explanation, getResult);
} else { } else {
return new ExplainResponse(shardId.getIndexName(), request.type(), request.id(), true, explanation); return new ExplainResponse(shardId.getIndexName(), request.type(), request.id(), true, explanation);

View File

@ -36,5 +36,5 @@ public interface WriteResponse {
* {@link RefreshPolicy#IMMEDIATE} should always mark this as true. Requests that set it to {@link RefreshPolicy#WAIT_UNTIL} will only * {@link RefreshPolicy#IMMEDIATE} should always mark this as true. Requests that set it to {@link RefreshPolicy#WAIT_UNTIL} will only
* set this to true if they run out of refresh listener slots (see {@link IndexSettings#MAX_REFRESH_LISTENERS_PER_SHARD}). * set this to true if they run out of refresh listener slots (see {@link IndexSettings#MAX_REFRESH_LISTENERS_PER_SHARD}).
*/ */
public abstract void setForcedRefresh(boolean forcedRefresh); void setForcedRefresh(boolean forcedRefresh);
} }

View File

@ -298,7 +298,6 @@ public class ReplicationResponse extends ActionResponse {
private static final String _SHARDS = "_shards"; private static final String _SHARDS = "_shards";
private static final String TOTAL = "total"; private static final String TOTAL = "total";
private static final String SUCCESSFUL = "successful"; private static final String SUCCESSFUL = "successful";
private static final String PENDING = "pending";
private static final String FAILED = "failed"; private static final String FAILED = "failed";
private static final String FAILURES = "failures"; private static final String FAILURES = "failures";

View File

@ -595,7 +595,7 @@ public class TermVectorsRequest extends SingleShardRequest<TermVectorsRequest> i
} else if (currentFieldName.equals("per_field_analyzer") || currentFieldName.equals("perFieldAnalyzer")) { } else if (currentFieldName.equals("per_field_analyzer") || currentFieldName.equals("perFieldAnalyzer")) {
termVectorsRequest.perFieldAnalyzer(readPerFieldAnalyzer(parser.map())); termVectorsRequest.perFieldAnalyzer(readPerFieldAnalyzer(parser.map()));
} else if (currentFieldName.equals("filter")) { } else if (currentFieldName.equals("filter")) {
termVectorsRequest.filterSettings(readFilterSettings(parser, termVectorsRequest)); termVectorsRequest.filterSettings(readFilterSettings(parser));
} else if ("_index".equals(currentFieldName)) { // the following is important for multi request parsing. } else if ("_index".equals(currentFieldName)) { // the following is important for multi request parsing.
termVectorsRequest.index = parser.text(); termVectorsRequest.index = parser.text();
} else if ("_type".equals(currentFieldName)) { } else if ("_type".equals(currentFieldName)) {
@ -641,7 +641,7 @@ public class TermVectorsRequest extends SingleShardRequest<TermVectorsRequest> i
return mapStrStr; return mapStrStr;
} }
private static FilterSettings readFilterSettings(XContentParser parser, TermVectorsRequest termVectorsRequest) throws IOException { private static FilterSettings readFilterSettings(XContentParser parser) throws IOException {
FilterSettings settings = new FilterSettings(); FilterSettings settings = new FilterSettings();
XContentParser.Token token; XContentParser.Token token;
String currentFieldName = null; String currentFieldName = null;

View File

@ -22,8 +22,6 @@ package org.elasticsearch.action.update;
import org.elasticsearch.action.delete.DeleteRequest; import org.elasticsearch.action.delete.DeleteRequest;
import org.elasticsearch.action.index.IndexRequest; import org.elasticsearch.action.index.IndexRequest;
import org.elasticsearch.client.Requests; import org.elasticsearch.client.Requests;
import org.elasticsearch.cluster.ClusterState;
import org.elasticsearch.cluster.service.ClusterService;
import org.elasticsearch.common.Nullable; import org.elasticsearch.common.Nullable;
import org.elasticsearch.common.bytes.BytesReference; import org.elasticsearch.common.bytes.BytesReference;
import org.elasticsearch.common.collect.Tuple; import org.elasticsearch.common.collect.Tuple;
@ -63,13 +61,11 @@ import java.util.Map;
public class UpdateHelper extends AbstractComponent { public class UpdateHelper extends AbstractComponent {
private final ScriptService scriptService; private final ScriptService scriptService;
private final ClusterService clusterService;
@Inject @Inject
public UpdateHelper(Settings settings, ScriptService scriptService, ClusterService clusterService) { public UpdateHelper(Settings settings, ScriptService scriptService) {
super(settings); super(settings);
this.scriptService = scriptService; this.scriptService = scriptService;
this.clusterService = clusterService;
} }
/** /**

View File

@ -26,16 +26,16 @@ package org.elasticsearch.cluster;
public interface ClusterInfoService { public interface ClusterInfoService {
/** The latest cluster information */ /** The latest cluster information */
public ClusterInfo getClusterInfo(); ClusterInfo getClusterInfo();
/** Add a listener that will be called every time new information is gathered */ /** Add a listener that will be called every time new information is gathered */
public void addListener(Listener listener); void addListener(Listener listener);
/** /**
* Interface for listeners to implement in order to perform actions when * Interface for listeners to implement in order to perform actions when
* new information about the cluster has been gathered * new information about the cluster has been gathered
*/ */
public interface Listener { interface Listener {
public void onNewInfo(ClusterInfo info); void onNewInfo(ClusterInfo info);
} }
} }

View File

@ -32,7 +32,7 @@ public interface IndexTemplateFilter {
*/ */
boolean apply(CreateIndexClusterStateUpdateRequest request, IndexTemplateMetaData template); boolean apply(CreateIndexClusterStateUpdateRequest request, IndexTemplateMetaData template);
static class Compound implements IndexTemplateFilter { class Compound implements IndexTemplateFilter {
private IndexTemplateFilter[] filters; private IndexTemplateFilter[] filters;

View File

@ -19,11 +19,8 @@
package org.elasticsearch.cluster.metadata; package org.elasticsearch.cluster.metadata;
import org.elasticsearch.Version;
import org.elasticsearch.action.TimestampParsingException; import org.elasticsearch.action.TimestampParsingException;
import org.elasticsearch.cluster.AbstractDiffable; import org.elasticsearch.cluster.AbstractDiffable;
import org.elasticsearch.common.Nullable;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.compress.CompressedXContent; import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.common.io.stream.StreamInput; import org.elasticsearch.common.io.stream.StreamInput;
import org.elasticsearch.common.io.stream.StreamOutput; import org.elasticsearch.common.io.stream.StreamOutput;
@ -34,11 +31,9 @@ import org.elasticsearch.common.xcontent.XContentFactory;
import org.elasticsearch.common.xcontent.XContentHelper; import org.elasticsearch.common.xcontent.XContentHelper;
import org.elasticsearch.common.xcontent.XContentParser; import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.index.mapper.DocumentMapper; import org.elasticsearch.index.mapper.DocumentMapper;
import org.elasticsearch.index.mapper.MapperParsingException;
import org.elasticsearch.index.mapper.internal.TimestampFieldMapper; import org.elasticsearch.index.mapper.internal.TimestampFieldMapper;
import java.io.IOException; import java.io.IOException;
import java.util.Arrays;
import java.util.Map; import java.util.Map;
import static org.elasticsearch.common.xcontent.support.XContentMapValues.lenientNodeBooleanValue; import static org.elasticsearch.common.xcontent.support.XContentMapValues.lenientNodeBooleanValue;
@ -82,8 +77,6 @@ public class MappingMetaData extends AbstractDiffable<MappingMetaData> {
public static class Timestamp { public static class Timestamp {
private static final FormatDateTimeFormatter EPOCH_MILLIS_PARSER = Joda.forPattern("epoch_millis");
public static String parseStringTimestamp(String timestampAsString, FormatDateTimeFormatter dateTimeFormatter) throws TimestampParsingException { public static String parseStringTimestamp(String timestampAsString, FormatDateTimeFormatter dateTimeFormatter) throws TimestampParsingException {
try { try {
return Long.toString(dateTimeFormatter.parser().parseMillis(timestampAsString)); return Long.toString(dateTimeFormatter.parser().parseMillis(timestampAsString));

View File

@ -35,8 +35,8 @@ import org.elasticsearch.common.regex.Regex;
import org.elasticsearch.common.settings.Settings; import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.common.unit.TimeValue; import org.elasticsearch.common.unit.TimeValue;
import org.elasticsearch.index.Index; import org.elasticsearch.index.Index;
import org.elasticsearch.index.NodeServicesProvider;
import org.elasticsearch.index.IndexService; import org.elasticsearch.index.IndexService;
import org.elasticsearch.index.NodeServicesProvider;
import org.elasticsearch.index.mapper.MapperParsingException; import org.elasticsearch.index.mapper.MapperParsingException;
import org.elasticsearch.index.mapper.MapperService; import org.elasticsearch.index.mapper.MapperService;
import org.elasticsearch.indices.IndexTemplateAlreadyExistsException; import org.elasticsearch.indices.IndexTemplateAlreadyExistsException;
@ -160,7 +160,7 @@ public class MetaDataIndexTemplateService extends AbstractComponent {
throw new IndexTemplateAlreadyExistsException(request.name); throw new IndexTemplateAlreadyExistsException(request.name);
} }
validateAndAddTemplate(request, templateBuilder, indicesService, nodeServicesProvider, metaDataCreateIndexService); validateAndAddTemplate(request, templateBuilder, indicesService, nodeServicesProvider);
for (Alias alias : request.aliases) { for (Alias alias : request.aliases) {
AliasMetaData aliasMetaData = AliasMetaData.builder(alias.name()).filter(alias.filter()) AliasMetaData aliasMetaData = AliasMetaData.builder(alias.name()).filter(alias.filter())
@ -185,7 +185,7 @@ public class MetaDataIndexTemplateService extends AbstractComponent {
} }
private static void validateAndAddTemplate(final PutRequest request, IndexTemplateMetaData.Builder templateBuilder, IndicesService indicesService, private static void validateAndAddTemplate(final PutRequest request, IndexTemplateMetaData.Builder templateBuilder, IndicesService indicesService,
NodeServicesProvider nodeServicesProvider, MetaDataCreateIndexService metaDataCreateIndexService) throws Exception { NodeServicesProvider nodeServicesProvider) throws Exception {
Index createdIndex = null; Index createdIndex = null;
final String temporaryIndexName = UUIDs.randomBase64UUID(); final String temporaryIndexName = UUIDs.randomBase64UUID();
try { try {
@ -276,7 +276,7 @@ public class MetaDataIndexTemplateService extends AbstractComponent {
} }
} }
public static interface PutListener { public interface PutListener {
void onResponse(PutResponse response); void onResponse(PutResponse response);
@ -391,7 +391,7 @@ public class MetaDataIndexTemplateService extends AbstractComponent {
} }
} }
public static interface RemoveListener { public interface RemoveListener {
void onResponse(RemoveResponse response); void onResponse(RemoveResponse response);

View File

@ -229,11 +229,4 @@ public class OperationRouting extends AbstractComponent {
// of original index to hash documents // of original index to hash documents
return Math.floorMod(hash, indexMetaData.getRoutingNumShards()) / indexMetaData.getRoutingFactor(); return Math.floorMod(hash, indexMetaData.getRoutingNumShards()) / indexMetaData.getRoutingFactor();
} }
private void ensureNodeIdExists(DiscoveryNodes nodes, String nodeId) {
if (!nodes.getDataNodes().keys().contains(nodeId)) {
throw new IllegalArgumentException("No data node with id[" + nodeId + "] found");
}
}
} }

View File

@ -505,7 +505,6 @@ public class DiskThresholdDecider extends AllocationDecider {
} }
private DiskUsage getDiskUsage(RoutingNode node, RoutingAllocation allocation, ImmutableOpenMap<String, DiskUsage> usages) { private DiskUsage getDiskUsage(RoutingNode node, RoutingAllocation allocation, ImmutableOpenMap<String, DiskUsage> usages) {
ClusterInfo clusterInfo = allocation.clusterInfo();
DiskUsage usage = usages.get(node.nodeId()); DiskUsage usage = usages.get(node.nodeId());
if (usage == null) { if (usage == null) {
// If there is no usage, and we have other nodes in the cluster, // If there is no usage, and we have other nodes in the cluster,

View File

@ -23,5 +23,5 @@ package org.elasticsearch.common;
* Generates opaque unique strings. * Generates opaque unique strings.
*/ */
interface UUIDGenerator { interface UUIDGenerator {
public String getBase64UUID(); String getBase64UUID();
} }

View File

@ -80,7 +80,7 @@ public enum GeoDistance implements Writeable {
@Override @Override
public FixedSourceDistance fixedSourceDistance(double sourceLatitude, double sourceLongitude, DistanceUnit unit) { public FixedSourceDistance fixedSourceDistance(double sourceLatitude, double sourceLongitude, DistanceUnit unit) {
return new FactorFixedSourceDistance(sourceLatitude, sourceLongitude, unit); return new FactorFixedSourceDistance(sourceLatitude, sourceLongitude);
} }
}, },
/** /**
@ -217,12 +217,12 @@ public enum GeoDistance implements Writeable {
throw new IllegalArgumentException("No geo distance for [" + name + "]"); throw new IllegalArgumentException("No geo distance for [" + name + "]");
} }
public static interface FixedSourceDistance { public interface FixedSourceDistance {
double calculate(double targetLatitude, double targetLongitude); double calculate(double targetLatitude, double targetLongitude);
} }
public static interface DistanceBoundingCheck { public interface DistanceBoundingCheck {
boolean isWithin(double targetLatitude, double targetLongitude); boolean isWithin(double targetLatitude, double targetLongitude);
@ -331,7 +331,7 @@ public enum GeoDistance implements Writeable {
private final double sinA; private final double sinA;
private final double cosA; private final double cosA;
public FactorFixedSourceDistance(double sourceLatitude, double sourceLongitude, DistanceUnit unit) { public FactorFixedSourceDistance(double sourceLatitude, double sourceLongitude) {
this.sourceLongitude = sourceLongitude; this.sourceLongitude = sourceLongitude;
this.a = Math.toRadians(90D - sourceLatitude); this.a = Math.toRadians(90D - sourceLatitude);
this.sinA = Math.sin(a); this.sinA = Math.sin(a);

View File

@ -19,20 +19,19 @@
package org.elasticsearch.common.geo.builders; package org.elasticsearch.common.geo.builders;
import org.locationtech.spatial4j.exception.InvalidShapeException;
import org.locationtech.spatial4j.shape.Shape;
import com.vividsolutions.jts.geom.Coordinate; import com.vividsolutions.jts.geom.Coordinate;
import com.vividsolutions.jts.geom.Geometry; import com.vividsolutions.jts.geom.Geometry;
import com.vividsolutions.jts.geom.GeometryFactory; import com.vividsolutions.jts.geom.GeometryFactory;
import com.vividsolutions.jts.geom.LinearRing; import com.vividsolutions.jts.geom.LinearRing;
import com.vividsolutions.jts.geom.MultiPolygon; import com.vividsolutions.jts.geom.MultiPolygon;
import com.vividsolutions.jts.geom.Polygon; import com.vividsolutions.jts.geom.Polygon;
import org.elasticsearch.common.collect.Tuple; import org.elasticsearch.common.collect.Tuple;
import org.elasticsearch.common.io.stream.StreamInput; import org.elasticsearch.common.io.stream.StreamInput;
import org.elasticsearch.common.io.stream.StreamOutput; import org.elasticsearch.common.io.stream.StreamOutput;
import org.elasticsearch.common.util.set.Sets; import org.elasticsearch.common.util.set.Sets;
import org.elasticsearch.common.xcontent.XContentBuilder; import org.elasticsearch.common.xcontent.XContentBuilder;
import org.locationtech.spatial4j.exception.InvalidShapeException;
import org.locationtech.spatial4j.shape.Shape;
import java.io.IOException; import java.io.IOException;
import java.util.ArrayList; import java.util.ArrayList;
@ -579,7 +578,7 @@ public class PolygonBuilder extends ShapeBuilder {
boolean direction = (component == 0 ^ orientation == Orientation.RIGHT); boolean direction = (component == 0 ^ orientation == Orientation.RIGHT);
// set the points array accordingly (shell or hole) // set the points array accordingly (shell or hole)
Coordinate[] points = (hole != null) ? hole.coordinates(false) : shell.coordinates(false); Coordinate[] points = (hole != null) ? hole.coordinates(false) : shell.coordinates(false);
ring(component, direction, orientation == Orientation.LEFT, shell, points, 0, edges, offset, points.length-1, translated); ring(component, direction, orientation == Orientation.LEFT, points, 0, edges, offset, points.length-1, translated);
return points.length-1; return points.length-1;
} }
@ -594,7 +593,7 @@ public class PolygonBuilder extends ShapeBuilder {
* number of points * number of points
* @return Array of edges * @return Array of edges
*/ */
private static Edge[] ring(int component, boolean direction, boolean handedness, LineStringBuilder shell, private static Edge[] ring(int component, boolean direction, boolean handedness,
Coordinate[] points, int offset, Edge[] edges, int toffset, int length, final AtomicBoolean translated) { Coordinate[] points, int offset, Edge[] edges, int toffset, int length, final AtomicBoolean translated) {
// calculate the direction of the points: // calculate the direction of the points:
// find the point a the top of the set and check its // find the point a the top of the set and check its

View File

@ -47,7 +47,7 @@ public interface Scope {
* when an instance of the requested object doesn't already exist in this * when an instance of the requested object doesn't already exist in this
* scope * scope
*/ */
public <T> Provider<T> scope(Key<T> key, Provider<T> unscoped); <T> Provider<T> scope(Key<T> key, Provider<T> unscoped);
/** /**
* A short but useful description of this scope. For comparison, the standard * A short but useful description of this scope. For comparison, the standard

View File

@ -36,7 +36,7 @@ import static java.util.Collections.emptySet;
*/ */
interface State { interface State {
static final State NONE = new State() { State NONE = new State() {
@Override @Override
public State parent() { public State parent() {
throw new UnsupportedOperationException(); throw new UnsupportedOperationException();

View File

@ -29,7 +29,7 @@ public interface InternalFactory<T> {
* ES: * ES:
* An factory that returns a pre created instance. * An factory that returns a pre created instance.
*/ */
public static class Instance<T> implements InternalFactory<T> { class Instance<T> implements InternalFactory<T> {
private final T object; private final T object;

View File

@ -22,7 +22,6 @@ package org.elasticsearch.common.lucene.all;
import org.apache.lucene.analysis.Analyzer; import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.analysis.TokenFilter; import org.apache.lucene.analysis.TokenFilter;
import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.TokenStream;
import org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
import org.apache.lucene.analysis.tokenattributes.PayloadAttribute; import org.apache.lucene.analysis.tokenattributes.PayloadAttribute;
import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.BytesRef;
import org.apache.lucene.util.SmallFloat; import org.apache.lucene.util.SmallFloat;
@ -38,12 +37,10 @@ public final class AllTokenStream extends TokenFilter {
} }
private final BytesRef payloadSpare = new BytesRef(new byte[1]); private final BytesRef payloadSpare = new BytesRef(new byte[1]);
private final OffsetAttribute offsetAttribute;
private final PayloadAttribute payloadAttribute; private final PayloadAttribute payloadAttribute;
AllTokenStream(TokenStream input, float boost) { AllTokenStream(TokenStream input, float boost) {
super(input); super(input);
offsetAttribute = addAttribute(OffsetAttribute.class);
payloadAttribute = addAttribute(PayloadAttribute.class); payloadAttribute = addAttribute(PayloadAttribute.class);
payloadSpare.bytes[0] = SmallFloat.floatToByte315(boost); payloadSpare.bytes[0] = SmallFloat.floatToByte315(boost);
} }

View File

@ -49,7 +49,6 @@ public final class SettingsFilter extends AbstractComponent {
public SettingsFilter(Settings settings, Collection<String> patterns) { public SettingsFilter(Settings settings, Collection<String> patterns) {
super(settings); super(settings);
HashSet<String> set = new HashSet<>();
for (String pattern : patterns) { for (String pattern : patterns) {
if (isValidPattern(pattern) == false) { if (isValidPattern(pattern) == false) {
throw new IllegalArgumentException("invalid pattern: " + pattern); throw new IllegalArgumentException("invalid pattern: " + pattern);

View File

@ -33,7 +33,7 @@ import java.util.Map;
*/ */
public interface SettingsLoader { public interface SettingsLoader {
static class Helper { class Helper {
public static Map<String, String> loadNestedFromMap(@Nullable Map map) { public static Map<String, String> loadNestedFromMap(@Nullable Map map) {
Map<String, String> settings = new HashMap<>(); Map<String, String> settings = new HashMap<>();

View File

@ -80,7 +80,7 @@ public class PortsRange {
return success; return success;
} }
public static interface PortCallback { public interface PortCallback {
boolean onPortNumber(int portNumber); boolean onPortNumber(int portNumber);
} }
} }

View File

@ -48,5 +48,5 @@ public interface TransportAddress extends Writeable {
boolean isLoopbackOrLinkLocalAddress(); boolean isLoopbackOrLinkLocalAddress();
public String toString(); String toString();
} }

View File

@ -26,6 +26,6 @@ import org.elasticsearch.common.lease.Releasable;
public interface BigArray extends Releasable, Accountable { public interface BigArray extends Releasable, Accountable {
/** Return the length of this array. */ /** Return the length of this array. */
public long size(); long size();
} }

View File

@ -27,21 +27,21 @@ public interface DoubleArray extends BigArray {
/** /**
* Get an element given its index. * Get an element given its index.
*/ */
public abstract double get(long index); double get(long index);
/** /**
* Set a value at the given index and return the previous value. * Set a value at the given index and return the previous value.
*/ */
public abstract double set(long index, double value); double set(long index, double value);
/** /**
* Increment value at the given index by <code>inc</code> and return the value. * Increment value at the given index by <code>inc</code> and return the value.
*/ */
public abstract double increment(long index, double inc); double increment(long index, double inc);
/** /**
* Fill slots between <code>fromIndex</code> inclusive to <code>toIndex</code> exclusive with <code>value</code>. * Fill slots between <code>fromIndex</code> inclusive to <code>toIndex</code> exclusive with <code>value</code>.
*/ */
public abstract void fill(long fromIndex, long toIndex, double value); void fill(long fromIndex, long toIndex, double value);
} }

View File

@ -27,21 +27,21 @@ public interface FloatArray extends BigArray {
/** /**
* Get an element given its index. * Get an element given its index.
*/ */
public abstract float get(long index); float get(long index);
/** /**
* Set a value at the given index and return the previous value. * Set a value at the given index and return the previous value.
*/ */
public abstract float set(long index, float value); float set(long index, float value);
/** /**
* Increment value at the given index by <code>inc</code> and return the value. * Increment value at the given index by <code>inc</code> and return the value.
*/ */
public abstract float increment(long index, float inc); float increment(long index, float inc);
/** /**
* Fill slots between <code>fromIndex</code> inclusive to <code>toIndex</code> exclusive with <code>value</code>. * Fill slots between <code>fromIndex</code> inclusive to <code>toIndex</code> exclusive with <code>value</code>.
*/ */
public abstract void fill(long fromIndex, long toIndex, float value); void fill(long fromIndex, long toIndex, float value);
} }

View File

@ -27,21 +27,21 @@ public interface IntArray extends BigArray {
/** /**
* Get an element given its index. * Get an element given its index.
*/ */
public abstract int get(long index); int get(long index);
/** /**
* Set a value at the given index and return the previous value. * Set a value at the given index and return the previous value.
*/ */
public abstract int set(long index, int value); int set(long index, int value);
/** /**
* Increment value at the given index by <code>inc</code> and return the value. * Increment value at the given index by <code>inc</code> and return the value.
*/ */
public abstract int increment(long index, int inc); int increment(long index, int inc);
/** /**
* Fill slots between <code>fromIndex</code> inclusive to <code>toIndex</code> exclusive with <code>value</code>. * Fill slots between <code>fromIndex</code> inclusive to <code>toIndex</code> exclusive with <code>value</code>.
*/ */
public abstract void fill(long fromIndex, long toIndex, int value); void fill(long fromIndex, long toIndex, int value);
} }

View File

@ -27,21 +27,21 @@ public interface LongArray extends BigArray {
/** /**
* Get an element given its index. * Get an element given its index.
*/ */
public abstract long get(long index); long get(long index);
/** /**
* Set a value at the given index and return the previous value. * Set a value at the given index and return the previous value.
*/ */
public abstract long set(long index, long value); long set(long index, long value);
/** /**
* Increment value at the given index by <code>inc</code> and return the value. * Increment value at the given index by <code>inc</code> and return the value.
*/ */
public abstract long increment(long index, long inc); long increment(long index, long inc);
/** /**
* Fill slots between <code>fromIndex</code> inclusive to <code>toIndex</code> exclusive with <code>value</code>. * Fill slots between <code>fromIndex</code> inclusive to <code>toIndex</code> exclusive with <code>value</code>.
*/ */
public abstract void fill(long fromIndex, long toIndex, long value); void fill(long fromIndex, long toIndex, long value);
} }

View File

@ -263,8 +263,7 @@ public final class ThreadContext implements Closeable, Writeable {
} }
private void putSingleHeader(String key, String value, Map<String, String> newHeaders) { private void putSingleHeader(String key, String value, Map<String, String> newHeaders) {
final String existingValue; if (newHeaders.putIfAbsent(key, value) != null) {
if ((existingValue = newHeaders.putIfAbsent(key, value)) != null) {
throw new IllegalArgumentException("value for key [" + key + "] already present"); throw new IllegalArgumentException("value for key [" + key + "] already present");
} }
} }

View File

@ -29,7 +29,7 @@ import java.util.Map;
*/ */
public interface ToXContent { public interface ToXContent {
public static interface Params { interface Params {
String param(String key); String param(String key);
String param(String key, String defaultValue); String param(String key, String defaultValue);
@ -39,7 +39,7 @@ public interface ToXContent {
Boolean paramAsBoolean(String key, Boolean defaultValue); Boolean paramAsBoolean(String key, Boolean defaultValue);
} }
public static final Params EMPTY_PARAMS = new Params() { Params EMPTY_PARAMS = new Params() {
@Override @Override
public String param(String key) { public String param(String key) {
return null; return null;
@ -62,7 +62,7 @@ public interface ToXContent {
}; };
public static class MapParams implements Params { class MapParams implements Params {
private final Map<String, String> params; private final Map<String, String> params;
@ -95,7 +95,7 @@ public interface ToXContent {
} }
} }
public static class DelegatingMapParams extends MapParams { class DelegatingMapParams extends MapParams {
private final Params delegate; private final Params delegate;

View File

@ -228,23 +228,13 @@ public abstract class AbstractXContentParser implements XContentParser {
return readListOrderedMap(this); return readListOrderedMap(this);
} }
static interface MapFactory { interface MapFactory {
Map<String, Object> newMap(); Map<String, Object> newMap();
} }
static final MapFactory SIMPLE_MAP_FACTORY = new MapFactory() { static final MapFactory SIMPLE_MAP_FACTORY = HashMap::new;
@Override
public Map<String, Object> newMap() {
return new HashMap<>();
}
};
static final MapFactory ORDERED_MAP_FACTORY = new MapFactory() { static final MapFactory ORDERED_MAP_FACTORY = LinkedHashMap::new;
@Override
public Map<String, Object> newMap() {
return new LinkedHashMap<>();
}
};
static Map<String, Object> readMap(XContentParser parser) throws IOException { static Map<String, Object> readMap(XContentParser parser) throws IOException {
return readMap(parser, SIMPLE_MAP_FACTORY); return readMap(parser, SIMPLE_MAP_FACTORY);

View File

@ -50,7 +50,6 @@ import java.util.HashSet;
import java.util.Queue; import java.util.Queue;
import java.util.Set; import java.util.Set;
import java.util.concurrent.ConcurrentMap; import java.util.concurrent.ConcurrentMap;
import java.util.concurrent.atomic.AtomicBoolean;
import static org.elasticsearch.cluster.ClusterState.Builder; import static org.elasticsearch.cluster.ClusterState.Builder;
@ -69,8 +68,6 @@ public class LocalDiscovery extends AbstractLifecycleComponent<Discovery> implem
private volatile boolean master = false; private volatile boolean master = false;
private final AtomicBoolean initialStateSent = new AtomicBoolean();
private static final ConcurrentMap<ClusterName, ClusterGroup> clusterGroups = ConcurrentCollections.newConcurrentMap(); private static final ConcurrentMap<ClusterName, ClusterGroup> clusterGroups = ConcurrentCollections.newConcurrentMap();
private volatile ClusterState lastProcessedClusterState; private volatile ClusterState lastProcessedClusterState;

View File

@ -197,7 +197,7 @@ public class ZenDiscovery extends AbstractLifecycleComponent<Discovery> implemen
discoverySettings, discoverySettings,
clusterService.getClusterName()); clusterService.getClusterName());
this.pingService.setPingContextProvider(this); this.pingService.setPingContextProvider(this);
this.membership = new MembershipAction(settings, clusterService, transportService, this, new MembershipListener()); this.membership = new MembershipAction(settings, transportService, this, new MembershipListener());
this.joinThreadControl = new JoinThreadControl(threadPool); this.joinThreadControl = new JoinThreadControl(threadPool);

View File

@ -21,7 +21,6 @@ package org.elasticsearch.discovery.zen.membership;
import org.elasticsearch.cluster.ClusterState; import org.elasticsearch.cluster.ClusterState;
import org.elasticsearch.cluster.node.DiscoveryNode; import org.elasticsearch.cluster.node.DiscoveryNode;
import org.elasticsearch.cluster.service.ClusterService;
import org.elasticsearch.common.component.AbstractComponent; import org.elasticsearch.common.component.AbstractComponent;
import org.elasticsearch.common.io.stream.StreamInput; import org.elasticsearch.common.io.stream.StreamInput;
import org.elasticsearch.common.io.stream.StreamOutput; import org.elasticsearch.common.io.stream.StreamOutput;
@ -48,13 +47,13 @@ public class MembershipAction extends AbstractComponent {
public static final String DISCOVERY_JOIN_VALIDATE_ACTION_NAME = "internal:discovery/zen/join/validate"; public static final String DISCOVERY_JOIN_VALIDATE_ACTION_NAME = "internal:discovery/zen/join/validate";
public static final String DISCOVERY_LEAVE_ACTION_NAME = "internal:discovery/zen/leave"; public static final String DISCOVERY_LEAVE_ACTION_NAME = "internal:discovery/zen/leave";
public static interface JoinCallback { public interface JoinCallback {
void onSuccess(); void onSuccess();
void onFailure(Throwable t); void onFailure(Throwable t);
} }
public static interface MembershipListener { public interface MembershipListener {
void onJoin(DiscoveryNode node, JoinCallback callback); void onJoin(DiscoveryNode node, JoinCallback callback);
void onLeave(DiscoveryNode node); void onLeave(DiscoveryNode node);
@ -66,14 +65,11 @@ public class MembershipAction extends AbstractComponent {
private final MembershipListener listener; private final MembershipListener listener;
private final ClusterService clusterService; public MembershipAction(Settings settings, TransportService transportService, DiscoveryNodesProvider nodesProvider, MembershipListener listener) {
public MembershipAction(Settings settings, ClusterService clusterService, TransportService transportService, DiscoveryNodesProvider nodesProvider, MembershipListener listener) {
super(settings); super(settings);
this.transportService = transportService; this.transportService = transportService;
this.nodesProvider = nodesProvider; this.nodesProvider = nodesProvider;
this.listener = listener; this.listener = listener;
this.clusterService = clusterService;
transportService.registerRequestHandler(DISCOVERY_JOIN_ACTION_NAME, JoinRequest::new, ThreadPool.Names.GENERIC, new JoinRequestRequestHandler()); transportService.registerRequestHandler(DISCOVERY_JOIN_ACTION_NAME, JoinRequest::new, ThreadPool.Names.GENERIC, new JoinRequestRequestHandler());
transportService.registerRequestHandler(DISCOVERY_JOIN_VALIDATE_ACTION_NAME, ValidateJoinRequest::new, ThreadPool.Names.GENERIC, new ValidateJoinRequestRequestHandler()); transportService.registerRequestHandler(DISCOVERY_JOIN_VALIDATE_ACTION_NAME, ValidateJoinRequest::new, ThreadPool.Names.GENERIC, new ValidateJoinRequestRequestHandler());

View File

@ -32,7 +32,6 @@ import org.elasticsearch.common.component.AbstractComponent;
import org.elasticsearch.common.settings.ClusterSettings; import org.elasticsearch.common.settings.ClusterSettings;
import org.elasticsearch.common.settings.Settings; import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.discovery.Discovery; import org.elasticsearch.discovery.Discovery;
import org.elasticsearch.env.NodeEnvironment;
import org.elasticsearch.index.Index; import org.elasticsearch.index.Index;
import org.elasticsearch.index.NodeServicesProvider; import org.elasticsearch.index.NodeServicesProvider;
import org.elasticsearch.indices.IndicesService; import org.elasticsearch.indices.IndicesService;
@ -47,8 +46,6 @@ public class Gateway extends AbstractComponent implements ClusterStateListener {
private final ClusterService clusterService; private final ClusterService clusterService;
private final NodeEnvironment nodeEnv;
private final GatewayMetaState metaState; private final GatewayMetaState metaState;
private final TransportNodesListGatewayMetaState listGatewayMetaState; private final TransportNodesListGatewayMetaState listGatewayMetaState;
@ -57,14 +54,13 @@ public class Gateway extends AbstractComponent implements ClusterStateListener {
private final IndicesService indicesService; private final IndicesService indicesService;
private final NodeServicesProvider nodeServicesProvider; private final NodeServicesProvider nodeServicesProvider;
public Gateway(Settings settings, ClusterService clusterService, NodeEnvironment nodeEnv, GatewayMetaState metaState, public Gateway(Settings settings, ClusterService clusterService, GatewayMetaState metaState,
TransportNodesListGatewayMetaState listGatewayMetaState, Discovery discovery, TransportNodesListGatewayMetaState listGatewayMetaState, Discovery discovery,
NodeServicesProvider nodeServicesProvider, IndicesService indicesService) { NodeServicesProvider nodeServicesProvider, IndicesService indicesService) {
super(settings); super(settings);
this.nodeServicesProvider = nodeServicesProvider; this.nodeServicesProvider = nodeServicesProvider;
this.indicesService = indicesService; this.indicesService = indicesService;
this.clusterService = clusterService; this.clusterService = clusterService;
this.nodeEnv = nodeEnv;
this.metaState = metaState; this.metaState = metaState;
this.listGatewayMetaState = listGatewayMetaState; this.listGatewayMetaState = listGatewayMetaState;
this.minimumMasterNodesProvider = discovery::getMinimumMasterNodes; this.minimumMasterNodesProvider = discovery::getMinimumMasterNodes;

View File

@ -42,7 +42,6 @@ import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.common.unit.TimeValue; import org.elasticsearch.common.unit.TimeValue;
import org.elasticsearch.common.util.concurrent.AbstractRunnable; import org.elasticsearch.common.util.concurrent.AbstractRunnable;
import org.elasticsearch.discovery.Discovery; import org.elasticsearch.discovery.Discovery;
import org.elasticsearch.env.NodeEnvironment;
import org.elasticsearch.index.NodeServicesProvider; import org.elasticsearch.index.NodeServicesProvider;
import org.elasticsearch.indices.IndicesService; import org.elasticsearch.indices.IndicesService;
import org.elasticsearch.rest.RestStatus; import org.elasticsearch.rest.RestStatus;
@ -96,11 +95,11 @@ public class GatewayService extends AbstractLifecycleComponent<GatewayService> i
@Inject @Inject
public GatewayService(Settings settings, AllocationService allocationService, ClusterService clusterService, public GatewayService(Settings settings, AllocationService allocationService, ClusterService clusterService,
ThreadPool threadPool, NodeEnvironment nodeEnvironment, GatewayMetaState metaState, ThreadPool threadPool, GatewayMetaState metaState,
TransportNodesListGatewayMetaState listGatewayMetaState, Discovery discovery, TransportNodesListGatewayMetaState listGatewayMetaState, Discovery discovery,
NodeServicesProvider nodeServicesProvider, IndicesService indicesService) { NodeServicesProvider nodeServicesProvider, IndicesService indicesService) {
super(settings); super(settings);
this.gateway = new Gateway(settings, clusterService, nodeEnvironment, metaState, listGatewayMetaState, discovery, this.gateway = new Gateway(settings, clusterService, metaState, listGatewayMetaState, discovery,
nodeServicesProvider, indicesService); nodeServicesProvider, indicesService);
this.allocationService = allocationService; this.allocationService = allocationService;
this.clusterService = clusterService; this.clusterService = clusterService;

View File

@ -157,7 +157,6 @@ public final class MergePolicyConfig {
MergePolicyConfig(ESLogger logger, IndexSettings indexSettings) { MergePolicyConfig(ESLogger logger, IndexSettings indexSettings) {
this.logger = logger; this.logger = logger;
IndexScopedSettings scopedSettings = indexSettings.getScopedSettings();
double forceMergeDeletesPctAllowed = indexSettings.getValue(INDEX_MERGE_POLICY_EXPUNGE_DELETES_ALLOWED_SETTING); // percentage double forceMergeDeletesPctAllowed = indexSettings.getValue(INDEX_MERGE_POLICY_EXPUNGE_DELETES_ALLOWED_SETTING); // percentage
ByteSizeValue floorSegment = indexSettings.getValue(INDEX_MERGE_POLICY_FLOOR_SEGMENT_SETTING); ByteSizeValue floorSegment = indexSettings.getValue(INDEX_MERGE_POLICY_FLOOR_SEGMENT_SETTING);
int maxMergeAtOnce = indexSettings.getValue(INDEX_MERGE_POLICY_MAX_MERGE_AT_ONCE_SETTING); int maxMergeAtOnce = indexSettings.getValue(INDEX_MERGE_POLICY_MAX_MERGE_AT_ONCE_SETTING);

View File

@ -41,7 +41,6 @@ import java.util.List;
public class MultiOrdinals extends Ordinals { public class MultiOrdinals extends Ordinals {
private static final int OFFSETS_PAGE_SIZE = 1024; private static final int OFFSETS_PAGE_SIZE = 1024;
private static final int OFFSET_INIT_PAGE_COUNT = 16;
/** /**
* Return true if this impl is going to be smaller than {@link SinglePackedOrdinals} by at least 20%. * Return true if this impl is going to be smaller than {@link SinglePackedOrdinals} by at least 20%.

View File

@ -110,7 +110,7 @@ public final class ShardGetService extends AbstractIndexShardComponent {
* <p> * <p>
* Note: Call <b>must</b> release engine searcher associated with engineGetResult! * Note: Call <b>must</b> release engine searcher associated with engineGetResult!
*/ */
public GetResult get(Engine.GetResult engineGetResult, String id, String type, String[] fields, FetchSourceContext fetchSourceContext, boolean ignoreErrorsOnGeneratedFields) { public GetResult get(Engine.GetResult engineGetResult, String id, String type, String[] fields, FetchSourceContext fetchSourceContext) {
if (!engineGetResult.exists()) { if (!engineGetResult.exists()) {
return new GetResult(shardId.getIndexName(), type, id, -1, false, null, null); return new GetResult(shardId.getIndexName(), type, id, -1, false, null, null);
} }
@ -119,7 +119,7 @@ public final class ShardGetService extends AbstractIndexShardComponent {
try { try {
long now = System.nanoTime(); long now = System.nanoTime();
fetchSourceContext = normalizeFetchSourceContent(fetchSourceContext, fields); fetchSourceContext = normalizeFetchSourceContent(fetchSourceContext, fields);
GetResult getResult = innerGetLoadFromStoredFields(type, id, fields, fetchSourceContext, engineGetResult, mapperService, ignoreErrorsOnGeneratedFields); GetResult getResult = innerGetLoadFromStoredFields(type, id, fields, fetchSourceContext, engineGetResult, mapperService);
if (getResult.isExists()) { if (getResult.isExists()) {
existsMetric.inc(System.nanoTime() - now); existsMetric.inc(System.nanoTime() - now);
} else { } else {
@ -183,7 +183,7 @@ public final class ShardGetService extends AbstractIndexShardComponent {
try { try {
// break between having loaded it from translog (so we only have _source), and having a document to load // break between having loaded it from translog (so we only have _source), and having a document to load
if (get.docIdAndVersion() != null) { if (get.docIdAndVersion() != null) {
return innerGetLoadFromStoredFields(type, id, gFields, fetchSourceContext, get, mapperService, ignoreErrorsOnGeneratedFields); return innerGetLoadFromStoredFields(type, id, gFields, fetchSourceContext, get, mapperService);
} else { } else {
Translog.Source source = get.source(); Translog.Source source = get.source();
@ -316,7 +316,7 @@ public final class ShardGetService extends AbstractIndexShardComponent {
} }
} }
private GetResult innerGetLoadFromStoredFields(String type, String id, String[] gFields, FetchSourceContext fetchSourceContext, Engine.GetResult get, MapperService mapperService, boolean ignoreErrorsOnGeneratedFields) { private GetResult innerGetLoadFromStoredFields(String type, String id, String[] gFields, FetchSourceContext fetchSourceContext, Engine.GetResult get, MapperService mapperService) {
Map<String, GetField> fields = null; Map<String, GetField> fields = null;
BytesReference source = null; BytesReference source = null;
Versions.DocIdAndVersion docIdAndVersion = get.docIdAndVersion(); Versions.DocIdAndVersion docIdAndVersion = get.docIdAndVersion();

View File

@ -76,7 +76,7 @@ final class DocumentParser {
docMapperParser, docMapper, source, parser); docMapperParser, docMapper, source, parser);
validateStart(parser); validateStart(parser);
internalParseDocument(mapping, context, parser); internalParseDocument(mapping, context, parser);
validateEnd(source, parser); validateEnd(parser);
} catch (Throwable t) { } catch (Throwable t) {
throw wrapInMapperParsingException(source, t); throw wrapInMapperParsingException(source, t);
} }
@ -128,7 +128,7 @@ final class DocumentParser {
} }
} }
private static void validateEnd(SourceToParse source, XContentParser parser) throws IOException { private static void validateEnd(XContentParser parser) throws IOException {
XContentParser.Token token;// only check for end of tokens if we created the parser here XContentParser.Token token;// only check for end of tokens if we created the parser here
// try to parse the next token, this should be null if the object is ended properly // try to parse the next token, this should be null if the object is ended properly
// but will throw a JSON exception if the extra tokens is not valid JSON (this will be handled by the catch) // but will throw a JSON exception if the extra tokens is not valid JSON (this will be handled by the catch)
@ -368,7 +368,6 @@ final class DocumentParser {
token = parser.nextToken(); token = parser.nextToken();
} }
ObjectMapper update = null;
innerParseObject(context, mapper, parser, currentFieldName, token); innerParseObject(context, mapper, parser, currentFieldName, token);
// restore the enable path flag // restore the enable path flag
if (nested.isNested()) { if (nested.isNested()) {

View File

@ -20,7 +20,6 @@
package org.elasticsearch.index.mapper; package org.elasticsearch.index.mapper;
import com.carrotsearch.hppc.ObjectHashSet; import com.carrotsearch.hppc.ObjectHashSet;
import com.carrotsearch.hppc.cursors.ObjectCursor; import com.carrotsearch.hppc.cursors.ObjectCursor;
import org.apache.lucene.analysis.Analyzer; import org.apache.lucene.analysis.Analyzer;
import org.apache.lucene.analysis.DelegatingAnalyzerWrapper; import org.apache.lucene.analysis.DelegatingAnalyzerWrapper;
@ -35,7 +34,6 @@ import org.elasticsearch.common.settings.Setting.Property;
import org.elasticsearch.common.xcontent.XContentFactory; import org.elasticsearch.common.xcontent.XContentFactory;
import org.elasticsearch.common.xcontent.XContentParser; import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.index.AbstractIndexComponent; import org.elasticsearch.index.AbstractIndexComponent;
import org.elasticsearch.index.IndexService;
import org.elasticsearch.index.IndexSettings; import org.elasticsearch.index.IndexSettings;
import org.elasticsearch.index.analysis.AnalysisService; import org.elasticsearch.index.analysis.AnalysisService;
import org.elasticsearch.index.mapper.Mapper.BuilderContext; import org.elasticsearch.index.mapper.Mapper.BuilderContext;
@ -316,7 +314,7 @@ public class MapperService extends AbstractIndexComponent {
Collections.addAll(fieldMappers, newMapper.mapping().metadataMappers); Collections.addAll(fieldMappers, newMapper.mapping().metadataMappers);
MapperUtils.collect(newMapper.mapping().root(), objectMappers, fieldMappers); MapperUtils.collect(newMapper.mapping().root(), objectMappers, fieldMappers);
checkFieldUniqueness(newMapper.type(), objectMappers, fieldMappers); checkFieldUniqueness(newMapper.type(), objectMappers, fieldMappers);
checkObjectsCompatibility(newMapper.type(), objectMappers, fieldMappers, updateAllTypes); checkObjectsCompatibility(objectMappers, updateAllTypes);
// 3. update lookup data-structures // 3. update lookup data-structures
// this will in particular make sure that the merged fields are compatible with other types // this will in particular make sure that the merged fields are compatible with other types
@ -381,7 +379,7 @@ public class MapperService extends AbstractIndexComponent {
for (DocumentMapper mapper : docMappers(false)) { for (DocumentMapper mapper : docMappers(false)) {
List<FieldMapper> fieldMappers = new ArrayList<>(); List<FieldMapper> fieldMappers = new ArrayList<>();
Collections.addAll(fieldMappers, mapper.mapping().metadataMappers); Collections.addAll(fieldMappers, mapper.mapping().metadataMappers);
MapperUtils.collect(mapper.root(), new ArrayList<ObjectMapper>(), fieldMappers); MapperUtils.collect(mapper.root(), new ArrayList<>(), fieldMappers);
for (FieldMapper fieldMapper : fieldMappers) { for (FieldMapper fieldMapper : fieldMappers) {
assert fieldMapper.fieldType() == fieldTypes.get(fieldMapper.name()) : fieldMapper.name(); assert fieldMapper.fieldType() == fieldTypes.get(fieldMapper.name()) : fieldMapper.name();
} }
@ -449,7 +447,7 @@ public class MapperService extends AbstractIndexComponent {
} }
} }
private void checkObjectsCompatibility(String type, Collection<ObjectMapper> objectMappers, Collection<FieldMapper> fieldMappers, boolean updateAllTypes) { private void checkObjectsCompatibility(Collection<ObjectMapper> objectMappers, boolean updateAllTypes) {
assert Thread.holdsLock(this); assert Thread.holdsLock(this);
for (ObjectMapper newObjectMapper : objectMappers) { for (ObjectMapper newObjectMapper : objectMappers) {

View File

@ -310,8 +310,6 @@ public abstract class ParseContext {
private Field version; private Field version;
private StringBuilder stringBuilder = new StringBuilder();
private final AllEntries allEntries; private final AllEntries allEntries;
private final List<Mapper> dynamicMappers; private final List<Mapper> dynamicMappers;

View File

@ -24,7 +24,6 @@ import org.apache.lucene.document.NumericDocValuesField;
import org.apache.lucene.index.IndexOptions; import org.apache.lucene.index.IndexOptions;
import org.elasticsearch.Version; import org.elasticsearch.Version;
import org.elasticsearch.action.TimestampParsingException; import org.elasticsearch.action.TimestampParsingException;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.joda.FormatDateTimeFormatter; import org.elasticsearch.common.joda.FormatDateTimeFormatter;
import org.elasticsearch.common.joda.Joda; import org.elasticsearch.common.joda.Joda;
import org.elasticsearch.common.lucene.Lucene; import org.elasticsearch.common.lucene.Lucene;
@ -82,7 +81,7 @@ public class TimestampFieldMapper extends MetadataFieldMapper {
private String defaultTimestamp = Defaults.DEFAULT_TIMESTAMP; private String defaultTimestamp = Defaults.DEFAULT_TIMESTAMP;
private Boolean ignoreMissing = null; private Boolean ignoreMissing = null;
public Builder(MappedFieldType existing, Settings settings) { public Builder(MappedFieldType existing) {
super(Defaults.NAME, existing == null ? Defaults.FIELD_TYPE : existing, Defaults.FIELD_TYPE); super(Defaults.NAME, existing == null ? Defaults.FIELD_TYPE : existing, Defaults.FIELD_TYPE);
} }
@ -130,7 +129,7 @@ public class TimestampFieldMapper extends MetadataFieldMapper {
if (parserContext.indexVersionCreated().onOrAfter(Version.V_5_0_0_alpha4)) { if (parserContext.indexVersionCreated().onOrAfter(Version.V_5_0_0_alpha4)) {
throw new IllegalArgumentException("[_timestamp] is removed in 5.0. As a replacement, you can use an ingest pipeline to add a field with the current timestamp to your documents."); throw new IllegalArgumentException("[_timestamp] is removed in 5.0. As a replacement, you can use an ingest pipeline to add a field with the current timestamp to your documents.");
} }
Builder builder = new Builder(parserContext.mapperService().fullName(NAME), parserContext.mapperService().getIndexSettings().getSettings()); Builder builder = new Builder(parserContext.mapperService().fullName(NAME));
boolean defaultSet = false; boolean defaultSet = false;
Boolean ignoreMissing = null; Boolean ignoreMissing = null;
for (Iterator<Map.Entry<String, Object>> iterator = node.entrySet().iterator(); iterator.hasNext();) { for (Iterator<Map.Entry<String, Object>> iterator = node.entrySet().iterator(); iterator.hasNext();) {

View File

@ -405,7 +405,6 @@ public class CommonTermsQueryBuilder extends AbstractQueryBuilder<CommonTermsQue
private static Query parseQueryString(ExtendedCommonTermsQuery query, Object queryString, String field, Analyzer analyzer, private static Query parseQueryString(ExtendedCommonTermsQuery query, Object queryString, String field, Analyzer analyzer,
String lowFreqMinimumShouldMatch, String highFreqMinimumShouldMatch) throws IOException { String lowFreqMinimumShouldMatch, String highFreqMinimumShouldMatch) throws IOException {
// Logic similar to QueryParser#getFieldQuery // Logic similar to QueryParser#getFieldQuery
int count = 0;
try (TokenStream source = analyzer.tokenStream(field, queryString.toString())) { try (TokenStream source = analyzer.tokenStream(field, queryString.toString())) {
source.reset(); source.reset();
CharTermAttribute termAtt = source.addAttribute(CharTermAttribute.class); CharTermAttribute termAtt = source.addAttribute(CharTermAttribute.class);
@ -414,7 +413,6 @@ public class CommonTermsQueryBuilder extends AbstractQueryBuilder<CommonTermsQue
// UTF-8 // UTF-8
builder.copyChars(termAtt); builder.copyChars(termAtt);
query.add(new Term(field, builder.toBytesRef())); query.add(new Term(field, builder.toBytesRef()));
count++;
} }
} }

View File

@ -43,8 +43,6 @@ public class SnapshotStatus {
private Translog translog = new Translog(); private Translog translog = new Translog();
private Throwable failure;
public Stage stage() { public Stage stage() {
return this.stage; return this.stage;
} }
@ -70,10 +68,6 @@ public class SnapshotStatus {
this.time = time; this.time = time;
} }
public void failed(Throwable failure) {
this.failure = failure;
}
public Index index() { public Index index() {
return index; return index;
} }

View File

@ -202,7 +202,7 @@ public class TermVectorsService {
/* generate term vectors from fetched document fields */ /* generate term vectors from fetched document fields */
GetResult getResult = indexShard.getService().get( GetResult getResult = indexShard.getService().get(
get, request.id(), request.type(), validFields.toArray(Strings.EMPTY_ARRAY), null, false); get, request.id(), request.type(), validFields.toArray(Strings.EMPTY_ARRAY), null);
Fields generatedTermVectors = generateTermVectors(indexShard, getResult.getFields().values(), request.offsets(), request.perFieldAnalyzer(), validFields); Fields generatedTermVectors = generateTermVectors(indexShard, getResult.getFields().values(), request.offsets(), request.perFieldAnalyzer(), validFields);
/* merge with existing Fields */ /* merge with existing Fields */

View File

@ -572,10 +572,6 @@ public class Translog extends AbstractIndexShardComponent implements IndexShardC
} }
} }
private boolean isReferencedGeneration(long generation) { // used to make decisions if a file can be deleted
return generation >= lastCommittedTranslogFileGeneration;
}
public TranslogConfig getConfig() { public TranslogConfig getConfig() {
return config; return config;
} }

View File

@ -20,8 +20,8 @@
package org.elasticsearch.indices; package org.elasticsearch.indices;
import org.elasticsearch.common.component.AbstractComponent; import org.elasticsearch.common.component.AbstractComponent;
import org.elasticsearch.common.settings.Setting.Property;
import org.elasticsearch.common.settings.Setting; import org.elasticsearch.common.settings.Setting;
import org.elasticsearch.common.settings.Setting.Property;
import org.elasticsearch.common.settings.Settings; import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.common.unit.ByteSizeUnit; import org.elasticsearch.common.unit.ByteSizeUnit;
import org.elasticsearch.common.unit.ByteSizeValue; import org.elasticsearch.common.unit.ByteSizeValue;
@ -34,8 +34,6 @@ import org.elasticsearch.index.engine.FlushNotAllowedEngineException;
import org.elasticsearch.index.shard.IndexShard; import org.elasticsearch.index.shard.IndexShard;
import org.elasticsearch.index.shard.IndexShardState; import org.elasticsearch.index.shard.IndexShardState;
import org.elasticsearch.index.shard.IndexingOperationListener; import org.elasticsearch.index.shard.IndexingOperationListener;
import org.elasticsearch.index.translog.Translog;
import org.elasticsearch.monitor.jvm.JvmInfo;
import org.elasticsearch.threadpool.ThreadPool; import org.elasticsearch.threadpool.ThreadPool;
import java.io.Closeable; import java.io.Closeable;
@ -94,10 +92,6 @@ public class IndexingMemoryController extends AbstractComponent implements Index
private final ShardsIndicesStatusChecker statusChecker; private final ShardsIndicesStatusChecker statusChecker;
IndexingMemoryController(Settings settings, ThreadPool threadPool, Iterable<IndexShard> indexServices) { IndexingMemoryController(Settings settings, ThreadPool threadPool, Iterable<IndexShard> indexServices) {
this(settings, threadPool, indexServices, JvmInfo.jvmInfo().getMem().getHeapMax().bytes());
}
IndexingMemoryController(Settings settings, ThreadPool threadPool, Iterable<IndexShard> indexServices, long jvmMemoryInBytes) {
super(settings); super(settings);
this.indexShards = indexServices; this.indexShards = indexServices;

View File

@ -20,13 +20,9 @@
package org.elasticsearch.monitor.jvm; package org.elasticsearch.monitor.jvm;
import org.apache.lucene.util.CollectionUtil; import org.apache.lucene.util.CollectionUtil;
import org.apache.lucene.util.Constants;
import org.elasticsearch.ElasticsearchException; import org.elasticsearch.ElasticsearchException;
import org.elasticsearch.ExceptionsHelper;
import org.elasticsearch.common.joda.FormatDateTimeFormatter; import org.elasticsearch.common.joda.FormatDateTimeFormatter;
import org.elasticsearch.common.joda.Joda; import org.elasticsearch.common.joda.Joda;
import org.elasticsearch.common.logging.ESLogger;
import org.elasticsearch.common.logging.Loggers;
import org.elasticsearch.common.unit.TimeValue; import org.elasticsearch.common.unit.TimeValue;
import java.lang.management.ManagementFactory; import java.lang.management.ManagementFactory;
@ -45,7 +41,6 @@ import java.util.concurrent.TimeUnit;
public class HotThreads { public class HotThreads {
private static final Object mutex = new Object(); private static final Object mutex = new Object();
private static final ESLogger logger = Loggers.getLogger(HotThreads.class);
private static final FormatDateTimeFormatter DATE_TIME_FORMATTER = Joda.forPattern("dateOptionalTime"); private static final FormatDateTimeFormatter DATE_TIME_FORMATTER = Joda.forPattern("dateOptionalTime");

View File

@ -409,7 +409,7 @@ public class RepositoriesService extends AbstractComponent implements ClusterSta
Repository repository = repositoryInjector.getInstance(Repository.class); Repository repository = repositoryInjector.getInstance(Repository.class);
IndexShardRepository indexShardRepository = repositoryInjector.getInstance(IndexShardRepository.class); IndexShardRepository indexShardRepository = repositoryInjector.getInstance(IndexShardRepository.class);
repository.start(); repository.start();
return new RepositoryHolder(repositoryMetaData.type(), repositoryMetaData.settings(), repositoryInjector, repository, indexShardRepository); return new RepositoryHolder(repositoryMetaData.type(), repositoryMetaData.settings(), repository, indexShardRepository);
} catch (Throwable t) { } catch (Throwable t) {
logger.warn("failed to create repository [{}][{}]", t, repositoryMetaData.type(), repositoryMetaData.name()); logger.warn("failed to create repository [{}][{}]", t, repositoryMetaData.type(), repositoryMetaData.name());
throw new RepositoryException(repositoryMetaData.name(), "failed to create repository", t); throw new RepositoryException(repositoryMetaData.name(), "failed to create repository", t);
@ -473,7 +473,7 @@ public class RepositoriesService extends AbstractComponent implements ClusterSta
private final Repository repository; private final Repository repository;
private final IndexShardRepository indexShardRepository; private final IndexShardRepository indexShardRepository;
public RepositoryHolder(String type, Settings settings, Injector injector, Repository repository, IndexShardRepository indexShardRepository) { public RepositoryHolder(String type, Settings settings,Repository repository, IndexShardRepository indexShardRepository) {
this.type = type; this.type = type;
this.settings = settings; this.settings = settings;
this.repository = repository; this.repository = repository;

View File

@ -26,7 +26,6 @@ import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.rest.BaseRestHandler; import org.elasticsearch.rest.BaseRestHandler;
import org.elasticsearch.rest.BytesRestResponse; import org.elasticsearch.rest.BytesRestResponse;
import org.elasticsearch.rest.RestChannel; import org.elasticsearch.rest.RestChannel;
import org.elasticsearch.rest.RestController;
import org.elasticsearch.rest.RestRequest; import org.elasticsearch.rest.RestRequest;
import org.elasticsearch.rest.RestStatus; import org.elasticsearch.rest.RestStatus;
@ -38,7 +37,7 @@ import static org.elasticsearch.rest.action.support.RestTable.pad;
*/ */
public abstract class AbstractCatAction extends BaseRestHandler { public abstract class AbstractCatAction extends BaseRestHandler {
public AbstractCatAction(Settings settings, RestController controller) { public AbstractCatAction(Settings settings) {
super(settings); super(settings);
} }

View File

@ -45,7 +45,7 @@ public class RestAliasAction extends AbstractCatAction {
@Inject @Inject
public RestAliasAction(Settings settings, RestController controller) { public RestAliasAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/aliases", this); controller.registerHandler(GET, "/_cat/aliases", this);
controller.registerHandler(GET, "/_cat/aliases/{alias}", this); controller.registerHandler(GET, "/_cat/aliases/{alias}", this);
} }

View File

@ -49,7 +49,7 @@ public class RestAllocationAction extends AbstractCatAction {
@Inject @Inject
public RestAllocationAction(Settings settings, RestController controller) { public RestAllocationAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/allocation", this); controller.registerHandler(GET, "/_cat/allocation", this);
controller.registerHandler(GET, "/_cat/allocation/{nodes}", this); controller.registerHandler(GET, "/_cat/allocation/{nodes}", this);
} }

View File

@ -46,7 +46,7 @@ public class RestCountAction extends AbstractCatAction {
@Inject @Inject
public RestCountAction(Settings settings, RestController restController, RestController controller, IndicesQueriesRegistry indicesQueriesRegistry) { public RestCountAction(Settings settings, RestController restController, RestController controller, IndicesQueriesRegistry indicesQueriesRegistry) {
super(settings, controller); super(settings);
restController.registerHandler(GET, "/_cat/count", this); restController.registerHandler(GET, "/_cat/count", this);
restController.registerHandler(GET, "/_cat/count/{index}", this); restController.registerHandler(GET, "/_cat/count/{index}", this);
this.indicesQueriesRegistry = indicesQueriesRegistry; this.indicesQueriesRegistry = indicesQueriesRegistry;

View File

@ -44,7 +44,7 @@ public class RestFielddataAction extends AbstractCatAction {
@Inject @Inject
public RestFielddataAction(Settings settings, RestController controller) { public RestFielddataAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/fielddata", this); controller.registerHandler(GET, "/_cat/fielddata", this);
controller.registerHandler(GET, "/_cat/fielddata/{fields}", this); controller.registerHandler(GET, "/_cat/fielddata/{fields}", this);
} }

View File

@ -40,7 +40,7 @@ public class RestHealthAction extends AbstractCatAction {
@Inject @Inject
public RestHealthAction(Settings settings, RestController controller) { public RestHealthAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/health", this); controller.registerHandler(GET, "/_cat/health", this);
} }

View File

@ -58,7 +58,7 @@ public class RestIndicesAction extends AbstractCatAction {
@Inject @Inject
public RestIndicesAction(Settings settings, RestController controller, IndexNameExpressionResolver indexNameExpressionResolver) { public RestIndicesAction(Settings settings, RestController controller, IndexNameExpressionResolver indexNameExpressionResolver) {
super(settings, controller); super(settings);
this.indexNameExpressionResolver = indexNameExpressionResolver; this.indexNameExpressionResolver = indexNameExpressionResolver;
controller.registerHandler(GET, "/_cat/indices", this); controller.registerHandler(GET, "/_cat/indices", this);
controller.registerHandler(GET, "/_cat/indices/{index}", this); controller.registerHandler(GET, "/_cat/indices/{index}", this);

View File

@ -40,7 +40,7 @@ public class RestMasterAction extends AbstractCatAction {
@Inject @Inject
public RestMasterAction(Settings settings, RestController controller) { public RestMasterAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/master", this); controller.registerHandler(GET, "/_cat/master", this);
} }

View File

@ -48,7 +48,7 @@ public class RestNodeAttrsAction extends AbstractCatAction {
@Inject @Inject
public RestNodeAttrsAction(Settings settings, RestController controller) { public RestNodeAttrsAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/nodeattrs", this); controller.registerHandler(GET, "/_cat/nodeattrs", this);
} }

View File

@ -73,7 +73,7 @@ public class RestNodesAction extends AbstractCatAction {
@Inject @Inject
public RestNodesAction(Settings settings, RestController controller) { public RestNodesAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/nodes", this); controller.registerHandler(GET, "/_cat/nodes", this);
} }

View File

@ -38,7 +38,7 @@ import static org.elasticsearch.rest.RestRequest.Method.GET;
public class RestPendingClusterTasksAction extends AbstractCatAction { public class RestPendingClusterTasksAction extends AbstractCatAction {
@Inject @Inject
public RestPendingClusterTasksAction(Settings settings, RestController controller) { public RestPendingClusterTasksAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/pending_tasks", this); controller.registerHandler(GET, "/_cat/pending_tasks", this);
} }

View File

@ -45,7 +45,7 @@ public class RestPluginsAction extends AbstractCatAction {
@Inject @Inject
public RestPluginsAction(Settings settings, RestController controller) { public RestPluginsAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/plugins", this); controller.registerHandler(GET, "/_cat/plugins", this);
} }

View File

@ -52,7 +52,7 @@ public class RestRecoveryAction extends AbstractCatAction {
@Inject @Inject
public RestRecoveryAction(Settings settings, RestController restController, RestController controller) { public RestRecoveryAction(Settings settings, RestController restController, RestController controller) {
super(settings, controller); super(settings);
restController.registerHandler(GET, "/_cat/recovery", this); restController.registerHandler(GET, "/_cat/recovery", this);
restController.registerHandler(GET, "/_cat/recovery/{index}", this); restController.registerHandler(GET, "/_cat/recovery/{index}", this);
} }

View File

@ -41,7 +41,7 @@ import static org.elasticsearch.rest.RestRequest.Method.GET;
public class RestRepositoriesAction extends AbstractCatAction { public class RestRepositoriesAction extends AbstractCatAction {
@Inject @Inject
public RestRepositoriesAction(Settings settings, RestController controller) { public RestRepositoriesAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/repositories", this); controller.registerHandler(GET, "/_cat/repositories", this);
} }

View File

@ -50,7 +50,7 @@ public class RestSegmentsAction extends AbstractCatAction {
@Inject @Inject
public RestSegmentsAction(Settings settings, RestController controller) { public RestSegmentsAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/segments", this); controller.registerHandler(GET, "/_cat/segments", this);
controller.registerHandler(GET, "/_cat/segments/{index}", this); controller.registerHandler(GET, "/_cat/segments/{index}", this);
} }

View File

@ -50,7 +50,7 @@ public class RestShardsAction extends AbstractCatAction {
@Inject @Inject
public RestShardsAction(Settings settings, RestController controller) { public RestShardsAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/shards", this); controller.registerHandler(GET, "/_cat/shards", this);
controller.registerHandler(GET, "/_cat/shards/{index}", this); controller.registerHandler(GET, "/_cat/shards/{index}", this);
} }

View File

@ -48,7 +48,7 @@ import static org.elasticsearch.rest.RestRequest.Method.GET;
public class RestSnapshotAction extends AbstractCatAction { public class RestSnapshotAction extends AbstractCatAction {
@Inject @Inject
public RestSnapshotAction(Settings settings, RestController controller) { public RestSnapshotAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/snapshots/{repository}", this); controller.registerHandler(GET, "/_cat/snapshots/{repository}", this);
} }

View File

@ -52,7 +52,7 @@ public class RestTasksAction extends AbstractCatAction {
@Inject @Inject
public RestTasksAction(Settings settings, RestController controller, ClusterService clusterService) { public RestTasksAction(Settings settings, RestController controller, ClusterService clusterService) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/tasks", this); controller.registerHandler(GET, "/_cat/tasks", this);
this.clusterService = clusterService; this.clusterService = clusterService;
} }

View File

@ -109,7 +109,7 @@ public class RestThreadPoolAction extends AbstractCatAction {
@Inject @Inject
public RestThreadPoolAction(Settings settings, RestController controller) { public RestThreadPoolAction(Settings settings, RestController controller) {
super(settings, controller); super(settings);
controller.registerHandler(GET, "/_cat/thread_pool", this); controller.registerHandler(GET, "/_cat/thread_pool", this);
} }

View File

@ -51,13 +51,13 @@ public class RestIndexAction extends BaseRestHandler {
controller.registerHandler(POST, "/{index}/{type}", this); // auto id creation controller.registerHandler(POST, "/{index}/{type}", this); // auto id creation
controller.registerHandler(PUT, "/{index}/{type}/{id}", this); controller.registerHandler(PUT, "/{index}/{type}/{id}", this);
controller.registerHandler(POST, "/{index}/{type}/{id}", this); controller.registerHandler(POST, "/{index}/{type}/{id}", this);
CreateHandler createHandler = new CreateHandler(settings, controller); CreateHandler createHandler = new CreateHandler(settings);
controller.registerHandler(PUT, "/{index}/{type}/{id}/_create", createHandler); controller.registerHandler(PUT, "/{index}/{type}/{id}/_create", createHandler);
controller.registerHandler(POST, "/{index}/{type}/{id}/_create", createHandler); controller.registerHandler(POST, "/{index}/{type}/{id}/_create", createHandler);
} }
final class CreateHandler extends BaseRestHandler { final class CreateHandler extends BaseRestHandler {
protected CreateHandler(Settings settings, RestController controller) { protected CreateHandler(Settings settings) {
super(settings); super(settings);
} }

View File

@ -236,7 +236,7 @@ public class ScriptService extends AbstractComponent implements Closeable, Clust
} }
ScriptEngineService scriptEngineService = getScriptEngineServiceForLang(lang); ScriptEngineService scriptEngineService = getScriptEngineServiceForLang(lang);
if (canExecuteScript(lang, scriptEngineService, script.getType(), scriptContext) == false) { if (canExecuteScript(lang, script.getType(), scriptContext) == false) {
throw new IllegalStateException("scripts of type [" + script.getType() + "], operation [" + scriptContext.getKey() + "] and lang [" + lang + "] are disabled"); throw new IllegalStateException("scripts of type [" + script.getType() + "], operation [" + scriptContext.getKey() + "] and lang [" + lang + "] are disabled");
} }
@ -357,7 +357,7 @@ public class ScriptService extends AbstractComponent implements Closeable, Clust
ScriptEngineService scriptEngineService = getScriptEngineServiceForLang(scriptLang); ScriptEngineService scriptEngineService = getScriptEngineServiceForLang(scriptLang);
//we don't know yet what the script will be used for, but if all of the operations for this lang with //we don't know yet what the script will be used for, but if all of the operations for this lang with
//indexed scripts are disabled, it makes no sense to even compile it. //indexed scripts are disabled, it makes no sense to even compile it.
if (isAnyScriptContextEnabled(scriptLang, scriptEngineService, ScriptType.STORED)) { if (isAnyScriptContextEnabled(scriptLang, ScriptType.STORED)) {
Object compiled = scriptEngineService.compile(id, template.getScript(), Collections.emptyMap()); Object compiled = scriptEngineService.compile(id, template.getScript(), Collections.emptyMap());
if (compiled == null) { if (compiled == null) {
throw new IllegalArgumentException("Unable to parse [" + template.getScript() + throw new IllegalArgumentException("Unable to parse [" + template.getScript() +
@ -466,16 +466,16 @@ public class ScriptService extends AbstractComponent implements Closeable, Clust
return getScriptEngineServiceForLang(compiledScript.lang()).search(compiledScript, lookup, script.getParams()); return getScriptEngineServiceForLang(compiledScript.lang()).search(compiledScript, lookup, script.getParams());
} }
private boolean isAnyScriptContextEnabled(String lang, ScriptEngineService scriptEngineService, ScriptType scriptType) { private boolean isAnyScriptContextEnabled(String lang, ScriptType scriptType) {
for (ScriptContext scriptContext : scriptContextRegistry.scriptContexts()) { for (ScriptContext scriptContext : scriptContextRegistry.scriptContexts()) {
if (canExecuteScript(lang, scriptEngineService, scriptType, scriptContext)) { if (canExecuteScript(lang, scriptType, scriptContext)) {
return true; return true;
} }
} }
return false; return false;
} }
private boolean canExecuteScript(String lang, ScriptEngineService scriptEngineService, ScriptType scriptType, ScriptContext scriptContext) { private boolean canExecuteScript(String lang, ScriptType scriptType, ScriptContext scriptContext) {
assert lang != null; assert lang != null;
if (scriptContextRegistry.isSupportedContext(scriptContext) == false) { if (scriptContextRegistry.isSupportedContext(scriptContext) == false) {
throw new IllegalArgumentException("script context [" + scriptContext.getKey() + "] not supported"); throw new IllegalArgumentException("script context [" + scriptContext.getKey() + "] not supported");
@ -556,7 +556,7 @@ public class ScriptService extends AbstractComponent implements Closeable, Clust
try { try {
//we don't know yet what the script will be used for, but if all of the operations for this lang //we don't know yet what the script will be used for, but if all of the operations for this lang
// with file scripts are disabled, it makes no sense to even compile it and cache it. // with file scripts are disabled, it makes no sense to even compile it and cache it.
if (isAnyScriptContextEnabled(engineService.getType(), engineService, ScriptType.FILE)) { if (isAnyScriptContextEnabled(engineService.getType(), ScriptType.FILE)) {
logger.info("compiling script file [{}]", file.toAbsolutePath()); logger.info("compiling script file [{}]", file.toAbsolutePath());
try (InputStreamReader reader = new InputStreamReader(Files.newInputStream(file), StandardCharsets.UTF_8)) { try (InputStreamReader reader = new InputStreamReader(Files.newInputStream(file), StandardCharsets.UTF_8)) {
String script = Streams.copyToString(reader); String script = Streams.copyToString(reader);

View File

@ -19,19 +19,7 @@
package org.elasticsearch.search; package org.elasticsearch.search;
import static java.util.Collections.unmodifiableMap; import com.carrotsearch.hppc.ObjectFloatHashMap;
import static org.elasticsearch.common.unit.TimeValue.timeValueMillis;
import static org.elasticsearch.common.unit.TimeValue.timeValueMinutes;
import java.io.IOException;
import java.util.Collections;
import java.util.HashMap;
import java.util.Map;
import java.util.Optional;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.ScheduledFuture;
import java.util.concurrent.atomic.AtomicLong;
import org.apache.lucene.search.FieldDoc; import org.apache.lucene.search.FieldDoc;
import org.apache.lucene.search.TopDocs; import org.apache.lucene.search.TopDocs;
import org.elasticsearch.ExceptionsHelper; import org.elasticsearch.ExceptionsHelper;
@ -57,8 +45,8 @@ import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.index.Index; import org.elasticsearch.index.Index;
import org.elasticsearch.index.IndexService; import org.elasticsearch.index.IndexService;
import org.elasticsearch.index.engine.Engine; import org.elasticsearch.index.engine.Engine;
import org.elasticsearch.index.query.QueryShardContext;
import org.elasticsearch.index.query.InnerHitBuilder; import org.elasticsearch.index.query.InnerHitBuilder;
import org.elasticsearch.index.query.QueryShardContext;
import org.elasticsearch.index.shard.IndexEventListener; import org.elasticsearch.index.shard.IndexEventListener;
import org.elasticsearch.index.shard.IndexShard; import org.elasticsearch.index.shard.IndexShard;
import org.elasticsearch.index.shard.SearchOperationListener; import org.elasticsearch.index.shard.SearchOperationListener;
@ -101,7 +89,18 @@ import org.elasticsearch.search.sort.SortAndFormats;
import org.elasticsearch.search.sort.SortBuilder; import org.elasticsearch.search.sort.SortBuilder;
import org.elasticsearch.threadpool.ThreadPool; import org.elasticsearch.threadpool.ThreadPool;
import com.carrotsearch.hppc.ObjectFloatHashMap; import java.io.IOException;
import java.util.Collections;
import java.util.HashMap;
import java.util.Map;
import java.util.Optional;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.ScheduledFuture;
import java.util.concurrent.atomic.AtomicLong;
import static java.util.Collections.unmodifiableMap;
import static org.elasticsearch.common.unit.TimeValue.timeValueMillis;
import static org.elasticsearch.common.unit.TimeValue.timeValueMinutes;
/** /**
* *
@ -844,20 +843,6 @@ public class SearchService extends AbstractLifecycleComponent<SearchService> imp
} }
} }
private void shortcutDocIdsToLoadForScanning(SearchContext context) {
TopDocs topDocs = context.queryResult().topDocs();
if (topDocs.scoreDocs.length == 0) {
// no more docs...
context.docIdsToLoad(EMPTY_DOC_IDS, 0, 0);
return;
}
int[] docIdsToLoad = new int[topDocs.scoreDocs.length];
for (int i = 0; i < docIdsToLoad.length; i++) {
docIdsToLoad[i] = topDocs.scoreDocs[i].doc;
}
context.docIdsToLoad(docIdsToLoad, 0, docIdsToLoad.length);
}
private void processScroll(InternalScrollSearchRequest request, SearchContext context) { private void processScroll(InternalScrollSearchRequest request, SearchContext context) {
// process scroll // process scroll
context.from(context.from() + context.size()); context.from(context.from() + context.size());

View File

@ -82,6 +82,7 @@ public class BestBucketsDeferringCollector extends DeferringBucketCollector {
} }
/** Set the deferred collectors. */ /** Set the deferred collectors. */
@Override
public void setDeferredCollector(Iterable<BucketCollector> deferredCollectors) { public void setDeferredCollector(Iterable<BucketCollector> deferredCollectors) {
this.collector = BucketCollector.wrap(deferredCollectors); this.collector = BucketCollector.wrap(deferredCollectors);
} }

View File

@ -79,6 +79,7 @@ public class BestDocsDeferringCollector extends DeferringBucketCollector impleme
} }
/** Set the deferred collectors. */ /** Set the deferred collectors. */
@Override
public void setDeferredCollector(Iterable<BucketCollector> deferredCollectors) { public void setDeferredCollector(Iterable<BucketCollector> deferredCollectors) {
this.deferred = BucketCollector.wrap(deferredCollectors); this.deferred = BucketCollector.wrap(deferredCollectors);
} }

View File

@ -34,18 +34,13 @@ import java.io.IOException;
*/ */
public abstract class DeferringBucketCollector extends BucketCollector { public abstract class DeferringBucketCollector extends BucketCollector {
private BucketCollector collector;
/** Sole constructor. */ /** Sole constructor. */
public DeferringBucketCollector() {} public DeferringBucketCollector() {}
/** Set the deferred collectors. */ /** Set the deferred collectors. */
public void setDeferredCollector(Iterable<BucketCollector> deferredCollectors) { public abstract void setDeferredCollector(Iterable<BucketCollector> deferredCollectors);
this.collector = BucketCollector.wrap(deferredCollectors);
}
public final void replay(long... selectedBuckets) throws IOException {
public final void replay(long... selectedBuckets) throws IOException
{
prepareSelectedBuckets(selectedBuckets); prepareSelectedBuckets(selectedBuckets);
} }

View File

@ -260,7 +260,6 @@ public class InternalRange<B extends InternalRange.Bucket, R extends InternalRan
} }
private List<B> ranges; private List<B> ranges;
private Map<String, B> rangeMap;
protected DocValueFormat format; protected DocValueFormat format;
protected boolean keyed; protected boolean keyed;
@ -333,7 +332,6 @@ public class InternalRange<B extends InternalRange.Bucket, R extends InternalRan
ranges.add(getFactory().createBucket(key, in.readDouble(), in.readDouble(), in.readVLong(), InternalAggregations.readAggregations(in), keyed, format)); ranges.add(getFactory().createBucket(key, in.readDouble(), in.readDouble(), in.readVLong(), InternalAggregations.readAggregations(in), keyed, format));
} }
this.ranges = ranges; this.ranges = ranges;
this.rangeMap = null;
} }
@Override @Override

View File

@ -20,7 +20,6 @@
package org.elasticsearch.search.aggregations.metrics.scripted; package org.elasticsearch.search.aggregations.metrics.scripted;
import org.apache.lucene.index.LeafReaderContext; import org.apache.lucene.index.LeafReaderContext;
import org.elasticsearch.cluster.ClusterState;
import org.elasticsearch.script.ExecutableScript; import org.elasticsearch.script.ExecutableScript;
import org.elasticsearch.script.LeafSearchScript; import org.elasticsearch.script.LeafSearchScript;
import org.elasticsearch.script.Script; import org.elasticsearch.script.Script;
@ -34,6 +33,7 @@ import org.elasticsearch.search.aggregations.LeafBucketCollectorBase;
import org.elasticsearch.search.aggregations.metrics.MetricsAggregator; import org.elasticsearch.search.aggregations.metrics.MetricsAggregator;
import org.elasticsearch.search.aggregations.pipeline.PipelineAggregator; import org.elasticsearch.search.aggregations.pipeline.PipelineAggregator;
import org.elasticsearch.search.aggregations.support.AggregationContext; import org.elasticsearch.search.aggregations.support.AggregationContext;
import java.io.IOException; import java.io.IOException;
import java.util.Collections; import java.util.Collections;
import java.util.List; import java.util.List;
@ -52,7 +52,6 @@ public class ScriptedMetricAggregator extends MetricsAggregator {
super(name, context, parent, pipelineAggregators, metaData); super(name, context, parent, pipelineAggregators, metaData);
this.params = params; this.params = params;
ScriptService scriptService = context.searchContext().scriptService(); ScriptService scriptService = context.searchContext().scriptService();
ClusterState state = context.searchContext().getQueryShardContext().getClusterState();
if (initScript != null) { if (initScript != null) {
scriptService.executable(initScript, ScriptContext.Standard.AGGS, Collections.emptyMap()).run(); scriptService.executable(initScript, ScriptContext.Standard.AGGS, Collections.emptyMap()).run();
} }

View File

@ -32,7 +32,6 @@ public class ValuesSourceConfig<VS extends ValuesSource> {
private SearchScript script; private SearchScript script;
private ValueType scriptValueType; private ValueType scriptValueType;
private boolean unmapped = false; private boolean unmapped = false;
private String formatPattern;
private DocValueFormat format = DocValueFormat.RAW; private DocValueFormat format = DocValueFormat.RAW;
private Object missing; private Object missing;
private DateTimeZone timeZone; private DateTimeZone timeZone;

View File

@ -110,7 +110,7 @@ public class FastVectorHighlighter implements Highlighter {
if (!forceSource && mapper.fieldType().stored()) { if (!forceSource && mapper.fieldType().stored()) {
fragmentsBuilder = new SimpleFragmentsBuilder(mapper, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner); fragmentsBuilder = new SimpleFragmentsBuilder(mapper, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner);
} else { } else {
fragmentsBuilder = new SourceSimpleFragmentsBuilder(mapper, context, hitContext, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner); fragmentsBuilder = new SourceSimpleFragmentsBuilder(mapper, context, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner);
} }
} else { } else {
fragListBuilder = field.fieldOptions().fragmentOffset() == -1 ? new SimpleFragListBuilder() : new SimpleFragListBuilder(field.fieldOptions().fragmentOffset()); fragListBuilder = field.fieldOptions().fragmentOffset() == -1 ? new SimpleFragListBuilder() : new SimpleFragListBuilder(field.fieldOptions().fragmentOffset());
@ -118,13 +118,13 @@ public class FastVectorHighlighter implements Highlighter {
if (!forceSource && mapper.fieldType().stored()) { if (!forceSource && mapper.fieldType().stored()) {
fragmentsBuilder = new ScoreOrderFragmentsBuilder(field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner); fragmentsBuilder = new ScoreOrderFragmentsBuilder(field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner);
} else { } else {
fragmentsBuilder = new SourceScoreOrderFragmentsBuilder(mapper, context, hitContext, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner); fragmentsBuilder = new SourceScoreOrderFragmentsBuilder(mapper, context, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner);
} }
} else { } else {
if (!forceSource && mapper.fieldType().stored()) { if (!forceSource && mapper.fieldType().stored()) {
fragmentsBuilder = new SimpleFragmentsBuilder(mapper, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner); fragmentsBuilder = new SimpleFragmentsBuilder(mapper, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner);
} else { } else {
fragmentsBuilder = new SourceSimpleFragmentsBuilder(mapper, context, hitContext, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner); fragmentsBuilder = new SourceSimpleFragmentsBuilder(mapper, context, field.fieldOptions().preTags(), field.fieldOptions().postTags(), boundaryScanner);
} }
} }
} }

View File

@ -27,7 +27,6 @@ import org.apache.lucene.search.vectorhighlight.BoundaryScanner;
import org.apache.lucene.search.vectorhighlight.FieldFragList.WeightedFragInfo; import org.apache.lucene.search.vectorhighlight.FieldFragList.WeightedFragInfo;
import org.apache.lucene.search.vectorhighlight.ScoreOrderFragmentsBuilder; import org.apache.lucene.search.vectorhighlight.ScoreOrderFragmentsBuilder;
import org.elasticsearch.index.mapper.FieldMapper; import org.elasticsearch.index.mapper.FieldMapper;
import org.elasticsearch.search.fetch.FetchSubPhase;
import org.elasticsearch.search.internal.SearchContext; import org.elasticsearch.search.internal.SearchContext;
import org.elasticsearch.search.lookup.SourceLookup; import org.elasticsearch.search.lookup.SourceLookup;
@ -43,14 +42,11 @@ public class SourceScoreOrderFragmentsBuilder extends ScoreOrderFragmentsBuilder
private final SearchContext searchContext; private final SearchContext searchContext;
private final FetchSubPhase.HitContext hitContext; public SourceScoreOrderFragmentsBuilder(FieldMapper mapper, SearchContext searchContext, String[] preTags, String[] postTags,
BoundaryScanner boundaryScanner) {
public SourceScoreOrderFragmentsBuilder(FieldMapper mapper, SearchContext searchContext,
FetchSubPhase.HitContext hitContext, String[] preTags, String[] postTags, BoundaryScanner boundaryScanner) {
super(preTags, postTags, boundaryScanner); super(preTags, postTags, boundaryScanner);
this.mapper = mapper; this.mapper = mapper;
this.searchContext = searchContext; this.searchContext = searchContext;
this.hitContext = hitContext;
} }
@Override @Override

View File

@ -24,7 +24,6 @@ import org.apache.lucene.index.IndexReader;
import org.apache.lucene.index.LeafReaderContext; import org.apache.lucene.index.LeafReaderContext;
import org.apache.lucene.search.vectorhighlight.BoundaryScanner; import org.apache.lucene.search.vectorhighlight.BoundaryScanner;
import org.elasticsearch.index.mapper.FieldMapper; import org.elasticsearch.index.mapper.FieldMapper;
import org.elasticsearch.search.fetch.FetchSubPhase;
import org.elasticsearch.search.internal.SearchContext; import org.elasticsearch.search.internal.SearchContext;
import org.elasticsearch.search.lookup.SourceLookup; import org.elasticsearch.search.lookup.SourceLookup;
@ -38,13 +37,10 @@ public class SourceSimpleFragmentsBuilder extends SimpleFragmentsBuilder {
private final SearchContext searchContext; private final SearchContext searchContext;
private final FetchSubPhase.HitContext hitContext; public SourceSimpleFragmentsBuilder(FieldMapper mapper, SearchContext searchContext, String[] preTags, String[] postTags,
BoundaryScanner boundaryScanner) {
public SourceSimpleFragmentsBuilder(FieldMapper mapper, SearchContext searchContext,
FetchSubPhase.HitContext hitContext, String[] preTags, String[] postTags, BoundaryScanner boundaryScanner) {
super(mapper, preTags, postTags, boundaryScanner); super(mapper, preTags, postTags, boundaryScanner);
this.searchContext = searchContext; this.searchContext = searchContext;
this.hitContext = hitContext;
} }
public static final Field[] EMPTY_FIELDS = new Field[0]; public static final Field[] EMPTY_FIELDS = new Field[0];

View File

@ -157,11 +157,4 @@ public class ShardSearchTransportRequest extends TransportRequest implements Sha
public void rewrite(QueryShardContext context) throws IOException { public void rewrite(QueryShardContext context) throws IOException {
shardSearchLocalRequest.rewrite(context); shardSearchLocalRequest.rewrite(context);
} }
private ShardSearchTransportRequest shallowCopy(ShardSearchLocalRequest rewritten) {
ShardSearchTransportRequest newRequest = new ShardSearchTransportRequest();
newRequest.originalIndices = originalIndices;
newRequest.shardSearchLocalRequest = rewritten;
return newRequest;
}
} }

View File

@ -98,7 +98,6 @@ public final class CompletionSuggestion extends Suggest.Suggestion<CompletionSug
// combine suggestion entries from participating shards on the coordinating node // combine suggestion entries from participating shards on the coordinating node
// the global top <code>size</code> entries are collected from the shard results // the global top <code>size</code> entries are collected from the shard results
// using a priority queue // using a priority queue
Comparator<Suggest.Suggestion.Entry.Option> optionComparator = sortComparator();
OptionPriorityQueue priorityQueue = new OptionPriorityQueue(size, sortComparator()); OptionPriorityQueue priorityQueue = new OptionPriorityQueue(size, sortComparator());
for (Suggest.Suggestion<Entry> entries : toReduce) { for (Suggest.Suggestion<Entry> entries : toReduce) {
assert entries.getEntries().size() == 1 : "CompletionSuggestion must have only one entry"; assert entries.getEntries().size() == 1 : "CompletionSuggestion must have only one entry";

View File

@ -177,7 +177,7 @@ public class StrictISODateTimeFormat {
if (workingFields.contains(DateTimeFieldType.monthOfYear())) { if (workingFields.contains(DateTimeFieldType.monthOfYear())) {
reducedPrec = dateByMonth(bld, workingFields, extended, strictISO); reducedPrec = dateByMonth(bld, workingFields, extended, strictISO);
} else if (workingFields.contains(DateTimeFieldType.dayOfYear())) { } else if (workingFields.contains(DateTimeFieldType.dayOfYear())) {
reducedPrec = dateByOrdinal(bld, workingFields, extended, strictISO); reducedPrec = dateByOrdinal(bld, workingFields, extended);
} else if (workingFields.contains(DateTimeFieldType.weekOfWeekyear())) { } else if (workingFields.contains(DateTimeFieldType.weekOfWeekyear())) {
reducedPrec = dateByWeek(bld, workingFields, extended, strictISO); reducedPrec = dateByWeek(bld, workingFields, extended, strictISO);
} else if (workingFields.contains(DateTimeFieldType.dayOfMonth())) { } else if (workingFields.contains(DateTimeFieldType.dayOfMonth())) {
@ -288,14 +288,12 @@ public class StrictISODateTimeFormat {
* @param bld the builder * @param bld the builder
* @param fields the fields * @param fields the fields
* @param extended true to use extended format * @param extended true to use extended format
* @param strictISO true to only allow ISO formats
* @since 1.1 * @since 1.1
*/ */
private static boolean dateByOrdinal( private static boolean dateByOrdinal(
DateTimeFormatterBuilder bld, DateTimeFormatterBuilder bld,
Collection<DateTimeFieldType> fields, Collection<DateTimeFieldType> fields,
boolean extended, boolean extended) {
boolean strictISO) {
boolean reducedPrec = false; boolean reducedPrec = false;
if (fields.remove(DateTimeFieldType.year())) { if (fields.remove(DateTimeFieldType.year())) {

View File

@ -48,12 +48,9 @@ public final class ClusterAllocationExplainIT extends ESIntegTestCase {
// Wait for all 3 nodes to be up // Wait for all 3 nodes to be up
logger.info("--> waiting for 3 nodes to be up"); logger.info("--> waiting for 3 nodes to be up");
assertBusy(new Runnable() { assertBusy(() -> {
@Override NodesStatsResponse resp = client().admin().cluster().prepareNodesStats().get();
public void run() { assertThat(resp.getNodes().size(), equalTo(3));
NodesStatsResponse resp = client().admin().cluster().prepareNodesStats().get();
assertThat(resp.getNodes().size(), equalTo(3));
}
}); });
logger.info("--> creating 'test' index"); logger.info("--> creating 'test' index");
@ -126,7 +123,6 @@ public final class ClusterAllocationExplainIT extends ESIntegTestCase {
Map<DiscoveryNode, NodeExplanation> explanations = cae.getNodeExplanations(); Map<DiscoveryNode, NodeExplanation> explanations = cae.getNodeExplanations();
Float noAttrWeight = -1f;
Float barAttrWeight = -1f; Float barAttrWeight = -1f;
Float fooBarAttrWeight = -1f; Float fooBarAttrWeight = -1f;
for (Map.Entry<DiscoveryNode, NodeExplanation> entry : explanations.entrySet()) { for (Map.Entry<DiscoveryNode, NodeExplanation> entry : explanations.entrySet()) {
@ -134,7 +130,6 @@ public final class ClusterAllocationExplainIT extends ESIntegTestCase {
String nodeName = node.getName(); String nodeName = node.getName();
NodeExplanation explanation = entry.getValue(); NodeExplanation explanation = entry.getValue();
ClusterAllocationExplanation.FinalDecision finalDecision = explanation.getFinalDecision(); ClusterAllocationExplanation.FinalDecision finalDecision = explanation.getFinalDecision();
String finalExplanation = explanation.getFinalExplanation();
ClusterAllocationExplanation.StoreCopy storeCopy = explanation.getStoreCopy(); ClusterAllocationExplanation.StoreCopy storeCopy = explanation.getStoreCopy();
Decision d = explanation.getDecision(); Decision d = explanation.getDecision();
float weight = explanation.getWeight(); float weight = explanation.getWeight();
@ -143,7 +138,6 @@ public final class ClusterAllocationExplainIT extends ESIntegTestCase {
assertEquals(d.type(), Decision.Type.NO); assertEquals(d.type(), Decision.Type.NO);
if (noAttrNode.equals(nodeName)) { if (noAttrNode.equals(nodeName)) {
assertThat(d.toString(), containsString("node does not match index include filters [foo:\"bar\"]")); assertThat(d.toString(), containsString("node does not match index include filters [foo:\"bar\"]"));
noAttrWeight = weight;
assertNull(storeStatus); assertNull(storeStatus);
assertEquals("the shard cannot be assigned because one or more allocation decider returns a 'NO' decision", assertEquals("the shard cannot be assigned because one or more allocation decider returns a 'NO' decision",
explanation.getFinalExplanation()); explanation.getFinalExplanation());

View File

@ -47,7 +47,6 @@ public final class ClusterAllocationExplainTests extends ESSingleNodeTestCase {
NodeExplanation explanation = cae.getNodeExplanations().values().iterator().next(); NodeExplanation explanation = cae.getNodeExplanations().values().iterator().next();
ClusterAllocationExplanation.FinalDecision fd = explanation.getFinalDecision(); ClusterAllocationExplanation.FinalDecision fd = explanation.getFinalDecision();
ClusterAllocationExplanation.StoreCopy storeCopy = explanation.getStoreCopy(); ClusterAllocationExplanation.StoreCopy storeCopy = explanation.getStoreCopy();
String finalExplanation = explanation.getFinalExplanation();
Decision d = explanation.getDecision(); Decision d = explanation.getDecision();
assertNotNull("should have a decision", d); assertNotNull("should have a decision", d);
assertEquals(Decision.Type.NO, d.type()); assertEquals(Decision.Type.NO, d.type());
@ -76,7 +75,6 @@ public final class ClusterAllocationExplainTests extends ESSingleNodeTestCase {
d = explanation.getDecision(); d = explanation.getDecision();
fd = explanation.getFinalDecision(); fd = explanation.getFinalDecision();
storeCopy = explanation.getStoreCopy(); storeCopy = explanation.getStoreCopy();
finalExplanation = explanation.getFinalExplanation();
assertNotNull("should have a decision", d); assertNotNull("should have a decision", d);
assertEquals(Decision.Type.NO, d.type()); assertEquals(Decision.Type.NO, d.type());
assertEquals(ClusterAllocationExplanation.FinalDecision.ALREADY_ASSIGNED, fd); assertEquals(ClusterAllocationExplanation.FinalDecision.ALREADY_ASSIGNED, fd);

View File

@ -41,7 +41,6 @@ import org.elasticsearch.index.shard.ShardId;
import org.elasticsearch.test.ESTestCase; import org.elasticsearch.test.ESTestCase;
import java.io.IOException; import java.io.IOException;
import java.util.Arrays;
import java.util.HashMap; import java.util.HashMap;
import java.util.HashSet; import java.util.HashSet;
import java.util.Map; import java.util.Map;
@ -215,9 +214,7 @@ public final class ClusterAllocationExplanationTests extends ESTestCase {
assertEquals(allocationDelay, cae2.getAllocationDelayMillis()); assertEquals(allocationDelay, cae2.getAllocationDelayMillis());
assertEquals(remainingDelay, cae2.getRemainingDelayMillis()); assertEquals(remainingDelay, cae2.getRemainingDelayMillis());
for (Map.Entry<DiscoveryNode, NodeExplanation> entry : cae2.getNodeExplanations().entrySet()) { for (Map.Entry<DiscoveryNode, NodeExplanation> entry : cae2.getNodeExplanations().entrySet()) {
DiscoveryNode node = entry.getKey();
NodeExplanation explanation = entry.getValue(); NodeExplanation explanation = entry.getValue();
IndicesShardStoresResponse.StoreStatus status = explanation.getStoreStatus();
assertNotNull(explanation.getStoreStatus()); assertNotNull(explanation.getStoreStatus());
assertNotNull(explanation.getDecision()); assertNotNull(explanation.getDecision());
assertEquals(nodeWeight, explanation.getWeight()); assertEquals(nodeWeight, explanation.getWeight());

View File

@ -29,11 +29,11 @@ public class BulkShardRequestTests extends ESTestCase {
public void testToString() { public void testToString() {
String index = randomSimpleString(random(), 10); String index = randomSimpleString(random(), 10);
int count = between(1, 100); int count = between(1, 100);
BulkShardRequest r = new BulkShardRequest(null, new ShardId(index, "ignored", 0), RefreshPolicy.NONE, new BulkItemRequest[count]); BulkShardRequest r = new BulkShardRequest(new ShardId(index, "ignored", 0), RefreshPolicy.NONE, new BulkItemRequest[count]);
assertEquals("BulkShardRequest to [" + index + "] containing [" + count + "] requests", r.toString()); assertEquals("BulkShardRequest to [" + index + "] containing [" + count + "] requests", r.toString());
r = new BulkShardRequest(null, new ShardId(index, "ignored", 0), RefreshPolicy.IMMEDIATE, new BulkItemRequest[count]); r = new BulkShardRequest(new ShardId(index, "ignored", 0), RefreshPolicy.IMMEDIATE, new BulkItemRequest[count]);
assertEquals("BulkShardRequest to [" + index + "] containing [" + count + "] requests and a refresh", r.toString()); assertEquals("BulkShardRequest to [" + index + "] containing [" + count + "] requests and a refresh", r.toString());
r = new BulkShardRequest(null, new ShardId(index, "ignored", 0), RefreshPolicy.WAIT_UNTIL, new BulkItemRequest[count]); r = new BulkShardRequest(new ShardId(index, "ignored", 0), RefreshPolicy.WAIT_UNTIL, new BulkItemRequest[count]);
assertEquals("BulkShardRequest to [" + index + "] containing [" + count + "] requests blocking until refresh", r.toString()); assertEquals("BulkShardRequest to [" + index + "] containing [" + count + "] requests blocking until refresh", r.toString());
} }
} }

View File

@ -220,7 +220,6 @@ public class ClusterStateCreationUtils {
* Creates a cluster state with no index * Creates a cluster state with no index
*/ */
public static ClusterState stateWithNoShard() { public static ClusterState stateWithNoShard() {
int numberOfNodes = 2;
DiscoveryNodes.Builder discoBuilder = DiscoveryNodes.builder(); DiscoveryNodes.Builder discoBuilder = DiscoveryNodes.builder();
discoBuilder.localNodeId(newNode(0).getId()); discoBuilder.localNodeId(newNode(0).getId());
discoBuilder.masterNodeId(newNode(1).getId()); discoBuilder.masterNodeId(newNode(1).getId());

View File

@ -39,7 +39,6 @@ import org.elasticsearch.common.xcontent.ToXContent;
import org.elasticsearch.common.xcontent.XContentBuilder; import org.elasticsearch.common.xcontent.XContentBuilder;
import org.elasticsearch.index.engine.VersionConflictEngineException; import org.elasticsearch.index.engine.VersionConflictEngineException;
import org.elasticsearch.index.mapper.FieldMapper; import org.elasticsearch.index.mapper.FieldMapper;
import org.hamcrest.Matcher;
import java.io.IOException; import java.io.IOException;
import java.util.ArrayList; import java.util.ArrayList;
@ -55,7 +54,6 @@ import static org.elasticsearch.common.xcontent.XContentFactory.jsonBuilder;
import static org.elasticsearch.test.hamcrest.ElasticsearchAssertions.assertAcked; import static org.elasticsearch.test.hamcrest.ElasticsearchAssertions.assertAcked;
import static org.elasticsearch.test.hamcrest.ElasticsearchAssertions.assertThrows; import static org.elasticsearch.test.hamcrest.ElasticsearchAssertions.assertThrows;
import static org.hamcrest.Matchers.equalTo; import static org.hamcrest.Matchers.equalTo;
import static org.hamcrest.Matchers.lessThan;
import static org.hamcrest.Matchers.notNullValue; import static org.hamcrest.Matchers.notNullValue;
import static org.hamcrest.Matchers.nullValue; import static org.hamcrest.Matchers.nullValue;
@ -963,21 +961,6 @@ public class GetTermVectorsIT extends AbstractTermVectorsTestCase {
return randomBoolean() ? "test" : "alias"; return randomBoolean() ? "test" : "alias";
} }
private Map<String, Integer> getFieldStatistics(Map<String, Object> stats, String fieldName) throws IOException {
return (Map<String, Integer>) ((Map<String, Object>) stats.get(fieldName)).get("field_statistics");
}
private Map<String, Integer> getTermStatistics(Map<String, Object> stats, String fieldName, String term) {
return (Map<String, Integer>) ((Map<String, Object>) ((Map<String, Object>) stats.get(fieldName)).get("terms")).get(term);
}
private Matcher<Integer> equalOrLessThanTo(Integer value, boolean isEqual) {
if (isEqual) {
return equalTo(value);
}
return lessThan(value);
}
public void testTermVectorsWithVersion() { public void testTermVectorsWithVersion() {
assertAcked(prepareCreate("test").addAlias(new Alias("alias")) assertAcked(prepareCreate("test").addAlias(new Alias("alias"))
.setSettings(Settings.builder().put("index.refresh_interval", -1))); .setSettings(Settings.builder().put("index.refresh_interval", -1)));

View File

@ -135,7 +135,7 @@ public class UpdateRequestTests extends ESTestCase {
TimeValue providedTTLValue = TimeValue.parseTimeValue(randomTimeValue(), null, "ttl"); TimeValue providedTTLValue = TimeValue.parseTimeValue(randomTimeValue(), null, "ttl");
Settings settings = settings(Version.CURRENT).build(); Settings settings = settings(Version.CURRENT).build();
UpdateHelper updateHelper = new UpdateHelper(settings, null, null); UpdateHelper updateHelper = new UpdateHelper(settings, null);
// We just upsert one document with ttl // We just upsert one document with ttl
IndexRequest indexRequest = new IndexRequest("test", "type1", "1") IndexRequest indexRequest = new IndexRequest("test", "type1", "1")

View File

@ -141,15 +141,6 @@ public class SimpleBlocksIT extends ESIntegTestCase {
} }
} }
private void canNotIndexExists(String index) {
try {
IndicesExistsResponse r = client().admin().indices().prepareExists(index).execute().actionGet();
fail();
} catch (ClusterBlockException e) {
// all is well
}
}
private void setIndexReadOnly(String index, Object value) { private void setIndexReadOnly(String index, Object value) {
HashMap<String, Object> newSettings = new HashMap<>(); HashMap<String, Object> newSettings = new HashMap<>();
newSettings.put(IndexMetaData.SETTING_READ_ONLY, value); newSettings.put(IndexMetaData.SETTING_READ_ONLY, value);

View File

@ -22,7 +22,6 @@ package org.elasticsearch.bwcompat;
import org.elasticsearch.action.admin.cluster.node.info.NodeInfo; import org.elasticsearch.action.admin.cluster.node.info.NodeInfo;
import org.elasticsearch.action.admin.cluster.node.info.NodesInfoResponse; import org.elasticsearch.action.admin.cluster.node.info.NodesInfoResponse;
import org.elasticsearch.action.admin.cluster.node.stats.NodesStatsRequestBuilder; import org.elasticsearch.action.admin.cluster.node.stats.NodesStatsRequestBuilder;
import org.elasticsearch.action.admin.cluster.node.stats.NodesStatsResponse;
import org.elasticsearch.client.transport.TransportClient; import org.elasticsearch.client.transport.TransportClient;
import org.elasticsearch.common.settings.Settings; import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.test.ESBackcompatTestCase; import org.elasticsearch.test.ESBackcompatTestCase;
@ -46,7 +45,7 @@ public class NodesStatsBasicBackwardsCompatIT extends ESBackcompatTestCase {
for (NodeInfo n : nodesInfo.getNodes()) { for (NodeInfo n : nodesInfo.getNodes()) {
TransportClient tc = TransportClient.builder().settings(settings).build().addTransportAddress(n.getNode().getAddress()); TransportClient tc = TransportClient.builder().settings(settings).build().addTransportAddress(n.getNode().getAddress());
// Just verify that the NS can be sent and serialized/deserialized between nodes with basic indices // Just verify that the NS can be sent and serialized/deserialized between nodes with basic indices
NodesStatsResponse ns = tc.admin().cluster().prepareNodesStats().setIndices(true).execute().actionGet(); tc.admin().cluster().prepareNodesStats().setIndices(true).execute().actionGet();
tc.close(); tc.close();
} }
} }
@ -78,7 +77,7 @@ public class NodesStatsBasicBackwardsCompatIT extends ESBackcompatTestCase {
method.invoke(nsBuilder); method.invoke(nsBuilder);
} }
} }
NodesStatsResponse ns = nsBuilder.execute().actionGet(); nsBuilder.execute().actionGet();
tc.close(); tc.close();
} }

View File

@ -95,7 +95,7 @@ public class NetworkModuleTests extends ModuleTestCase {
static class FakeCatRestHandler extends AbstractCatAction { static class FakeCatRestHandler extends AbstractCatAction {
public FakeCatRestHandler() { public FakeCatRestHandler() {
super(null, null); super(null);
} }
@Override @Override
protected void doRequest(RestRequest request, RestChannel channel, NodeClient client) {} protected void doRequest(RestRequest request, RestChannel channel, NodeClient client) {}

View File

@ -19,7 +19,6 @@
package org.elasticsearch.gateway; package org.elasticsearch.gateway;
import org.elasticsearch.cluster.ClusterName;
import org.elasticsearch.cluster.service.ClusterService; import org.elasticsearch.cluster.service.ClusterService;
import org.elasticsearch.common.settings.ClusterSettings; import org.elasticsearch.common.settings.ClusterSettings;
import org.elasticsearch.common.settings.Settings; import org.elasticsearch.common.settings.Settings;
@ -40,7 +39,7 @@ public class GatewayServiceTests extends ESTestCase {
.put("http.enabled", "false") .put("http.enabled", "false")
.put("discovery.type", "local") .put("discovery.type", "local")
.put(settings.build()).build(), .put(settings.build()).build(),
null, clusterService, null, null, null, null, new NoopDiscovery(), null, null); null, clusterService, null, null, null, new NoopDiscovery(), null, null);
} }
public void testDefaultRecoverAfterTime() throws IOException { public void testDefaultRecoverAfterTime() throws IOException {

View File

@ -193,7 +193,6 @@ public class IndexModuleTests extends ESTestCase {
IndexSettings indexSettings = IndexSettingsModule.newIndexSettings(index, settings); IndexSettings indexSettings = IndexSettingsModule.newIndexSettings(index, settings);
IndexModule module = new IndexModule(indexSettings, null, IndexModule module = new IndexModule(indexSettings, null,
new AnalysisRegistry(environment, emptyMap(), emptyMap(), emptyMap(), emptyMap())); new AnalysisRegistry(environment, emptyMap(), emptyMap(), emptyMap(), emptyMap()));
Consumer<Settings> listener = (s) -> {};
module.addIndexEventListener(eventListener); module.addIndexEventListener(eventListener);
IndexService indexService = module.newIndexService(nodeEnvironment, deleter, nodeServicesProvider, indicesQueryCache, mapperRegistry, IndexService indexService = module.newIndexService(nodeEnvironment, deleter, nodeServicesProvider, indicesQueryCache, mapperRegistry,
new IndicesFieldDataCache(settings, this.listener)); new IndicesFieldDataCache(settings, this.listener));

View File

@ -77,7 +77,6 @@ public class IndexServiceTests extends ESSingleNodeTestCase {
public void testFilteringAliases() throws Exception { public void testFilteringAliases() throws Exception {
IndexService indexService = createIndex("test", Settings.EMPTY); IndexService indexService = createIndex("test", Settings.EMPTY);
IndexShard shard = indexService.getShard(0);
add(indexService, "cats", filter(termQuery("animal", "cat"))); add(indexService, "cats", filter(termQuery("animal", "cat")));
add(indexService, "dogs", filter(termQuery("animal", "dog"))); add(indexService, "dogs", filter(termQuery("animal", "dog")));
add(indexService, "all", null); add(indexService, "all", null);
@ -101,7 +100,6 @@ public class IndexServiceTests extends ESSingleNodeTestCase {
public void testAliasFilters() throws Exception { public void testAliasFilters() throws Exception {
IndexService indexService = createIndex("test", Settings.EMPTY); IndexService indexService = createIndex("test", Settings.EMPTY);
IndexShard shard = indexService.getShard(0);
add(indexService, "cats", filter(termQuery("animal", "cat"))); add(indexService, "cats", filter(termQuery("animal", "cat")));
add(indexService, "dogs", filter(termQuery("animal", "dog"))); add(indexService, "dogs", filter(termQuery("animal", "dog")));
@ -118,7 +116,6 @@ public class IndexServiceTests extends ESSingleNodeTestCase {
public void testRemovedAliasFilter() throws Exception { public void testRemovedAliasFilter() throws Exception {
IndexService indexService = createIndex("test", Settings.EMPTY); IndexService indexService = createIndex("test", Settings.EMPTY);
IndexShard shard = indexService.getShard(0);
add(indexService, "cats", filter(termQuery("animal", "cat"))); add(indexService, "cats", filter(termQuery("animal", "cat")));
remove(indexService, "cats"); remove(indexService, "cats");
@ -132,7 +129,6 @@ public class IndexServiceTests extends ESSingleNodeTestCase {
public void testUnknownAliasFilter() throws Exception { public void testUnknownAliasFilter() throws Exception {
IndexService indexService = createIndex("test", Settings.EMPTY); IndexService indexService = createIndex("test", Settings.EMPTY);
IndexShard shard = indexService.getShard(0);
add(indexService, "cats", filter(termQuery("animal", "cat"))); add(indexService, "cats", filter(termQuery("animal", "cat")));
add(indexService, "dogs", filter(termQuery("animal", "dog"))); add(indexService, "dogs", filter(termQuery("animal", "dog")));

View File

@ -45,8 +45,6 @@ public class SettingsListenerIT extends ESIntegTestCase {
public static class SettingsListenerPlugin extends Plugin { public static class SettingsListenerPlugin extends Plugin {
private final SettingsTestingService service = new SettingsTestingService(); private final SettingsTestingService service = new SettingsTestingService();
private static final Setting<Integer> SETTING = Setting.intSetting("index.test.new.setting", 0,
Property.Dynamic, Property.IndexScope);
@Override @Override
public List<Setting<?>> getSettings() { public List<Setting<?>> getSettings() {

Some files were not shown because too many files have changed in this diff Show More