Merge branch 'master' into remove_index_buffer_size
Conflicts: core/src/main/java/org/elasticsearch/index/engine/EngineConfig.java
This commit is contained in:
commit
2cc873b2d6
|
@ -18,8 +18,6 @@
|
|||
*/
|
||||
package org.apache.lucene.queries;
|
||||
|
||||
import com.google.common.primitives.Ints;
|
||||
|
||||
import org.apache.lucene.index.IndexReader;
|
||||
import org.apache.lucene.index.IndexReaderContext;
|
||||
import org.apache.lucene.index.LeafReaderContext;
|
||||
|
@ -141,7 +139,7 @@ public abstract class BlendedTermQuery extends Query {
|
|||
}
|
||||
@Override
|
||||
protected int compare(int i, int j) {
|
||||
return Ints.compare(contexts[tieBreak[j]].docFreq(), contexts[tieBreak[i]].docFreq());
|
||||
return Integer.compare(contexts[tieBreak[j]].docFreq(), contexts[tieBreak[i]].docFreq());
|
||||
}
|
||||
}.sort(0, tieBreak.length);
|
||||
int prev = contexts[tieBreak[0]].docFreq();
|
||||
|
|
|
@ -19,9 +19,9 @@
|
|||
|
||||
package org.elasticsearch.action.admin.cluster.repositories.get;
|
||||
|
||||
import com.google.common.collect.ObjectArrays;
|
||||
import org.elasticsearch.action.support.master.MasterNodeReadOperationRequestBuilder;
|
||||
import org.elasticsearch.client.ElasticsearchClient;
|
||||
import org.elasticsearch.common.util.ArrayUtils;
|
||||
|
||||
/**
|
||||
* Get repository request builder
|
||||
|
@ -60,7 +60,7 @@ public class GetRepositoriesRequestBuilder extends MasterNodeReadOperationReques
|
|||
* @return builder
|
||||
*/
|
||||
public GetRepositoriesRequestBuilder addRepositories(String... repositories) {
|
||||
request.repositories(ObjectArrays.concat(request.repositories(), repositories, String.class));
|
||||
request.repositories(ArrayUtils.concat(request.repositories(), repositories));
|
||||
return this;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,9 +19,9 @@
|
|||
|
||||
package org.elasticsearch.action.admin.cluster.snapshots.get;
|
||||
|
||||
import com.google.common.collect.ObjectArrays;
|
||||
import org.elasticsearch.action.support.master.MasterNodeOperationRequestBuilder;
|
||||
import org.elasticsearch.client.ElasticsearchClient;
|
||||
import org.elasticsearch.common.util.ArrayUtils;
|
||||
|
||||
/**
|
||||
* Get snapshots request builder
|
||||
|
@ -81,7 +81,7 @@ public class GetSnapshotsRequestBuilder extends MasterNodeOperationRequestBuilde
|
|||
* @return this builder
|
||||
*/
|
||||
public GetSnapshotsRequestBuilder addSnapshots(String... snapshots) {
|
||||
request.snapshots(ObjectArrays.concat(request.snapshots(), snapshots, String.class));
|
||||
request.snapshots(ArrayUtils.concat(request.snapshots(), snapshots));
|
||||
return this;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,10 +19,9 @@
|
|||
|
||||
package org.elasticsearch.action.admin.cluster.snapshots.status;
|
||||
|
||||
import com.google.common.collect.ObjectArrays;
|
||||
import org.elasticsearch.action.ActionListener;
|
||||
import org.elasticsearch.action.support.master.MasterNodeOperationRequestBuilder;
|
||||
import org.elasticsearch.client.ElasticsearchClient;
|
||||
import org.elasticsearch.common.util.ArrayUtils;
|
||||
|
||||
/**
|
||||
* Snapshots status request builder
|
||||
|
@ -72,7 +71,7 @@ public class SnapshotsStatusRequestBuilder extends MasterNodeOperationRequestBui
|
|||
* @return this builder
|
||||
*/
|
||||
public SnapshotsStatusRequestBuilder addSnapshots(String... snapshots) {
|
||||
request.snapshots(ObjectArrays.concat(request.snapshots(), snapshots, String.class));
|
||||
request.snapshots(ArrayUtils.concat(request.snapshots(), snapshots));
|
||||
return this;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,13 +19,12 @@
|
|||
|
||||
package org.elasticsearch.action.admin.indices.alias.get;
|
||||
|
||||
import com.google.common.collect.ObjectArrays;
|
||||
import org.elasticsearch.action.Action;
|
||||
import org.elasticsearch.action.ActionResponse;
|
||||
import org.elasticsearch.action.support.IndicesOptions;
|
||||
import org.elasticsearch.action.support.master.MasterNodeReadOperationRequestBuilder;
|
||||
import org.elasticsearch.client.ElasticsearchClient;
|
||||
import org.elasticsearch.client.IndicesAdminClient;
|
||||
import org.elasticsearch.common.util.ArrayUtils;
|
||||
|
||||
/**
|
||||
*/
|
||||
|
@ -43,7 +42,7 @@ public abstract class BaseAliasesRequestBuilder<Response extends ActionResponse,
|
|||
|
||||
@SuppressWarnings("unchecked")
|
||||
public Builder addAliases(String... aliases) {
|
||||
request.aliases(ObjectArrays.concat(request.aliases(), aliases, String.class));
|
||||
request.aliases(ArrayUtils.concat(request.aliases(), aliases));
|
||||
return (Builder) this;
|
||||
}
|
||||
|
||||
|
@ -55,7 +54,7 @@ public abstract class BaseAliasesRequestBuilder<Response extends ActionResponse,
|
|||
|
||||
@SuppressWarnings("unchecked")
|
||||
public Builder addIndices(String... indices) {
|
||||
request.indices(ObjectArrays.concat(request.indices(), indices, String.class));
|
||||
request.indices(ArrayUtils.concat(request.indices(), indices));
|
||||
return (Builder) this;
|
||||
}
|
||||
|
||||
|
|
|
@ -19,12 +19,11 @@
|
|||
|
||||
package org.elasticsearch.action.admin.indices.get;
|
||||
|
||||
import com.google.common.collect.ObjectArrays;
|
||||
|
||||
import org.elasticsearch.action.ActionRequestValidationException;
|
||||
import org.elasticsearch.action.support.master.info.ClusterInfoRequest;
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||
import org.elasticsearch.common.util.ArrayUtils;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Arrays;
|
||||
|
@ -115,7 +114,7 @@ public class GetIndexRequest extends ClusterInfoRequest<GetIndexRequest> {
|
|||
if (this.features == DEFAULT_FEATURES) {
|
||||
return features(features);
|
||||
} else {
|
||||
return features(ObjectArrays.concat(featuresAsEnums(), features, Feature.class));
|
||||
return features(ArrayUtils.concat(features(), features, Feature.class));
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -19,12 +19,10 @@
|
|||
|
||||
package org.elasticsearch.action.admin.indices.mapping.get;
|
||||
|
||||
import com.google.common.collect.ObjectArrays;
|
||||
import org.elasticsearch.action.ActionListener;
|
||||
import org.elasticsearch.action.ActionRequestBuilder;
|
||||
import org.elasticsearch.action.support.IndicesOptions;
|
||||
import org.elasticsearch.client.ElasticsearchClient;
|
||||
import org.elasticsearch.client.IndicesAdminClient;
|
||||
import org.elasticsearch.common.util.ArrayUtils;
|
||||
|
||||
/** A helper class to build {@link GetFieldMappingsRequest} objects */
|
||||
public class GetFieldMappingsRequestBuilder extends ActionRequestBuilder<GetFieldMappingsRequest, GetFieldMappingsResponse, GetFieldMappingsRequestBuilder> {
|
||||
|
@ -39,7 +37,7 @@ public class GetFieldMappingsRequestBuilder extends ActionRequestBuilder<GetFiel
|
|||
}
|
||||
|
||||
public GetFieldMappingsRequestBuilder addIndices(String... indices) {
|
||||
request.indices(ObjectArrays.concat(request.indices(), indices, String.class));
|
||||
request.indices(ArrayUtils.concat(request.indices(), indices));
|
||||
return this;
|
||||
}
|
||||
|
||||
|
@ -49,7 +47,7 @@ public class GetFieldMappingsRequestBuilder extends ActionRequestBuilder<GetFiel
|
|||
}
|
||||
|
||||
public GetFieldMappingsRequestBuilder addTypes(String... types) {
|
||||
request.types(ObjectArrays.concat(request.types(), types, String.class));
|
||||
request.types(ArrayUtils.concat(request.types(), types));
|
||||
return this;
|
||||
}
|
||||
|
||||
|
|
|
@ -19,10 +19,10 @@
|
|||
|
||||
package org.elasticsearch.action.admin.indices.settings.get;
|
||||
|
||||
import com.google.common.collect.ObjectArrays;
|
||||
import org.elasticsearch.action.support.IndicesOptions;
|
||||
import org.elasticsearch.action.support.master.MasterNodeReadOperationRequestBuilder;
|
||||
import org.elasticsearch.client.ElasticsearchClient;
|
||||
import org.elasticsearch.common.util.ArrayUtils;
|
||||
|
||||
/**
|
||||
*/
|
||||
|
@ -38,7 +38,7 @@ public class GetSettingsRequestBuilder extends MasterNodeReadOperationRequestBui
|
|||
}
|
||||
|
||||
public GetSettingsRequestBuilder addIndices(String... indices) {
|
||||
request.indices(ObjectArrays.concat(request.indices(), indices, String.class));
|
||||
request.indices(ArrayUtils.concat(request.indices(), indices));
|
||||
return this;
|
||||
}
|
||||
|
||||
|
|
|
@ -19,9 +19,9 @@
|
|||
|
||||
package org.elasticsearch.action.admin.indices.warmer.get;
|
||||
|
||||
import com.google.common.collect.ObjectArrays;
|
||||
import org.elasticsearch.action.support.master.info.ClusterInfoRequestBuilder;
|
||||
import org.elasticsearch.client.ElasticsearchClient;
|
||||
import org.elasticsearch.common.util.ArrayUtils;
|
||||
|
||||
/**
|
||||
* Builder for {@link GetWarmersRequest}
|
||||
|
@ -40,7 +40,7 @@ public class GetWarmersRequestBuilder extends ClusterInfoRequestBuilder<GetWarme
|
|||
}
|
||||
|
||||
public GetWarmersRequestBuilder addWarmers(String... warmers) {
|
||||
request.warmers(ObjectArrays.concat(request.warmers(), warmers, String.class));
|
||||
request.warmers(ArrayUtils.concat(request.warmers(), warmers));
|
||||
return this;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -18,14 +18,12 @@
|
|||
*/
|
||||
package org.elasticsearch.action.support.master.info;
|
||||
|
||||
import com.google.common.collect.ObjectArrays;
|
||||
import org.elasticsearch.action.Action;
|
||||
import org.elasticsearch.action.ActionResponse;
|
||||
import org.elasticsearch.action.support.IndicesOptions;
|
||||
import org.elasticsearch.action.support.master.MasterNodeReadOperationRequestBuilder;
|
||||
import org.elasticsearch.client.ClusterAdminClient;
|
||||
import org.elasticsearch.client.ElasticsearchClient;
|
||||
import org.elasticsearch.client.IndicesAdminClient;
|
||||
import org.elasticsearch.common.util.ArrayUtils;
|
||||
|
||||
/**
|
||||
*/
|
||||
|
@ -44,7 +42,7 @@ public abstract class ClusterInfoRequestBuilder<Request extends ClusterInfoReque
|
|||
|
||||
@SuppressWarnings("unchecked")
|
||||
public Builder addIndices(String... indices) {
|
||||
request.indices(ObjectArrays.concat(request.indices(), indices, String.class));
|
||||
request.indices(ArrayUtils.concat(request.indices(), indices));
|
||||
return (Builder) this;
|
||||
}
|
||||
|
||||
|
@ -56,7 +54,7 @@ public abstract class ClusterInfoRequestBuilder<Request extends ClusterInfoReque
|
|||
|
||||
@SuppressWarnings("unchecked")
|
||||
public Builder addTypes(String... types) {
|
||||
request.types(ObjectArrays.concat(request.types(), types, String.class));
|
||||
request.types(ArrayUtils.concat(request.types(), types));
|
||||
return (Builder) this;
|
||||
}
|
||||
|
||||
|
|
|
@ -18,7 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.action.termvectors;
|
||||
|
||||
import com.google.common.util.concurrent.AtomicLongMap;
|
||||
import org.apache.lucene.index.*;
|
||||
import org.apache.lucene.search.TermStatistics;
|
||||
import org.apache.lucene.search.similarities.DefaultSimilarity;
|
||||
|
@ -54,7 +53,7 @@ public class TermVectorsFilter {
|
|||
private final Set<String> selectedFields;
|
||||
private AggregatedDfs dfs;
|
||||
private Map<Term, ScoreTerm> scoreTerms;
|
||||
private AtomicLongMap<String> sizes;
|
||||
private Map<String, Integer> sizes = new HashMap<>();
|
||||
private TFIDFSimilarity similarity;
|
||||
|
||||
public TermVectorsFilter(Fields termVectorsByField, Fields topLevelFields, Set<String> selectedFields, @Nullable AggregatedDfs dfs) {
|
||||
|
@ -64,7 +63,6 @@ public class TermVectorsFilter {
|
|||
|
||||
this.dfs = dfs;
|
||||
this.scoreTerms = new HashMap<>();
|
||||
this.sizes = AtomicLongMap.create();
|
||||
this.similarity = new DefaultSimilarity();
|
||||
}
|
||||
|
||||
|
@ -228,10 +226,12 @@ public class TermVectorsFilter {
|
|||
|
||||
// retain the best terms for quick lookups
|
||||
ScoreTerm scoreTerm;
|
||||
int count = 0;
|
||||
while ((scoreTerm = queue.pop()) != null) {
|
||||
scoreTerms.put(new Term(scoreTerm.field, scoreTerm.word), scoreTerm);
|
||||
sizes.incrementAndGet(scoreTerm.field);
|
||||
count++;
|
||||
}
|
||||
sizes.put(fieldName, count);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.cluster.metadata;
|
||||
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
import org.elasticsearch.common.collect.Tuple;
|
||||
|
||||
import java.util.ArrayList;
|
||||
|
@ -106,7 +105,7 @@ public interface AliasOrIndex {
|
|||
return new Iterable<Tuple<String, AliasMetaData>>() {
|
||||
@Override
|
||||
public Iterator<Tuple<String, AliasMetaData>> iterator() {
|
||||
return new UnmodifiableIterator<Tuple<String,AliasMetaData>>() {
|
||||
return new Iterator<Tuple<String,AliasMetaData>>() {
|
||||
|
||||
int index = 0;
|
||||
|
||||
|
@ -121,6 +120,11 @@ public interface AliasOrIndex {
|
|||
return new Tuple<>(indexMetaData.getIndex(), indexMetaData.getAliases().get(aliasName));
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
|
||||
};
|
||||
}
|
||||
};
|
||||
|
|
|
@ -21,7 +21,6 @@ package org.elasticsearch.cluster.metadata;
|
|||
|
||||
import com.carrotsearch.hppc.cursors.ObjectCursor;
|
||||
import com.carrotsearch.hppc.cursors.ObjectObjectCursor;
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.elasticsearch.ElasticsearchException;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.cluster.Diff;
|
||||
|
@ -205,8 +204,12 @@ public class IndexMetaData implements Diffable<IndexMetaData>, FromXContentBuild
|
|||
private final boolean useTypeForRouting;
|
||||
|
||||
private IndexMetaData(String index, long version, State state, Settings settings, ImmutableOpenMap<String, MappingMetaData> mappings, ImmutableOpenMap<String, AliasMetaData> aliases, ImmutableOpenMap<String, Custom> customs) {
|
||||
Preconditions.checkArgument(settings.getAsInt(SETTING_NUMBER_OF_SHARDS, null) != null, "must specify numberOfShards for index [" + index + "]");
|
||||
Preconditions.checkArgument(settings.getAsInt(SETTING_NUMBER_OF_REPLICAS, null) != null, "must specify numberOfReplicas for index [" + index + "]");
|
||||
if (settings.getAsInt(SETTING_NUMBER_OF_SHARDS, null) == null) {
|
||||
throw new IllegalArgumentException("must specify numberOfShards for index [" + index + "]");
|
||||
}
|
||||
if (settings.getAsInt(SETTING_NUMBER_OF_REPLICAS, null) == null) {
|
||||
throw new IllegalArgumentException("must specify numberOfReplicas for index [" + index + "]");
|
||||
}
|
||||
this.index = index;
|
||||
this.version = version;
|
||||
this.state = state;
|
||||
|
|
|
@ -23,7 +23,6 @@ import com.carrotsearch.hppc.ObjectHashSet;
|
|||
import com.carrotsearch.hppc.cursors.ObjectCursor;
|
||||
import com.carrotsearch.hppc.cursors.ObjectObjectCursor;
|
||||
import com.google.common.collect.ImmutableSet;
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
import org.apache.lucene.util.CollectionUtil;
|
||||
import org.elasticsearch.cluster.Diff;
|
||||
import org.elasticsearch.cluster.Diffable;
|
||||
|
@ -60,18 +59,7 @@ import org.elasticsearch.rest.RestStatus;
|
|||
import org.elasticsearch.search.warmer.IndexWarmersMetaData;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
import java.util.Collections;
|
||||
import java.util.Comparator;
|
||||
import java.util.EnumSet;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import java.util.SortedMap;
|
||||
import java.util.TreeMap;
|
||||
import java.util.*;
|
||||
import java.util.stream.Collectors;
|
||||
|
||||
import static org.elasticsearch.common.settings.Settings.readSettingsFromStream;
|
||||
|
@ -569,7 +557,7 @@ public class MetaData implements Iterable<IndexMetaData>, Diffable<MetaData>, Fr
|
|||
}
|
||||
|
||||
@Override
|
||||
public UnmodifiableIterator<IndexMetaData> iterator() {
|
||||
public Iterator<IndexMetaData> iterator() {
|
||||
return indices.valuesIt();
|
||||
}
|
||||
|
||||
|
|
|
@ -22,7 +22,6 @@ package org.elasticsearch.cluster.node;
|
|||
import com.carrotsearch.hppc.ObjectHashSet;
|
||||
import com.carrotsearch.hppc.cursors.ObjectCursor;
|
||||
import com.carrotsearch.hppc.cursors.ObjectObjectCursor;
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.cluster.AbstractDiffable;
|
||||
import org.elasticsearch.common.Booleans;
|
||||
|
@ -34,11 +33,7 @@ import org.elasticsearch.common.regex.Regex;
|
|||
import org.elasticsearch.common.transport.TransportAddress;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import java.util.*;
|
||||
|
||||
/**
|
||||
* This class holds all {@link DiscoveryNode} in the cluster and provides convenience methods to
|
||||
|
@ -69,7 +64,7 @@ public class DiscoveryNodes extends AbstractDiffable<DiscoveryNodes> implements
|
|||
}
|
||||
|
||||
@Override
|
||||
public UnmodifiableIterator<DiscoveryNode> iterator() {
|
||||
public Iterator<DiscoveryNode> iterator() {
|
||||
return nodes.valuesIt();
|
||||
}
|
||||
|
||||
|
|
|
@ -22,7 +22,6 @@ package org.elasticsearch.cluster.routing;
|
|||
import com.carrotsearch.hppc.IntSet;
|
||||
import com.carrotsearch.hppc.cursors.IntCursor;
|
||||
import com.carrotsearch.hppc.cursors.IntObjectCursor;
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
import org.apache.lucene.util.CollectionUtil;
|
||||
import org.elasticsearch.cluster.AbstractDiffable;
|
||||
import org.elasticsearch.cluster.metadata.IndexMetaData;
|
||||
|
@ -33,12 +32,7 @@ import org.elasticsearch.common.io.stream.StreamOutput;
|
|||
import org.elasticsearch.index.shard.ShardId;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.Comparator;
|
||||
import java.util.HashSet;
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
import java.util.*;
|
||||
import java.util.concurrent.ThreadLocalRandom;
|
||||
|
||||
/**
|
||||
|
@ -163,7 +157,7 @@ public class IndexRoutingTable extends AbstractDiffable<IndexRoutingTable> imple
|
|||
}
|
||||
|
||||
@Override
|
||||
public UnmodifiableIterator<IndexShardRoutingTable> iterator() {
|
||||
public Iterator<IndexShardRoutingTable> iterator() {
|
||||
return shards.valuesIt();
|
||||
}
|
||||
|
||||
|
|
|
@ -21,7 +21,6 @@ package org.elasticsearch.cluster.routing;
|
|||
|
||||
import com.carrotsearch.hppc.ObjectIntHashMap;
|
||||
import com.carrotsearch.hppc.cursors.ObjectCursor;
|
||||
import com.google.common.collect.Iterables;
|
||||
import com.google.common.collect.Iterators;
|
||||
import org.apache.lucene.util.CollectionUtil;
|
||||
import org.elasticsearch.cluster.ClusterState;
|
||||
|
@ -31,15 +30,7 @@ import org.elasticsearch.cluster.node.DiscoveryNode;
|
|||
import org.elasticsearch.common.collect.ImmutableOpenMap;
|
||||
import org.elasticsearch.index.shard.ShardId;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.Comparator;
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
import java.util.Iterator;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import java.util.*;
|
||||
import java.util.function.Predicate;
|
||||
|
||||
/**
|
||||
|
@ -308,7 +299,7 @@ public class RoutingNodes implements Iterable<RoutingNode> {
|
|||
}
|
||||
for (ShardRoutingState s : state) {
|
||||
if (s == ShardRoutingState.UNASSIGNED) {
|
||||
Iterables.addAll(shards, unassigned());
|
||||
unassigned().forEach(shards::add);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -21,8 +21,6 @@ package org.elasticsearch.cluster.routing;
|
|||
|
||||
import com.carrotsearch.hppc.IntSet;
|
||||
import com.google.common.collect.ImmutableMap;
|
||||
import com.google.common.collect.Iterables;
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
import org.elasticsearch.cluster.Diff;
|
||||
import org.elasticsearch.cluster.Diffable;
|
||||
import org.elasticsearch.cluster.DiffableUtils;
|
||||
|
@ -30,14 +28,11 @@ import org.elasticsearch.cluster.metadata.IndexMetaData;
|
|||
import org.elasticsearch.cluster.metadata.MetaData;
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.index.IndexNotFoundException;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.*;
|
||||
import java.util.function.Predicate;
|
||||
|
||||
/**
|
||||
|
@ -72,7 +67,7 @@ public class RoutingTable implements Iterable<IndexRoutingTable>, Diffable<Routi
|
|||
}
|
||||
|
||||
@Override
|
||||
public UnmodifiableIterator<IndexRoutingTable> iterator() {
|
||||
public Iterator<IndexRoutingTable> iterator() {
|
||||
return indicesRouting.values().iterator();
|
||||
}
|
||||
|
||||
|
@ -379,7 +374,10 @@ public class RoutingTable implements Iterable<IndexRoutingTable>, Diffable<Routi
|
|||
indexBuilder.addShard(refData, shardRoutingEntry);
|
||||
}
|
||||
}
|
||||
for (ShardRouting shardRoutingEntry : Iterables.concat(routingNodes.unassigned(), routingNodes.unassigned().ignored())) {
|
||||
|
||||
Iterable<ShardRouting> shardRoutingEntries = Iterables.concat(routingNodes.unassigned(), routingNodes.unassigned().ignored());
|
||||
|
||||
for (ShardRouting shardRoutingEntry : shardRoutingEntries) {
|
||||
String index = shardRoutingEntry.index();
|
||||
IndexRoutingTable.Builder indexBuilder = indexRoutingTableBuilders.get(index);
|
||||
if (indexBuilder == null) {
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.cluster.service;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.cluster.*;
|
||||
import org.elasticsearch.cluster.ClusterState.Builder;
|
||||
|
@ -42,6 +41,7 @@ import org.elasticsearch.common.settings.Settings;
|
|||
import org.elasticsearch.common.text.StringText;
|
||||
import org.elasticsearch.common.transport.TransportAddress;
|
||||
import org.elasticsearch.common.unit.TimeValue;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.common.util.concurrent.*;
|
||||
import org.elasticsearch.discovery.Discovery;
|
||||
import org.elasticsearch.discovery.DiscoveryService;
|
||||
|
@ -89,10 +89,7 @@ public class InternalClusterService extends AbstractLifecycleComponent<ClusterSe
|
|||
private final Collection<ClusterStateListener> lastClusterStateListeners = new CopyOnWriteArrayList<>();
|
||||
// TODO this is rather frequently changing I guess a Synced Set would be better here and a dedicated remove API
|
||||
private final Collection<ClusterStateListener> postAppliedListeners = new CopyOnWriteArrayList<>();
|
||||
private final Iterable<ClusterStateListener> preAppliedListeners = Iterables.concat(
|
||||
priorityClusterStateListeners,
|
||||
clusterStateListeners,
|
||||
lastClusterStateListeners);
|
||||
private final Iterable<ClusterStateListener> preAppliedListeners = Iterables.concat(priorityClusterStateListeners, clusterStateListeners, lastClusterStateListeners);
|
||||
|
||||
private final LocalNodeMasterListeners localNodeMasterListeners;
|
||||
|
||||
|
|
|
@ -20,8 +20,6 @@
|
|||
package org.elasticsearch.common;
|
||||
|
||||
import com.google.common.collect.ImmutableSet;
|
||||
import com.google.common.collect.Iterables;
|
||||
|
||||
import org.apache.lucene.util.BytesRefBuilder;
|
||||
import org.elasticsearch.common.bytes.BytesReference;
|
||||
import org.elasticsearch.common.io.FastStringReader;
|
||||
|
@ -32,18 +30,7 @@ import org.elasticsearch.common.xcontent.json.JsonXContent;
|
|||
|
||||
import java.io.BufferedReader;
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
import java.util.HashSet;
|
||||
import java.util.Iterator;
|
||||
import java.util.LinkedList;
|
||||
import java.util.List;
|
||||
import java.util.Properties;
|
||||
import java.util.Random;
|
||||
import java.util.Set;
|
||||
import java.util.StringTokenizer;
|
||||
import java.util.TreeSet;
|
||||
import java.util.*;
|
||||
|
||||
/**
|
||||
*
|
||||
|
@ -840,9 +827,6 @@ public class Strings {
|
|||
}
|
||||
|
||||
public static String collectionToDelimitedString(Iterable<?> coll, String delim, String prefix, String suffix, StringBuilder sb) {
|
||||
if (Iterables.isEmpty(coll)) {
|
||||
return "";
|
||||
}
|
||||
Iterator<?> it = coll.iterator();
|
||||
while (it.hasNext()) {
|
||||
sb.append(prefix).append(it.next()).append(suffix);
|
||||
|
|
|
@ -19,11 +19,9 @@
|
|||
|
||||
package org.elasticsearch.common.cli;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.commons.cli.CommandLine;
|
||||
import org.apache.commons.cli.CommandLineParser;
|
||||
import org.apache.commons.cli.DefaultParser;
|
||||
import org.elasticsearch.common.collect.Tuple;
|
||||
import org.elasticsearch.common.settings.Settings;
|
||||
import org.elasticsearch.env.Environment;
|
||||
import org.elasticsearch.node.internal.InternalSettingsPreparer;
|
||||
|
@ -101,7 +99,9 @@ public abstract class CliTool {
|
|||
}
|
||||
|
||||
protected CliTool(CliToolConfig config, Terminal terminal) {
|
||||
Preconditions.checkArgument(config.cmds().size() != 0, "At least one command must be configured");
|
||||
if (config.cmds().size() == 0) {
|
||||
throw new IllegalArgumentException("At least one command must be configured");
|
||||
}
|
||||
this.config = config;
|
||||
this.terminal = terminal;
|
||||
env = InternalSettingsPreparer.prepareEnvironment(EMPTY_SETTINGS, terminal);
|
||||
|
|
|
@ -19,13 +19,14 @@
|
|||
|
||||
package org.elasticsearch.common.cli;
|
||||
|
||||
import com.google.common.collect.ImmutableMap;
|
||||
import org.apache.commons.cli.CommandLine;
|
||||
import org.apache.commons.cli.Option;
|
||||
import org.apache.commons.cli.OptionGroup;
|
||||
import org.apache.commons.cli.Options;
|
||||
|
||||
import java.util.Collection;
|
||||
import java.util.Collections;
|
||||
import java.util.HashMap;
|
||||
import java.util.Map;
|
||||
|
||||
/**
|
||||
*
|
||||
|
@ -38,18 +39,18 @@ public class CliToolConfig {
|
|||
|
||||
private final Class<? extends CliTool> toolType;
|
||||
private final String name;
|
||||
private final ImmutableMap<String, Cmd> cmds;
|
||||
private final Map<String, Cmd> cmds;
|
||||
|
||||
private static final HelpPrinter helpPrinter = new HelpPrinter();
|
||||
|
||||
private CliToolConfig(String name, Class<? extends CliTool> toolType, Cmd[] cmds) {
|
||||
this.name = name;
|
||||
this.toolType = toolType;
|
||||
ImmutableMap.Builder<String, Cmd> cmdsBuilder = ImmutableMap.builder();
|
||||
final Map<String, Cmd> cmdsMapping = new HashMap<>();
|
||||
for (int i = 0; i < cmds.length; i++) {
|
||||
cmdsBuilder.put(cmds[i].name, cmds[i]);
|
||||
cmdsMapping.put(cmds[i].name, cmds[i]);
|
||||
}
|
||||
this.cmds = cmdsBuilder.build();
|
||||
this.cmds = Collections.unmodifiableMap(cmdsMapping);
|
||||
}
|
||||
|
||||
public boolean isSingle() {
|
||||
|
|
|
@ -18,26 +18,10 @@
|
|||
*/
|
||||
|
||||
package org.elasticsearch.common.collect;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import com.google.common.base.Supplier;
|
||||
import com.google.common.base.Suppliers;
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
import org.apache.lucene.util.mutable.MutableValueInt;
|
||||
|
||||
import java.lang.reflect.Array;
|
||||
import java.util.AbstractMap;
|
||||
import java.util.AbstractSet;
|
||||
import java.util.ArrayDeque;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
import java.util.Deque;
|
||||
import java.util.Iterator;
|
||||
import java.util.Map;
|
||||
import java.util.NoSuchElementException;
|
||||
import java.util.Set;
|
||||
import java.util.function.Consumer;
|
||||
import java.util.stream.Collectors;
|
||||
import java.util.*;
|
||||
import java.util.stream.Stream;
|
||||
|
||||
/**
|
||||
|
@ -434,7 +418,7 @@ public final class CopyOnWriteHashMap<K, V> extends AbstractMap<K, V> {
|
|||
|
||||
}
|
||||
|
||||
private static class EntryIterator<K, V> extends UnmodifiableIterator<Map.Entry<K, V>> {
|
||||
private static class EntryIterator<K, V> implements Iterator<Map.Entry<K, V>> {
|
||||
|
||||
private final Deque<Map.Entry<K, V>> entries;
|
||||
private final Deque<Node<K, V>> nodes;
|
||||
|
@ -462,6 +446,11 @@ public final class CopyOnWriteHashMap<K, V> extends AbstractMap<K, V> {
|
|||
return entries.pop();
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
private final InnerNode<K, V> root;
|
||||
|
@ -487,7 +476,9 @@ public final class CopyOnWriteHashMap<K, V> extends AbstractMap<K, V> {
|
|||
|
||||
@Override
|
||||
public V get(Object key) {
|
||||
Preconditions.checkArgument(key != null, "Null keys are not supported");
|
||||
if (key == null) {
|
||||
throw new IllegalArgumentException("null keys are not supported");
|
||||
}
|
||||
final int hash = key.hashCode();
|
||||
return root.get(key, hash);
|
||||
}
|
||||
|
@ -503,8 +494,12 @@ public final class CopyOnWriteHashMap<K, V> extends AbstractMap<K, V> {
|
|||
* of the hash table. The current hash table is not modified.
|
||||
*/
|
||||
public CopyOnWriteHashMap<K, V> copyAndPut(K key, V value) {
|
||||
Preconditions.checkArgument(key != null, "null keys are not supported");
|
||||
Preconditions.checkArgument(value != null, "null values are not supported");
|
||||
if (key == null) {
|
||||
throw new IllegalArgumentException("null keys are not supported");
|
||||
}
|
||||
if (value == null) {
|
||||
throw new IllegalArgumentException("null values are not supported");
|
||||
}
|
||||
final int hash = key.hashCode();
|
||||
final MutableValueInt newValue = new MutableValueInt();
|
||||
final InnerNode<K, V> newRoot = root.put(key, hash, TOTAL_HASH_BITS, value, newValue);
|
||||
|
@ -535,7 +530,9 @@ public final class CopyOnWriteHashMap<K, V> extends AbstractMap<K, V> {
|
|||
* Remove the given key from this map. The current hash table is not modified.
|
||||
*/
|
||||
public CopyOnWriteHashMap<K, V> copyAndRemove(Object key) {
|
||||
Preconditions.checkArgument(key != null, "Null keys are not supported");
|
||||
if (key == null) {
|
||||
throw new IllegalArgumentException("null keys are not supported");
|
||||
}
|
||||
final int hash = key.hashCode();
|
||||
final InnerNode<K, V> newRoot = root.remove(key, hash);
|
||||
if (root == newRoot) {
|
||||
|
|
|
@ -26,7 +26,6 @@ import com.carrotsearch.hppc.cursors.ObjectCursor;
|
|||
import com.carrotsearch.hppc.predicates.IntObjectPredicate;
|
||||
import com.carrotsearch.hppc.predicates.IntPredicate;
|
||||
import com.carrotsearch.hppc.procedures.IntObjectProcedure;
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
|
||||
import java.util.Iterator;
|
||||
import java.util.Map;
|
||||
|
@ -113,9 +112,9 @@ public final class ImmutableOpenIntMap<VType> implements Iterable<IntObjectCurso
|
|||
/**
|
||||
* Returns a direct iterator over the keys.
|
||||
*/
|
||||
public UnmodifiableIterator<Integer> keysIt() {
|
||||
public Iterator<Integer> keysIt() {
|
||||
final Iterator<IntCursor> iterator = map.keys().iterator();
|
||||
return new UnmodifiableIterator<Integer>() {
|
||||
return new Iterator<Integer>() {
|
||||
@Override
|
||||
public boolean hasNext() {
|
||||
return iterator.hasNext();
|
||||
|
@ -125,6 +124,11 @@ public final class ImmutableOpenIntMap<VType> implements Iterable<IntObjectCurso
|
|||
public Integer next() {
|
||||
return iterator.next().value;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
|
@ -138,9 +142,9 @@ public final class ImmutableOpenIntMap<VType> implements Iterable<IntObjectCurso
|
|||
/**
|
||||
* Returns a direct iterator over the keys.
|
||||
*/
|
||||
public UnmodifiableIterator<VType> valuesIt() {
|
||||
public Iterator<VType> valuesIt() {
|
||||
final Iterator<ObjectCursor<VType>> iterator = map.values().iterator();
|
||||
return new UnmodifiableIterator<VType>() {
|
||||
return new Iterator<VType>() {
|
||||
@Override
|
||||
public boolean hasNext() {
|
||||
return iterator.hasNext();
|
||||
|
@ -150,6 +154,11 @@ public final class ImmutableOpenIntMap<VType> implements Iterable<IntObjectCurso
|
|||
public VType next() {
|
||||
return iterator.next().value;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
|
|
|
@ -25,7 +25,6 @@ import com.carrotsearch.hppc.cursors.ObjectObjectCursor;
|
|||
import com.carrotsearch.hppc.predicates.ObjectObjectPredicate;
|
||||
import com.carrotsearch.hppc.predicates.ObjectPredicate;
|
||||
import com.carrotsearch.hppc.procedures.ObjectObjectProcedure;
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
|
||||
import java.util.Iterator;
|
||||
import java.util.Map;
|
||||
|
@ -120,18 +119,21 @@ public final class ImmutableOpenMap<KType, VType> implements Iterable<ObjectObje
|
|||
/**
|
||||
* Returns a direct iterator over the keys.
|
||||
*/
|
||||
public UnmodifiableIterator<KType> keysIt() {
|
||||
public Iterator<KType> keysIt() {
|
||||
final Iterator<ObjectCursor<KType>> iterator = map.keys().iterator();
|
||||
return new UnmodifiableIterator<KType>() {
|
||||
return new Iterator<KType>() {
|
||||
@Override
|
||||
public boolean hasNext() {
|
||||
return iterator.hasNext();
|
||||
}
|
||||
public boolean hasNext() { return iterator.hasNext(); }
|
||||
|
||||
@Override
|
||||
public KType next() {
|
||||
return iterator.next().value;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
|
@ -145,18 +147,21 @@ public final class ImmutableOpenMap<KType, VType> implements Iterable<ObjectObje
|
|||
/**
|
||||
* Returns a direct iterator over the keys.
|
||||
*/
|
||||
public UnmodifiableIterator<VType> valuesIt() {
|
||||
public Iterator<VType> valuesIt() {
|
||||
final Iterator<ObjectCursor<VType>> iterator = map.values().iterator();
|
||||
return new UnmodifiableIterator<VType>() {
|
||||
return new Iterator<VType>() {
|
||||
@Override
|
||||
public boolean hasNext() {
|
||||
return iterator.hasNext();
|
||||
}
|
||||
public boolean hasNext() { return iterator.hasNext(); }
|
||||
|
||||
@Override
|
||||
public VType next() {
|
||||
return iterator.next().value;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
|
|
|
@ -27,8 +27,6 @@ import org.elasticsearch.common.inject.spi.TypeListener;
|
|||
import java.lang.annotation.Annotation;
|
||||
import java.util.Objects;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
|
||||
/**
|
||||
* A support class for {@link Module}s which reduces repetition and results in
|
||||
* a more readable configuration. Simply extend this class, implement {@link
|
||||
|
@ -54,8 +52,9 @@ public abstract class AbstractModule implements Module {
|
|||
|
||||
@Override
|
||||
public final synchronized void configure(Binder builder) {
|
||||
checkState(this.binder == null, "Re-entry is not allowed.");
|
||||
|
||||
if (this.binder != null) {
|
||||
throw new IllegalStateException("Re-entry is not allowed.");
|
||||
}
|
||||
this.binder = Objects.requireNonNull(builder, "builder");
|
||||
try {
|
||||
configure();
|
||||
|
|
|
@ -21,8 +21,7 @@ import org.elasticsearch.common.inject.internal.Errors;
|
|||
import org.elasticsearch.common.inject.spi.Message;
|
||||
|
||||
import java.util.Collection;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
import java.util.Locale;
|
||||
|
||||
/**
|
||||
* Thrown when a programming error such as a misplaced annotation, illegal binding, or unsupported
|
||||
|
@ -48,8 +47,10 @@ public final class ConfigurationException extends RuntimeException {
|
|||
* Returns a copy of this configuration exception with the specified partial value.
|
||||
*/
|
||||
public ConfigurationException withPartialValue(Object partialValue) {
|
||||
checkState(this.partialValue == null,
|
||||
"Can't clobber existing partial value %s with %s", this.partialValue, partialValue);
|
||||
if (this.partialValue != null) {
|
||||
String message = String.format(Locale.ROOT, "Can't clobber existing partial value %s with %s", this.partialValue, partialValue);
|
||||
throw new IllegalStateException(message);
|
||||
}
|
||||
ConfigurationException result = new ConfigurationException(messages);
|
||||
result.partialValue = partialValue;
|
||||
return result;
|
||||
|
|
|
@ -25,8 +25,6 @@ import org.elasticsearch.common.inject.spi.InjectionPoint;
|
|||
|
||||
import java.util.Set;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
|
||||
class ConstructorBindingImpl<T> extends BindingImpl<T> implements ConstructorBinding<T> {
|
||||
|
||||
private final Factory<T> factory;
|
||||
|
@ -52,19 +50,25 @@ class ConstructorBindingImpl<T> extends BindingImpl<T> implements ConstructorBin
|
|||
|
||||
@Override
|
||||
public <V> V acceptTargetVisitor(BindingTargetVisitor<? super T, V> visitor) {
|
||||
checkState(factory.constructorInjector != null, "not initialized");
|
||||
if (factory.constructorInjector == null) {
|
||||
throw new IllegalStateException("not initialized");
|
||||
}
|
||||
return visitor.visit(this);
|
||||
}
|
||||
|
||||
@Override
|
||||
public InjectionPoint getConstructor() {
|
||||
checkState(factory.constructorInjector != null, "Binding is not ready");
|
||||
if (factory.constructorInjector == null) {
|
||||
throw new IllegalStateException("Binding is not ready");
|
||||
}
|
||||
return factory.constructorInjector.getConstructionProxy().getInjectionPoint();
|
||||
}
|
||||
|
||||
@Override
|
||||
public Set<InjectionPoint> getInjectableMembers() {
|
||||
checkState(factory.constructorInjector != null, "Binding is not ready");
|
||||
if (factory.constructorInjector == null) {
|
||||
throw new IllegalStateException("Binding is not ready");
|
||||
}
|
||||
return factory.constructorInjector.getInjectableMembers();
|
||||
}
|
||||
|
||||
|
@ -97,7 +101,9 @@ class ConstructorBindingImpl<T> extends BindingImpl<T> implements ConstructorBin
|
|||
@SuppressWarnings("unchecked")
|
||||
public T get(Errors errors, InternalContext context, Dependency<?> dependency)
|
||||
throws ErrorsException {
|
||||
checkState(constructorInjector != null, "Constructor not ready");
|
||||
if (constructorInjector == null) {
|
||||
throw new IllegalStateException("Constructor not ready");
|
||||
}
|
||||
|
||||
// This may not actually be safe because it could return a super type of T (if that's all the
|
||||
// client needs), but it should be OK in practice thanks to the wonders of erasure.
|
||||
|
|
|
@ -22,8 +22,6 @@ import org.elasticsearch.common.inject.spi.Message;
|
|||
|
||||
import java.util.Collection;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkArgument;
|
||||
|
||||
/**
|
||||
* Thrown when errors occur while creating a {@link Injector}. Includes a list of encountered
|
||||
* errors. Clients should catch this exception, log it, and stop execution.
|
||||
|
@ -39,7 +37,9 @@ public class CreationException extends RuntimeException {
|
|||
*/
|
||||
public CreationException(Collection<Message> messages) {
|
||||
this.messages = ImmutableSet.copyOf(messages);
|
||||
checkArgument(!this.messages.isEmpty());
|
||||
if (this.messages.isEmpty()) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
initCause(Errors.getOnlyCause(this.messages));
|
||||
}
|
||||
|
||||
|
|
|
@ -25,8 +25,6 @@ import java.util.ArrayList;
|
|||
import java.util.Collections;
|
||||
import java.util.List;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
|
||||
/**
|
||||
* @author jessewilson@google.com (Jesse Wilson)
|
||||
*/
|
||||
|
@ -61,7 +59,9 @@ final class EncounterImpl<T> implements TypeEncounter<T> {
|
|||
|
||||
@Override
|
||||
public void register(MembersInjector<? super T> membersInjector) {
|
||||
checkState(valid, "Encounters may not be used after hear() returns.");
|
||||
if (!valid) {
|
||||
throw new IllegalStateException("Encounters may not be used after hear() returns.");
|
||||
}
|
||||
|
||||
if (membersInjectors == null) {
|
||||
membersInjectors = new ArrayList<>();
|
||||
|
@ -72,7 +72,9 @@ final class EncounterImpl<T> implements TypeEncounter<T> {
|
|||
|
||||
@Override
|
||||
public void register(InjectionListener<? super T> injectionListener) {
|
||||
checkState(valid, "Encounters may not be used after hear() returns.");
|
||||
if (!valid) {
|
||||
throw new IllegalStateException("Encounters may not be used after hear() returns.");
|
||||
}
|
||||
|
||||
if (injectionListeners == null) {
|
||||
injectionListeners = new ArrayList<>();
|
||||
|
@ -83,25 +85,33 @@ final class EncounterImpl<T> implements TypeEncounter<T> {
|
|||
|
||||
@Override
|
||||
public void addError(String message, Object... arguments) {
|
||||
checkState(valid, "Encounters may not be used after hear() returns.");
|
||||
if (!valid) {
|
||||
throw new IllegalStateException("Encounters may not be used after hear() returns.");
|
||||
}
|
||||
errors.addMessage(message, arguments);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void addError(Throwable t) {
|
||||
checkState(valid, "Encounters may not be used after hear() returns.");
|
||||
if (!valid) {
|
||||
throw new IllegalStateException("Encounters may not be used after hear() returns.");
|
||||
}
|
||||
errors.errorInUserCode(t, "An exception was caught and reported. Message: %s", t.getMessage());
|
||||
}
|
||||
|
||||
@Override
|
||||
public void addError(Message message) {
|
||||
checkState(valid, "Encounters may not be used after hear() returns.");
|
||||
if (!valid) {
|
||||
throw new IllegalStateException("Encounters may not be used after hear() returns.");
|
||||
}
|
||||
errors.addMessage(message);
|
||||
}
|
||||
|
||||
@Override
|
||||
public <T> Provider<T> getProvider(Key<T> key) {
|
||||
checkState(valid, "Encounters may not be used after hear() returns.");
|
||||
if (!valid) {
|
||||
throw new IllegalStateException("Encounters may not be used after hear() returns.");
|
||||
}
|
||||
return lookups.getProvider(key);
|
||||
}
|
||||
|
||||
|
@ -112,7 +122,9 @@ final class EncounterImpl<T> implements TypeEncounter<T> {
|
|||
|
||||
@Override
|
||||
public <T> MembersInjector<T> getMembersInjector(TypeLiteral<T> typeLiteral) {
|
||||
checkState(valid, "Encounters may not be used after hear() returns.");
|
||||
if (!valid) {
|
||||
throw new IllegalStateException("Encounters may not be used after hear() returns.");
|
||||
}
|
||||
return lookups.getMembersInjector(typeLiteral);
|
||||
}
|
||||
|
||||
|
|
|
@ -17,9 +17,9 @@
|
|||
package org.elasticsearch.common.inject;
|
||||
|
||||
import com.google.common.collect.ImmutableSet;
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.elasticsearch.common.inject.internal.*;
|
||||
import org.elasticsearch.common.inject.spi.Dependency;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
|
||||
import java.util.Collection;
|
||||
import java.util.List;
|
||||
|
|
|
@ -17,28 +17,14 @@
|
|||
package org.elasticsearch.common.inject;
|
||||
|
||||
import com.google.common.collect.ImmutableSet;
|
||||
import org.elasticsearch.common.inject.internal.Errors;
|
||||
import org.elasticsearch.common.inject.internal.ErrorsException;
|
||||
import org.elasticsearch.common.inject.internal.InternalContext;
|
||||
import org.elasticsearch.common.inject.internal.InternalFactory;
|
||||
import org.elasticsearch.common.inject.internal.PrivateElementsImpl;
|
||||
import org.elasticsearch.common.inject.internal.ProviderInstanceBindingImpl;
|
||||
import org.elasticsearch.common.inject.internal.Scoping;
|
||||
import org.elasticsearch.common.inject.internal.SourceProvider;
|
||||
import org.elasticsearch.common.inject.internal.Stopwatch;
|
||||
import org.elasticsearch.common.inject.spi.Dependency;
|
||||
import org.elasticsearch.common.inject.spi.Element;
|
||||
import org.elasticsearch.common.inject.spi.Elements;
|
||||
import org.elasticsearch.common.inject.spi.InjectionPoint;
|
||||
import org.elasticsearch.common.inject.spi.PrivateElements;
|
||||
import org.elasticsearch.common.inject.spi.TypeListenerBinding;
|
||||
import org.elasticsearch.common.inject.internal.*;
|
||||
import org.elasticsearch.common.inject.spi.*;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Objects;
|
||||
import java.util.logging.Logger;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
import static org.elasticsearch.common.inject.Scopes.SINGLETON;
|
||||
|
||||
/**
|
||||
|
@ -125,9 +111,15 @@ class InjectorShell {
|
|||
*/
|
||||
List<InjectorShell> build(Initializer initializer, BindingProcessor bindingProcessor,
|
||||
Stopwatch stopwatch, Errors errors) {
|
||||
checkState(stage != null, "Stage not initialized");
|
||||
checkState(privateElements == null || parent != null, "PrivateElements with no parent");
|
||||
checkState(state != null, "no state. Did you remember to lock() ?");
|
||||
if (stage == null) {
|
||||
throw new IllegalStateException("Stage not initialized");
|
||||
}
|
||||
if (privateElements != null && parent == null) {
|
||||
throw new IllegalStateException("PrivateElements with no parent");
|
||||
}
|
||||
if (state == null) {
|
||||
throw new IllegalStateException("no state. Did you remember to lock() ?");
|
||||
}
|
||||
|
||||
InjectorImpl injector = new InjectorImpl(parent, state, initializer);
|
||||
if (privateElements != null) {
|
||||
|
|
|
@ -24,8 +24,6 @@ import java.lang.annotation.Annotation;
|
|||
import java.lang.reflect.Type;
|
||||
import java.util.Objects;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkArgument;
|
||||
|
||||
/**
|
||||
* Binding key consisting of an injection type and an optional annotation.
|
||||
* Matches the type and annotation at a point of injection.
|
||||
|
@ -367,16 +365,20 @@ public class Key<T> {
|
|||
|
||||
private static void ensureRetainedAtRuntime(
|
||||
Class<? extends Annotation> annotationType) {
|
||||
checkArgument(Annotations.isRetainedAtRuntime(annotationType),
|
||||
"%s is not retained at runtime. Please annotate it with @Retention(RUNTIME).",
|
||||
annotationType.getName());
|
||||
if (!Annotations.isRetainedAtRuntime(annotationType)) {
|
||||
throw new IllegalArgumentException(
|
||||
annotationType.getName() + " is not retained at runtime. Please annotate it with @Retention(RUNTIME)."
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
private static void ensureIsBindingAnnotation(
|
||||
Class<? extends Annotation> annotationType) {
|
||||
checkArgument(isBindingAnnotation(annotationType),
|
||||
"%s is not a binding annotation. Please annotate it with @BindingAnnotation.",
|
||||
annotationType.getName());
|
||||
if (!isBindingAnnotation(annotationType)) {
|
||||
throw new IllegalArgumentException(
|
||||
annotationType.getName() + " is not a binding annotation. Please annotate it with @BindingAnnotation."
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
static enum NullAnnotationStrategy implements AnnotationStrategy {
|
||||
|
|
|
@ -27,8 +27,6 @@ import org.elasticsearch.common.inject.spi.TypeListener;
|
|||
|
||||
import java.lang.annotation.Annotation;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
|
||||
/**
|
||||
* A module whose configuration information is hidden from its environment by default. Only bindings
|
||||
* that are explicitly exposed will be available to other modules and to the users of the injector.
|
||||
|
@ -93,7 +91,9 @@ public abstract class PrivateModule implements Module {
|
|||
|
||||
@Override
|
||||
public final synchronized void configure(Binder binder) {
|
||||
checkState(this.binder == null, "Re-entry is not allowed.");
|
||||
if (this.binder != null) {
|
||||
throw new IllegalStateException("Re-entry is not allowed.");
|
||||
}
|
||||
|
||||
// Guice treats PrivateModules specially and passes in a PrivateBinder automatically.
|
||||
this.binder = (PrivateBinder) binder.skipSources(PrivateModule.class);
|
||||
|
|
|
@ -23,8 +23,6 @@ import org.elasticsearch.common.inject.spi.Message;
|
|||
import java.util.Collection;
|
||||
import java.util.Collections;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkArgument;
|
||||
|
||||
/**
|
||||
* Indicates that there was a runtime failure while providing an instance.
|
||||
*
|
||||
|
@ -41,7 +39,9 @@ public final class ProvisionException extends RuntimeException {
|
|||
*/
|
||||
public ProvisionException(Iterable<Message> messages) {
|
||||
this.messages = ImmutableSet.copyOf(messages);
|
||||
checkArgument(!this.messages.isEmpty());
|
||||
if (this.messages.isEmpty()) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
initCause(Errors.getOnlyCause(this.messages));
|
||||
}
|
||||
|
||||
|
|
|
@ -24,7 +24,6 @@ import java.util.Arrays;
|
|||
import java.util.List;
|
||||
import java.util.Objects;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkArgument;
|
||||
import static org.elasticsearch.common.inject.internal.MoreTypes.canonicalize;
|
||||
|
||||
/**
|
||||
|
@ -257,8 +256,9 @@ public class TypeLiteral<T> {
|
|||
* @since 2.0
|
||||
*/
|
||||
public TypeLiteral<?> getSupertype(Class<?> supertype) {
|
||||
checkArgument(supertype.isAssignableFrom(rawType),
|
||||
"%s is not a supertype of %s", supertype, this.type);
|
||||
if (!supertype.isAssignableFrom(rawType)) {
|
||||
throw new IllegalArgumentException(supertype + " is not a supertype of " + type);
|
||||
}
|
||||
return resolve(MoreTypes.getGenericSupertype(type, rawType, supertype));
|
||||
}
|
||||
|
||||
|
@ -269,8 +269,9 @@ public class TypeLiteral<T> {
|
|||
* @since 2.0
|
||||
*/
|
||||
public TypeLiteral<?> getFieldType(Field field) {
|
||||
checkArgument(field.getDeclaringClass().isAssignableFrom(rawType),
|
||||
"%s is not defined by a supertype of %s", field, type);
|
||||
if (!field.getDeclaringClass().isAssignableFrom(rawType)) {
|
||||
throw new IllegalArgumentException(field + " is not defined by a supertype of " + type);
|
||||
}
|
||||
return resolve(field.getGenericType());
|
||||
}
|
||||
|
||||
|
@ -285,14 +286,17 @@ public class TypeLiteral<T> {
|
|||
|
||||
if (methodOrConstructor instanceof Method) {
|
||||
Method method = (Method) methodOrConstructor;
|
||||
checkArgument(method.getDeclaringClass().isAssignableFrom(rawType),
|
||||
"%s is not defined by a supertype of %s", method, type);
|
||||
if (!method.getDeclaringClass().isAssignableFrom(rawType)) {
|
||||
throw new IllegalArgumentException(method + " is not defined by a supertype of " + type);
|
||||
}
|
||||
genericParameterTypes = method.getGenericParameterTypes();
|
||||
|
||||
} else if (methodOrConstructor instanceof Constructor) {
|
||||
Constructor constructor = (Constructor) methodOrConstructor;
|
||||
checkArgument(constructor.getDeclaringClass().isAssignableFrom(rawType),
|
||||
"%s does not construct a supertype of %s", constructor, type);
|
||||
if (!constructor.getDeclaringClass().isAssignableFrom(rawType)) {
|
||||
throw new IllegalArgumentException(constructor + " does not construct a supertype of " + type);
|
||||
}
|
||||
|
||||
genericParameterTypes = constructor.getGenericParameterTypes();
|
||||
|
||||
} else {
|
||||
|
@ -313,14 +317,17 @@ public class TypeLiteral<T> {
|
|||
|
||||
if (methodOrConstructor instanceof Method) {
|
||||
Method method = (Method) methodOrConstructor;
|
||||
checkArgument(method.getDeclaringClass().isAssignableFrom(rawType),
|
||||
"%s is not defined by a supertype of %s", method, type);
|
||||
if (!method.getDeclaringClass().isAssignableFrom(rawType)) {
|
||||
throw new IllegalArgumentException(method + " is not defined by a supertype of " + type);
|
||||
}
|
||||
|
||||
genericExceptionTypes = method.getGenericExceptionTypes();
|
||||
|
||||
} else if (methodOrConstructor instanceof Constructor) {
|
||||
Constructor<?> constructor = (Constructor<?>) methodOrConstructor;
|
||||
checkArgument(constructor.getDeclaringClass().isAssignableFrom(rawType),
|
||||
"%s does not construct a supertype of %s", constructor, type);
|
||||
if (!constructor.getDeclaringClass().isAssignableFrom(rawType)) {
|
||||
throw new IllegalArgumentException(constructor + " does not construct a supertype of " + type);
|
||||
}
|
||||
genericExceptionTypes = constructor.getGenericExceptionTypes();
|
||||
|
||||
} else {
|
||||
|
@ -337,8 +344,10 @@ public class TypeLiteral<T> {
|
|||
* @since 2.0
|
||||
*/
|
||||
public TypeLiteral<?> getReturnType(Method method) {
|
||||
checkArgument(method.getDeclaringClass().isAssignableFrom(rawType),
|
||||
"%s is not defined by a supertype of %s", method, type);
|
||||
if (!method.getDeclaringClass().isAssignableFrom(rawType)) {
|
||||
throw new IllegalArgumentException(method + " is not defined by a supertype of " + type);
|
||||
}
|
||||
|
||||
return resolve(method.getGenericReturnType());
|
||||
}
|
||||
}
|
||||
|
|
|
@ -17,18 +17,7 @@
|
|||
package org.elasticsearch.common.inject.assistedinject;
|
||||
|
||||
import com.google.common.collect.ImmutableMap;
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.elasticsearch.common.inject.AbstractModule;
|
||||
import org.elasticsearch.common.inject.Binder;
|
||||
import org.elasticsearch.common.inject.Binding;
|
||||
import org.elasticsearch.common.inject.ConfigurationException;
|
||||
import org.elasticsearch.common.inject.Inject;
|
||||
import org.elasticsearch.common.inject.Injector;
|
||||
import org.elasticsearch.common.inject.Key;
|
||||
import org.elasticsearch.common.inject.Module;
|
||||
import org.elasticsearch.common.inject.Provider;
|
||||
import org.elasticsearch.common.inject.ProvisionException;
|
||||
import org.elasticsearch.common.inject.TypeLiteral;
|
||||
import org.elasticsearch.common.inject.*;
|
||||
import org.elasticsearch.common.inject.internal.Errors;
|
||||
import org.elasticsearch.common.inject.internal.ErrorsException;
|
||||
import org.elasticsearch.common.inject.spi.Message;
|
||||
|
@ -43,7 +32,6 @@ import java.util.Arrays;
|
|||
import java.util.Collections;
|
||||
import java.util.List;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
import static org.elasticsearch.common.inject.internal.Annotations.getKey;
|
||||
|
||||
/**
|
||||
|
@ -192,8 +180,9 @@ public final class FactoryProvider2<F> implements InvocationHandler, Provider<F>
|
|||
* Creates a child injector that binds the args, and returns the binding for the method's result.
|
||||
*/
|
||||
public Binding<?> getBindingFromNewInjector(final Method method, final Object[] args) {
|
||||
checkState(injector != null,
|
||||
"Factories.create() factories cannot be used until they're initialized by Guice.");
|
||||
if (injector == null) {
|
||||
throw new IllegalStateException("Factories.create() factories cannot be used until they're initialized by Guice.");
|
||||
}
|
||||
|
||||
final Key<?> returnType = returnTypesByMethod.get(method);
|
||||
|
||||
|
@ -237,7 +226,7 @@ public final class FactoryProvider2<F> implements InvocationHandler, Provider<F>
|
|||
} catch (ProvisionException e) {
|
||||
// if this is an exception declared by the factory method, throw it as-is
|
||||
if (e.getErrorMessages().size() == 1) {
|
||||
Message onlyError = Iterables.getOnlyElement(e.getErrorMessages());
|
||||
Message onlyError = e.getErrorMessages().iterator().next();
|
||||
Throwable cause = onlyError.getCause();
|
||||
if (cause != null && canRethrow(method, cause)) {
|
||||
throw cause;
|
||||
|
|
|
@ -22,8 +22,6 @@ import java.lang.annotation.Annotation;
|
|||
import java.lang.reflect.ParameterizedType;
|
||||
import java.lang.reflect.Type;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkArgument;
|
||||
|
||||
/**
|
||||
* Models a method or constructor parameter.
|
||||
*
|
||||
|
@ -144,8 +142,9 @@ class Parameter {
|
|||
Annotation bindingAnnotation = null;
|
||||
for (Annotation a : annotations) {
|
||||
if (a.annotationType().getAnnotation(BindingAnnotation.class) != null) {
|
||||
checkArgument(bindingAnnotation == null,
|
||||
"Parameter has multiple binding annotations: %s and %s", bindingAnnotation, a);
|
||||
if (bindingAnnotation != null) {
|
||||
throw new IllegalArgumentException("Parameter has multiple binding annotations: " + bindingAnnotation + " and " + a);
|
||||
}
|
||||
bindingAnnotation = a;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -30,8 +30,6 @@ import java.util.Map;
|
|||
import java.util.NoSuchElementException;
|
||||
import java.util.Objects;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkArgument;
|
||||
|
||||
/**
|
||||
* Static methods for working with types that we aren't publishing in the
|
||||
* public {@code Types} API.
|
||||
|
@ -152,8 +150,11 @@ public class MoreTypes {
|
|||
// Neal isn't either but suspects some pathological case related
|
||||
// to nested classes exists.
|
||||
Type rawType = parameterizedType.getRawType();
|
||||
checkArgument(rawType instanceof Class,
|
||||
"Expected a Class, but <%s> is of type %s", type, type.getClass().getName());
|
||||
if (!(rawType instanceof Class)) {
|
||||
throw new IllegalArgumentException(
|
||||
"Expected a Class, but <" + type +"> is of type " + type.getClass().getName()
|
||||
);
|
||||
}
|
||||
return (Class<?>) rawType;
|
||||
|
||||
} else if (type instanceof GenericArrayType) {
|
||||
|
@ -445,10 +446,13 @@ public class MoreTypes {
|
|||
// require an owner type if the raw type needs it
|
||||
if (rawType instanceof Class<?>) {
|
||||
Class rawTypeAsClass = (Class) rawType;
|
||||
checkArgument(ownerType != null || rawTypeAsClass.getEnclosingClass() == null,
|
||||
"No owner type for enclosed %s", rawType);
|
||||
checkArgument(ownerType == null || rawTypeAsClass.getEnclosingClass() != null,
|
||||
"Owner type for unenclosed %s", rawType);
|
||||
if (ownerType == null && rawTypeAsClass.getEnclosingClass() != null) {
|
||||
throw new IllegalArgumentException("No owner type for enclosed " + rawType);
|
||||
}
|
||||
if (ownerType != null && rawTypeAsClass.getEnclosingClass() == null) {
|
||||
throw new IllegalArgumentException("Owner type for unenclosed " + rawType);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
this.ownerType = ownerType == null ? null : canonicalize(ownerType);
|
||||
|
@ -561,13 +565,18 @@ public class MoreTypes {
|
|||
private final Type lowerBound;
|
||||
|
||||
public WildcardTypeImpl(Type[] upperBounds, Type[] lowerBounds) {
|
||||
checkArgument(lowerBounds.length <= 1, "Must have at most one lower bound.");
|
||||
checkArgument(upperBounds.length == 1, "Must have exactly one upper bound.");
|
||||
|
||||
if (lowerBounds.length > 1) {
|
||||
throw new IllegalArgumentException("Must have at most one lower bound.");
|
||||
}
|
||||
if (upperBounds.length != 1) {
|
||||
throw new IllegalArgumentException("Must have exactly one upper bound.");
|
||||
}
|
||||
if (lowerBounds.length == 1) {
|
||||
Objects.requireNonNull(lowerBounds[0], "lowerBound");
|
||||
checkNotPrimitive(lowerBounds[0], "wildcard bounds");
|
||||
checkArgument(upperBounds[0] == Object.class, "bounded both ways");
|
||||
if (upperBounds[0] != Object.class) {
|
||||
throw new IllegalArgumentException("bounded both ways");
|
||||
}
|
||||
this.lowerBound = canonicalize(lowerBounds[0]);
|
||||
this.upperBound = Object.class;
|
||||
|
||||
|
@ -615,8 +624,9 @@ public class MoreTypes {
|
|||
}
|
||||
|
||||
private static void checkNotPrimitive(Type type, String use) {
|
||||
checkArgument(!(type instanceof Class<?>) || !((Class) type).isPrimitive(),
|
||||
"Primitive types are not allowed in %s: %s", use, type);
|
||||
if (type instanceof Class<?> && ((Class) type).isPrimitive()) {
|
||||
throw new IllegalArgumentException("Primitive types are not allowed in " + use + ": " + type);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -25,16 +25,7 @@ import org.elasticsearch.common.inject.spi.Element;
|
|||
import org.elasticsearch.common.inject.spi.ElementVisitor;
|
||||
import org.elasticsearch.common.inject.spi.PrivateElements;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.LinkedHashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Objects;
|
||||
import java.util.Set;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkArgument;
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
import java.util.*;
|
||||
|
||||
/**
|
||||
* @author jessewilson@google.com (Jesse Wilson)
|
||||
|
@ -88,7 +79,9 @@ public final class PrivateElementsImpl implements PrivateElements {
|
|||
}
|
||||
|
||||
public void initInjector(Injector injector) {
|
||||
checkState(this.injector == null, "injector already initialized");
|
||||
if (this.injector != null) {
|
||||
throw new IllegalStateException("injector already initialized");
|
||||
}
|
||||
this.injector = Objects.requireNonNull(injector, "injector");
|
||||
}
|
||||
|
||||
|
@ -137,7 +130,9 @@ public final class PrivateElementsImpl implements PrivateElements {
|
|||
public Object getExposedSource(Key<?> key) {
|
||||
getExposedKeys(); // ensure exposedKeysToSources is populated
|
||||
Object source = exposedKeysToSources.get(key);
|
||||
checkArgument(source != null, "%s not exposed by %s.", key, this);
|
||||
if (source == null) {
|
||||
throw new IllegalArgumentException(key + " not exposed by " + ".");
|
||||
}
|
||||
return source;
|
||||
}
|
||||
|
||||
|
|
|
@ -17,7 +17,7 @@
|
|||
package org.elasticsearch.common.inject.internal;
|
||||
|
||||
import com.google.common.collect.ImmutableSet;
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
|
|
@ -24,8 +24,6 @@ import java.lang.reflect.AnnotatedElement;
|
|||
import java.lang.reflect.Method;
|
||||
import java.util.Objects;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkArgument;
|
||||
|
||||
/**
|
||||
* Matcher implementations. Supports matching classes and methods.
|
||||
*
|
||||
|
@ -103,8 +101,9 @@ public class Matchers {
|
|||
private static void checkForRuntimeRetention(
|
||||
Class<? extends Annotation> annotationType) {
|
||||
Retention retention = annotationType.getAnnotation(Retention.class);
|
||||
checkArgument(retention != null && retention.value() == RetentionPolicy.RUNTIME,
|
||||
"Annotation " + annotationType.getSimpleName() + " is missing RUNTIME retention");
|
||||
if (retention == null || retention.value() != RetentionPolicy.RUNTIME) {
|
||||
throw new IllegalArgumentException("Annotation " + annotationType.getSimpleName() + " is missing RUNTIME retention");
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -16,43 +16,17 @@
|
|||
|
||||
package org.elasticsearch.common.inject.spi;
|
||||
|
||||
import org.elasticsearch.common.inject.AbstractModule;
|
||||
import org.elasticsearch.common.inject.Binder;
|
||||
import org.elasticsearch.common.inject.Binding;
|
||||
import org.elasticsearch.common.inject.Key;
|
||||
import org.elasticsearch.common.inject.MembersInjector;
|
||||
import org.elasticsearch.common.inject.Module;
|
||||
import org.elasticsearch.common.inject.PrivateBinder;
|
||||
import org.elasticsearch.common.inject.PrivateModule;
|
||||
import org.elasticsearch.common.inject.Provider;
|
||||
import org.elasticsearch.common.inject.Scope;
|
||||
import org.elasticsearch.common.inject.Stage;
|
||||
import org.elasticsearch.common.inject.TypeLiteral;
|
||||
import org.elasticsearch.common.inject.*;
|
||||
import org.elasticsearch.common.inject.binder.AnnotatedBindingBuilder;
|
||||
import org.elasticsearch.common.inject.binder.AnnotatedConstantBindingBuilder;
|
||||
import org.elasticsearch.common.inject.binder.AnnotatedElementBuilder;
|
||||
import org.elasticsearch.common.inject.internal.AbstractBindingBuilder;
|
||||
import org.elasticsearch.common.inject.internal.BindingBuilder;
|
||||
import org.elasticsearch.common.inject.internal.ConstantBindingBuilderImpl;
|
||||
import org.elasticsearch.common.inject.internal.Errors;
|
||||
import org.elasticsearch.common.inject.internal.ExposureBuilder;
|
||||
import org.elasticsearch.common.inject.internal.PrivateElementsImpl;
|
||||
import org.elasticsearch.common.inject.internal.ProviderMethodsModule;
|
||||
import org.elasticsearch.common.inject.internal.SourceProvider;
|
||||
import org.elasticsearch.common.inject.internal.*;
|
||||
import org.elasticsearch.common.inject.matcher.Matcher;
|
||||
import org.elasticsearch.common.logging.ESLogger;
|
||||
import org.elasticsearch.common.logging.Loggers;
|
||||
|
||||
import java.lang.annotation.Annotation;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
import java.util.Collections;
|
||||
import java.util.HashSet;
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkArgument;
|
||||
import java.util.*;
|
||||
|
||||
/**
|
||||
* Exposes elements of a module so they can be inspected, validated or {@link
|
||||
|
@ -156,7 +130,9 @@ public final class Elements {
|
|||
*/
|
||||
private RecordingBinder(
|
||||
RecordingBinder prototype, Object source, SourceProvider sourceProvider) {
|
||||
checkArgument(source == null ^ sourceProvider == null);
|
||||
if (!(source == null ^ sourceProvider == null)) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
|
||||
this.stage = prototype.stage;
|
||||
this.modules = prototype.modules;
|
||||
|
|
|
@ -22,8 +22,6 @@ import org.elasticsearch.common.inject.TypeLiteral;
|
|||
|
||||
import java.util.Objects;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
|
||||
/**
|
||||
* A lookup of the members injector for a type. Lookups are created explicitly in a module using
|
||||
* {@link org.elasticsearch.common.inject.Binder#getMembersInjector(Class) getMembersInjector()} statements:
|
||||
|
@ -68,7 +66,9 @@ public final class MembersInjectorLookup<T> implements Element {
|
|||
* @throws IllegalStateException if the delegate is already set
|
||||
*/
|
||||
public void initializeDelegate(MembersInjector<T> delegate) {
|
||||
checkState(this.delegate == null, "delegate already initialized");
|
||||
if (this.delegate != null) {
|
||||
throw new IllegalStateException("delegate already initialized");
|
||||
}
|
||||
this.delegate = Objects.requireNonNull(delegate, "delegate");
|
||||
}
|
||||
|
||||
|
@ -95,8 +95,9 @@ public final class MembersInjectorLookup<T> implements Element {
|
|||
return new MembersInjector<T>() {
|
||||
@Override
|
||||
public void injectMembers(T instance) {
|
||||
checkState(delegate != null,
|
||||
"This MembersInjector cannot be used until the Injector has been created.");
|
||||
if (delegate == null) {
|
||||
throw new IllegalStateException("This MembersInjector cannot be used until the Injector has been created.");
|
||||
}
|
||||
delegate.injectMembers(instance);
|
||||
}
|
||||
|
||||
|
|
|
@ -22,8 +22,6 @@ import org.elasticsearch.common.inject.Provider;
|
|||
|
||||
import java.util.Objects;
|
||||
|
||||
import static com.google.common.base.Preconditions.checkState;
|
||||
|
||||
/**
|
||||
* A lookup of the provider for a type. Lookups are created explicitly in a module using
|
||||
* {@link org.elasticsearch.common.inject.Binder#getProvider(Class) getProvider()} statements:
|
||||
|
@ -46,8 +44,9 @@ public final class ProviderLookup<T> implements Element {
|
|||
|
||||
@Override
|
||||
public T get() {
|
||||
checkState(lookup.delegate != null,
|
||||
"This Provider cannot be used until the Injector has been created.");
|
||||
if (lookup.delegate == null) {
|
||||
throw new IllegalStateException( "This Provider cannot be used until the Injector has been created.");
|
||||
}
|
||||
return lookup.delegate.get();
|
||||
}
|
||||
|
||||
|
@ -89,7 +88,9 @@ public final class ProviderLookup<T> implements Element {
|
|||
* @throws IllegalStateException if the delegate is already set
|
||||
*/
|
||||
public void initializeDelegate(Provider<T> delegate) {
|
||||
checkState(this.delegate == null, "delegate already initialized");
|
||||
if (this.delegate != null) {
|
||||
throw new IllegalStateException("delegate already initialized");
|
||||
}
|
||||
this.delegate = Objects.requireNonNull(delegate, "delegate");
|
||||
}
|
||||
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.common.lucene;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.apache.lucene.analysis.core.KeywordAnalyzer;
|
||||
import org.apache.lucene.analysis.standard.StandardAnalyzer;
|
||||
import org.apache.lucene.codecs.CodecUtil;
|
||||
|
@ -27,7 +26,10 @@ import org.apache.lucene.codecs.DocValuesFormat;
|
|||
import org.apache.lucene.codecs.PostingsFormat;
|
||||
import org.apache.lucene.index.*;
|
||||
import org.apache.lucene.search.*;
|
||||
import org.apache.lucene.store.*;
|
||||
import org.apache.lucene.store.Directory;
|
||||
import org.apache.lucene.store.IOContext;
|
||||
import org.apache.lucene.store.IndexInput;
|
||||
import org.apache.lucene.store.Lock;
|
||||
import org.apache.lucene.util.Bits;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.Counter;
|
||||
|
@ -40,6 +42,7 @@ import org.elasticsearch.common.SuppressForbidden;
|
|||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||
import org.elasticsearch.common.logging.ESLogger;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.index.analysis.AnalyzerScope;
|
||||
import org.elasticsearch.index.analysis.NamedAnalyzer;
|
||||
import org.elasticsearch.index.fielddata.IndexFieldData;
|
||||
|
@ -107,7 +110,7 @@ public class Lucene {
|
|||
for (SegmentCommitInfo info : infos) {
|
||||
list.add(info.files());
|
||||
}
|
||||
return Iterables.concat(list);
|
||||
return Iterables.flatten(list);
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -19,19 +19,11 @@
|
|||
|
||||
package org.elasticsearch.common.lucene;
|
||||
|
||||
import com.google.common.collect.HashMultimap;
|
||||
import com.google.common.collect.ImmutableSet;
|
||||
import com.google.common.collect.Multimap;
|
||||
|
||||
import org.apache.lucene.index.LeafReader;
|
||||
import org.apache.lucene.index.LeafReader.CoreClosedListener;
|
||||
import org.elasticsearch.index.shard.ShardId;
|
||||
import org.elasticsearch.index.shard.ShardUtils;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.IdentityHashMap;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import java.util.*;
|
||||
|
||||
/**
|
||||
* A map between segment core cache keys and the shard that these segments
|
||||
|
@ -47,16 +39,16 @@ import java.util.Set;
|
|||
public final class ShardCoreKeyMap {
|
||||
|
||||
private final Map<Object, ShardId> coreKeyToShard;
|
||||
private final Multimap<String, Object> indexToCoreKey;
|
||||
private final Map<String, Set<Object>> indexToCoreKey;
|
||||
|
||||
public ShardCoreKeyMap() {
|
||||
coreKeyToShard = new IdentityHashMap<>();
|
||||
indexToCoreKey = HashMultimap.create();
|
||||
indexToCoreKey = new HashMap<>();
|
||||
}
|
||||
|
||||
/**
|
||||
* Register a {@link LeafReader}. This is necessary so that the core cache
|
||||
* key of this reader can be found later using {@link #getCoreCacheKeys(ShardId)}.
|
||||
* key of this reader can be found later using {@link #getCoreKeysForIndex(String)}.
|
||||
*/
|
||||
public void add(LeafReader reader) {
|
||||
final ShardId shardId = ShardUtils.extractShardId(reader);
|
||||
|
@ -67,15 +59,22 @@ public final class ShardCoreKeyMap {
|
|||
final String index = shardId.getIndex();
|
||||
synchronized (this) {
|
||||
if (coreKeyToShard.put(coreKey, shardId) == null) {
|
||||
final boolean added = indexToCoreKey.put(index, coreKey);
|
||||
Set<Object> objects = indexToCoreKey.get(index);
|
||||
if (objects == null) {
|
||||
objects = new HashSet<>();
|
||||
indexToCoreKey.put(index, objects);
|
||||
}
|
||||
final boolean added = objects.add(coreKey);
|
||||
assert added;
|
||||
reader.addCoreClosedListener(new CoreClosedListener() {
|
||||
@Override
|
||||
public void onClose(Object ownerCoreCacheKey) throws IOException {
|
||||
assert coreKey == ownerCoreCacheKey;
|
||||
synchronized (ShardCoreKeyMap.this) {
|
||||
coreKeyToShard.remove(ownerCoreCacheKey);
|
||||
indexToCoreKey.remove(index, coreKey);
|
||||
reader.addCoreClosedListener(ownerCoreCacheKey -> {
|
||||
assert coreKey == ownerCoreCacheKey;
|
||||
synchronized (ShardCoreKeyMap.this) {
|
||||
coreKeyToShard.remove(ownerCoreCacheKey);
|
||||
final Set<Object> coreKeys = indexToCoreKey.get(index);
|
||||
final boolean removed = coreKeys.remove(coreKey);
|
||||
assert removed;
|
||||
if (coreKeys.isEmpty()) {
|
||||
indexToCoreKey.remove(index);
|
||||
}
|
||||
}
|
||||
});
|
||||
|
@ -95,15 +94,35 @@ public final class ShardCoreKeyMap {
|
|||
* Get the set of core cache keys associated with the given index.
|
||||
*/
|
||||
public synchronized Set<Object> getCoreKeysForIndex(String index) {
|
||||
return ImmutableSet.copyOf(indexToCoreKey.get(index));
|
||||
final Set<Object> objects = indexToCoreKey.get(index);
|
||||
if (objects == null) {
|
||||
return Collections.emptySet();
|
||||
}
|
||||
// we have to copy otherwise we risk ConcurrentModificationException
|
||||
return Collections.unmodifiableSet(new HashSet<>(objects));
|
||||
}
|
||||
|
||||
/**
|
||||
* Return the number of tracked segments.
|
||||
*/
|
||||
public synchronized int size() {
|
||||
assert indexToCoreKey.size() == coreKeyToShard.size();
|
||||
assert assertSize();
|
||||
return coreKeyToShard.size();
|
||||
}
|
||||
|
||||
private synchronized boolean assertSize() {
|
||||
// this is heavy and should only used in assertions
|
||||
boolean assertionsEnabled = false;
|
||||
assert assertionsEnabled = true;
|
||||
if (assertionsEnabled == false) {
|
||||
throw new AssertionError("only run this if assertions are enabled");
|
||||
}
|
||||
Collection<Set<Object>> values = indexToCoreKey.values();
|
||||
int size = 0;
|
||||
for (Set<Object> value : values) {
|
||||
size += value.size();
|
||||
}
|
||||
return size == coreKeyToShard.size();
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -18,9 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.common.unit;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.lucene.search.FuzzyQuery;
|
||||
import org.apache.lucene.util.automaton.LevenshteinAutomata;
|
||||
import org.elasticsearch.common.ParseField;
|
||||
import org.elasticsearch.common.xcontent.ToXContent;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
|
@ -46,7 +43,9 @@ public final class Fuzziness implements ToXContent {
|
|||
private final String fuzziness;
|
||||
|
||||
private Fuzziness(int fuzziness) {
|
||||
Preconditions.checkArgument(fuzziness >= 0 && fuzziness <= 2, "Valid edit distances are [0, 1, 2] but was [" + fuzziness + "]");
|
||||
if (fuzziness != 0 && fuzziness != 1 && fuzziness != 2) {
|
||||
throw new IllegalArgumentException("Valid edit distances are [0, 1, 2] but was [" + fuzziness + "]");
|
||||
}
|
||||
this.fuzziness = Integer.toString(fuzziness);
|
||||
}
|
||||
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.common.unit;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.elasticsearch.ElasticsearchParseException;
|
||||
import org.elasticsearch.common.Strings;
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
|
@ -46,7 +45,9 @@ public class SizeValue implements Streamable {
|
|||
}
|
||||
|
||||
public SizeValue(long size, SizeUnit sizeUnit) {
|
||||
Preconditions.checkArgument(size >= 0, "size in SizeValue may not be negative");
|
||||
if (size < 0) {
|
||||
throw new IllegalArgumentException("size in SizeValue may not be negative");
|
||||
}
|
||||
this.size = size;
|
||||
this.sizeUnit = sizeUnit;
|
||||
}
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.RamUsageEstimator;
|
||||
import org.elasticsearch.cache.recycler.PageCacheRecycler;
|
||||
|
@ -42,8 +41,12 @@ abstract class AbstractBigArray extends AbstractArray {
|
|||
protected AbstractBigArray(int pageSize, BigArrays bigArrays, boolean clearOnResize) {
|
||||
super(bigArrays, clearOnResize);
|
||||
this.recycler = bigArrays.recycler;
|
||||
Preconditions.checkArgument(pageSize >= 128, "pageSize must be >= 128");
|
||||
Preconditions.checkArgument((pageSize & (pageSize - 1)) == 0, "pageSize must be a power of two");
|
||||
if (pageSize < 128) {
|
||||
throw new IllegalArgumentException("pageSize must be >= 128");
|
||||
}
|
||||
if ((pageSize & (pageSize - 1)) != 0) {
|
||||
throw new IllegalArgumentException("pageSize must be a power of two");
|
||||
}
|
||||
this.pageShift = Integer.numberOfTrailingZeros(pageSize);
|
||||
this.pageMask = pageSize - 1;
|
||||
size = 0;
|
||||
|
@ -56,7 +59,9 @@ abstract class AbstractBigArray extends AbstractArray {
|
|||
|
||||
final int numPages(long capacity) {
|
||||
final long numPages = (capacity + pageMask) >>> pageShift;
|
||||
Preconditions.checkArgument(numPages <= Integer.MAX_VALUE, "pageSize=" + (pageMask + 1) + " is too small for such as capacity: " + capacity);
|
||||
if (numPages > Integer.MAX_VALUE) {
|
||||
throw new IllegalArgumentException("pageSize=" + (pageMask + 1) + " is too small for such as capacity: " + capacity);
|
||||
}
|
||||
return (int) numPages;
|
||||
}
|
||||
|
||||
|
|
|
@ -20,7 +20,6 @@
|
|||
package org.elasticsearch.common.util;
|
||||
|
||||
import com.carrotsearch.hppc.BitMixer;
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.elasticsearch.common.lease.Releasable;
|
||||
|
||||
/**
|
||||
|
@ -48,8 +47,12 @@ abstract class AbstractPagedHashMap implements Releasable {
|
|||
long mask;
|
||||
|
||||
AbstractPagedHashMap(long capacity, float maxLoadFactor, BigArrays bigArrays) {
|
||||
Preconditions.checkArgument(capacity >= 0, "capacity must be >= 0");
|
||||
Preconditions.checkArgument(maxLoadFactor > 0 && maxLoadFactor < 1, "maxLoadFactor must be > 0 and < 1");
|
||||
if (capacity < 0) {
|
||||
throw new IllegalArgumentException("capacity must be >= 0");
|
||||
}
|
||||
if (maxLoadFactor <= 0 || maxLoadFactor >= 1) {
|
||||
throw new IllegalArgumentException("maxLoadFactor must be > 0 and < 1");
|
||||
}
|
||||
this.bigArrays = bigArrays;
|
||||
this.maxLoadFactor = maxLoadFactor;
|
||||
long buckets = 1L + (long) (capacity / maxLoadFactor);
|
||||
|
|
|
@ -19,6 +19,7 @@
|
|||
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import java.lang.reflect.Array;
|
||||
import java.util.Arrays;
|
||||
|
||||
/**
|
||||
|
@ -69,4 +70,22 @@ public class ArrayUtils {
|
|||
}
|
||||
return index;
|
||||
}
|
||||
|
||||
/**
|
||||
* Concatenates 2 arrays
|
||||
*/
|
||||
public static String[] concat(String[] one, String[] other) {
|
||||
return concat(one, other, String.class);
|
||||
}
|
||||
|
||||
/**
|
||||
* Concatenates 2 arrays
|
||||
*/
|
||||
public static <T> T[] concat(T[] one, T[] other, Class<T> clazz) {
|
||||
T[] target = (T[]) Array.newInstance(clazz, one.length + other.length);
|
||||
System.arraycopy(one, 0, target, 0, one.length);
|
||||
System.arraycopy(other, 0, target, one.length, other.length);
|
||||
return target;
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.RamUsageEstimator;
|
||||
|
@ -55,9 +54,15 @@ public class BigArrays {
|
|||
/** Return the next size to grow to that is >= <code>minTargetSize</code>.
|
||||
* Inspired from {@link ArrayUtil#oversize(int, int)} and adapted to play nicely with paging. */
|
||||
public static long overSize(long minTargetSize, int pageSize, int bytesPerElement) {
|
||||
Preconditions.checkArgument(minTargetSize >= 0, "minTargetSize must be >= 0");
|
||||
Preconditions.checkArgument(pageSize >= 0, "pageSize must be > 0");
|
||||
Preconditions.checkArgument(bytesPerElement > 0, "bytesPerElement must be > 0");
|
||||
if (minTargetSize < 0) {
|
||||
throw new IllegalArgumentException("minTargetSize must be >= 0");
|
||||
}
|
||||
if (pageSize < 0) {
|
||||
throw new IllegalArgumentException("pageSize must be > 0");
|
||||
}
|
||||
if (bytesPerElement <= 0) {
|
||||
throw new IllegalArgumentException("bytesPerElement must be > 0");
|
||||
}
|
||||
|
||||
long newSize;
|
||||
if (minTargetSize < pageSize) {
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.RamUsageEstimator;
|
||||
|
@ -110,7 +109,9 @@ final class BigByteArray extends AbstractBigArray implements ByteArray {
|
|||
|
||||
@Override
|
||||
public void fill(long fromIndex, long toIndex, byte value) {
|
||||
Preconditions.checkArgument(fromIndex <= toIndex);
|
||||
if (fromIndex > toIndex) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
final int fromPage = pageIndex(fromIndex);
|
||||
final int toPage = pageIndex(toIndex - 1);
|
||||
if (fromPage == toPage) {
|
||||
|
|
|
@ -19,8 +19,6 @@
|
|||
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.RamUsageEstimator;
|
||||
|
||||
|
@ -95,7 +93,9 @@ final class BigDoubleArray extends AbstractBigArray implements DoubleArray {
|
|||
|
||||
@Override
|
||||
public void fill(long fromIndex, long toIndex, double value) {
|
||||
Preconditions.checkArgument(fromIndex <= toIndex);
|
||||
if (fromIndex > toIndex) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
final long longBits = Double.doubleToRawLongBits(value);
|
||||
final int fromPage = pageIndex(fromIndex);
|
||||
final int toPage = pageIndex(toIndex - 1);
|
||||
|
|
|
@ -19,8 +19,6 @@
|
|||
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.RamUsageEstimator;
|
||||
|
||||
|
@ -95,7 +93,9 @@ final class BigFloatArray extends AbstractBigArray implements FloatArray {
|
|||
|
||||
@Override
|
||||
public void fill(long fromIndex, long toIndex, float value) {
|
||||
Preconditions.checkArgument(fromIndex <= toIndex);
|
||||
if (fromIndex > toIndex) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
final int intBits = Float.floatToRawIntBits(value);
|
||||
final int fromPage = pageIndex(fromIndex);
|
||||
final int toPage = pageIndex(toIndex - 1);
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.RamUsageEstimator;
|
||||
|
||||
|
@ -71,7 +70,9 @@ final class BigIntArray extends AbstractBigArray implements IntArray {
|
|||
|
||||
@Override
|
||||
public void fill(long fromIndex, long toIndex, int value) {
|
||||
Preconditions.checkArgument(fromIndex <= toIndex);
|
||||
if (fromIndex > toIndex) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
final int fromPage = pageIndex(fromIndex);
|
||||
final int toPage = pageIndex(toIndex - 1);
|
||||
if (fromPage == toPage) {
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.RamUsageEstimator;
|
||||
|
||||
|
@ -93,7 +92,9 @@ final class BigLongArray extends AbstractBigArray implements LongArray {
|
|||
|
||||
@Override
|
||||
public void fill(long fromIndex, long toIndex, long value) {
|
||||
Preconditions.checkArgument(fromIndex <= toIndex);
|
||||
if (fromIndex > toIndex) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
if (fromIndex == toIndex) {
|
||||
return; // empty range
|
||||
}
|
||||
|
|
|
@ -18,7 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import com.google.common.math.LongMath;
|
||||
import com.google.common.primitives.Ints;
|
||||
import org.apache.lucene.store.DataInput;
|
||||
import org.apache.lucene.store.DataOutput;
|
||||
|
@ -33,7 +32,6 @@ import org.elasticsearch.common.io.stream.StreamOutput;
|
|||
import org.elasticsearch.common.unit.SizeValue;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.math.RoundingMode;
|
||||
import java.util.Arrays;
|
||||
import java.util.Comparator;
|
||||
|
||||
|
@ -321,7 +319,13 @@ public class BloomFilter {
|
|||
long bitCount;
|
||||
|
||||
BitArray(long bits) {
|
||||
this(new long[Ints.checkedCast(LongMath.divide(bits, 64, RoundingMode.CEILING))]);
|
||||
this(new long[size(bits)]);
|
||||
}
|
||||
|
||||
private static int size(long bits) {
|
||||
long quotient = bits / 64;
|
||||
long remainder = bits - quotient * 64;
|
||||
return Ints.checkedCast(remainder == 0 ? quotient : 1 + quotient);
|
||||
}
|
||||
|
||||
// Used by serialization
|
||||
|
|
|
@ -23,24 +23,10 @@ import com.carrotsearch.hppc.DoubleArrayList;
|
|||
import com.carrotsearch.hppc.FloatArrayList;
|
||||
import com.carrotsearch.hppc.LongArrayList;
|
||||
import com.carrotsearch.hppc.ObjectArrayList;
|
||||
import com.google.common.base.Preconditions;
|
||||
import com.google.common.collect.Iterators;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.BytesRefArray;
|
||||
import org.apache.lucene.util.BytesRefBuilder;
|
||||
import org.apache.lucene.util.InPlaceMergeSorter;
|
||||
import org.apache.lucene.util.IntroSorter;
|
||||
import org.apache.lucene.util.*;
|
||||
|
||||
import java.util.AbstractList;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
import java.util.Collections;
|
||||
import java.util.Comparator;
|
||||
import java.util.Iterator;
|
||||
import java.util.LinkedList;
|
||||
import java.util.List;
|
||||
import java.util.RandomAccess;
|
||||
import java.util.*;
|
||||
|
||||
/** Collections-related utility methods. */
|
||||
public enum CollectionUtils {
|
||||
|
@ -297,8 +283,12 @@ public enum CollectionUtils {
|
|||
private final int distance;
|
||||
|
||||
public RotatedList(List<T> list, int distance) {
|
||||
Preconditions.checkArgument(distance >= 0 && distance < list.size());
|
||||
Preconditions.checkArgument(list instanceof RandomAccess);
|
||||
if (distance < 0 || distance >= list.size()) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
if (!(list instanceof RandomAccess)) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
this.in = list;
|
||||
this.distance = distance;
|
||||
}
|
||||
|
@ -462,4 +452,5 @@ public enum CollectionUtils {
|
|||
|
||||
return result;
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -20,7 +20,6 @@
|
|||
package org.elasticsearch.common.util;
|
||||
|
||||
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
import org.elasticsearch.ElasticsearchException;
|
||||
import org.elasticsearch.common.lease.Releasables;
|
||||
|
||||
|
@ -125,7 +124,7 @@ public class LongObjectPagedHashMap<T> extends AbstractPagedHashMap implements I
|
|||
|
||||
@Override
|
||||
public Iterator<Cursor<T>> iterator() {
|
||||
return new UnmodifiableIterator<Cursor<T>>() {
|
||||
return new Iterator<Cursor<T>>() {
|
||||
|
||||
boolean cached;
|
||||
final Cursor<T> cursor;
|
||||
|
@ -162,6 +161,11 @@ public class LongObjectPagedHashMap<T> extends AbstractPagedHashMap implements I
|
|||
return cursor;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
|
||||
};
|
||||
}
|
||||
|
||||
|
|
|
@ -19,11 +19,12 @@
|
|||
|
||||
package org.elasticsearch.common.util.concurrent;
|
||||
|
||||
import com.google.common.base.Joiner;
|
||||
import org.elasticsearch.common.settings.Settings;
|
||||
|
||||
import java.util.Arrays;
|
||||
import java.util.concurrent.*;
|
||||
import java.util.concurrent.atomic.AtomicInteger;
|
||||
import java.util.stream.Collectors;
|
||||
|
||||
/**
|
||||
*
|
||||
|
@ -81,7 +82,12 @@ public class EsExecutors {
|
|||
}
|
||||
|
||||
public static String threadName(Settings settings, String ... names) {
|
||||
return threadName(settings, "[" + Joiner.on(".").skipNulls().join(names) + "]");
|
||||
String namePrefix =
|
||||
Arrays
|
||||
.stream(names)
|
||||
.filter(name -> name != null)
|
||||
.collect(Collectors.joining(".", "[", "]"));
|
||||
return threadName(settings, namePrefix);
|
||||
}
|
||||
|
||||
public static String threadName(Settings settings, String namePrefix) {
|
||||
|
|
|
@ -0,0 +1,142 @@
|
|||
/*
|
||||
* Licensed to Elasticsearch under one or more contributor
|
||||
* license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright
|
||||
* ownership. Elasticsearch licenses this file to you under
|
||||
* the Apache License, Version 2.0 (the "License"); you may
|
||||
* not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing,
|
||||
* software distributed under the License is distributed on an
|
||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
* KIND, either express or implied. See the License for the
|
||||
* specific language governing permissions and limitations
|
||||
* under the License.
|
||||
*/
|
||||
|
||||
package org.elasticsearch.common.util.iterable;
|
||||
|
||||
import org.elasticsearch.common.lucene.store.IndexOutputOutputStream;
|
||||
|
||||
import java.util.*;
|
||||
import java.util.stream.Stream;
|
||||
import java.util.stream.StreamSupport;
|
||||
|
||||
public class Iterables {
|
||||
public Iterables() {
|
||||
}
|
||||
|
||||
public static <T> Iterable<T> concat(Iterable<T>... inputs) {
|
||||
Objects.requireNonNull(inputs);
|
||||
return new ConcatenatedIterable(inputs);
|
||||
}
|
||||
|
||||
static class ConcatenatedIterable<T> implements Iterable<T> {
|
||||
private final Iterable<T>[] inputs;
|
||||
|
||||
ConcatenatedIterable(Iterable<T>[] inputs) {
|
||||
this.inputs = Arrays.copyOf(inputs, inputs.length);
|
||||
}
|
||||
|
||||
@Override
|
||||
public Iterator<T> iterator() {
|
||||
return Stream
|
||||
.of(inputs)
|
||||
.map(it -> StreamSupport.stream(it.spliterator(), false))
|
||||
.reduce(Stream::concat)
|
||||
.orElseGet(Stream::empty).iterator();
|
||||
}
|
||||
}
|
||||
|
||||
public static <T> Iterable<T> flatten(Iterable<? extends Iterable<T>> inputs) {
|
||||
Objects.requireNonNull(inputs);
|
||||
return new FlattenedIterables<>(inputs);
|
||||
}
|
||||
|
||||
static class FlattenedIterables<T> implements Iterable<T> {
|
||||
private final Iterable<? extends Iterable<T>> inputs;
|
||||
|
||||
FlattenedIterables(Iterable<? extends Iterable<T>> inputs) {
|
||||
List<Iterable<T>> list = new ArrayList<>();
|
||||
for (Iterable<T> iterable : inputs) {
|
||||
list.add(iterable);
|
||||
}
|
||||
this.inputs = list;
|
||||
}
|
||||
|
||||
@Override
|
||||
public Iterator<T> iterator() {
|
||||
return StreamSupport
|
||||
.stream(inputs.spliterator(), false)
|
||||
.flatMap(s -> StreamSupport.stream(s.spliterator(), false)).iterator();
|
||||
}
|
||||
}
|
||||
|
||||
public static boolean allElementsAreEqual(Iterable<?> left, Iterable<?> right) {
|
||||
Objects.requireNonNull(left);
|
||||
Objects.requireNonNull(right);
|
||||
if (left instanceof Collection && right instanceof Collection) {
|
||||
Collection collection1 = (Collection) left;
|
||||
Collection collection2 = (Collection) right;
|
||||
if (collection1.size() != collection2.size()) {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
Iterator<?> leftIt = left.iterator();
|
||||
Iterator<?> rightIt = right.iterator();
|
||||
|
||||
while (true) {
|
||||
if (leftIt.hasNext()) {
|
||||
if (!rightIt.hasNext()) {
|
||||
return false;
|
||||
}
|
||||
|
||||
Object o1 = leftIt.next();
|
||||
Object o2 = rightIt.next();
|
||||
if (Objects.equals(o1, o2)) {
|
||||
continue;
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
return !rightIt.hasNext();
|
||||
}
|
||||
}
|
||||
|
||||
public static <T> T getFirst(Iterable<T> collection, T defaultValue) {
|
||||
Objects.requireNonNull(collection);
|
||||
Iterator<T> iterator = collection.iterator();
|
||||
return iterator.hasNext() ? iterator.next() : defaultValue;
|
||||
}
|
||||
|
||||
public static <T> T get(Iterable<T> iterable, int position) {
|
||||
Objects.requireNonNull(iterable);
|
||||
if (position < 0) {
|
||||
throw new IllegalArgumentException("position >= 0");
|
||||
}
|
||||
if (iterable instanceof List) {
|
||||
List<T> list = (List<T>)iterable;
|
||||
if (position >= list.size()) {
|
||||
throw new IndexOutOfBoundsException(Integer.toString(position));
|
||||
}
|
||||
return list.get(position);
|
||||
} else {
|
||||
Iterator<T> it = iterable.iterator();
|
||||
for (int index = 0; index < position; index++) {
|
||||
if (!it.hasNext()) {
|
||||
throw new IndexOutOfBoundsException(Integer.toString(position));
|
||||
}
|
||||
it.next();
|
||||
}
|
||||
if (!it.hasNext()) {
|
||||
throw new IndexOutOfBoundsException(Integer.toString(position));
|
||||
}
|
||||
return it.next();
|
||||
}
|
||||
}
|
||||
}
|
|
@ -18,28 +18,18 @@
|
|||
*/
|
||||
package org.elasticsearch.gateway;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
|
||||
import org.apache.lucene.codecs.CodecUtil;
|
||||
import org.apache.lucene.index.CorruptIndexException;
|
||||
import org.apache.lucene.index.IndexFormatTooNewException;
|
||||
import org.apache.lucene.index.IndexFormatTooOldException;
|
||||
import org.apache.lucene.store.Directory;
|
||||
import org.apache.lucene.store.IOContext;
|
||||
import org.apache.lucene.store.IndexInput;
|
||||
import org.apache.lucene.store.OutputStreamIndexOutput;
|
||||
import org.apache.lucene.store.SimpleFSDirectory;
|
||||
import org.apache.lucene.store.*;
|
||||
import org.apache.lucene.util.IOUtils;
|
||||
import org.elasticsearch.ExceptionsHelper;
|
||||
import org.elasticsearch.common.bytes.BytesArray;
|
||||
import org.elasticsearch.common.logging.ESLogger;
|
||||
import org.elasticsearch.common.lucene.store.IndexOutputOutputStream;
|
||||
import org.elasticsearch.common.lucene.store.InputStreamIndexInput;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
import org.elasticsearch.common.xcontent.XContentFactory;
|
||||
import org.elasticsearch.common.xcontent.XContentHelper;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.common.xcontent.XContentType;
|
||||
import org.elasticsearch.common.xcontent.*;
|
||||
|
||||
import java.io.FileNotFoundException;
|
||||
import java.io.IOException;
|
||||
|
@ -99,8 +89,12 @@ public abstract class MetaDataStateFormat<T> {
|
|||
* @throws IOException if an IOException occurs
|
||||
*/
|
||||
public final void write(final T state, final long version, final Path... locations) throws IOException {
|
||||
Preconditions.checkArgument(locations != null, "Locations must not be null");
|
||||
Preconditions.checkArgument(locations.length > 0, "One or more locations required");
|
||||
if (locations == null) {
|
||||
throw new IllegalArgumentException("Locations must not be null");
|
||||
}
|
||||
if (locations.length <= 0) {
|
||||
throw new IllegalArgumentException("One or more locations required");
|
||||
}
|
||||
final long maxStateId = findMaxStateId(prefix, locations)+1;
|
||||
assert maxStateId >= 0 : "maxStateId must be positive but was: [" + maxStateId + "]";
|
||||
final String fileName = prefix + maxStateId + STATE_FILE_EXTENSION;
|
||||
|
|
|
@ -20,16 +20,17 @@
|
|||
package org.elasticsearch.http;
|
||||
|
||||
import com.google.common.collect.ImmutableMap;
|
||||
import com.google.common.io.ByteStreams;
|
||||
|
||||
import org.elasticsearch.common.component.AbstractLifecycleComponent;
|
||||
import org.elasticsearch.common.inject.Inject;
|
||||
import org.elasticsearch.common.io.FileSystemUtils;
|
||||
import org.elasticsearch.common.io.Streams;
|
||||
import org.elasticsearch.common.settings.Settings;
|
||||
import org.elasticsearch.env.Environment;
|
||||
import org.elasticsearch.node.service.NodeService;
|
||||
import org.elasticsearch.rest.*;
|
||||
|
||||
import java.io.ByteArrayOutputStream;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStream;
|
||||
import java.nio.file.*;
|
||||
|
@ -141,8 +142,9 @@ public class HttpServer extends AbstractLifecycleComponent<HttpServer> {
|
|||
if (request.method() == RestRequest.Method.GET) {
|
||||
try {
|
||||
try (InputStream stream = getClass().getResourceAsStream("/config/favicon.ico")) {
|
||||
byte[] content = ByteStreams.toByteArray(stream);
|
||||
BytesRestResponse restResponse = new BytesRestResponse(RestStatus.OK, "image/x-icon", content);
|
||||
ByteArrayOutputStream out = new ByteArrayOutputStream();
|
||||
Streams.copy(stream, out);
|
||||
BytesRestResponse restResponse = new BytesRestResponse(RestStatus.OK, "image/x-icon", out.toByteArray());
|
||||
channel.sendResponse(restResponse);
|
||||
}
|
||||
} catch (IOException e) {
|
||||
|
|
|
@ -666,7 +666,6 @@ public class InternalEngine extends Engine {
|
|||
// since it flushes the index as well (though, in terms of concurrency, we are allowed to do it)
|
||||
try (ReleasableLock lock = readLock.acquire()) {
|
||||
ensureOpen();
|
||||
updateIndexWriterSettings();
|
||||
searcherManager.maybeRefreshBlocking();
|
||||
} catch (AlreadyClosedException e) {
|
||||
ensureOpen();
|
||||
|
@ -736,7 +735,6 @@ public class InternalEngine extends Engine {
|
|||
*/
|
||||
try (ReleasableLock lock = readLock.acquire()) {
|
||||
ensureOpen();
|
||||
updateIndexWriterSettings();
|
||||
if (flushLock.tryLock() == false) {
|
||||
// if we can't get the lock right away we block if needed otherwise barf
|
||||
if (waitIfOngoing) {
|
||||
|
@ -954,7 +952,6 @@ public class InternalEngine extends Engine {
|
|||
}
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Closes the engine without acquiring the write lock. This should only be
|
||||
* called while the write lock is hold or in a disaster condition ie. if the engine
|
||||
|
@ -1168,8 +1165,6 @@ public class InternalEngine extends Engine {
|
|||
return indexWriter.getConfig();
|
||||
}
|
||||
|
||||
|
||||
|
||||
private final class EngineMergeScheduler extends ElasticsearchConcurrentMergeScheduler {
|
||||
private final AtomicInteger numMergesInFlight = new AtomicInteger(0);
|
||||
private final AtomicBoolean isThrottling = new AtomicBoolean();
|
||||
|
@ -1245,11 +1240,14 @@ public class InternalEngine extends Engine {
|
|||
|
||||
public void onSettingsChanged() {
|
||||
mergeScheduler.refreshConfig();
|
||||
updateIndexWriterSettings();
|
||||
// config().getVersionMapSize() may have changed:
|
||||
checkVersionMapRefresh();
|
||||
// config().isEnableGcDeletes() or config.getGcDeletesInMillis() may have changed:
|
||||
maybePruneDeletedTombstones();
|
||||
}
|
||||
|
||||
public MergeStats getMergeStats() {
|
||||
return mergeScheduler.stats();
|
||||
}
|
||||
|
||||
|
||||
}
|
||||
|
|
|
@ -19,8 +19,6 @@
|
|||
|
||||
package org.elasticsearch.index.fielddata.plain;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
|
||||
import org.apache.lucene.index.BinaryDocValues;
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.LeafReaderContext;
|
||||
|
@ -52,7 +50,9 @@ public class BinaryDVNumericIndexFieldData extends DocValuesIndexFieldData imple
|
|||
|
||||
public BinaryDVNumericIndexFieldData(Index index, Names fieldNames, NumericType numericType, FieldDataType fieldDataType) {
|
||||
super(index, fieldNames, fieldDataType);
|
||||
Preconditions.checkArgument(numericType != null, "numericType must be non-null");
|
||||
if (numericType == null) {
|
||||
throw new IllegalArgumentException("numericType must be non-null");
|
||||
}
|
||||
this.numericType = numericType;
|
||||
}
|
||||
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.index.fielddata.plain;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.lucene.index.*;
|
||||
import org.apache.lucene.util.*;
|
||||
import org.apache.lucene.util.BitSet;
|
||||
|
@ -73,7 +72,9 @@ public class PackedArrayIndexFieldData extends AbstractIndexFieldData<AtomicNume
|
|||
CircuitBreakerService breakerService) {
|
||||
super(index, indexSettings, fieldNames, fieldDataType, cache);
|
||||
Objects.requireNonNull(numericType);
|
||||
Preconditions.checkArgument(EnumSet.of(NumericType.BOOLEAN, NumericType.BYTE, NumericType.SHORT, NumericType.INT, NumericType.LONG).contains(numericType), getClass().getSimpleName() + " only supports integer types, not " + numericType);
|
||||
if (!EnumSet.of(NumericType.BOOLEAN, NumericType.BYTE, NumericType.SHORT, NumericType.INT, NumericType.LONG).contains(numericType)) {
|
||||
throw new IllegalArgumentException(getClass().getSimpleName() + " only supports integer types, not " + numericType);
|
||||
}
|
||||
this.numericType = numericType;
|
||||
this.breakerService = breakerService;
|
||||
}
|
||||
|
|
|
@ -19,24 +19,12 @@
|
|||
|
||||
package org.elasticsearch.index.fielddata.plain;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.FieldInfo;
|
||||
import org.apache.lucene.index.LeafReader;
|
||||
import org.apache.lucene.index.LeafReaderContext;
|
||||
import org.apache.lucene.index.NumericDocValues;
|
||||
import org.apache.lucene.index.SortedNumericDocValues;
|
||||
import org.apache.lucene.index.*;
|
||||
import org.apache.lucene.util.Accountable;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.elasticsearch.index.Index;
|
||||
import org.elasticsearch.index.fielddata.AtomicNumericFieldData;
|
||||
import org.elasticsearch.index.fielddata.FieldData;
|
||||
import org.elasticsearch.index.fielddata.FieldDataType;
|
||||
import org.elasticsearch.index.fielddata.*;
|
||||
import org.elasticsearch.index.fielddata.IndexFieldData.XFieldComparatorSource.Nested;
|
||||
import org.elasticsearch.index.fielddata.IndexNumericFieldData;
|
||||
import org.elasticsearch.index.fielddata.NumericDoubleValues;
|
||||
import org.elasticsearch.index.fielddata.SortedNumericDoubleValues;
|
||||
import org.elasticsearch.index.fielddata.fieldcomparator.DoubleValuesComparatorSource;
|
||||
import org.elasticsearch.index.fielddata.fieldcomparator.FloatValuesComparatorSource;
|
||||
import org.elasticsearch.index.fielddata.fieldcomparator.LongValuesComparatorSource;
|
||||
|
@ -56,7 +44,9 @@ public class SortedNumericDVIndexFieldData extends DocValuesIndexFieldData imple
|
|||
|
||||
public SortedNumericDVIndexFieldData(Index index, Names fieldNames, NumericType numericType, FieldDataType fieldDataType) {
|
||||
super(index, fieldNames, fieldDataType);
|
||||
Preconditions.checkArgument(numericType != null, "numericType must be non-null");
|
||||
if (numericType == null) {
|
||||
throw new IllegalArgumentException("numericType must be non-null");
|
||||
}
|
||||
this.numericType = numericType;
|
||||
}
|
||||
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.index.mapper.internal;
|
||||
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
import org.apache.lucene.document.Field;
|
||||
import org.apache.lucene.index.IndexOptions;
|
||||
import org.apache.lucene.index.IndexableField;
|
||||
|
@ -240,7 +239,7 @@ public class FieldNamesFieldMapper extends MetadataFieldMapper {
|
|||
return new Iterable<String>() {
|
||||
@Override
|
||||
public Iterator<String> iterator() {
|
||||
return new UnmodifiableIterator<String>() {
|
||||
return new Iterator<String>() {
|
||||
|
||||
int endIndex = nextEndIndex(0);
|
||||
|
||||
|
@ -263,6 +262,11 @@ public class FieldNamesFieldMapper extends MetadataFieldMapper {
|
|||
return result;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
|
||||
};
|
||||
}
|
||||
};
|
||||
|
|
|
@ -19,18 +19,12 @@
|
|||
|
||||
package org.elasticsearch.index.mapper.internal;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.apache.lucene.document.BinaryDocValuesField;
|
||||
import org.apache.lucene.document.Field;
|
||||
import org.apache.lucene.index.IndexOptions;
|
||||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.queries.TermsQuery;
|
||||
import org.apache.lucene.search.BooleanClause;
|
||||
import org.apache.lucene.search.BooleanQuery;
|
||||
import org.apache.lucene.search.MultiTermQuery;
|
||||
import org.apache.lucene.search.PrefixQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.RegexpQuery;
|
||||
import org.apache.lucene.search.*;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.common.Nullable;
|
||||
|
@ -38,17 +32,11 @@ import org.elasticsearch.common.Strings;
|
|||
import org.elasticsearch.common.lucene.BytesRefs;
|
||||
import org.elasticsearch.common.lucene.Lucene;
|
||||
import org.elasticsearch.common.settings.Settings;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.index.fielddata.FieldDataType;
|
||||
import org.elasticsearch.index.mapper.MappedFieldType;
|
||||
import org.elasticsearch.index.mapper.Mapper;
|
||||
import org.elasticsearch.index.mapper.MapperParsingException;
|
||||
import org.elasticsearch.index.mapper.MergeMappingException;
|
||||
import org.elasticsearch.index.mapper.MergeResult;
|
||||
import org.elasticsearch.index.mapper.MetadataFieldMapper;
|
||||
import org.elasticsearch.index.mapper.ParseContext;
|
||||
import org.elasticsearch.index.mapper.Uid;
|
||||
import org.elasticsearch.index.mapper.*;
|
||||
import org.elasticsearch.index.query.QueryParseContext;
|
||||
|
||||
import java.io.IOException;
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.index.mapper.object;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.search.Filter;
|
||||
import org.apache.lucene.search.QueryWrapperFilter;
|
||||
|
@ -33,30 +32,13 @@ import org.elasticsearch.common.collect.CopyOnWriteHashMap;
|
|||
import org.elasticsearch.common.settings.Settings;
|
||||
import org.elasticsearch.common.xcontent.ToXContent;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
import org.elasticsearch.index.mapper.ContentPath;
|
||||
import org.elasticsearch.index.mapper.DocumentMapper;
|
||||
import org.elasticsearch.index.mapper.DocumentMapperParser;
|
||||
import org.elasticsearch.index.mapper.FieldMapper;
|
||||
import org.elasticsearch.index.mapper.Mapper;
|
||||
import org.elasticsearch.index.mapper.MapperParsingException;
|
||||
import org.elasticsearch.index.mapper.MapperUtils;
|
||||
import org.elasticsearch.index.mapper.MergeMappingException;
|
||||
import org.elasticsearch.index.mapper.MergeResult;
|
||||
import org.elasticsearch.index.mapper.MetadataFieldMapper;
|
||||
import org.elasticsearch.index.mapper.*;
|
||||
import org.elasticsearch.index.mapper.internal.AllFieldMapper;
|
||||
import org.elasticsearch.index.mapper.internal.TypeFieldMapper;
|
||||
import org.elasticsearch.index.settings.IndexSettings;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
import java.util.Comparator;
|
||||
import java.util.HashMap;
|
||||
import java.util.Iterator;
|
||||
import java.util.List;
|
||||
import java.util.Locale;
|
||||
import java.util.Map;
|
||||
import java.util.*;
|
||||
|
||||
import static org.elasticsearch.common.xcontent.support.XContentMapValues.nodeBooleanValue;
|
||||
import static org.elasticsearch.index.mapper.MapperBuilders.object;
|
||||
|
@ -583,7 +565,7 @@ public class ObjectMapper extends Mapper implements AllFieldMapper.IncludeInAll,
|
|||
doXContent(builder, params);
|
||||
|
||||
// sort the mappers so we get consistent serialization format
|
||||
Mapper[] sortedMappers = Iterables.toArray(mappers.values(), Mapper.class);
|
||||
Mapper[] sortedMappers = mappers.values().stream().toArray(size -> new Mapper[size]);
|
||||
Arrays.sort(sortedMappers, new Comparator<Mapper>() {
|
||||
@Override
|
||||
public int compare(Mapper o1, Mapper o2) {
|
||||
|
|
|
@ -19,13 +19,12 @@
|
|||
|
||||
package org.elasticsearch.index.query;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
|
||||
import org.apache.lucene.queries.TermsQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.elasticsearch.common.inject.Inject;
|
||||
import org.elasticsearch.common.lucene.search.Queries;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.index.mapper.Uid;
|
||||
import org.elasticsearch.index.mapper.internal.UidFieldMapper;
|
||||
|
|
|
@ -20,14 +20,12 @@
|
|||
|
||||
package org.elasticsearch.index.query.functionscore.random;
|
||||
|
||||
import com.google.common.primitives.Longs;
|
||||
|
||||
import org.elasticsearch.common.inject.Inject;
|
||||
import org.elasticsearch.common.lucene.search.function.RandomScoreFunction;
|
||||
import org.elasticsearch.common.lucene.search.function.ScoreFunction;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.index.fielddata.IndexFieldData;
|
||||
import org.elasticsearch.index.mapper.FieldMapper;
|
||||
import org.elasticsearch.index.mapper.MappedFieldType;
|
||||
import org.elasticsearch.index.query.QueryParseContext;
|
||||
import org.elasticsearch.index.query.QueryParsingException;
|
||||
|
@ -66,7 +64,7 @@ public class RandomScoreFunctionParser implements ScoreFunctionParser {
|
|||
if (parser.numberType() == XContentParser.NumberType.INT) {
|
||||
seed = parser.intValue();
|
||||
} else if (parser.numberType() == XContentParser.NumberType.LONG) {
|
||||
seed = Longs.hashCode(parser.longValue());
|
||||
seed = hash(parser.longValue());
|
||||
} else {
|
||||
throw new QueryParsingException(parseContext, "random_score seed must be an int, long or string, not '"
|
||||
+ token.toString() + "'");
|
||||
|
@ -90,7 +88,7 @@ public class RandomScoreFunctionParser implements ScoreFunctionParser {
|
|||
}
|
||||
|
||||
if (seed == -1) {
|
||||
seed = Longs.hashCode(parseContext.nowInMillis());
|
||||
seed = hash(parseContext.nowInMillis());
|
||||
}
|
||||
final ShardId shardId = SearchContext.current().indexShard().shardId();
|
||||
final int salt = (shardId.index().name().hashCode() << 10) | shardId.id();
|
||||
|
@ -98,4 +96,8 @@ public class RandomScoreFunctionParser implements ScoreFunctionParser {
|
|||
|
||||
return new RandomScoreFunction(seed, salt, uidFieldData);
|
||||
}
|
||||
|
||||
private static final int hash(long value) {
|
||||
return (int) (value ^ (value >>> 32));
|
||||
}
|
||||
}
|
|
@ -115,9 +115,6 @@ import java.util.concurrent.ScheduledFuture;
|
|||
import java.util.concurrent.TimeUnit;
|
||||
import java.util.concurrent.atomic.AtomicReference;
|
||||
|
||||
/**
|
||||
*
|
||||
*/
|
||||
public class IndexShard extends AbstractIndexShardComponent {
|
||||
|
||||
private final ThreadPool threadPool;
|
||||
|
@ -985,15 +982,27 @@ public class IndexShard extends AbstractIndexShardComponent {
|
|||
}
|
||||
|
||||
public void updateBufferSize(ByteSizeValue shardIndexingBufferSize, ByteSizeValue shardTranslogBufferSize) {
|
||||
|
||||
final EngineConfig config = engineConfig;
|
||||
final ByteSizeValue preValue = config.getIndexingBufferSize();
|
||||
|
||||
config.setIndexingBufferSize(shardIndexingBufferSize);
|
||||
|
||||
Engine engine = engineUnsafe();
|
||||
if (engine == null) {
|
||||
logger.debug("updateBufferSize: engine is closed; skipping");
|
||||
return;
|
||||
}
|
||||
|
||||
// update engine if it is already started.
|
||||
if (preValue.bytes() != shardIndexingBufferSize.bytes() && engineUnsafe() != null) {
|
||||
// its inactive, make sure we do a refresh / full IW flush in this case, since the memory
|
||||
// changes only after a "data" change has happened to the writer
|
||||
// the index writer lazily allocates memory and a refresh will clean it all up.
|
||||
if (shardIndexingBufferSize == EngineConfig.INACTIVE_SHARD_INDEXING_BUFFER && preValue != EngineConfig.INACTIVE_SHARD_INDEXING_BUFFER) {
|
||||
if (preValue.bytes() != shardIndexingBufferSize.bytes()) {
|
||||
// so we push changes these changes down to IndexWriter:
|
||||
engine.onSettingsChanged();
|
||||
|
||||
if (shardIndexingBufferSize == EngineConfig.INACTIVE_SHARD_INDEXING_BUFFER) {
|
||||
// it's inactive: make sure we do a refresh / full IW flush in this case, since the memory
|
||||
// changes only after a "data" change has happened to the writer
|
||||
// the index writer lazily allocates memory and a refresh will clean it all up.
|
||||
logger.debug("updating index_buffer_size from [{}] to (inactive) [{}]", preValue, shardIndexingBufferSize);
|
||||
try {
|
||||
refresh("update index buffer");
|
||||
|
@ -1004,10 +1013,8 @@ public class IndexShard extends AbstractIndexShardComponent {
|
|||
logger.debug("updating index_buffer_size from [{}] to [{}]", preValue, shardIndexingBufferSize);
|
||||
}
|
||||
}
|
||||
Engine engine = engineUnsafe();
|
||||
if (engine != null) {
|
||||
engine.getTranslog().updateBuffer(shardTranslogBufferSize);
|
||||
}
|
||||
|
||||
engine.getTranslog().updateBuffer(shardTranslogBufferSize);
|
||||
}
|
||||
|
||||
public void markAsInactive() {
|
||||
|
@ -1129,7 +1136,7 @@ public class IndexShard extends AbstractIndexShardComponent {
|
|||
searchService.onRefreshSettings(settings);
|
||||
indexingService.onRefreshSettings(settings);
|
||||
if (change) {
|
||||
refresh("apply settings");
|
||||
engine().onSettingsChanged();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -1267,6 +1274,8 @@ public class IndexShard extends AbstractIndexShardComponent {
|
|||
return engine;
|
||||
}
|
||||
|
||||
/** NOTE: returns null if engine is not yet started (e.g. recovery phase 1, copying over index files, is still running), or if engine is
|
||||
* closed. */
|
||||
protected Engine engineUnsafe() {
|
||||
return this.currentEngineReference.get();
|
||||
}
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.index.snapshots.blobstore;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.apache.lucene.index.CorruptIndexException;
|
||||
import org.apache.lucene.index.IndexFormatTooNewException;
|
||||
import org.apache.lucene.index.IndexFormatTooOldException;
|
||||
|
@ -48,14 +47,11 @@ import org.elasticsearch.common.lucene.Lucene;
|
|||
import org.elasticsearch.common.lucene.store.InputStreamIndexInput;
|
||||
import org.elasticsearch.common.settings.Settings;
|
||||
import org.elasticsearch.common.unit.ByteSizeValue;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.index.IndexService;
|
||||
import org.elasticsearch.index.deletionpolicy.SnapshotIndexCommit;
|
||||
import org.elasticsearch.index.shard.ShardId;
|
||||
import org.elasticsearch.index.snapshots.IndexShardRepository;
|
||||
import org.elasticsearch.index.snapshots.IndexShardRestoreFailedException;
|
||||
import org.elasticsearch.index.snapshots.IndexShardSnapshotException;
|
||||
import org.elasticsearch.index.snapshots.IndexShardSnapshotFailedException;
|
||||
import org.elasticsearch.index.snapshots.IndexShardSnapshotStatus;
|
||||
import org.elasticsearch.index.snapshots.*;
|
||||
import org.elasticsearch.index.snapshots.blobstore.BlobStoreIndexShardSnapshot.FileInfo;
|
||||
import org.elasticsearch.index.store.Store;
|
||||
import org.elasticsearch.index.store.StoreFileMetaData;
|
||||
|
@ -71,11 +67,7 @@ import org.elasticsearch.repositories.blobstore.LegacyBlobStoreFormat;
|
|||
import java.io.FilterInputStream;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStream;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.*;
|
||||
|
||||
import static org.elasticsearch.repositories.blobstore.BlobStoreRepository.testBlobPrefix;
|
||||
|
||||
|
|
|
@ -20,7 +20,6 @@
|
|||
package org.elasticsearch.index.store;
|
||||
|
||||
import com.google.common.collect.ImmutableMap;
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.apache.lucene.codecs.CodecUtil;
|
||||
import org.apache.lucene.index.*;
|
||||
import org.apache.lucene.store.*;
|
||||
|
@ -42,6 +41,7 @@ import org.elasticsearch.common.lucene.store.InputStreamIndexInput;
|
|||
import org.elasticsearch.common.settings.Settings;
|
||||
import org.elasticsearch.common.unit.TimeValue;
|
||||
import org.elasticsearch.common.util.Callback;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.common.util.SingleObjectCache;
|
||||
import org.elasticsearch.common.util.concurrent.AbstractRefCounted;
|
||||
import org.elasticsearch.common.util.concurrent.RefCounted;
|
||||
|
|
|
@ -19,8 +19,8 @@
|
|||
package org.elasticsearch.indices.flush;
|
||||
|
||||
import com.google.common.collect.ImmutableMap;
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.elasticsearch.cluster.routing.ShardRouting;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.common.xcontent.ToXContent;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilderString;
|
||||
|
@ -41,7 +41,7 @@ public class IndicesSyncedFlushResult implements ToXContent {
|
|||
|
||||
public IndicesSyncedFlushResult(Map<String, List<ShardsSyncedFlushResult>> shardsResultPerIndex) {
|
||||
this.shardsResultPerIndex = ImmutableMap.copyOf(shardsResultPerIndex);
|
||||
this.shardCounts = calculateShardCounts(Iterables.concat(shardsResultPerIndex.values()));
|
||||
this.shardCounts = calculateShardCounts(Iterables.flatten(shardsResultPerIndex.values()));
|
||||
}
|
||||
|
||||
/** total number shards, including replicas, both assigned and unassigned */
|
||||
|
|
|
@ -51,6 +51,42 @@ import java.util.concurrent.ScheduledFuture;
|
|||
*/
|
||||
public class IndexingMemoryController extends AbstractLifecycleComponent<IndexingMemoryController> {
|
||||
|
||||
/** How much heap (% or bytes) we will share across all actively indexing shards on this node (default: 10%). */
|
||||
public static final String INDEX_BUFFER_SIZE_SETTING = "indices.memory.index_buffer_size";
|
||||
|
||||
/** Only applies when <code>indices.memory.index_buffer_size</code> is a %, to set a floor on the actual size in bytes (default: 48 MB). */
|
||||
public static final String MIN_INDEX_BUFFER_SIZE_SETTING = "indices.memory.min_index_buffer_size";
|
||||
|
||||
/** Only applies when <code>indices.memory.index_buffer_size</code> is a %, to set a ceiling on the actual size in bytes (default: not set). */
|
||||
public static final String MAX_INDEX_BUFFER_SIZE_SETTING = "indices.memory.max_index_buffer_size";
|
||||
|
||||
/** Sets a floor on the per-shard index buffer size (default: 4 MB). */
|
||||
public static final String MIN_SHARD_INDEX_BUFFER_SIZE_SETTING = "indices.memory.min_shard_index_buffer_size";
|
||||
|
||||
/** Sets a ceiling on the per-shard index buffer size (default: 512 MB). */
|
||||
public static final String MAX_SHARD_INDEX_BUFFER_SIZE_SETTING = "indices.memory.max_shard_index_buffer_size";
|
||||
|
||||
/** How much heap (% or bytes) we will share across all actively indexing shards for the translog buffer (default: 1%). */
|
||||
public static final String TRANSLOG_BUFFER_SIZE_SETTING = "indices.memory.translog_buffer_size";
|
||||
|
||||
/** Only applies when <code>indices.memory.translog_buffer_size</code> is a %, to set a floor on the actual size in bytes (default: 256 KB). */
|
||||
public static final String MIN_TRANSLOG_BUFFER_SIZE_SETTING = "indices.memory.min_translog_buffer_size";
|
||||
|
||||
/** Only applies when <code>indices.memory.translog_buffer_size</code> is a %, to set a ceiling on the actual size in bytes (default: not set). */
|
||||
public static final String MAX_TRANSLOG_BUFFER_SIZE_SETTING = "indices.memory.max_translog_buffer_size";
|
||||
|
||||
/** Sets a floor on the per-shard translog buffer size (default: 2 KB). */
|
||||
public static final String MIN_SHARD_TRANSLOG_BUFFER_SIZE_SETTING = "indices.memory.min_shard_translog_buffer_size";
|
||||
|
||||
/** Sets a ceiling on the per-shard translog buffer size (default: 64 KB). */
|
||||
public static final String MAX_SHARD_TRANSLOG_BUFFER_SIZE_SETTING = "indices.memory.max_shard_translog_buffer_size";
|
||||
|
||||
/** If we see no indexing operations after this much time for a given shard, we consider that shard inactive (default: 5 minutes). */
|
||||
public static final String SHARD_INACTIVE_TIME_SETTING = "indices.memory.shard_inactive_time";
|
||||
|
||||
/** How frequently we check shards to find inactive ones (default: 30 seconds). */
|
||||
public static final String SHARD_INACTIVE_INTERVAL_TIME_SETTING = "indices.memory.interval";
|
||||
|
||||
private final ThreadPool threadPool;
|
||||
private final IndicesService indicesService;
|
||||
|
||||
|
@ -77,12 +113,12 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
this.indicesService = indicesService;
|
||||
|
||||
ByteSizeValue indexingBuffer;
|
||||
String indexingBufferSetting = this.settings.get("indices.memory.index_buffer_size", "10%");
|
||||
String indexingBufferSetting = this.settings.get(INDEX_BUFFER_SIZE_SETTING, "10%");
|
||||
if (indexingBufferSetting.endsWith("%")) {
|
||||
double percent = Double.parseDouble(indexingBufferSetting.substring(0, indexingBufferSetting.length() - 1));
|
||||
indexingBuffer = new ByteSizeValue((long) (((double) JvmInfo.jvmInfo().getMem().getHeapMax().bytes()) * (percent / 100)));
|
||||
ByteSizeValue minIndexingBuffer = this.settings.getAsBytesSize("indices.memory.min_index_buffer_size", new ByteSizeValue(48, ByteSizeUnit.MB));
|
||||
ByteSizeValue maxIndexingBuffer = this.settings.getAsBytesSize("indices.memory.max_index_buffer_size", null);
|
||||
ByteSizeValue minIndexingBuffer = this.settings.getAsBytesSize(MIN_INDEX_BUFFER_SIZE_SETTING, new ByteSizeValue(48, ByteSizeUnit.MB));
|
||||
ByteSizeValue maxIndexingBuffer = this.settings.getAsBytesSize(MAX_INDEX_BUFFER_SIZE_SETTING, null);
|
||||
|
||||
if (indexingBuffer.bytes() < minIndexingBuffer.bytes()) {
|
||||
indexingBuffer = minIndexingBuffer;
|
||||
|
@ -91,20 +127,20 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
indexingBuffer = maxIndexingBuffer;
|
||||
}
|
||||
} else {
|
||||
indexingBuffer = ByteSizeValue.parseBytesSizeValue(indexingBufferSetting, null);
|
||||
indexingBuffer = ByteSizeValue.parseBytesSizeValue(indexingBufferSetting, INDEX_BUFFER_SIZE_SETTING);
|
||||
}
|
||||
this.indexingBuffer = indexingBuffer;
|
||||
this.minShardIndexBufferSize = this.settings.getAsBytesSize("indices.memory.min_shard_index_buffer_size", new ByteSizeValue(4, ByteSizeUnit.MB));
|
||||
this.minShardIndexBufferSize = this.settings.getAsBytesSize(MIN_SHARD_INDEX_BUFFER_SIZE_SETTING, new ByteSizeValue(4, ByteSizeUnit.MB));
|
||||
// LUCENE MONITOR: Based on this thread, currently (based on Mike), having a large buffer does not make a lot of sense: https://issues.apache.org/jira/browse/LUCENE-2324?focusedCommentId=13005155&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13005155
|
||||
this.maxShardIndexBufferSize = this.settings.getAsBytesSize("indices.memory.max_shard_index_buffer_size", new ByteSizeValue(512, ByteSizeUnit.MB));
|
||||
this.maxShardIndexBufferSize = this.settings.getAsBytesSize(MAX_SHARD_INDEX_BUFFER_SIZE_SETTING, new ByteSizeValue(512, ByteSizeUnit.MB));
|
||||
|
||||
ByteSizeValue translogBuffer;
|
||||
String translogBufferSetting = this.settings.get("indices.memory.translog_buffer_size", "1%");
|
||||
String translogBufferSetting = this.settings.get(TRANSLOG_BUFFER_SIZE_SETTING, "1%");
|
||||
if (translogBufferSetting.endsWith("%")) {
|
||||
double percent = Double.parseDouble(translogBufferSetting.substring(0, translogBufferSetting.length() - 1));
|
||||
translogBuffer = new ByteSizeValue((long) (((double) JvmInfo.jvmInfo().getMem().getHeapMax().bytes()) * (percent / 100)));
|
||||
ByteSizeValue minTranslogBuffer = this.settings.getAsBytesSize("indices.memory.min_translog_buffer_size", new ByteSizeValue(256, ByteSizeUnit.KB));
|
||||
ByteSizeValue maxTranslogBuffer = this.settings.getAsBytesSize("indices.memory.max_translog_buffer_size", null);
|
||||
ByteSizeValue minTranslogBuffer = this.settings.getAsBytesSize(MIN_TRANSLOG_BUFFER_SIZE_SETTING, new ByteSizeValue(256, ByteSizeUnit.KB));
|
||||
ByteSizeValue maxTranslogBuffer = this.settings.getAsBytesSize(MAX_TRANSLOG_BUFFER_SIZE_SETTING, null);
|
||||
|
||||
if (translogBuffer.bytes() < minTranslogBuffer.bytes()) {
|
||||
translogBuffer = minTranslogBuffer;
|
||||
|
@ -116,15 +152,19 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
translogBuffer = ByteSizeValue.parseBytesSizeValue(translogBufferSetting, null);
|
||||
}
|
||||
this.translogBuffer = translogBuffer;
|
||||
this.minShardTranslogBufferSize = this.settings.getAsBytesSize("indices.memory.min_shard_translog_buffer_size", new ByteSizeValue(2, ByteSizeUnit.KB));
|
||||
this.maxShardTranslogBufferSize = this.settings.getAsBytesSize("indices.memory.max_shard_translog_buffer_size", new ByteSizeValue(64, ByteSizeUnit.KB));
|
||||
this.minShardTranslogBufferSize = this.settings.getAsBytesSize(MIN_SHARD_TRANSLOG_BUFFER_SIZE_SETTING, new ByteSizeValue(2, ByteSizeUnit.KB));
|
||||
this.maxShardTranslogBufferSize = this.settings.getAsBytesSize(MAX_SHARD_TRANSLOG_BUFFER_SIZE_SETTING, new ByteSizeValue(64, ByteSizeUnit.KB));
|
||||
|
||||
this.inactiveTime = this.settings.getAsTime("indices.memory.shard_inactive_time", TimeValue.timeValueMinutes(5));
|
||||
this.inactiveTime = this.settings.getAsTime(SHARD_INACTIVE_TIME_SETTING, TimeValue.timeValueMinutes(5));
|
||||
// we need to have this relatively small to move a shard from inactive to active fast (enough)
|
||||
this.interval = this.settings.getAsTime("indices.memory.interval", TimeValue.timeValueSeconds(30));
|
||||
|
||||
logger.debug("using index_buffer_size [{}], with min_shard_index_buffer_size [{}], max_shard_index_buffer_size [{}], shard_inactive_time [{}]", this.indexingBuffer, this.minShardIndexBufferSize, this.maxShardIndexBufferSize, this.inactiveTime);
|
||||
this.interval = this.settings.getAsTime(SHARD_INACTIVE_INTERVAL_TIME_SETTING, TimeValue.timeValueSeconds(30));
|
||||
|
||||
logger.debug("using indexing buffer size [{}], with {} [{}], {} [{}], {} [{}], {} [{}]",
|
||||
this.indexingBuffer,
|
||||
MIN_SHARD_INDEX_BUFFER_SIZE_SETTING, this.minShardIndexBufferSize,
|
||||
MAX_SHARD_INDEX_BUFFER_SIZE_SETTING, this.maxShardIndexBufferSize,
|
||||
SHARD_INACTIVE_TIME_SETTING, this.inactiveTime,
|
||||
SHARD_INACTIVE_INTERVAL_TIME_SETTING, this.interval);
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -155,12 +195,9 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
|
||||
private final Map<ShardId, ShardIndexingStatus> shardsIndicesStatus = new HashMap<>();
|
||||
|
||||
|
||||
@Override
|
||||
public void run() {
|
||||
EnumSet<ShardStatusChangeType> changes = EnumSet.noneOf(ShardStatusChangeType.class);
|
||||
|
||||
changes.addAll(purgeDeletedAndClosedShards());
|
||||
EnumSet<ShardStatusChangeType> changes = purgeDeletedAndClosedShards();
|
||||
|
||||
final List<IndexShard> activeToInactiveIndexingShards = new ArrayList<>();
|
||||
final int activeShards = updateShardStatuses(changes, activeToInactiveIndexingShards);
|
||||
|
@ -170,11 +207,15 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
indexShard.markAsInactive();
|
||||
} catch (EngineClosedException e) {
|
||||
// ignore
|
||||
logger.trace("ignore EngineClosedException while marking shard [{}][{}] as inactive", indexShard.shardId().index().name(), indexShard.shardId().id());
|
||||
} catch (FlushNotAllowedEngineException e) {
|
||||
// ignore
|
||||
logger.trace("ignore FlushNotAllowedException while marking shard [{}][{}] as inactive", indexShard.shardId().index().name(), indexShard.shardId().id());
|
||||
}
|
||||
}
|
||||
if (!changes.isEmpty()) {
|
||||
|
||||
if (changes.isEmpty() == false) {
|
||||
// Something changed: recompute indexing buffers:
|
||||
calcAndSetShardBuffers(activeShards, "[" + changes + "]");
|
||||
}
|
||||
}
|
||||
|
@ -190,23 +231,24 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
for (IndexShard indexShard : indexService) {
|
||||
|
||||
if (!CAN_UPDATE_INDEX_BUFFER_STATES.contains(indexShard.state())) {
|
||||
// not ready to be updated yet.
|
||||
// not ready to be updated yet
|
||||
continue;
|
||||
}
|
||||
|
||||
if (indexShard.canIndex() == false) {
|
||||
// not relevant for memory related issues.
|
||||
// shadow replica doesn't have an indexing buffer
|
||||
continue;
|
||||
}
|
||||
|
||||
final Translog translog;
|
||||
try {
|
||||
translog = indexShard.engine().getTranslog();
|
||||
} catch (EngineClosedException e) {
|
||||
// not ready yet to be checked for in activity
|
||||
// not ready yet to be checked for activity
|
||||
continue;
|
||||
}
|
||||
|
||||
final long time = threadPool.estimatedTimeInMillis();
|
||||
final long timeMS = threadPool.estimatedTimeInMillis();
|
||||
|
||||
ShardIndexingStatus status = shardsIndicesStatus.get(indexShard.shardId());
|
||||
if (status == null) {
|
||||
|
@ -214,21 +256,22 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
shardsIndicesStatus.put(indexShard.shardId(), status);
|
||||
changes.add(ShardStatusChangeType.ADDED);
|
||||
}
|
||||
// check if it is deemed to be inactive (sam translogFileGeneration and numberOfOperations over a long period of time)
|
||||
|
||||
// consider shard inactive if it has same translogFileGeneration and no operations for a long time
|
||||
if (status.translogId == translog.currentFileGeneration() && translog.totalOperations() == 0) {
|
||||
if (status.time == -1) { // first time
|
||||
status.time = time;
|
||||
if (status.timeMS == -1) {
|
||||
// first time we noticed the shard become idle
|
||||
status.timeMS = timeMS;
|
||||
}
|
||||
// inactive?
|
||||
if (status.activeIndexing) {
|
||||
// mark it as inactive only if enough time has passed and there are no ongoing merges going on...
|
||||
if ((time - status.time) > inactiveTime.millis() && indexShard.mergeStats().getCurrent() == 0) {
|
||||
// inactive for this amount of time, mark it
|
||||
activeToInactiveIndexingShards.add(indexShard);
|
||||
status.activeIndexing = false;
|
||||
changes.add(ShardStatusChangeType.BECAME_INACTIVE);
|
||||
logger.debug("marking shard [{}][{}] as inactive (inactive_time[{}]) indexing wise, setting size to [{}]", indexShard.shardId().index().name(), indexShard.shardId().id(), inactiveTime, EngineConfig.INACTIVE_SHARD_INDEXING_BUFFER);
|
||||
}
|
||||
// mark it as inactive only if enough time has passed
|
||||
if (status.activeIndexing && (timeMS - status.timeMS) > inactiveTime.millis()) {
|
||||
// inactive for this amount of time, mark it
|
||||
activeToInactiveIndexingShards.add(indexShard);
|
||||
status.activeIndexing = false;
|
||||
changes.add(ShardStatusChangeType.BECAME_INACTIVE);
|
||||
logger.debug("marking shard [{}][{}] as inactive (inactive_time[{}]) indexing wise, setting size to [{}]",
|
||||
indexShard.shardId().index().name(), indexShard.shardId().id(),
|
||||
inactiveTime, EngineConfig.INACTIVE_SHARD_INDEXING_BUFFER);
|
||||
}
|
||||
} else {
|
||||
if (!status.activeIndexing) {
|
||||
|
@ -236,10 +279,9 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
changes.add(ShardStatusChangeType.BECAME_ACTIVE);
|
||||
logger.debug("marking shard [{}][{}] as active indexing wise", indexShard.shardId().index().name(), indexShard.shardId().id());
|
||||
}
|
||||
status.time = -1;
|
||||
status.timeMS = -1;
|
||||
}
|
||||
status.translogId = translog.currentFileGeneration();
|
||||
status.translogNumberOfOperations = translog.totalOperations();
|
||||
|
||||
if (status.activeIndexing) {
|
||||
activeShards++;
|
||||
|
@ -261,31 +303,28 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
while (statusShardIdIterator.hasNext()) {
|
||||
ShardId statusShardId = statusShardIdIterator.next();
|
||||
IndexService indexService = indicesService.indexService(statusShardId.getIndex());
|
||||
boolean remove = false;
|
||||
try {
|
||||
if (indexService == null) {
|
||||
remove = true;
|
||||
continue;
|
||||
}
|
||||
boolean remove;
|
||||
if (indexService == null) {
|
||||
remove = true;
|
||||
} else {
|
||||
IndexShard indexShard = indexService.shard(statusShardId.id());
|
||||
if (indexShard == null) {
|
||||
remove = true;
|
||||
continue;
|
||||
}
|
||||
remove = !CAN_UPDATE_INDEX_BUFFER_STATES.contains(indexShard.state());
|
||||
|
||||
} finally {
|
||||
if (remove) {
|
||||
changes.add(ShardStatusChangeType.DELETED);
|
||||
statusShardIdIterator.remove();
|
||||
} else {
|
||||
remove = !CAN_UPDATE_INDEX_BUFFER_STATES.contains(indexShard.state());
|
||||
}
|
||||
}
|
||||
if (remove) {
|
||||
changes.add(ShardStatusChangeType.DELETED);
|
||||
statusShardIdIterator.remove();
|
||||
}
|
||||
}
|
||||
return changes;
|
||||
}
|
||||
|
||||
private void calcAndSetShardBuffers(int activeShards, String reason) {
|
||||
if (activeShards == 0) {
|
||||
logger.debug("no active shards (reason={})", reason);
|
||||
return;
|
||||
}
|
||||
ByteSizeValue shardIndexingBufferSize = new ByteSizeValue(indexingBuffer.bytes() / activeShards);
|
||||
|
@ -335,11 +374,9 @@ public class IndexingMemoryController extends AbstractLifecycleComponent<Indexin
|
|||
ADDED, DELETED, BECAME_ACTIVE, BECAME_INACTIVE
|
||||
}
|
||||
|
||||
|
||||
static class ShardIndexingStatus {
|
||||
private static class ShardIndexingStatus {
|
||||
long translogId = -1;
|
||||
int translogNumberOfOperations = -1;
|
||||
boolean activeIndexing = true;
|
||||
long time = -1; // contains the first time we saw this shard with no operations done on it
|
||||
long timeMS = -1; // contains the first time we saw this shard with no operations done on it
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.indices.recovery;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.apache.lucene.index.CorruptIndexException;
|
||||
import org.apache.lucene.index.IndexFormatTooNewException;
|
||||
import org.apache.lucene.index.IndexFormatTooOldException;
|
||||
|
@ -39,6 +38,7 @@ import org.elasticsearch.common.logging.ESLogger;
|
|||
import org.elasticsearch.common.unit.ByteSizeValue;
|
||||
import org.elasticsearch.common.util.CancellableThreads;
|
||||
import org.elasticsearch.common.util.CancellableThreads.Interruptable;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.common.util.concurrent.AbstractRunnable;
|
||||
import org.elasticsearch.index.deletionpolicy.SnapshotIndexCommit;
|
||||
import org.elasticsearch.index.engine.Engine;
|
||||
|
@ -64,6 +64,7 @@ import java.util.concurrent.CountDownLatch;
|
|||
import java.util.concurrent.ThreadPoolExecutor;
|
||||
import java.util.concurrent.atomic.AtomicLong;
|
||||
import java.util.concurrent.atomic.AtomicReference;
|
||||
import java.util.stream.StreamSupport;
|
||||
|
||||
/**
|
||||
* RecoverySourceHandler handles the three phases of shard recovery, which is
|
||||
|
@ -411,7 +412,8 @@ public class RecoverySourceHandler {
|
|||
if ((corruptIndexException = ExceptionsHelper.unwrapCorruption(remoteException)) != null) {
|
||||
try {
|
||||
final Store.MetadataSnapshot recoverySourceMetadata = store.getMetadata(snapshot);
|
||||
StoreFileMetaData[] metadata = Iterables.toArray(recoverySourceMetadata, StoreFileMetaData.class);
|
||||
StoreFileMetaData[] metadata =
|
||||
StreamSupport.stream(recoverySourceMetadata.spliterator(), false).toArray(size -> new StoreFileMetaData[size]);
|
||||
ArrayUtil.timSort(metadata, new Comparator<StoreFileMetaData>() {
|
||||
@Override
|
||||
public int compare(StoreFileMetaData o1, StoreFileMetaData o2) {
|
||||
|
|
|
@ -19,14 +19,9 @@
|
|||
|
||||
package org.elasticsearch.repositories;
|
||||
|
||||
import com.google.common.base.Joiner;
|
||||
import com.google.common.collect.ImmutableMap;
|
||||
import org.elasticsearch.action.ActionListener;
|
||||
import org.elasticsearch.cluster.AckedClusterStateUpdateTask;
|
||||
import org.elasticsearch.cluster.ClusterChangedEvent;
|
||||
import org.elasticsearch.cluster.ClusterService;
|
||||
import org.elasticsearch.cluster.ClusterState;
|
||||
import org.elasticsearch.cluster.ClusterStateListener;
|
||||
import org.elasticsearch.cluster.*;
|
||||
import org.elasticsearch.cluster.ack.ClusterStateUpdateRequest;
|
||||
import org.elasticsearch.cluster.ack.ClusterStateUpdateResponse;
|
||||
import org.elasticsearch.cluster.metadata.MetaData;
|
||||
|
@ -45,10 +40,8 @@ import org.elasticsearch.snapshots.SnapshotsService;
|
|||
import org.elasticsearch.transport.TransportService;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.*;
|
||||
import java.util.stream.Collectors;
|
||||
|
||||
import static org.elasticsearch.common.settings.Settings.Builder.EMPTY_SETTINGS;
|
||||
|
||||
|
@ -571,9 +564,10 @@ public class RepositoriesService extends AbstractComponent implements ClusterSta
|
|||
}
|
||||
|
||||
public String failureDescription() {
|
||||
StringBuilder builder = new StringBuilder('[');
|
||||
Joiner.on(", ").appendTo(builder, failures);
|
||||
return builder.append(']').toString();
|
||||
return Arrays
|
||||
.stream(failures)
|
||||
.map(failure -> failure.toString())
|
||||
.collect(Collectors.joining(", ", "[", "]"));
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.repositories.blobstore;
|
||||
|
||||
import com.google.common.io.ByteStreams;
|
||||
import org.apache.lucene.store.RateLimiter;
|
||||
import org.elasticsearch.ElasticsearchParseException;
|
||||
import org.elasticsearch.Version;
|
||||
|
@ -36,6 +35,7 @@ import org.elasticsearch.common.bytes.BytesArray;
|
|||
import org.elasticsearch.common.bytes.BytesReference;
|
||||
import org.elasticsearch.common.component.AbstractLifecycleComponent;
|
||||
import org.elasticsearch.common.compress.NotXContentException;
|
||||
import org.elasticsearch.common.io.Streams;
|
||||
import org.elasticsearch.common.io.stream.BytesStreamOutput;
|
||||
import org.elasticsearch.common.io.stream.OutputStreamStreamOutput;
|
||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||
|
@ -66,6 +66,7 @@ import java.io.FileNotFoundException;
|
|||
import java.io.IOException;
|
||||
import java.io.InputStream;
|
||||
import java.io.OutputStream;
|
||||
import java.nio.file.Files;
|
||||
import java.nio.file.NoSuchFileException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
|
@ -590,9 +591,10 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent<Rep
|
|||
*/
|
||||
protected List<SnapshotId> readSnapshotList() throws IOException {
|
||||
try (InputStream blob = snapshotsBlobContainer.readBlob(SNAPSHOTS_FILE)) {
|
||||
final byte[] data = ByteStreams.toByteArray(blob);
|
||||
BytesStreamOutput out = new BytesStreamOutput();
|
||||
Streams.copy(blob, out);
|
||||
ArrayList<SnapshotId> snapshots = new ArrayList<>();
|
||||
try (XContentParser parser = XContentHelper.createParser(new BytesArray(data))) {
|
||||
try (XContentParser parser = XContentHelper.createParser(out.bytes())) {
|
||||
if (parser.nextToken() == XContentParser.Token.START_OBJECT) {
|
||||
if (parser.nextToken() == XContentParser.Token.FIELD_NAME) {
|
||||
String currentFieldName = parser.currentName();
|
||||
|
|
|
@ -18,7 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.repositories.blobstore;
|
||||
|
||||
import com.google.common.io.ByteStreams;
|
||||
import org.apache.lucene.codecs.CodecUtil;
|
||||
import org.apache.lucene.index.CorruptIndexException;
|
||||
import org.apache.lucene.index.IndexFormatTooNewException;
|
||||
|
@ -29,6 +28,7 @@ import org.elasticsearch.common.blobstore.BlobContainer;
|
|||
import org.elasticsearch.common.bytes.BytesArray;
|
||||
import org.elasticsearch.common.bytes.BytesReference;
|
||||
import org.elasticsearch.common.compress.CompressorFactory;
|
||||
import org.elasticsearch.common.io.Streams;
|
||||
import org.elasticsearch.common.io.stream.BytesStreamOutput;
|
||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||
import org.elasticsearch.common.lucene.store.ByteArrayIndexInput;
|
||||
|
@ -93,7 +93,9 @@ public class ChecksumBlobStoreFormat<T extends ToXContent> extends BlobStoreForm
|
|||
*/
|
||||
public T readBlob(BlobContainer blobContainer, String blobName) throws IOException {
|
||||
try (InputStream inputStream = blobContainer.readBlob(blobName)) {
|
||||
byte[] bytes = ByteStreams.toByteArray(inputStream);
|
||||
ByteArrayOutputStream out = new ByteArrayOutputStream();
|
||||
Streams.copy(inputStream, out);
|
||||
final byte[] bytes = out.toByteArray();
|
||||
final String resourceDesc = "ChecksumBlobStoreFormat.readBlob(blob=\"" + blobName + "\")";
|
||||
try (ByteArrayIndexInput indexInput = new ByteArrayIndexInput(resourceDesc, bytes)) {
|
||||
CodecUtil.checksumEntireFile(indexInput);
|
||||
|
|
|
@ -18,10 +18,10 @@
|
|||
*/
|
||||
package org.elasticsearch.repositories.blobstore;
|
||||
|
||||
import com.google.common.io.ByteStreams;
|
||||
import org.elasticsearch.common.ParseFieldMatcher;
|
||||
import org.elasticsearch.common.blobstore.BlobContainer;
|
||||
import org.elasticsearch.common.bytes.BytesArray;
|
||||
import org.elasticsearch.common.io.Streams;
|
||||
import org.elasticsearch.common.io.stream.BytesStreamOutput;
|
||||
import org.elasticsearch.common.xcontent.FromXContentBuilder;
|
||||
import org.elasticsearch.common.xcontent.ToXContent;
|
||||
|
||||
|
@ -53,7 +53,9 @@ public class LegacyBlobStoreFormat<T extends ToXContent> extends BlobStoreFormat
|
|||
*/
|
||||
public T readBlob(BlobContainer blobContainer, String blobName) throws IOException {
|
||||
try (InputStream inputStream = blobContainer.readBlob(blobName)) {
|
||||
return read(new BytesArray(ByteStreams.toByteArray(inputStream)));
|
||||
BytesStreamOutput out = new BytesStreamOutput();
|
||||
Streams.copy(inputStream, out);
|
||||
return read(out.bytes());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -20,14 +20,13 @@
|
|||
package org.elasticsearch.search.aggregations;
|
||||
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
|
||||
import org.apache.lucene.index.LeafReaderContext;
|
||||
import org.apache.lucene.search.Collector;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.stream.StreamSupport;
|
||||
|
||||
/**
|
||||
* A Collector that can collect data in separate buckets.
|
||||
|
@ -58,7 +57,8 @@ public abstract class BucketCollector implements Collector {
|
|||
* Wrap the given collectors into a single instance.
|
||||
*/
|
||||
public static BucketCollector wrap(Iterable<? extends BucketCollector> collectorList) {
|
||||
final BucketCollector[] collectors = Iterables.toArray(collectorList, BucketCollector.class);
|
||||
final BucketCollector[] collectors =
|
||||
StreamSupport.stream(collectorList.spliterator(), false).toArray(size -> new BucketCollector[size]);
|
||||
switch (collectors.length) {
|
||||
case 0:
|
||||
return NO_OP_COLLECTOR;
|
||||
|
|
|
@ -19,12 +19,11 @@
|
|||
|
||||
package org.elasticsearch.search.aggregations;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
|
||||
import org.apache.lucene.search.LeafCollector;
|
||||
import org.apache.lucene.search.Scorer;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.stream.Stream;
|
||||
import java.util.stream.StreamSupport;
|
||||
|
||||
/**
|
||||
|
@ -44,9 +43,9 @@ public abstract class LeafBucketCollector implements LeafCollector {
|
|||
};
|
||||
|
||||
public static LeafBucketCollector wrap(Iterable<LeafBucketCollector> collectors) {
|
||||
final Iterable<LeafBucketCollector> actualCollectors =
|
||||
StreamSupport.stream(collectors.spliterator(), false).filter(c -> c != NO_OP_COLLECTOR)::iterator;
|
||||
final LeafBucketCollector[] colls = Iterables.toArray(actualCollectors, LeafBucketCollector.class);
|
||||
final Stream<LeafBucketCollector> actualCollectors =
|
||||
StreamSupport.stream(collectors.spliterator(), false).filter(c -> c != NO_OP_COLLECTOR);
|
||||
final LeafBucketCollector[] colls = actualCollectors.toArray(size -> new LeafBucketCollector[size]);
|
||||
switch (colls.length) {
|
||||
case 0:
|
||||
return NO_OP_COLLECTOR;
|
||||
|
|
|
@ -18,8 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.search.aggregations.bucket.terms;
|
||||
|
||||
import com.google.common.collect.ArrayListMultimap;
|
||||
import com.google.common.collect.Multimap;
|
||||
import org.elasticsearch.common.io.stream.Streamable;
|
||||
import org.elasticsearch.common.xcontent.ToXContent;
|
||||
import org.elasticsearch.search.aggregations.AggregationExecutionException;
|
||||
|
@ -33,7 +31,6 @@ import org.elasticsearch.search.aggregations.support.format.ValueFormatter;
|
|||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
@ -167,7 +164,7 @@ public abstract class InternalTerms<A extends InternalTerms, B extends InternalT
|
|||
@Override
|
||||
public InternalAggregation doReduce(List<InternalAggregation> aggregations, ReduceContext reduceContext) {
|
||||
|
||||
Multimap<Object, InternalTerms.Bucket> buckets = ArrayListMultimap.create();
|
||||
Map<Object, List<InternalTerms.Bucket>> buckets = new HashMap<>();
|
||||
long sumDocCountError = 0;
|
||||
long otherDocCount = 0;
|
||||
InternalTerms<A, B> referenceTerms = null;
|
||||
|
@ -208,14 +205,18 @@ public abstract class InternalTerms<A extends InternalTerms, B extends InternalT
|
|||
terms.docCountError = thisAggDocCountError;
|
||||
for (Bucket bucket : terms.buckets) {
|
||||
bucket.docCountError = thisAggDocCountError;
|
||||
buckets.put(bucket.getKey(), bucket);
|
||||
List<Bucket> bucketList = buckets.get(bucket.getKey());
|
||||
if (bucketList == null) {
|
||||
bucketList = new ArrayList<>();
|
||||
buckets.put(bucket.getKey(), bucketList);
|
||||
}
|
||||
bucketList.add(bucket);
|
||||
}
|
||||
}
|
||||
|
||||
final int size = Math.min(requiredSize, buckets.size());
|
||||
BucketPriorityQueue ordered = new BucketPriorityQueue(size, order.comparator(null));
|
||||
for (Collection<Bucket> l : buckets.asMap().values()) {
|
||||
List<Bucket> sameTermBuckets = (List<Bucket>) l; // cast is ok according to javadocs
|
||||
for (List<Bucket> sameTermBuckets : buckets.values()) {
|
||||
final Bucket b = sameTermBuckets.get(0).reduce(sameTermBuckets, reduceContext);
|
||||
if (b.docCountError != -1) {
|
||||
if (sumDocCountError == -1) {
|
||||
|
|
|
@ -20,8 +20,6 @@
|
|||
package org.elasticsearch.search.aggregations.metrics.cardinality;
|
||||
|
||||
import com.carrotsearch.hppc.BitMixer;
|
||||
import com.google.common.base.Preconditions;
|
||||
|
||||
import org.apache.lucene.index.LeafReaderContext;
|
||||
import org.apache.lucene.index.RandomAccessOrds;
|
||||
import org.apache.lucene.index.SortedNumericDocValues;
|
||||
|
@ -234,7 +232,9 @@ public class CardinalityAggregator extends NumericMetricsAggregator.SingleValue
|
|||
private ObjectArray<FixedBitSet> visitedOrds;
|
||||
|
||||
OrdinalsCollector(HyperLogLogPlusPlus counts, RandomAccessOrds values, BigArrays bigArrays) {
|
||||
Preconditions.checkArgument(values.getValueCount() <= Integer.MAX_VALUE);
|
||||
if (values.getValueCount() > Integer.MAX_VALUE) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
maxOrd = (int) values.getValueCount();
|
||||
this.bigArrays = bigArrays;
|
||||
this.counts = counts;
|
||||
|
|
|
@ -19,12 +19,10 @@
|
|||
|
||||
package org.elasticsearch.search.aggregations.metrics.cardinality;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.LongBitSet;
|
||||
import org.apache.lucene.util.RamUsageEstimator;
|
||||
import org.apache.lucene.util.packed.PackedInts;
|
||||
import org.elasticsearch.ElasticsearchException;
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||
import org.elasticsearch.common.lease.Releasable;
|
||||
|
@ -37,7 +35,6 @@ import org.elasticsearch.common.util.IntArray;
|
|||
import java.io.IOException;
|
||||
import java.nio.ByteBuffer;
|
||||
import java.nio.ByteOrder;
|
||||
import java.util.Arrays;
|
||||
|
||||
/**
|
||||
* Hyperloglog++ counter, implemented based on pseudo code from
|
||||
|
@ -162,8 +159,12 @@ public final class HyperLogLogPlusPlus implements Releasable {
|
|||
private final double alphaMM;
|
||||
|
||||
public HyperLogLogPlusPlus(int precision, BigArrays bigArrays, long initialBucketCount) {
|
||||
Preconditions.checkArgument(precision >= 4, "precision must be >= 4");
|
||||
Preconditions.checkArgument(precision <= 18, "precision must be <= 18");
|
||||
if (precision < 4) {
|
||||
throw new IllegalArgumentException("precision must be >= 4");
|
||||
}
|
||||
if (precision > 18) {
|
||||
throw new IllegalArgumentException("precision must be <= 18");
|
||||
}
|
||||
p = precision;
|
||||
m = 1 << p;
|
||||
this.bigArrays = bigArrays;
|
||||
|
@ -198,7 +199,9 @@ public final class HyperLogLogPlusPlus implements Releasable {
|
|||
}
|
||||
|
||||
public void merge(long thisBucket, HyperLogLogPlusPlus other, long otherBucket) {
|
||||
Preconditions.checkArgument(p == other.p);
|
||||
if (p != other.p) {
|
||||
throw new IllegalArgumentException();
|
||||
}
|
||||
ensureCapacity(thisBucket + 1);
|
||||
if (other.algorithm.get(otherBucket) == LINEAR_COUNTING) {
|
||||
final IntArray values = other.hashSet.values(otherBucket);
|
||||
|
|
|
@ -18,8 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.search.aggregations.metrics.percentiles.hdr;
|
||||
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
|
||||
import org.HdrHistogram.DoubleHistogram;
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.search.aggregations.AggregationStreams;
|
||||
|
@ -106,7 +104,7 @@ public class InternalHDRPercentileRanks extends AbstractInternalHDRPercentiles i
|
|||
return percentileRank;
|
||||
}
|
||||
|
||||
public static class Iter extends UnmodifiableIterator<Percentile> {
|
||||
public static class Iter implements Iterator<Percentile> {
|
||||
|
||||
private final double[] values;
|
||||
private final DoubleHistogram state;
|
||||
|
@ -129,5 +127,10 @@ public class InternalHDRPercentileRanks extends AbstractInternalHDRPercentiles i
|
|||
++i;
|
||||
return next;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -18,8 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.search.aggregations.metrics.percentiles.hdr;
|
||||
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
|
||||
import org.HdrHistogram.DoubleHistogram;
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.search.aggregations.AggregationStreams;
|
||||
|
@ -96,7 +94,7 @@ public class InternalHDRPercentiles extends AbstractInternalHDRPercentiles imple
|
|||
return TYPE;
|
||||
}
|
||||
|
||||
public static class Iter extends UnmodifiableIterator<Percentile> {
|
||||
public static class Iter implements Iterator<Percentile> {
|
||||
|
||||
private final double[] percents;
|
||||
private final DoubleHistogram state;
|
||||
|
@ -119,5 +117,10 @@ public class InternalHDRPercentiles extends AbstractInternalHDRPercentiles imple
|
|||
++i;
|
||||
return next;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -18,8 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.search.aggregations.metrics.percentiles.tdigest;
|
||||
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.search.aggregations.AggregationStreams;
|
||||
import org.elasticsearch.search.aggregations.metrics.percentiles.InternalPercentile;
|
||||
|
@ -102,7 +100,7 @@ public class InternalTDigestPercentileRanks extends AbstractInternalTDigestPerce
|
|||
return percentileRank * 100;
|
||||
}
|
||||
|
||||
public static class Iter extends UnmodifiableIterator<Percentile> {
|
||||
public static class Iter implements Iterator<Percentile> {
|
||||
|
||||
private final double[] values;
|
||||
private final TDigestState state;
|
||||
|
@ -125,5 +123,10 @@ public class InternalTDigestPercentileRanks extends AbstractInternalTDigestPerce
|
|||
++i;
|
||||
return next;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -18,8 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.search.aggregations.metrics.percentiles.tdigest;
|
||||
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.search.aggregations.AggregationStreams;
|
||||
import org.elasticsearch.search.aggregations.metrics.percentiles.InternalPercentile;
|
||||
|
@ -92,7 +90,7 @@ public class InternalTDigestPercentiles extends AbstractInternalTDigestPercentil
|
|||
return TYPE;
|
||||
}
|
||||
|
||||
public static class Iter extends UnmodifiableIterator<Percentile> {
|
||||
public static class Iter implements Iterator<Percentile> {
|
||||
|
||||
private final double[] percents;
|
||||
private final TDigestState state;
|
||||
|
@ -115,5 +113,10 @@ public class InternalTDigestPercentiles extends AbstractInternalTDigestPercentil
|
|||
++i;
|
||||
return next;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.search.aggregations.pipeline.bucketmetrics.percentile;
|
||||
|
||||
import com.google.common.collect.UnmodifiableIterator;
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
|
@ -136,7 +135,7 @@ public class InternalPercentilesBucket extends InternalNumericMetricsAggregation
|
|||
return builder;
|
||||
}
|
||||
|
||||
public static class Iter extends UnmodifiableIterator<Percentile> {
|
||||
public static class Iter implements Iterator<Percentile> {
|
||||
|
||||
private final double[] percents;
|
||||
private final double[] percentiles;
|
||||
|
@ -159,5 +158,10 @@ public class InternalPercentilesBucket extends InternalNumericMetricsAggregation
|
|||
++i;
|
||||
return next;
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void remove() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -18,9 +18,6 @@
|
|||
*/
|
||||
package org.elasticsearch.search.internal;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
import com.google.common.collect.Multimap;
|
||||
import com.google.common.collect.MultimapBuilder;
|
||||
|
||||
import org.apache.lucene.search.Collector;
|
||||
import org.apache.lucene.search.Query;
|
||||
|
@ -35,6 +32,7 @@ import org.elasticsearch.common.ParseFieldMatcher;
|
|||
import org.elasticsearch.common.lease.Releasable;
|
||||
import org.elasticsearch.common.lease.Releasables;
|
||||
import org.elasticsearch.common.util.BigArrays;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.index.analysis.AnalysisService;
|
||||
import org.elasticsearch.index.cache.bitset.BitsetFilterCache;
|
||||
import org.elasticsearch.index.fielddata.IndexFieldDataService;
|
||||
|
@ -62,10 +60,7 @@ import org.elasticsearch.search.query.QuerySearchResult;
|
|||
import org.elasticsearch.search.rescore.RescoreSearchContext;
|
||||
import org.elasticsearch.search.suggest.SuggestionSearchContext;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collection;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.*;
|
||||
import java.util.concurrent.atomic.AtomicBoolean;
|
||||
|
||||
public abstract class SearchContext extends DelegatingHasContextAndHeaders implements Releasable {
|
||||
|
@ -87,7 +82,7 @@ public abstract class SearchContext extends DelegatingHasContextAndHeaders imple
|
|||
return current.get();
|
||||
}
|
||||
|
||||
private Multimap<Lifetime, Releasable> clearables = null;
|
||||
private Map<Lifetime, List<Releasable>> clearables = null;
|
||||
private final AtomicBoolean closed = new AtomicBoolean(false);
|
||||
|
||||
protected final ParseFieldMatcher parseFieldMatcher;
|
||||
|
@ -316,21 +311,29 @@ public abstract class SearchContext extends DelegatingHasContextAndHeaders imple
|
|||
*/
|
||||
public void addReleasable(Releasable releasable, Lifetime lifetime) {
|
||||
if (clearables == null) {
|
||||
clearables = MultimapBuilder.enumKeys(Lifetime.class).arrayListValues().build();
|
||||
clearables = new HashMap<>();
|
||||
}
|
||||
clearables.put(lifetime, releasable);
|
||||
List<Releasable> releasables = clearables.get(lifetime);
|
||||
if (releasables == null) {
|
||||
releasables = new ArrayList<>();
|
||||
clearables.put(lifetime, releasables);
|
||||
}
|
||||
releasables.add(releasable);
|
||||
}
|
||||
|
||||
public void clearReleasables(Lifetime lifetime) {
|
||||
if (clearables != null) {
|
||||
List<Collection<Releasable>> releasables = new ArrayList<>();
|
||||
List<List<Releasable>>releasables = new ArrayList<>();
|
||||
for (Lifetime lc : Lifetime.values()) {
|
||||
if (lc.compareTo(lifetime) > 0) {
|
||||
break;
|
||||
}
|
||||
releasables.add(clearables.removeAll(lc));
|
||||
List<Releasable> remove = clearables.remove(lc);
|
||||
if (remove != null) {
|
||||
releasables.add(remove);
|
||||
}
|
||||
}
|
||||
Releasables.close(Iterables.concat(releasables));
|
||||
Releasables.close(Iterables.flatten(releasables));
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -19,8 +19,6 @@
|
|||
|
||||
package org.elasticsearch.search.suggest.context;
|
||||
|
||||
import com.google.common.base.Joiner;
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.apache.lucene.analysis.PrefixAnalyzer;
|
||||
import org.apache.lucene.analysis.TokenStream;
|
||||
import org.apache.lucene.index.IndexableField;
|
||||
|
@ -28,6 +26,7 @@ import org.apache.lucene.util.automaton.Automata;
|
|||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
import org.elasticsearch.ElasticsearchParseException;
|
||||
import org.elasticsearch.common.util.iterable.Iterables;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.common.xcontent.XContentParser.Token;
|
||||
|
@ -35,11 +34,9 @@ import org.elasticsearch.index.mapper.ParseContext;
|
|||
import org.elasticsearch.index.mapper.ParseContext.Document;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collections;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.*;
|
||||
import java.util.stream.Collectors;
|
||||
import java.util.stream.StreamSupport;
|
||||
|
||||
/**
|
||||
* The {@link CategoryContextMapping} is used to define a {@link ContextMapping} that
|
||||
|
@ -213,7 +210,7 @@ public class CategoryContextMapping extends ContextMapping {
|
|||
if (obj instanceof CategoryContextMapping) {
|
||||
CategoryContextMapping other = (CategoryContextMapping) obj;
|
||||
if (this.fieldName.equals(other.fieldName)) {
|
||||
return Iterables.elementsEqual(this.defaultValues, other.defaultValues);
|
||||
return Iterables.allElementsAreEqual(this.defaultValues, other.defaultValues);
|
||||
}
|
||||
}
|
||||
return false;
|
||||
|
@ -262,14 +259,18 @@ public class CategoryContextMapping extends ContextMapping {
|
|||
public String toString() {
|
||||
StringBuilder sb = new StringBuilder("FieldConfig(" + fieldname + " = [");
|
||||
if (this.values != null && this.values.iterator().hasNext()) {
|
||||
sb.append("(").append(Joiner.on(", ").join(this.values.iterator())).append(")");
|
||||
sb.append(delimitValues(this.values));
|
||||
}
|
||||
if (this.defaultValues != null && this.defaultValues.iterator().hasNext()) {
|
||||
sb.append(" default(").append(Joiner.on(", ").join(this.defaultValues.iterator())).append(")");
|
||||
sb.append(" default").append(delimitValues(this.defaultValues));
|
||||
}
|
||||
return sb.append("])").toString();
|
||||
}
|
||||
|
||||
private String delimitValues(Iterable<? extends CharSequence> values) {
|
||||
return StreamSupport.stream(values.spliterator(), false).collect(Collectors.joining(", ", "(", ")"));
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
private static class FieldQuery extends ContextQuery {
|
||||
|
|
|
@ -19,7 +19,6 @@
|
|||
|
||||
package org.elasticsearch.search.suggest.context;
|
||||
|
||||
import com.google.common.collect.Iterables;
|
||||
import org.apache.lucene.analysis.TokenStream;
|
||||
import org.apache.lucene.search.suggest.analyzing.XAnalyzingSuggester;
|
||||
import org.apache.lucene.util.automaton.Automata;
|
||||
|
@ -36,13 +35,7 @@ import org.elasticsearch.index.mapper.ParseContext;
|
|||
import org.elasticsearch.index.mapper.ParseContext.Document;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.SortedMap;
|
||||
import java.util.TreeMap;
|
||||
import java.util.*;
|
||||
|
||||
/**
|
||||
* A {@link ContextMapping} is used t define a context that may used
|
||||
|
@ -157,7 +150,7 @@ public abstract class ContextMapping implements ToXContent {
|
|||
* @return true if both arguments are equal
|
||||
*/
|
||||
public static boolean mappingsAreEqual(SortedMap<String, ? extends ContextMapping> thisMappings, SortedMap<String, ? extends ContextMapping> otherMappings) {
|
||||
return Iterables.elementsEqual(thisMappings.entrySet(), otherMappings.entrySet());
|
||||
return thisMappings.entrySet().equals(otherMappings.entrySet());
|
||||
}
|
||||
|
||||
@Override
|
||||
|
|
|
@ -19,9 +19,6 @@
|
|||
package org.elasticsearch;
|
||||
|
||||
import junit.framework.TestCase;
|
||||
|
||||
import com.google.common.base.Joiner;
|
||||
|
||||
import org.apache.lucene.util.LuceneTestCase;
|
||||
import org.elasticsearch.common.io.PathUtils;
|
||||
import org.elasticsearch.test.ESIntegTestCase;
|
||||
|
@ -40,6 +37,7 @@ import java.nio.file.Path;
|
|||
import java.nio.file.attribute.BasicFileAttributes;
|
||||
import java.util.HashSet;
|
||||
import java.util.Set;
|
||||
import java.util.stream.Collectors;
|
||||
|
||||
/**
|
||||
* Simple class that ensures that all subclasses concrete of ESTestCase end with either Test | Tests
|
||||
|
@ -149,24 +147,27 @@ public class NamingConventionTests extends ESTestCase {
|
|||
assertTrue(pureUnitTest.remove(PlainUnit.class));
|
||||
assertTrue(pureUnitTest.remove(PlainUnitTheSecond.class));
|
||||
|
||||
String classesToSubclass = Joiner.on(',').join(
|
||||
ESTestCase.class.getSimpleName(),
|
||||
ESTestCase.class.getSimpleName(),
|
||||
ESTokenStreamTestCase.class.getSimpleName(),
|
||||
LuceneTestCase.class.getSimpleName());
|
||||
assertTrue("Not all subclasses of " + ESTestCase.class.getSimpleName() +
|
||||
" match the naming convention. Concrete classes must end with [Tests]:\n" + Joiner.on('\n').join(missingSuffix),
|
||||
missingSuffix.isEmpty());
|
||||
assertTrue("Classes ending with [Tests] are abstract or interfaces:\n" + Joiner.on('\n').join(notRunnable),
|
||||
notRunnable.isEmpty());
|
||||
assertTrue("Found inner classes that are tests, which are excluded from the test runner:\n" + Joiner.on('\n').join(innerClasses),
|
||||
innerClasses.isEmpty());
|
||||
assertTrue("Pure Unit-Test found must subclass one of [" + classesToSubclass +"]:\n" + Joiner.on('\n').join(pureUnitTest),
|
||||
pureUnitTest.isEmpty());
|
||||
assertTrue("Classes ending with [Tests] must subclass [" + classesToSubclass + "]:\n" + Joiner.on('\n').join(notImplementing),
|
||||
notImplementing.isEmpty());
|
||||
assertTrue("Subclasses of ESIntegTestCase should end with IT as they are integration tests:\n" + Joiner.on('\n').join(integTestsInDisguise),
|
||||
integTestsInDisguise.isEmpty());
|
||||
String classesToSubclass = String.join(
|
||||
",",
|
||||
ESTestCase.class.getSimpleName(),
|
||||
ESTestCase.class.getSimpleName(),
|
||||
ESTokenStreamTestCase.class.getSimpleName(),
|
||||
LuceneTestCase.class.getSimpleName()
|
||||
);
|
||||
assertNoViolations("Not all subclasses of " + ESTestCase.class.getSimpleName() + " match the naming convention. Concrete classes must end with [Tests]:\n", missingSuffix);
|
||||
assertNoViolations("Classes ending with [Tests] are abstract or interfaces:\n", notRunnable);
|
||||
assertNoViolations("Found inner classes that are tests, which are excluded from the test runner:\n", innerClasses);
|
||||
assertNoViolations("Pure Unit-Test found must subclass one of [" + classesToSubclass + "]:\n", pureUnitTest);
|
||||
assertNoViolations("Classes ending with [Tests] must subclass [" + classesToSubclass + "]:\n", notImplementing);
|
||||
assertNoViolations("Subclasses of ESIntegTestCase should end with IT as they are integration tests:\n", integTestsInDisguise);
|
||||
}
|
||||
|
||||
private String join(Set<Class> set) {
|
||||
return set.stream().map(Object::toString).collect(Collectors.joining("\n"));
|
||||
}
|
||||
|
||||
private void assertNoViolations(String message, Set<Class> set) {
|
||||
assertTrue(message + join(set), set.isEmpty());
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
@ -161,7 +161,7 @@ public class ClusterInfoServiceIT extends ESIntegTestCase {
|
|||
}
|
||||
for (DiskUsage usage : mostUsages.values()) {
|
||||
logger.info("--> usage: {}", usage);
|
||||
assertThat("usage has be retrieved", usage.getFreeBytes(), greaterThan(0L));
|
||||
assertThat("usage has be retrieved", usage.getFreeBytes(), greaterThanOrEqualTo(0L));
|
||||
}
|
||||
for (Long size : shardSizes.values()) {
|
||||
logger.info("--> shard size: {}", size);
|
||||
|
|
|
@ -19,9 +19,12 @@
|
|||
|
||||
package org.elasticsearch.common.util;
|
||||
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.elasticsearch.test.ESTestCase;
|
||||
import org.junit.Test;
|
||||
|
||||
import java.lang.reflect.Array;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.BitSet;
|
||||
|
||||
|
@ -90,4 +93,22 @@ public class ArrayUtilsTests extends ESTestCase {
|
|||
return min + delta;
|
||||
}
|
||||
|
||||
public void testConcat() {
|
||||
assertArrayEquals(new String[]{"a", "b", "c", "d"}, ArrayUtils.concat(new String[]{"a", "b"}, new String[]{"c", "d"}));
|
||||
int firstSize = randomIntBetween(0, 10);
|
||||
String[] first = new String[firstSize];
|
||||
ArrayList<String> sourceOfTruth = new ArrayList<>();
|
||||
for (int i = 0; i < firstSize; i++) {
|
||||
first[i] = randomRealisticUnicodeOfCodepointLengthBetween(0,10);
|
||||
sourceOfTruth.add(first[i]);
|
||||
}
|
||||
int secondSize = randomIntBetween(0, 10);
|
||||
String[] second = new String[secondSize];
|
||||
for (int i = 0; i < secondSize; i++) {
|
||||
second[i] = randomRealisticUnicodeOfCodepointLengthBetween(0, 10);
|
||||
sourceOfTruth.add(second[i]);
|
||||
}
|
||||
assertArrayEquals(sourceOfTruth.toArray(new String[0]), ArrayUtils.concat(first, second));
|
||||
}
|
||||
|
||||
}
|
||||
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue