mirror of
https://github.com/honeymoose/OpenSearch.git
synced 2025-03-26 01:48:45 +00:00
1023 lines
55 KiB
Java
1023 lines
55 KiB
Java
/*
|
|
* Licensed to Elasticsearch under one or more contributor
|
|
* license agreements. See the NOTICE file distributed with
|
|
* this work for additional information regarding copyright
|
|
* ownership. Elasticsearch licenses this file to you under
|
|
* the Apache License, Version 2.0 (the "License"); you may
|
|
* not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing,
|
|
* software distributed under the License is distributed on an
|
|
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
* KIND, either express or implied. See the License for the
|
|
* specific language governing permissions and limitations
|
|
* under the License.
|
|
*/
|
|
|
|
package org.elasticsearch.node;
|
|
|
|
import org.apache.logging.log4j.LogManager;
|
|
import org.apache.logging.log4j.Logger;
|
|
import org.apache.lucene.util.Constants;
|
|
import org.apache.lucene.util.SetOnce;
|
|
import org.elasticsearch.Build;
|
|
import org.elasticsearch.ElasticsearchException;
|
|
import org.elasticsearch.ElasticsearchTimeoutException;
|
|
import org.elasticsearch.action.Action;
|
|
import org.elasticsearch.action.ActionModule;
|
|
import org.elasticsearch.action.search.SearchExecutionStatsCollector;
|
|
import org.elasticsearch.action.search.SearchPhaseController;
|
|
import org.elasticsearch.action.search.SearchTransportService;
|
|
import org.elasticsearch.action.support.TransportAction;
|
|
import org.elasticsearch.action.update.UpdateHelper;
|
|
import org.elasticsearch.bootstrap.BootstrapCheck;
|
|
import org.elasticsearch.bootstrap.BootstrapContext;
|
|
import org.elasticsearch.client.Client;
|
|
import org.elasticsearch.client.node.NodeClient;
|
|
import org.elasticsearch.cluster.ClusterInfo;
|
|
import org.elasticsearch.cluster.ClusterInfoService;
|
|
import org.elasticsearch.cluster.ClusterModule;
|
|
import org.elasticsearch.cluster.ClusterState;
|
|
import org.elasticsearch.cluster.ClusterStateObserver;
|
|
import org.elasticsearch.cluster.InternalClusterInfoService;
|
|
import org.elasticsearch.cluster.NodeConnectionsService;
|
|
import org.elasticsearch.cluster.action.index.MappingUpdatedAction;
|
|
import org.elasticsearch.cluster.metadata.AliasValidator;
|
|
import org.elasticsearch.cluster.metadata.IndexMetaData;
|
|
import org.elasticsearch.cluster.metadata.IndexTemplateMetaData;
|
|
import org.elasticsearch.cluster.metadata.MetaData;
|
|
import org.elasticsearch.cluster.metadata.MetaDataCreateIndexService;
|
|
import org.elasticsearch.cluster.metadata.MetaDataIndexUpgradeService;
|
|
import org.elasticsearch.cluster.metadata.TemplateUpgradeService;
|
|
import org.elasticsearch.cluster.node.DiscoveryNode;
|
|
import org.elasticsearch.cluster.routing.RoutingService;
|
|
import org.elasticsearch.cluster.routing.allocation.DiskThresholdMonitor;
|
|
import org.elasticsearch.cluster.service.ClusterService;
|
|
import org.elasticsearch.common.StopWatch;
|
|
import org.elasticsearch.common.breaker.CircuitBreaker;
|
|
import org.elasticsearch.common.component.Lifecycle;
|
|
import org.elasticsearch.common.component.LifecycleComponent;
|
|
import org.elasticsearch.common.inject.Injector;
|
|
import org.elasticsearch.common.inject.Key;
|
|
import org.elasticsearch.common.inject.Module;
|
|
import org.elasticsearch.common.inject.ModulesBuilder;
|
|
import org.elasticsearch.common.io.stream.NamedWriteableRegistry;
|
|
import org.elasticsearch.common.lease.Releasables;
|
|
import org.elasticsearch.common.logging.DeprecationLogger;
|
|
import org.elasticsearch.common.logging.NodeAndClusterIdStateListener;
|
|
import org.elasticsearch.common.network.NetworkAddress;
|
|
import org.elasticsearch.common.network.NetworkModule;
|
|
import org.elasticsearch.common.network.NetworkService;
|
|
import org.elasticsearch.common.settings.ClusterSettings;
|
|
import org.elasticsearch.common.settings.Setting;
|
|
import org.elasticsearch.common.settings.Setting.Property;
|
|
import org.elasticsearch.common.settings.SettingUpgrader;
|
|
import org.elasticsearch.common.settings.Settings;
|
|
import org.elasticsearch.common.settings.SettingsModule;
|
|
import org.elasticsearch.common.transport.BoundTransportAddress;
|
|
import org.elasticsearch.common.transport.TransportAddress;
|
|
import org.elasticsearch.common.unit.TimeValue;
|
|
import org.elasticsearch.common.util.BigArrays;
|
|
import org.elasticsearch.common.util.PageCacheRecycler;
|
|
import org.elasticsearch.common.xcontent.NamedXContentRegistry;
|
|
import org.elasticsearch.core.internal.io.IOUtils;
|
|
import org.elasticsearch.discovery.Discovery;
|
|
import org.elasticsearch.discovery.DiscoveryModule;
|
|
import org.elasticsearch.discovery.DiscoverySettings;
|
|
import org.elasticsearch.env.Environment;
|
|
import org.elasticsearch.env.NodeEnvironment;
|
|
import org.elasticsearch.gateway.GatewayAllocator;
|
|
import org.elasticsearch.gateway.GatewayMetaState;
|
|
import org.elasticsearch.gateway.GatewayModule;
|
|
import org.elasticsearch.gateway.GatewayService;
|
|
import org.elasticsearch.gateway.MetaStateService;
|
|
import org.elasticsearch.http.HttpServerTransport;
|
|
import org.elasticsearch.index.IndexSettings;
|
|
import org.elasticsearch.index.analysis.AnalysisRegistry;
|
|
import org.elasticsearch.index.engine.EngineFactory;
|
|
import org.elasticsearch.index.store.IndexStore;
|
|
import org.elasticsearch.indices.IndicesModule;
|
|
import org.elasticsearch.indices.IndicesService;
|
|
import org.elasticsearch.indices.analysis.AnalysisModule;
|
|
import org.elasticsearch.indices.breaker.CircuitBreakerService;
|
|
import org.elasticsearch.indices.breaker.HierarchyCircuitBreakerService;
|
|
import org.elasticsearch.indices.breaker.NoneCircuitBreakerService;
|
|
import org.elasticsearch.indices.cluster.IndicesClusterStateService;
|
|
import org.elasticsearch.indices.recovery.PeerRecoverySourceService;
|
|
import org.elasticsearch.indices.recovery.PeerRecoveryTargetService;
|
|
import org.elasticsearch.indices.recovery.RecoverySettings;
|
|
import org.elasticsearch.indices.store.IndicesStore;
|
|
import org.elasticsearch.ingest.IngestService;
|
|
import org.elasticsearch.monitor.MonitorService;
|
|
import org.elasticsearch.monitor.jvm.JvmInfo;
|
|
import org.elasticsearch.persistent.PersistentTasksClusterService;
|
|
import org.elasticsearch.persistent.PersistentTasksExecutor;
|
|
import org.elasticsearch.persistent.PersistentTasksExecutorRegistry;
|
|
import org.elasticsearch.persistent.PersistentTasksService;
|
|
import org.elasticsearch.plugins.ActionPlugin;
|
|
import org.elasticsearch.plugins.AnalysisPlugin;
|
|
import org.elasticsearch.plugins.ClusterPlugin;
|
|
import org.elasticsearch.plugins.DiscoveryPlugin;
|
|
import org.elasticsearch.plugins.EnginePlugin;
|
|
import org.elasticsearch.plugins.IndexStorePlugin;
|
|
import org.elasticsearch.plugins.IngestPlugin;
|
|
import org.elasticsearch.plugins.MapperPlugin;
|
|
import org.elasticsearch.plugins.MetaDataUpgrader;
|
|
import org.elasticsearch.plugins.NetworkPlugin;
|
|
import org.elasticsearch.plugins.PersistentTaskPlugin;
|
|
import org.elasticsearch.plugins.Plugin;
|
|
import org.elasticsearch.plugins.PluginsService;
|
|
import org.elasticsearch.plugins.RepositoryPlugin;
|
|
import org.elasticsearch.plugins.ScriptPlugin;
|
|
import org.elasticsearch.plugins.SearchPlugin;
|
|
import org.elasticsearch.repositories.RepositoriesModule;
|
|
import org.elasticsearch.rest.RestController;
|
|
import org.elasticsearch.script.ScriptModule;
|
|
import org.elasticsearch.script.ScriptService;
|
|
import org.elasticsearch.search.SearchModule;
|
|
import org.elasticsearch.search.SearchService;
|
|
import org.elasticsearch.search.fetch.FetchPhase;
|
|
import org.elasticsearch.snapshots.SnapshotShardsService;
|
|
import org.elasticsearch.snapshots.SnapshotsService;
|
|
import org.elasticsearch.tasks.Task;
|
|
import org.elasticsearch.tasks.TaskResultsService;
|
|
import org.elasticsearch.threadpool.ExecutorBuilder;
|
|
import org.elasticsearch.threadpool.ThreadPool;
|
|
import org.elasticsearch.transport.Transport;
|
|
import org.elasticsearch.transport.TransportInterceptor;
|
|
import org.elasticsearch.transport.TransportService;
|
|
import org.elasticsearch.usage.UsageService;
|
|
import org.elasticsearch.watcher.ResourceWatcherService;
|
|
|
|
import javax.net.ssl.SNIHostName;
|
|
import java.io.BufferedWriter;
|
|
import java.io.Closeable;
|
|
import java.io.IOException;
|
|
import java.net.InetAddress;
|
|
import java.net.InetSocketAddress;
|
|
import java.nio.charset.Charset;
|
|
import java.nio.file.Files;
|
|
import java.nio.file.Path;
|
|
import java.nio.file.StandardCopyOption;
|
|
import java.util.ArrayList;
|
|
import java.util.Arrays;
|
|
import java.util.Collection;
|
|
import java.util.Collections;
|
|
import java.util.List;
|
|
import java.util.Map;
|
|
import java.util.Optional;
|
|
import java.util.Set;
|
|
import java.util.concurrent.CountDownLatch;
|
|
import java.util.concurrent.TimeUnit;
|
|
import java.util.function.Consumer;
|
|
import java.util.function.Function;
|
|
import java.util.function.UnaryOperator;
|
|
import java.util.stream.Collectors;
|
|
import java.util.stream.Stream;
|
|
|
|
import static java.util.stream.Collectors.toList;
|
|
|
|
/**
|
|
* A node represent a node within a cluster ({@code cluster.name}). The {@link #client()} can be used
|
|
* in order to use a {@link Client} to perform actions/operations against the cluster.
|
|
*/
|
|
public class Node implements Closeable {
|
|
public static final Setting<Boolean> WRITE_PORTS_FILE_SETTING =
|
|
Setting.boolSetting("node.portsfile", false, Property.NodeScope);
|
|
public static final Setting<Boolean> NODE_DATA_SETTING = Setting.boolSetting("node.data", true, Property.NodeScope);
|
|
public static final Setting<Boolean> NODE_MASTER_SETTING =
|
|
Setting.boolSetting("node.master", true, Property.NodeScope);
|
|
public static final Setting<Boolean> NODE_INGEST_SETTING =
|
|
Setting.boolSetting("node.ingest", true, Property.NodeScope);
|
|
|
|
/**
|
|
* controls whether the node is allowed to persist things like metadata to disk
|
|
* Note that this does not control whether the node stores actual indices (see
|
|
* {@link #NODE_DATA_SETTING}). However, if this is false, {@link #NODE_DATA_SETTING}
|
|
* and {@link #NODE_MASTER_SETTING} must also be false.
|
|
*
|
|
*/
|
|
public static final Setting<Boolean> NODE_LOCAL_STORAGE_SETTING = Setting.boolSetting("node.local_storage", true, Property.NodeScope);
|
|
public static final Setting<String> NODE_NAME_SETTING = Setting.simpleString("node.name", Property.NodeScope);
|
|
public static final Setting.AffixSetting<String> NODE_ATTRIBUTES = Setting.prefixKeySetting("node.attr.", (key) ->
|
|
new Setting<>(key, "", (value) -> {
|
|
if (value.length() > 0
|
|
&& (Character.isWhitespace(value.charAt(0)) || Character.isWhitespace(value.charAt(value.length() - 1)))) {
|
|
throw new IllegalArgumentException(key + " cannot have leading or trailing whitespace " +
|
|
"[" + value + "]");
|
|
}
|
|
if (value.length() > 0 && "node.attr.server_name".equals(key)) {
|
|
try {
|
|
new SNIHostName(value);
|
|
} catch (IllegalArgumentException e) {
|
|
throw new IllegalArgumentException("invalid node.attr.server_name [" + value + "]", e );
|
|
}
|
|
}
|
|
return value;
|
|
}, Property.NodeScope));
|
|
public static final Setting<String> BREAKER_TYPE_KEY = new Setting<>("indices.breaker.type", "hierarchy", (s) -> {
|
|
switch (s) {
|
|
case "hierarchy":
|
|
case "none":
|
|
return s;
|
|
default:
|
|
throw new IllegalArgumentException("indices.breaker.type must be one of [hierarchy, none] but was: " + s);
|
|
}
|
|
}, Setting.Property.NodeScope);
|
|
|
|
private static final String CLIENT_TYPE = "node";
|
|
|
|
private final Lifecycle lifecycle = new Lifecycle();
|
|
|
|
/**
|
|
* Logger initialized in the ctor because if it were initialized statically
|
|
* then it wouldn't get the node name.
|
|
*/
|
|
private final Logger logger;
|
|
private final Injector injector;
|
|
private final Environment environment;
|
|
private final NodeEnvironment nodeEnvironment;
|
|
private final PluginsService pluginsService;
|
|
private final NodeClient client;
|
|
private final Collection<LifecycleComponent> pluginLifecycleComponents;
|
|
private final LocalNodeFactory localNodeFactory;
|
|
private final NodeService nodeService;
|
|
|
|
public Node(Environment environment) {
|
|
this(environment, Collections.emptyList(), true);
|
|
}
|
|
|
|
/**
|
|
* Constructs a node
|
|
*
|
|
* @param environment the environment for this node
|
|
* @param classpathPlugins the plugins to be loaded from the classpath
|
|
* @param forbidPrivateIndexSettings whether or not private index settings are forbidden when creating an index; this is used in the
|
|
* test framework for tests that rely on being able to set private settings
|
|
*/
|
|
protected Node(
|
|
final Environment environment, Collection<Class<? extends Plugin>> classpathPlugins, boolean forbidPrivateIndexSettings) {
|
|
logger = LogManager.getLogger(Node.class);
|
|
final List<Closeable> resourcesToClose = new ArrayList<>(); // register everything we need to release in the case of an error
|
|
boolean success = false;
|
|
try {
|
|
Settings tmpSettings = Settings.builder().put(environment.settings())
|
|
.put(Client.CLIENT_TYPE_SETTING_S.getKey(), CLIENT_TYPE).build();
|
|
|
|
nodeEnvironment = new NodeEnvironment(tmpSettings, environment);
|
|
resourcesToClose.add(nodeEnvironment);
|
|
logger.info("node name [{}], node ID [{}]",
|
|
NODE_NAME_SETTING.get(tmpSettings), nodeEnvironment.nodeId());
|
|
|
|
final JvmInfo jvmInfo = JvmInfo.jvmInfo();
|
|
logger.info(
|
|
"version[{}], pid[{}], build[{}/{}/{}/{}], OS[{}/{}/{}], JVM[{}/{}/{}/{}]",
|
|
Build.CURRENT.getQualifiedVersion(),
|
|
jvmInfo.pid(),
|
|
Build.CURRENT.flavor().displayName(),
|
|
Build.CURRENT.type().displayName(),
|
|
Build.CURRENT.shortHash(),
|
|
Build.CURRENT.date(),
|
|
Constants.OS_NAME,
|
|
Constants.OS_VERSION,
|
|
Constants.OS_ARCH,
|
|
Constants.JVM_VENDOR,
|
|
Constants.JVM_NAME,
|
|
Constants.JAVA_VERSION,
|
|
Constants.JVM_VERSION);
|
|
logger.info("JVM home [{}]", System.getProperty("java.home"));
|
|
logger.info("JVM arguments {}", Arrays.toString(jvmInfo.getInputArguments()));
|
|
if (Build.CURRENT.isProductionRelease() == false) {
|
|
logger.warn(
|
|
"version [{}] is a pre-release version of Elasticsearch and is not suitable for production",
|
|
Build.CURRENT.getQualifiedVersion());
|
|
}
|
|
|
|
if (logger.isDebugEnabled()) {
|
|
logger.debug("using config [{}], data [{}], logs [{}], plugins [{}]",
|
|
environment.configFile(), Arrays.toString(environment.dataFiles()), environment.logsFile(), environment.pluginsFile());
|
|
}
|
|
|
|
this.pluginsService = new PluginsService(tmpSettings, environment.configFile(), environment.modulesFile(),
|
|
environment.pluginsFile(), classpathPlugins);
|
|
final Settings settings = pluginsService.updatedSettings();
|
|
localNodeFactory = new LocalNodeFactory(settings, nodeEnvironment.nodeId());
|
|
|
|
// create the environment based on the finalized (processed) view of the settings
|
|
// this is just to makes sure that people get the same settings, no matter where they ask them from
|
|
this.environment = new Environment(settings, environment.configFile());
|
|
Environment.assertEquivalent(environment, this.environment);
|
|
|
|
final List<ExecutorBuilder<?>> executorBuilders = pluginsService.getExecutorBuilders(settings);
|
|
|
|
final ThreadPool threadPool = new ThreadPool(settings, executorBuilders.toArray(new ExecutorBuilder[0]));
|
|
resourcesToClose.add(() -> ThreadPool.terminate(threadPool, 10, TimeUnit.SECONDS));
|
|
// adds the context to the DeprecationLogger so that it does not need to be injected everywhere
|
|
DeprecationLogger.setThreadContext(threadPool.getThreadContext());
|
|
resourcesToClose.add(() -> DeprecationLogger.removeThreadContext(threadPool.getThreadContext()));
|
|
|
|
final List<Setting<?>> additionalSettings = new ArrayList<>(pluginsService.getPluginSettings());
|
|
final List<String> additionalSettingsFilter = new ArrayList<>(pluginsService.getPluginSettingsFilter());
|
|
for (final ExecutorBuilder<?> builder : threadPool.builders()) {
|
|
additionalSettings.addAll(builder.getRegisteredSettings());
|
|
}
|
|
client = new NodeClient(settings, threadPool);
|
|
final ResourceWatcherService resourceWatcherService = new ResourceWatcherService(settings, threadPool);
|
|
final ScriptModule scriptModule = new ScriptModule(settings, pluginsService.filterPlugins(ScriptPlugin.class));
|
|
AnalysisModule analysisModule = new AnalysisModule(this.environment, pluginsService.filterPlugins(AnalysisPlugin.class));
|
|
// this is as early as we can validate settings at this point. we already pass them to ScriptModule as well as ThreadPool
|
|
// so we might be late here already
|
|
|
|
final Set<SettingUpgrader<?>> settingsUpgraders = pluginsService.filterPlugins(Plugin.class)
|
|
.stream()
|
|
.map(Plugin::getSettingUpgraders)
|
|
.flatMap(List::stream)
|
|
.collect(Collectors.toSet());
|
|
|
|
final SettingsModule settingsModule =
|
|
new SettingsModule(settings, additionalSettings, additionalSettingsFilter, settingsUpgraders);
|
|
scriptModule.registerClusterSettingsListeners(settingsModule.getClusterSettings());
|
|
resourcesToClose.add(resourceWatcherService);
|
|
final NetworkService networkService = new NetworkService(
|
|
getCustomNameResolvers(pluginsService.filterPlugins(DiscoveryPlugin.class)));
|
|
|
|
List<ClusterPlugin> clusterPlugins = pluginsService.filterPlugins(ClusterPlugin.class);
|
|
final ClusterService clusterService = new ClusterService(settings, settingsModule.getClusterSettings(), threadPool);
|
|
clusterService.addStateApplier(scriptModule.getScriptService());
|
|
resourcesToClose.add(clusterService);
|
|
final DiskThresholdMonitor listener = new DiskThresholdMonitor(settings, clusterService::state,
|
|
clusterService.getClusterSettings(), client);
|
|
final ClusterInfoService clusterInfoService = newClusterInfoService(settings, clusterService, threadPool, client,
|
|
listener::onNewInfo);
|
|
final UsageService usageService = new UsageService();
|
|
|
|
ModulesBuilder modules = new ModulesBuilder();
|
|
// plugin modules must be added here, before others or we can get crazy injection errors...
|
|
for (Module pluginModule : pluginsService.createGuiceModules()) {
|
|
modules.add(pluginModule);
|
|
}
|
|
final MonitorService monitorService = new MonitorService(settings, nodeEnvironment, threadPool, clusterInfoService);
|
|
ClusterModule clusterModule = new ClusterModule(settings, clusterService, clusterPlugins, clusterInfoService);
|
|
modules.add(clusterModule);
|
|
IndicesModule indicesModule = new IndicesModule(pluginsService.filterPlugins(MapperPlugin.class));
|
|
modules.add(indicesModule);
|
|
|
|
SearchModule searchModule = new SearchModule(settings, false, pluginsService.filterPlugins(SearchPlugin.class));
|
|
CircuitBreakerService circuitBreakerService = createCircuitBreakerService(settingsModule.getSettings(),
|
|
settingsModule.getClusterSettings());
|
|
resourcesToClose.add(circuitBreakerService);
|
|
modules.add(new GatewayModule());
|
|
|
|
|
|
PageCacheRecycler pageCacheRecycler = createPageCacheRecycler(settings);
|
|
BigArrays bigArrays = createBigArrays(pageCacheRecycler, circuitBreakerService);
|
|
resourcesToClose.add(pageCacheRecycler);
|
|
modules.add(settingsModule);
|
|
List<NamedWriteableRegistry.Entry> namedWriteables = Stream.of(
|
|
NetworkModule.getNamedWriteables().stream(),
|
|
indicesModule.getNamedWriteables().stream(),
|
|
searchModule.getNamedWriteables().stream(),
|
|
pluginsService.filterPlugins(Plugin.class).stream()
|
|
.flatMap(p -> p.getNamedWriteables().stream()),
|
|
ClusterModule.getNamedWriteables().stream())
|
|
.flatMap(Function.identity()).collect(Collectors.toList());
|
|
final NamedWriteableRegistry namedWriteableRegistry = new NamedWriteableRegistry(namedWriteables);
|
|
NamedXContentRegistry xContentRegistry = new NamedXContentRegistry(Stream.of(
|
|
NetworkModule.getNamedXContents().stream(),
|
|
indicesModule.getNamedXContents().stream(),
|
|
searchModule.getNamedXContents().stream(),
|
|
pluginsService.filterPlugins(Plugin.class).stream()
|
|
.flatMap(p -> p.getNamedXContent().stream()),
|
|
ClusterModule.getNamedXWriteables().stream())
|
|
.flatMap(Function.identity()).collect(toList()));
|
|
final MetaStateService metaStateService = new MetaStateService(nodeEnvironment, xContentRegistry);
|
|
|
|
// collect engine factory providers from server and from plugins
|
|
final Collection<EnginePlugin> enginePlugins = pluginsService.filterPlugins(EnginePlugin.class);
|
|
final Collection<Function<IndexSettings, Optional<EngineFactory>>> engineFactoryProviders =
|
|
Stream.concat(
|
|
indicesModule.getEngineFactories().stream(),
|
|
enginePlugins.stream().map(plugin -> plugin::getEngineFactory))
|
|
.collect(Collectors.toList());
|
|
|
|
|
|
final Map<String, Function<IndexSettings, IndexStore>> indexStoreFactories =
|
|
pluginsService.filterPlugins(IndexStorePlugin.class)
|
|
.stream()
|
|
.map(IndexStorePlugin::getIndexStoreFactories)
|
|
.flatMap(m -> m.entrySet().stream())
|
|
.collect(Collectors.toMap(Map.Entry::getKey, Map.Entry::getValue));
|
|
|
|
final IndicesService indicesService =
|
|
new IndicesService(settings, pluginsService, nodeEnvironment, xContentRegistry, analysisModule.getAnalysisRegistry(),
|
|
clusterModule.getIndexNameExpressionResolver(), indicesModule.getMapperRegistry(), namedWriteableRegistry,
|
|
threadPool, settingsModule.getIndexScopedSettings(), circuitBreakerService, bigArrays,
|
|
scriptModule.getScriptService(), client, metaStateService, engineFactoryProviders, indexStoreFactories);
|
|
|
|
final IngestService ingestService = new IngestService(clusterService, threadPool, this.environment,
|
|
scriptModule.getScriptService(), analysisModule.getAnalysisRegistry(),
|
|
pluginsService.filterPlugins(IngestPlugin.class), indicesService);
|
|
|
|
final AliasValidator aliasValidator = new AliasValidator();
|
|
|
|
final MetaDataCreateIndexService metaDataCreateIndexService = new MetaDataCreateIndexService(
|
|
settings,
|
|
clusterService,
|
|
indicesService,
|
|
clusterModule.getAllocationService(),
|
|
aliasValidator,
|
|
environment,
|
|
settingsModule.getIndexScopedSettings(),
|
|
threadPool,
|
|
xContentRegistry,
|
|
forbidPrivateIndexSettings);
|
|
|
|
Collection<Object> pluginComponents = pluginsService.filterPlugins(Plugin.class).stream()
|
|
.flatMap(p -> p.createComponents(client, clusterService, threadPool, resourceWatcherService,
|
|
scriptModule.getScriptService(), xContentRegistry, environment, nodeEnvironment,
|
|
namedWriteableRegistry).stream())
|
|
.collect(Collectors.toList());
|
|
|
|
ActionModule actionModule = new ActionModule(false, settings, clusterModule.getIndexNameExpressionResolver(),
|
|
settingsModule.getIndexScopedSettings(), settingsModule.getClusterSettings(), settingsModule.getSettingsFilter(),
|
|
threadPool, pluginsService.filterPlugins(ActionPlugin.class), client, circuitBreakerService, usageService);
|
|
modules.add(actionModule);
|
|
|
|
final RestController restController = actionModule.getRestController();
|
|
final NetworkModule networkModule = new NetworkModule(settings, false, pluginsService.filterPlugins(NetworkPlugin.class),
|
|
threadPool, bigArrays, pageCacheRecycler, circuitBreakerService, namedWriteableRegistry, xContentRegistry,
|
|
networkService, restController);
|
|
Collection<UnaryOperator<Map<String, MetaData.Custom>>> customMetaDataUpgraders =
|
|
pluginsService.filterPlugins(Plugin.class).stream()
|
|
.map(Plugin::getCustomMetaDataUpgrader)
|
|
.collect(Collectors.toList());
|
|
Collection<UnaryOperator<Map<String, IndexTemplateMetaData>>> indexTemplateMetaDataUpgraders =
|
|
pluginsService.filterPlugins(Plugin.class).stream()
|
|
.map(Plugin::getIndexTemplateMetaDataUpgrader)
|
|
.collect(Collectors.toList());
|
|
Collection<UnaryOperator<IndexMetaData>> indexMetaDataUpgraders = pluginsService.filterPlugins(Plugin.class).stream()
|
|
.map(Plugin::getIndexMetaDataUpgrader).collect(Collectors.toList());
|
|
final MetaDataUpgrader metaDataUpgrader = new MetaDataUpgrader(customMetaDataUpgraders, indexTemplateMetaDataUpgraders);
|
|
final MetaDataIndexUpgradeService metaDataIndexUpgradeService = new MetaDataIndexUpgradeService(settings, xContentRegistry,
|
|
indicesModule.getMapperRegistry(), settingsModule.getIndexScopedSettings(), indexMetaDataUpgraders);
|
|
new TemplateUpgradeService(client, clusterService, threadPool, indexTemplateMetaDataUpgraders);
|
|
final Transport transport = networkModule.getTransportSupplier().get();
|
|
Set<String> taskHeaders = Stream.concat(
|
|
pluginsService.filterPlugins(ActionPlugin.class).stream().flatMap(p -> p.getTaskHeaders().stream()),
|
|
Stream.of(Task.X_OPAQUE_ID)
|
|
).collect(Collectors.toSet());
|
|
final TransportService transportService = newTransportService(settings, transport, threadPool,
|
|
networkModule.getTransportInterceptor(), localNodeFactory, settingsModule.getClusterSettings(), taskHeaders);
|
|
final GatewayMetaState gatewayMetaState = new GatewayMetaState(settings, nodeEnvironment, metaStateService,
|
|
metaDataIndexUpgradeService, metaDataUpgrader, transportService, clusterService, indicesService);
|
|
final ResponseCollectorService responseCollectorService = new ResponseCollectorService(clusterService);
|
|
final SearchTransportService searchTransportService = new SearchTransportService(transportService,
|
|
SearchExecutionStatsCollector.makeWrapper(responseCollectorService));
|
|
final HttpServerTransport httpServerTransport = newHttpTransport(networkModule);
|
|
|
|
|
|
modules.add(new RepositoriesModule(this.environment, pluginsService.filterPlugins(RepositoryPlugin.class), transportService,
|
|
clusterService, threadPool, xContentRegistry));
|
|
|
|
final DiscoveryModule discoveryModule = new DiscoveryModule(settings, threadPool, transportService, namedWriteableRegistry,
|
|
networkService, clusterService.getMasterService(), clusterService.getClusterApplierService(),
|
|
clusterService.getClusterSettings(), pluginsService.filterPlugins(DiscoveryPlugin.class),
|
|
clusterModule.getAllocationService(), environment.configFile(), gatewayMetaState);
|
|
this.nodeService = new NodeService(settings, threadPool, monitorService, discoveryModule.getDiscovery(),
|
|
transportService, indicesService, pluginsService, circuitBreakerService, scriptModule.getScriptService(),
|
|
httpServerTransport, ingestService, clusterService, settingsModule.getSettingsFilter(), responseCollectorService,
|
|
searchTransportService);
|
|
|
|
final SearchService searchService = newSearchService(clusterService, indicesService,
|
|
threadPool, scriptModule.getScriptService(), bigArrays, searchModule.getFetchPhase(),
|
|
responseCollectorService);
|
|
|
|
final List<PersistentTasksExecutor<?>> tasksExecutors = pluginsService
|
|
.filterPlugins(PersistentTaskPlugin.class).stream()
|
|
.map(p -> p.getPersistentTasksExecutor(clusterService, threadPool, client, settingsModule))
|
|
.flatMap(List::stream)
|
|
.collect(toList());
|
|
|
|
final PersistentTasksExecutorRegistry registry = new PersistentTasksExecutorRegistry(tasksExecutors);
|
|
final PersistentTasksClusterService persistentTasksClusterService =
|
|
new PersistentTasksClusterService(settings, registry, clusterService, threadPool);
|
|
resourcesToClose.add(persistentTasksClusterService);
|
|
final PersistentTasksService persistentTasksService = new PersistentTasksService(clusterService, threadPool, client);
|
|
|
|
modules.add(b -> {
|
|
b.bind(Node.class).toInstance(this);
|
|
b.bind(NodeService.class).toInstance(nodeService);
|
|
b.bind(NamedXContentRegistry.class).toInstance(xContentRegistry);
|
|
b.bind(PluginsService.class).toInstance(pluginsService);
|
|
b.bind(Client.class).toInstance(client);
|
|
b.bind(NodeClient.class).toInstance(client);
|
|
b.bind(Environment.class).toInstance(this.environment);
|
|
b.bind(ThreadPool.class).toInstance(threadPool);
|
|
b.bind(NodeEnvironment.class).toInstance(nodeEnvironment);
|
|
b.bind(ResourceWatcherService.class).toInstance(resourceWatcherService);
|
|
b.bind(CircuitBreakerService.class).toInstance(circuitBreakerService);
|
|
b.bind(BigArrays.class).toInstance(bigArrays);
|
|
b.bind(PageCacheRecycler.class).toInstance(pageCacheRecycler);
|
|
b.bind(ScriptService.class).toInstance(scriptModule.getScriptService());
|
|
b.bind(AnalysisRegistry.class).toInstance(analysisModule.getAnalysisRegistry());
|
|
b.bind(IngestService.class).toInstance(ingestService);
|
|
b.bind(UsageService.class).toInstance(usageService);
|
|
b.bind(NamedWriteableRegistry.class).toInstance(namedWriteableRegistry);
|
|
b.bind(MetaDataUpgrader.class).toInstance(metaDataUpgrader);
|
|
b.bind(MetaStateService.class).toInstance(metaStateService);
|
|
b.bind(IndicesService.class).toInstance(indicesService);
|
|
b.bind(AliasValidator.class).toInstance(aliasValidator);
|
|
b.bind(MetaDataCreateIndexService.class).toInstance(metaDataCreateIndexService);
|
|
b.bind(SearchService.class).toInstance(searchService);
|
|
b.bind(SearchTransportService.class).toInstance(searchTransportService);
|
|
b.bind(SearchPhaseController.class).toInstance(new SearchPhaseController(searchService::createReduceContext));
|
|
b.bind(Transport.class).toInstance(transport);
|
|
b.bind(TransportService.class).toInstance(transportService);
|
|
b.bind(NetworkService.class).toInstance(networkService);
|
|
b.bind(UpdateHelper.class).toInstance(new UpdateHelper(scriptModule.getScriptService()));
|
|
b.bind(MetaDataIndexUpgradeService.class).toInstance(metaDataIndexUpgradeService);
|
|
b.bind(ClusterInfoService.class).toInstance(clusterInfoService);
|
|
b.bind(GatewayMetaState.class).toInstance(gatewayMetaState);
|
|
b.bind(Discovery.class).toInstance(discoveryModule.getDiscovery());
|
|
{
|
|
RecoverySettings recoverySettings = new RecoverySettings(settings, settingsModule.getClusterSettings());
|
|
processRecoverySettings(settingsModule.getClusterSettings(), recoverySettings);
|
|
b.bind(PeerRecoverySourceService.class).toInstance(new PeerRecoverySourceService(transportService,
|
|
indicesService, recoverySettings));
|
|
b.bind(PeerRecoveryTargetService.class).toInstance(new PeerRecoveryTargetService(threadPool,
|
|
transportService, recoverySettings, clusterService));
|
|
}
|
|
b.bind(HttpServerTransport.class).toInstance(httpServerTransport);
|
|
pluginComponents.stream().forEach(p -> b.bind((Class) p.getClass()).toInstance(p));
|
|
b.bind(PersistentTasksService.class).toInstance(persistentTasksService);
|
|
b.bind(PersistentTasksClusterService.class).toInstance(persistentTasksClusterService);
|
|
b.bind(PersistentTasksExecutorRegistry.class).toInstance(registry);
|
|
}
|
|
);
|
|
injector = modules.createInjector();
|
|
|
|
// TODO hack around circular dependencies problems in AllocationService
|
|
clusterModule.getAllocationService().setGatewayAllocator(injector.getInstance(GatewayAllocator.class));
|
|
|
|
List<LifecycleComponent> pluginLifecycleComponents = pluginComponents.stream()
|
|
.filter(p -> p instanceof LifecycleComponent)
|
|
.map(p -> (LifecycleComponent) p).collect(Collectors.toList());
|
|
pluginLifecycleComponents.addAll(pluginsService.getGuiceServiceClasses().stream()
|
|
.map(injector::getInstance).collect(Collectors.toList()));
|
|
resourcesToClose.addAll(pluginLifecycleComponents);
|
|
this.pluginLifecycleComponents = Collections.unmodifiableList(pluginLifecycleComponents);
|
|
client.initialize(injector.getInstance(new Key<Map<Action, TransportAction>>() {}),
|
|
() -> clusterService.localNode().getId(), transportService.getRemoteClusterService());
|
|
|
|
logger.debug("initializing HTTP handlers ...");
|
|
actionModule.initRestHandlers(() -> clusterService.state().nodes());
|
|
logger.info("initialized");
|
|
|
|
success = true;
|
|
} catch (IOException ex) {
|
|
throw new ElasticsearchException("failed to bind service", ex);
|
|
} finally {
|
|
if (!success) {
|
|
IOUtils.closeWhileHandlingException(resourcesToClose);
|
|
}
|
|
}
|
|
}
|
|
|
|
protected TransportService newTransportService(Settings settings, Transport transport, ThreadPool threadPool,
|
|
TransportInterceptor interceptor,
|
|
Function<BoundTransportAddress, DiscoveryNode> localNodeFactory,
|
|
ClusterSettings clusterSettings, Set<String> taskHeaders) {
|
|
return new TransportService(settings, transport, threadPool, interceptor, localNodeFactory, clusterSettings, taskHeaders);
|
|
}
|
|
|
|
protected void processRecoverySettings(ClusterSettings clusterSettings, RecoverySettings recoverySettings) {
|
|
// Noop in production, overridden by tests
|
|
}
|
|
|
|
/**
|
|
* The settings that are used by this node. Contains original settings as well as additional settings provided by plugins.
|
|
*/
|
|
public Settings settings() {
|
|
return this.environment.settings();
|
|
}
|
|
|
|
/**
|
|
* A client that can be used to execute actions (operations) against the cluster.
|
|
*/
|
|
public Client client() {
|
|
return client;
|
|
}
|
|
|
|
/**
|
|
* Returns the environment of the node
|
|
*/
|
|
public Environment getEnvironment() {
|
|
return environment;
|
|
}
|
|
|
|
/**
|
|
* Returns the {@link NodeEnvironment} instance of this node
|
|
*/
|
|
public NodeEnvironment getNodeEnvironment() {
|
|
return nodeEnvironment;
|
|
}
|
|
|
|
|
|
/**
|
|
* Start the node. If the node is already started, this method is no-op.
|
|
*/
|
|
public Node start() throws NodeValidationException {
|
|
if (!lifecycle.moveToStarted()) {
|
|
return this;
|
|
}
|
|
|
|
logger.info("starting ...");
|
|
pluginLifecycleComponents.forEach(LifecycleComponent::start);
|
|
|
|
injector.getInstance(MappingUpdatedAction.class).setClient(client);
|
|
injector.getInstance(IndicesService.class).start();
|
|
injector.getInstance(IndicesClusterStateService.class).start();
|
|
injector.getInstance(SnapshotsService.class).start();
|
|
injector.getInstance(SnapshotShardsService.class).start();
|
|
injector.getInstance(RoutingService.class).start();
|
|
injector.getInstance(SearchService.class).start();
|
|
nodeService.getMonitorService().start();
|
|
|
|
final ClusterService clusterService = injector.getInstance(ClusterService.class);
|
|
|
|
final NodeConnectionsService nodeConnectionsService = injector.getInstance(NodeConnectionsService.class);
|
|
nodeConnectionsService.start();
|
|
clusterService.setNodeConnectionsService(nodeConnectionsService);
|
|
|
|
injector.getInstance(ResourceWatcherService.class).start();
|
|
injector.getInstance(GatewayService.class).start();
|
|
Discovery discovery = injector.getInstance(Discovery.class);
|
|
clusterService.getMasterService().setClusterStatePublisher(discovery::publish);
|
|
|
|
// Start the transport service now so the publish address will be added to the local disco node in ClusterService
|
|
TransportService transportService = injector.getInstance(TransportService.class);
|
|
transportService.getTaskManager().setTaskResultsService(injector.getInstance(TaskResultsService.class));
|
|
transportService.start();
|
|
assert localNodeFactory.getNode() != null;
|
|
assert transportService.getLocalNode().equals(localNodeFactory.getNode())
|
|
: "transportService has a different local node than the factory provided";
|
|
final MetaData onDiskMetadata;
|
|
// we load the global state here (the persistent part of the cluster state stored on disk) to
|
|
// pass it to the bootstrap checks to allow plugins to enforce certain preconditions based on the recovered state.
|
|
if (DiscoveryNode.isMasterNode(settings()) || DiscoveryNode.isDataNode(settings())) {
|
|
onDiskMetadata = injector.getInstance(GatewayMetaState.class).getMetaData();
|
|
} else {
|
|
onDiskMetadata = MetaData.EMPTY_META_DATA;
|
|
}
|
|
assert onDiskMetadata != null : "metadata is null but shouldn't"; // this is never null
|
|
validateNodeBeforeAcceptingRequests(new BootstrapContext(environment, onDiskMetadata), transportService.boundAddress(),
|
|
pluginsService.filterPlugins(Plugin.class).stream()
|
|
.flatMap(p -> p.getBootstrapChecks().stream()).collect(Collectors.toList()));
|
|
|
|
clusterService.addStateApplier(transportService.getTaskManager());
|
|
// start after transport service so the local disco is known
|
|
discovery.start(); // start before cluster service so that it can set initial state on ClusterApplierService
|
|
clusterService.start();
|
|
assert clusterService.localNode().equals(localNodeFactory.getNode())
|
|
: "clusterService has a different local node than the factory provided";
|
|
transportService.acceptIncomingRequests();
|
|
discovery.startInitialJoin();
|
|
final TimeValue initialStateTimeout = DiscoverySettings.INITIAL_STATE_TIMEOUT_SETTING.get(settings());
|
|
configureNodeAndClusterIdStateListener(clusterService);
|
|
|
|
if (initialStateTimeout.millis() > 0) {
|
|
final ThreadPool thread = injector.getInstance(ThreadPool.class);
|
|
ClusterState clusterState = clusterService.state();
|
|
ClusterStateObserver observer =
|
|
new ClusterStateObserver(clusterState, clusterService, null, logger, thread.getThreadContext());
|
|
|
|
if (clusterState.nodes().getMasterNodeId() == null) {
|
|
logger.debug("waiting to join the cluster. timeout [{}]", initialStateTimeout);
|
|
final CountDownLatch latch = new CountDownLatch(1);
|
|
observer.waitForNextChange(new ClusterStateObserver.Listener() {
|
|
@Override
|
|
public void onNewClusterState(ClusterState state) { latch.countDown(); }
|
|
|
|
@Override
|
|
public void onClusterServiceClose() {
|
|
latch.countDown();
|
|
}
|
|
|
|
@Override
|
|
public void onTimeout(TimeValue timeout) {
|
|
logger.warn("timed out while waiting for initial discovery state - timeout: {}",
|
|
initialStateTimeout);
|
|
latch.countDown();
|
|
}
|
|
}, state -> state.nodes().getMasterNodeId() != null, initialStateTimeout);
|
|
|
|
try {
|
|
latch.await();
|
|
} catch (InterruptedException e) {
|
|
throw new ElasticsearchTimeoutException("Interrupted while waiting for initial discovery state");
|
|
}
|
|
}
|
|
}
|
|
|
|
injector.getInstance(HttpServerTransport.class).start();
|
|
|
|
if (WRITE_PORTS_FILE_SETTING.get(settings())) {
|
|
TransportService transport = injector.getInstance(TransportService.class);
|
|
writePortsFile("transport", transport.boundAddress());
|
|
HttpServerTransport http = injector.getInstance(HttpServerTransport.class);
|
|
writePortsFile("http", http.boundAddress());
|
|
}
|
|
|
|
logger.info("started");
|
|
|
|
pluginsService.filterPlugins(ClusterPlugin.class).forEach(ClusterPlugin::onNodeStarted);
|
|
|
|
return this;
|
|
}
|
|
|
|
protected void configureNodeAndClusterIdStateListener(ClusterService clusterService) {
|
|
NodeAndClusterIdStateListener.getAndSetNodeIdAndClusterId(clusterService,
|
|
injector.getInstance(ThreadPool.class).getThreadContext());
|
|
}
|
|
|
|
private Node stop() {
|
|
if (!lifecycle.moveToStopped()) {
|
|
return this;
|
|
}
|
|
logger.info("stopping ...");
|
|
|
|
injector.getInstance(ResourceWatcherService.class).stop();
|
|
injector.getInstance(HttpServerTransport.class).stop();
|
|
|
|
injector.getInstance(SnapshotsService.class).stop();
|
|
injector.getInstance(SnapshotShardsService.class).stop();
|
|
// stop any changes happening as a result of cluster state changes
|
|
injector.getInstance(IndicesClusterStateService.class).stop();
|
|
// close discovery early to not react to pings anymore.
|
|
// This can confuse other nodes and delay things - mostly if we're the master and we're running tests.
|
|
injector.getInstance(Discovery.class).stop();
|
|
// we close indices first, so operations won't be allowed on it
|
|
injector.getInstance(RoutingService.class).stop();
|
|
injector.getInstance(ClusterService.class).stop();
|
|
injector.getInstance(NodeConnectionsService.class).stop();
|
|
nodeService.getMonitorService().stop();
|
|
injector.getInstance(GatewayService.class).stop();
|
|
injector.getInstance(SearchService.class).stop();
|
|
injector.getInstance(TransportService.class).stop();
|
|
|
|
pluginLifecycleComponents.forEach(LifecycleComponent::stop);
|
|
// we should stop this last since it waits for resources to get released
|
|
// if we had scroll searchers etc or recovery going on we wait for to finish.
|
|
injector.getInstance(IndicesService.class).stop();
|
|
logger.info("stopped");
|
|
|
|
return this;
|
|
}
|
|
|
|
// During concurrent close() calls we want to make sure that all of them return after the node has completed it's shutdown cycle.
|
|
// If not, the hook that is added in Bootstrap#setup() will be useless:
|
|
// close() might not be executed, in case another (for example api) call to close() has already set some lifecycles to stopped.
|
|
// In this case the process will be terminated even if the first call to close() has not finished yet.
|
|
@Override
|
|
public synchronized void close() throws IOException {
|
|
synchronized (lifecycle) {
|
|
if (lifecycle.started()) {
|
|
stop();
|
|
}
|
|
if (!lifecycle.moveToClosed()) {
|
|
return;
|
|
}
|
|
}
|
|
|
|
logger.info("closing ...");
|
|
List<Closeable> toClose = new ArrayList<>();
|
|
StopWatch stopWatch = new StopWatch("node_close");
|
|
toClose.add(() -> stopWatch.start("node_service"));
|
|
toClose.add(nodeService);
|
|
toClose.add(() -> stopWatch.stop().start("http"));
|
|
toClose.add(injector.getInstance(HttpServerTransport.class));
|
|
toClose.add(() -> stopWatch.stop().start("snapshot_service"));
|
|
toClose.add(injector.getInstance(SnapshotsService.class));
|
|
toClose.add(injector.getInstance(SnapshotShardsService.class));
|
|
toClose.add(() -> stopWatch.stop().start("client"));
|
|
Releasables.close(injector.getInstance(Client.class));
|
|
toClose.add(() -> stopWatch.stop().start("indices_cluster"));
|
|
toClose.add(injector.getInstance(IndicesClusterStateService.class));
|
|
toClose.add(() -> stopWatch.stop().start("indices"));
|
|
toClose.add(injector.getInstance(IndicesService.class));
|
|
// close filter/fielddata caches after indices
|
|
toClose.add(injector.getInstance(IndicesStore.class));
|
|
toClose.add(() -> stopWatch.stop().start("routing"));
|
|
toClose.add(injector.getInstance(RoutingService.class));
|
|
toClose.add(() -> stopWatch.stop().start("cluster"));
|
|
toClose.add(injector.getInstance(ClusterService.class));
|
|
toClose.add(() -> stopWatch.stop().start("node_connections_service"));
|
|
toClose.add(injector.getInstance(NodeConnectionsService.class));
|
|
toClose.add(() -> stopWatch.stop().start("discovery"));
|
|
toClose.add(injector.getInstance(Discovery.class));
|
|
toClose.add(() -> stopWatch.stop().start("monitor"));
|
|
toClose.add(nodeService.getMonitorService());
|
|
toClose.add(() -> stopWatch.stop().start("gateway"));
|
|
toClose.add(injector.getInstance(GatewayService.class));
|
|
toClose.add(() -> stopWatch.stop().start("search"));
|
|
toClose.add(injector.getInstance(SearchService.class));
|
|
toClose.add(() -> stopWatch.stop().start("transport"));
|
|
toClose.add(injector.getInstance(TransportService.class));
|
|
|
|
for (LifecycleComponent plugin : pluginLifecycleComponents) {
|
|
toClose.add(() -> stopWatch.stop().start("plugin(" + plugin.getClass().getName() + ")"));
|
|
toClose.add(plugin);
|
|
}
|
|
toClose.addAll(pluginsService.filterPlugins(Plugin.class));
|
|
|
|
toClose.add(() -> stopWatch.stop().start("script"));
|
|
toClose.add(injector.getInstance(ScriptService.class));
|
|
|
|
toClose.add(() -> stopWatch.stop().start("thread_pool"));
|
|
toClose.add(() -> injector.getInstance(ThreadPool.class).shutdown());
|
|
// Don't call shutdownNow here, it might break ongoing operations on Lucene indices.
|
|
// See https://issues.apache.org/jira/browse/LUCENE-7248. We call shutdownNow in
|
|
// awaitClose if the node doesn't finish closing within the specified time.
|
|
toClose.add(() -> stopWatch.stop());
|
|
|
|
toClose.add(injector.getInstance(NodeEnvironment.class));
|
|
toClose.add(injector.getInstance(PageCacheRecycler.class));
|
|
|
|
if (logger.isTraceEnabled()) {
|
|
logger.trace("Close times for each service:\n{}", stopWatch.prettyPrint());
|
|
}
|
|
IOUtils.close(toClose);
|
|
logger.info("closed");
|
|
}
|
|
|
|
/**
|
|
* Wait for this node to be effectively closed.
|
|
*/
|
|
// synchronized to prevent running concurrently with close()
|
|
public synchronized boolean awaitClose(long timeout, TimeUnit timeUnit) throws InterruptedException {
|
|
if (lifecycle.closed() == false) {
|
|
// We don't want to shutdown the threadpool or interrupt threads on a node that is not
|
|
// closed yet.
|
|
throw new IllegalStateException("Call close() first");
|
|
}
|
|
|
|
|
|
ThreadPool threadPool = injector.getInstance(ThreadPool.class);
|
|
final boolean terminated = ThreadPool.terminate(threadPool, timeout, timeUnit);
|
|
if (terminated) {
|
|
// All threads terminated successfully. Because search, recovery and all other operations
|
|
// that run on shards run in the threadpool, indices should be effectively closed by now.
|
|
if (nodeService.awaitClose(0, TimeUnit.MILLISECONDS) == false) {
|
|
throw new IllegalStateException("Some shards are still open after the threadpool terminated. " +
|
|
"Something is leaking index readers or store references.");
|
|
}
|
|
}
|
|
return terminated;
|
|
}
|
|
|
|
/**
|
|
* Returns {@code true} if the node is closed.
|
|
*/
|
|
public boolean isClosed() {
|
|
return lifecycle.closed();
|
|
}
|
|
|
|
public Injector injector() {
|
|
return this.injector;
|
|
}
|
|
|
|
/**
|
|
* Hook for validating the node after network
|
|
* services are started but before the cluster service is started
|
|
* and before the network service starts accepting incoming network
|
|
* requests.
|
|
*
|
|
* @param context the bootstrap context for this node
|
|
* @param boundTransportAddress the network addresses the node is
|
|
* bound and publishing to
|
|
*/
|
|
@SuppressWarnings("unused")
|
|
protected void validateNodeBeforeAcceptingRequests(
|
|
final BootstrapContext context,
|
|
final BoundTransportAddress boundTransportAddress, List<BootstrapCheck> bootstrapChecks) throws NodeValidationException {
|
|
}
|
|
|
|
/** Writes a file to the logs dir containing the ports for the given transport type */
|
|
private void writePortsFile(String type, BoundTransportAddress boundAddress) {
|
|
Path tmpPortsFile = environment.logsFile().resolve(type + ".ports.tmp");
|
|
try (BufferedWriter writer = Files.newBufferedWriter(tmpPortsFile, Charset.forName("UTF-8"))) {
|
|
for (TransportAddress address : boundAddress.boundAddresses()) {
|
|
InetAddress inetAddress = InetAddress.getByName(address.getAddress());
|
|
writer.write(NetworkAddress.format(new InetSocketAddress(inetAddress, address.getPort())) + "\n");
|
|
}
|
|
} catch (IOException e) {
|
|
throw new RuntimeException("Failed to write ports file", e);
|
|
}
|
|
Path portsFile = environment.logsFile().resolve(type + ".ports");
|
|
try {
|
|
Files.move(tmpPortsFile, portsFile, StandardCopyOption.ATOMIC_MOVE);
|
|
} catch (IOException e) {
|
|
throw new RuntimeException("Failed to rename ports file", e);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* The {@link PluginsService} used to build this node's components.
|
|
*/
|
|
protected PluginsService getPluginsService() {
|
|
return pluginsService;
|
|
}
|
|
|
|
/**
|
|
* Creates a new {@link CircuitBreakerService} based on the settings provided.
|
|
* @see #BREAKER_TYPE_KEY
|
|
*/
|
|
public static CircuitBreakerService createCircuitBreakerService(Settings settings, ClusterSettings clusterSettings) {
|
|
String type = BREAKER_TYPE_KEY.get(settings);
|
|
if (type.equals("hierarchy")) {
|
|
return new HierarchyCircuitBreakerService(settings, clusterSettings);
|
|
} else if (type.equals("none")) {
|
|
return new NoneCircuitBreakerService();
|
|
} else {
|
|
throw new IllegalArgumentException("Unknown circuit breaker type [" + type + "]");
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Creates a new {@link BigArrays} instance used for this node.
|
|
* This method can be overwritten by subclasses to change their {@link BigArrays} implementation for instance for testing
|
|
*/
|
|
BigArrays createBigArrays(PageCacheRecycler pageCacheRecycler, CircuitBreakerService circuitBreakerService) {
|
|
return new BigArrays(pageCacheRecycler, circuitBreakerService, CircuitBreaker.REQUEST);
|
|
}
|
|
|
|
/**
|
|
* Creates a new {@link BigArrays} instance used for this node.
|
|
* This method can be overwritten by subclasses to change their {@link BigArrays} implementation for instance for testing
|
|
*/
|
|
PageCacheRecycler createPageCacheRecycler(Settings settings) {
|
|
return new PageCacheRecycler(settings);
|
|
}
|
|
|
|
/**
|
|
* Creates a new the SearchService. This method can be overwritten by tests to inject mock implementations.
|
|
*/
|
|
protected SearchService newSearchService(ClusterService clusterService, IndicesService indicesService,
|
|
ThreadPool threadPool, ScriptService scriptService, BigArrays bigArrays,
|
|
FetchPhase fetchPhase, ResponseCollectorService responseCollectorService) {
|
|
return new SearchService(clusterService, indicesService, threadPool,
|
|
scriptService, bigArrays, fetchPhase, responseCollectorService);
|
|
}
|
|
|
|
/**
|
|
* Get Custom Name Resolvers list based on a Discovery Plugins list
|
|
* @param discoveryPlugins Discovery plugins list
|
|
*/
|
|
private List<NetworkService.CustomNameResolver> getCustomNameResolvers(List<DiscoveryPlugin> discoveryPlugins) {
|
|
List<NetworkService.CustomNameResolver> customNameResolvers = new ArrayList<>();
|
|
for (DiscoveryPlugin discoveryPlugin : discoveryPlugins) {
|
|
NetworkService.CustomNameResolver customNameResolver = discoveryPlugin.getCustomNameResolver(settings());
|
|
if (customNameResolver != null) {
|
|
customNameResolvers.add(customNameResolver);
|
|
}
|
|
}
|
|
return customNameResolvers;
|
|
}
|
|
|
|
/** Constructs a ClusterInfoService which may be mocked for tests. */
|
|
protected ClusterInfoService newClusterInfoService(Settings settings, ClusterService clusterService,
|
|
ThreadPool threadPool, NodeClient client, Consumer<ClusterInfo> listeners) {
|
|
return new InternalClusterInfoService(settings, clusterService, threadPool, client, listeners);
|
|
}
|
|
|
|
/** Constructs a {@link org.elasticsearch.http.HttpServerTransport} which may be mocked for tests. */
|
|
protected HttpServerTransport newHttpTransport(NetworkModule networkModule) {
|
|
return networkModule.getHttpServerTransportSupplier().get();
|
|
}
|
|
|
|
private static class LocalNodeFactory implements Function<BoundTransportAddress, DiscoveryNode> {
|
|
private final SetOnce<DiscoveryNode> localNode = new SetOnce<>();
|
|
private final String persistentNodeId;
|
|
private final Settings settings;
|
|
|
|
private LocalNodeFactory(Settings settings, String persistentNodeId) {
|
|
this.persistentNodeId = persistentNodeId;
|
|
this.settings = settings;
|
|
}
|
|
|
|
@Override
|
|
public DiscoveryNode apply(BoundTransportAddress boundTransportAddress) {
|
|
localNode.set(DiscoveryNode.createLocal(settings, boundTransportAddress.publishAddress(), persistentNodeId));
|
|
return localNode.get();
|
|
}
|
|
|
|
DiscoveryNode getNode() {
|
|
assert localNode.get() != null;
|
|
return localNode.get();
|
|
}
|
|
}
|
|
}
|