diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services-api/src/main/resources/definition/YARN-Services-Examples.md b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services-api/src/main/resources/definition/YARN-Services-Examples.md index 3cd3d48620f..00b21dd60ed 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services-api/src/main/resources/definition/YARN-Services-Examples.md +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services-api/src/main/resources/definition/YARN-Services-Examples.md @@ -176,14 +176,14 @@ POST URL - http://localhost:9191:/ws/v1/services/hbase-app-1 { "type": "XML", "dest_file": "/etc/hadoop/conf/core-site.xml", - "props": { + "properties": { "fs.defaultFS": "${CLUSTER_FS_URI}" } }, { "type": "XML", "dest_file": "/etc/hbase/conf/hbase-site.xml", - "props": { + "properties": { "hbase.cluster.distributed": "true", "hbase.zookeeper.quorum": "${CLUSTER_ZK_QUORUM}", "hbase.rootdir": "${SERVICE_HDFS_DIR}/hbase", @@ -216,14 +216,14 @@ POST URL - http://localhost:9191:/ws/v1/services/hbase-app-1 { "type": "XML", "dest_file": "/etc/hadoop/conf/core-site.xml", - "props": { + "properties": { "fs.defaultFS": "${CLUSTER_FS_URI}" } }, { "type": "XML", "dest_file": "/etc/hbase/conf/hbase-site.xml", - "props": { + "properties": { "hbase.cluster.distributed": "true", "hbase.zookeeper.quorum": "${CLUSTER_ZK_QUORUM}", "hbase.rootdir": "${SERVICE_HDFS_DIR}/hbase", diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services-api/src/main/resources/definition/YARN-Simplified-V1-API-Layer-For-Services.yaml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services-api/src/main/resources/definition/YARN-Simplified-V1-API-Layer-For-Services.yaml index cc7625915ac..088b50cca8b 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services-api/src/main/resources/definition/YARN-Simplified-V1-API-Layer-For-Services.yaml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services-api/src/main/resources/definition/YARN-Simplified-V1-API-Layer-For-Services.yaml @@ -340,7 +340,7 @@ definitions: enum: - HTTP - PORT - props: + properties: type: object description: A blob of key value pairs that will be used to configure the check. additionalProperties: @@ -353,7 +353,7 @@ definitions: properties: properties: type: object - description: A blob of key-value pairs of common service properties. + description: A blob of key-value pairs for configuring the YARN service AM additionalProperties: type: string env: @@ -386,7 +386,7 @@ definitions: src_file: type: string description: This provides the source location of the configuration file, the content of which is dumped to dest_file post property substitutions, in the format as specified in type. Typically the src_file would point to a source controlled network accessible file maintained by tools like puppet, chef, or hdfs etc. Currently, only hdfs is supported. - props: + properties: type: object description: A blob of key value pairs that will be dumped in the dest_file in the format as specified in type. If src_file is specified, src_file content are dumped in the dest_file and these properties will overwrite, if any, existing properties in src_file or be added as new properties in src_file. Container: diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/examples/httpd-no-dns/httpd-no-dns.json b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/examples/httpd-no-dns/httpd-no-dns.json index 6b35538d4e5..7bf2f68f2bc 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/examples/httpd-no-dns/httpd-no-dns.json +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/examples/httpd-no-dns/httpd-no-dns.json @@ -16,16 +16,16 @@ }, "readiness_check": { "type": "HTTP", - "props": { + "properties": { "url": "http://${THIS_HOST}:8080" } }, "configuration": { "files": [ { - "type": "ENV", + "type": "TEMPLATE", "dest_file": "/var/www/html/index.html", - "props": { + "properties": { "content": "
Title
Hello from ${COMPONENT_INSTANCE_NAME}!" } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/examples/httpd/httpd.json b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/examples/httpd/httpd.json index e63376dcf86..87862144090 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/examples/httpd/httpd.json +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/examples/httpd/httpd.json @@ -17,9 +17,9 @@ "configuration": { "files": [ { - "type": "ENV", + "type": "TEMPLATE", "dest_file": "/var/www/html/index.html", - "props": { + "properties": { "content": "
Title
Hello from ${COMPONENT_INSTANCE_NAME}!" } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceMaster.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceMaster.java index d58e32b3293..2d86a5a6cd6 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceMaster.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceMaster.java @@ -78,6 +78,12 @@ protected void serviceInit(Configuration conf) throws Exception { fs.setAppDir(appDir); loadApplicationJson(context, fs); + // Take yarn config from YarnFile and merge them into YarnConfiguration + for (Map.Entry entry : context.service + .getConfiguration().getProperties().entrySet()) { + conf.set(entry.getKey(), entry.getValue()); + } + ContainerId amContainerId = getAMContainerId(); ApplicationAttemptId attemptId = amContainerId.getApplicationAttemptId(); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceMetrics.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceMetrics.java index 9fc886e65af..b5bbb7d0347 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceMetrics.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceMetrics.java @@ -35,9 +35,6 @@ public class ServiceMetrics implements MetricsSource { @Metric("containers requested") public MutableGaugeInt containersRequested; - @Metric("anti-affinity containers pending") - public MutableGaugeInt pendingAAContainers; - @Metric("containers running") public MutableGaugeInt containersRunning; @@ -86,7 +83,6 @@ public void tag(String name, String description, String value) { @Override public String toString() { return "ServiceMetrics{" + "containersRequested=" + containersRequested.value() - + ", pendingAAContainers=" + pendingAAContainers.value() + ", containersRunning=" + containersRunning.value() + ", containersDesired=" + containersDesired.value() + ", containersSucceeded=" + containersSucceeded.value() diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceScheduler.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceScheduler.java index 631f89e7161..a7b7e22d42c 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceScheduler.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/ServiceScheduler.java @@ -397,10 +397,6 @@ private void registerServiceInstance(ApplicationAttemptId attemptId, PersistencePolicies.APPLICATION); serviceRecord.description = "YarnServiceMaster"; - // set any provided attributes - setUserProvidedServiceRecordAttributes(service.getConfiguration(), - serviceRecord); - executorService.submit(new Runnable() { @Override public void run() { try { @@ -426,17 +422,6 @@ private void registerServiceInstance(ApplicationAttemptId attemptId, } } - private void setUserProvidedServiceRecordAttributes( - org.apache.hadoop.yarn.service.api.records.Configuration conf, ServiceRecord record) { - String prefix = "service.record.attribute"; - for (Map.Entry entry : conf.getProperties().entrySet()) { - if (entry.getKey().startsWith(prefix)) { - String key = entry.getKey().substring(prefix.length() + 1); - record.set(key, entry.getValue().trim()); - } - } - } - private void createAllComponents() { long allocateId = 0; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/ServiceApiConstants.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/ServiceApiConstants.java index a85191c2cd5..a4f2243d227 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/ServiceApiConstants.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/ServiceApiConstants.java @@ -53,9 +53,9 @@ public interface ServiceApiConstants { String CONTAINER_ID = $("CONTAINER_ID"); // Templates for component instance host/IP - String COMPONENT_HOST = $("%s_HOST"); + String COMPONENT_INSTANCE_HOST = $("%s_HOST"); - String COMPONENT_IP = $("%s_IP"); + String COMPONENT_INSTANCE_IP = $("%s_IP"); // Constants for default cluster ZK String CLUSTER_ZK_QUORUM = $("CLUSTER_ZK_QUORUM"); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/ConfigFile.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/ConfigFile.java index fb088e2c559..984e6f7cfb4 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/ConfigFile.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/ConfigFile.java @@ -49,13 +49,13 @@ public class ConfigFile implements Serializable { private static final long serialVersionUID = -7009402089417704612L; /** - * Config Type. XML, JSON, YAML, TEMPLATE, ENV and HADOOP_XML are supported. + * Config Type. XML, JSON, YAML, TEMPLATE and HADOOP_XML are supported. **/ @XmlType(name = "config_type") @XmlEnum public enum TypeEnum { XML("XML"), PROPERTIES("PROPERTIES"), JSON("JSON"), YAML("YAML"), TEMPLATE( - "TEMPLATE"), ENV("ENV"), HADOOP_XML("HADOOP_XML"),; + "TEMPLATE"), HADOOP_XML("HADOOP_XML"),; private String value; @@ -73,15 +73,15 @@ public String toString() { private TypeEnum type = null; private String destFile = null; private String srcFile = null; - private Map props = new HashMap<>(); + private Map properties = new HashMap<>(); public ConfigFile copy() { ConfigFile copy = new ConfigFile(); copy.setType(this.getType()); copy.setSrcFile(this.getSrcFile()); copy.setDestFile(this.getDestFile()); - if (this.getProps() != null && !this.getProps().isEmpty()) { - copy.getProps().putAll(this.getProps()); + if (this.getProperties() != null && !this.getProperties().isEmpty()) { + copy.getProperties().putAll(this.getProperties()); } return copy; } @@ -156,28 +156,28 @@ public void setSrcFile(String srcFile) { in the dest_file and these properties will overwrite, if any, existing properties in src_file or be added as new properties in src_file. **/ - public ConfigFile props(Map props) { - this.props = props; + public ConfigFile properties(Map properties) { + this.properties = properties; return this; } @ApiModelProperty(example = "null", value = "A blob of key value pairs that will be dumped in the dest_file in the format as specified in type." + " If src_file is specified, src_file content are dumped in the dest_file and these properties will overwrite, if any," + " existing properties in src_file or be added as new properties in src_file.") - @JsonProperty("props") - public Map getProps() { - return props; + @JsonProperty("properties") + public Map getProperties() { + return properties; } - public void setProps(Map props) { - this.props = props; + public void setProperties(Map properties) { + this.properties = properties; } public long getLong(String name, long defaultValue) { if (name == null) { return defaultValue; } - String value = props.get(name.trim()); + String value = properties.get(name.trim()); return Long.parseLong(value); } @@ -185,7 +185,7 @@ public boolean getBoolean(String name, boolean defaultValue) { if (name == null) { return defaultValue; } - return Boolean.valueOf(props.get(name.trim())); + return Boolean.valueOf(properties.get(name.trim())); } @Override @@ -204,7 +204,7 @@ public boolean equals(java.lang.Object o) { @Override public int hashCode() { - return Objects.hash(type, destFile, srcFile, props); + return Objects.hash(type, destFile, srcFile, properties); } @Override @@ -215,7 +215,7 @@ public String toString() { sb.append(" type: ").append(toIndentedString(type)).append("\n"); sb.append(" destFile: ").append(toIndentedString(destFile)).append("\n"); sb.append(" srcFile: ").append(toIndentedString(srcFile)).append("\n"); - sb.append(" props: ").append(toIndentedString(props)).append("\n"); + sb.append(" properties: ").append(toIndentedString(properties)).append("\n"); sb.append("}"); return sb.toString(); } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/Configuration.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/Configuration.java index 2f8ca9641fb..dc6dff8bbf8 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/Configuration.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/Configuration.java @@ -212,8 +212,8 @@ public synchronized void mergeFrom(Configuration that) { for (ConfigFile thisFile : files) { if(thatMap.containsKey(thisFile.getDestFile())) { ConfigFile thatFile = thatMap.get(thisFile.getDestFile()); - SliderUtils.mergeMapsIgnoreDuplicateKeys(thisFile.getProps(), - thatFile.getProps()); + SliderUtils.mergeMapsIgnoreDuplicateKeys(thisFile.getProperties(), + thatFile.getProperties()); thatMap.remove(thisFile.getDestFile()); } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/ReadinessCheck.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/ReadinessCheck.java index b25828fd4ff..0a3713c6d0a 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/ReadinessCheck.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/api/records/ReadinessCheck.java @@ -69,7 +69,7 @@ public String toString() { } private TypeEnum type = null; - private Map props = new HashMap(); + private Map properties = new HashMap(); private Artifact artifact = null; /** @@ -91,27 +91,27 @@ public void setType(TypeEnum type) { this.type = type; } - public ReadinessCheck props(Map props) { - this.props = props; + public ReadinessCheck properties(Map properties) { + this.properties = properties; return this; } public ReadinessCheck putPropsItem(String key, String propsItem) { - this.props.put(key, propsItem); + this.properties.put(key, propsItem); return this; } /** * A blob of key value pairs that will be used to configure the check. - * @return props + * @return properties **/ @ApiModelProperty(example = "null", value = "A blob of key value pairs that will be used to configure the check.") - public Map getProps() { - return props; + public Map getProperties() { + return properties; } - public void setProps(Map props) { - this.props = props; + public void setProperties(Map properties) { + this.properties = properties; } /** @@ -148,13 +148,13 @@ public boolean equals(java.lang.Object o) { } ReadinessCheck readinessCheck = (ReadinessCheck) o; return Objects.equals(this.type, readinessCheck.type) && - Objects.equals(this.props, readinessCheck.props) && + Objects.equals(this.properties, readinessCheck.properties) && Objects.equals(this.artifact, readinessCheck.artifact); } @Override public int hashCode() { - return Objects.hash(type, props, artifact); + return Objects.hash(type, properties, artifact); } @@ -164,7 +164,7 @@ public String toString() { sb.append("class ReadinessCheck {\n"); sb.append(" type: ").append(toIndentedString(type)).append("\n"); - sb.append(" props: ").append(toIndentedString(props)).append("\n"); + sb.append(" properties: ").append(toIndentedString(properties)).append("\n"); sb.append(" artifact: ").append(toIndentedString(artifact)).append("\n"); sb.append("}"); return sb.toString(); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/client/ServiceClient.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/client/ServiceClient.java index 636b127b930..775a642e962 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/client/ServiceClient.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/client/ServiceClient.java @@ -38,17 +38,8 @@ import org.apache.hadoop.yarn.api.ApplicationConstants; import org.apache.hadoop.yarn.api.protocolrecords.GetApplicationsRequest; import org.apache.hadoop.yarn.api.protocolrecords.UpdateApplicationTimeoutsRequest; -import org.apache.hadoop.yarn.api.records.ApplicationId; -import org.apache.hadoop.yarn.api.records.ApplicationReport; -import org.apache.hadoop.yarn.api.records.ApplicationSubmissionContext; -import org.apache.hadoop.yarn.api.records.ApplicationTimeout; -import org.apache.hadoop.yarn.api.records.ApplicationTimeoutType; -import org.apache.hadoop.yarn.api.records.ContainerLaunchContext; -import org.apache.hadoop.yarn.api.records.FinalApplicationStatus; -import org.apache.hadoop.yarn.api.records.LocalResource; -import org.apache.hadoop.yarn.api.records.LocalResourceType; -import org.apache.hadoop.yarn.api.records.Resource; -import org.apache.hadoop.yarn.api.records.YarnApplicationState; + +import org.apache.hadoop.yarn.api.records.*; import org.apache.hadoop.yarn.client.api.AppAdminClient; import org.apache.hadoop.yarn.client.api.YarnClient; import org.apache.hadoop.yarn.client.api.YarnClientApplication; @@ -99,7 +90,7 @@ import java.util.concurrent.ConcurrentHashMap; import static org.apache.hadoop.yarn.api.records.YarnApplicationState.*; -import static org.apache.hadoop.yarn.service.conf.YarnServiceConf.YARN_QUEUE; +import static org.apache.hadoop.yarn.service.conf.YarnServiceConf.*; import static org.apache.hadoop.yarn.service.utils.ServiceApiUtil.jsonSerDeser; import static org.apache.hadoop.yarn.service.utils.SliderUtils.*; @@ -553,8 +544,11 @@ private ApplicationId submitApp(Service app) appTimeout.put(ApplicationTimeoutType.LIFETIME, app.getLifetime()); submissionContext.setApplicationTimeouts(appTimeout); } - submissionContext.setMaxAppAttempts(conf.getInt( - YarnServiceConf.AM_RESTART_MAX, 2)); + submissionContext.setMaxAppAttempts(YarnServiceConf + .getInt(YarnServiceConf.AM_RESTART_MAX, 20, app.getConfiguration(), + conf)); + + setLogAggregationContext(app, conf, submissionContext); Map localResources = new HashMap<>(); @@ -568,14 +562,14 @@ private ApplicationId submitApp(Service app) if (LOG.isDebugEnabled()) { printLocalResources(localResources); } - Map env = addAMEnv(conf); + Map env = addAMEnv(); // create AM CLI - String cmdStr = - buildCommandLine(serviceName, conf, appRootDir, hasAMLog4j); + String cmdStr = buildCommandLine(serviceName, conf, appRootDir, hasAMLog4j); submissionContext.setResource(Resource.newInstance(YarnServiceConf - .getLong(YarnServiceConf.AM_RESOURCE_MEM, YarnServiceConf.DEFAULT_KEY_AM_RESOURCE_MEM, - app.getConfiguration(), conf), 1)); + .getLong(YarnServiceConf.AM_RESOURCE_MEM, + YarnServiceConf.DEFAULT_KEY_AM_RESOURCE_MEM, app.getConfiguration(), + conf), 1)); String queue = app.getQueue(); if (StringUtils.isEmpty(queue)) { queue = conf.get(YARN_QUEUE, "default"); @@ -598,6 +592,33 @@ private ApplicationId submitApp(Service app) return submissionContext.getApplicationId(); } + private void setLogAggregationContext(Service app, Configuration conf, + ApplicationSubmissionContext submissionContext) { + LogAggregationContext context = Records.newRecord(LogAggregationContext + .class); + String finalLogInclude = YarnServiceConf.get + (FINAL_LOG_INCLUSION_PATTERN, null, app.getConfiguration(), conf); + if (!StringUtils.isEmpty(finalLogInclude)) { + context.setIncludePattern(finalLogInclude); + } + String finalLogExclude = YarnServiceConf.get + (FINAL_LOG_EXCLUSION_PATTERN, null, app.getConfiguration(), conf); + if (!StringUtils.isEmpty(finalLogExclude)) { + context.setExcludePattern(finalLogExclude); + } + String rollingLogInclude = YarnServiceConf.get + (ROLLING_LOG_INCLUSION_PATTERN, null, app.getConfiguration(), conf); + if (!StringUtils.isEmpty(rollingLogInclude)) { + context.setRolledLogsIncludePattern(rollingLogInclude); + } + String rollingLogExclude = YarnServiceConf.get + (ROLLING_LOG_EXCLUSION_PATTERN, null, app.getConfiguration(), conf); + if (!StringUtils.isEmpty(rollingLogExclude)) { + context.setRolledLogsExcludePattern(rollingLogExclude); + } + submissionContext.setLogAggregationContext(context); + } + private void printLocalResources(Map map) { LOG.debug("Added LocalResource for localization: "); StringBuilder builder = new StringBuilder(); @@ -635,7 +656,7 @@ private String buildCommandLine(String serviceName, Configuration conf, return cmdStr; } - private Map addAMEnv(Configuration conf) throws IOException { + private Map addAMEnv() throws IOException { Map env = new HashMap<>(); ClasspathConstructor classpath = buildClasspath(YarnServiceConstants.SUBMITTED_CONF_DIR, "lib", fs, getConfig() diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/component/Component.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/component/Component.java index a5dd39c0456..7667a53e59e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/component/Component.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/component/Component.java @@ -116,7 +116,7 @@ CONTAINER_STARTED, new ContainerStartedTransition()) .addTransition(FLEXING, FLEXING, CONTAINER_COMPLETED, new ContainerCompletedTransition()) // Flex while previous flex is still in progress - .addTransition(FLEXING, EnumSet.of(FLEXING), FLEX, + .addTransition(FLEXING, EnumSet.of(FLEXING, STABLE), FLEX, new FlexComponentTransition()) // container failed while stable @@ -214,14 +214,12 @@ public ComponentState transition(Component component, new ArrayList<>(component.getAllComponentInstances()); // sort in Most recent -> oldest order, destroy most recent ones. - Collections.sort(list, Collections.reverseOrder()); + list.sort(Collections.reverseOrder()); for (int i = 0; i < delta; i++) { ComponentInstance instance = list.get(i); // remove the instance component.compInstances.remove(instance.getCompInstanceName()); component.pendingInstances.remove(instance); - component.componentMetrics.containersFailed.incr(); - component.componentMetrics.containersRunning.decr(); // decrement id counter component.instanceIdCounter.decrementAndGet(); instance.destroy(); @@ -282,7 +280,6 @@ public void transition(Component component, ComponentEvent event) { component.compInstanceDispatcher.getEventHandler().handle( new ComponentInstanceEvent(instance.getContainerId(), START)); - component.incRunningContainers(); } } @@ -294,7 +291,6 @@ private static class ContainerStartedTransition implements component.compInstanceDispatcher.getEventHandler().handle( new ComponentInstanceEvent(event.getInstance().getContainerId(), START)); - component.incRunningContainers(); return checkIfStable(component); } } @@ -413,9 +409,6 @@ private void updateMetrics(ContainerStatus status) { componentMetrics.containersFailed.incr(); scheduler.getServiceMetrics().containersFailed.incr(); - // dec running container - decRunningContainers(); - if (Apps.shouldCountTowardsNodeBlacklisting(status.getExitStatus())) { String host = scheduler.getLiveInstances().get(status.getContainerId()) .getNodeId().getHost(); @@ -468,31 +461,33 @@ public Map getDependencyHostIpTokens() { } String ip = instance.getContainerStatus().getIPs().get(0); String host = instance.getContainerStatus().getHost(); - tokens.put(String.format(COMPONENT_IP, + tokens.put(String.format(COMPONENT_INSTANCE_IP, instance.getCompInstanceName().toUpperCase()), ip); - tokens.put(String.format(COMPONENT_HOST, + tokens.put(String.format(COMPONENT_INSTANCE_HOST, instance.getCompInstanceName().toUpperCase()), host); } } return tokens; } - private void incRunningContainers() { + public void incRunningContainers() { componentMetrics.containersRunning.incr(); scheduler.getServiceMetrics().containersRunning.incr(); } + public void decRunningContainers() { + componentMetrics.containersRunning.decr(); + scheduler.getServiceMetrics().containersRunning.decr(); + } + public void incContainersReady() { componentMetrics.containersReady.incr(); + scheduler.getServiceMetrics().containersReady.incr(); } public void decContainersReady() { componentMetrics.containersReady.decr(); - } - - private void decRunningContainers() { - componentMetrics.containersRunning.decr(); - scheduler.getServiceMetrics().containersRunning.decr(); + scheduler.getServiceMetrics().containersReady.decr(); } public int getNumReadyInstances() { diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/component/instance/ComponentInstance.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/component/instance/ComponentInstance.java index 6e585aac498..2492454ec26 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/component/instance/ComponentInstance.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/component/instance/ComponentInstance.java @@ -148,7 +148,7 @@ private static class ContainerStartedTransition extends BaseTransition { new ContainerStatusRetriever(compInstance.scheduler, compInstance.getContainerId(), compInstance), 0, 1, TimeUnit.SECONDS); - + compInstance.component.incRunningContainers(); long containerStartTime = System.currentTimeMillis(); try { ContainerTokenIdentifier containerTokenIdentifier = BuilderUtils @@ -186,6 +186,10 @@ public void transition(ComponentInstance compInstance, ComponentInstanceEvent event) { compInstance.component.incContainersReady(); compInstance.containerSpec.setState(ContainerState.READY); + if (compInstance.timelineServiceEnabled) { + compInstance.serviceTimelinePublisher + .componentInstanceBecomeReady(compInstance.containerSpec); + } } } @@ -225,6 +229,10 @@ public void transition(ComponentInstance compInstance, .getDiagnostics(); compInstance.diagnostics.append(containerDiag + System.lineSeparator()); + if (compInstance.getState().equals(READY)) { + compInstance.component.decContainersReady(); + } + compInstance.component.decRunningContainers(); boolean shouldExit = false; // check if it exceeds the failure threshold if (comp.currentContainerFailure.get() > comp.maxContainerFailurePerComp) { @@ -250,10 +258,8 @@ public void transition(ComponentInstance compInstance, .submit(compInstance::cleanupRegistry); if (compInstance.timelineServiceEnabled) { // record in ATS - compInstance.serviceTimelinePublisher - .componentInstanceFinished(compInstance, - event.getStatus().getExitStatus(), event.getStatus().getState(), - containerDiag); + compInstance.serviceTimelinePublisher.componentInstanceFinished + (compInstance, event.getStatus().getExitStatus(), containerDiag); } compInstance.containerSpec.setState(ContainerState.STOPPED); } @@ -336,7 +342,7 @@ public void updateContainerStatus(ContainerStatus status) { container.setIp(StringUtils.join(",", status.getIPs())); container.setHostname(status.getHost()); if (timelineServiceEnabled) { - serviceTimelinePublisher.componentInstanceUpdated(container); + serviceTimelinePublisher.componentInstanceIPHostUpdated(container); } } updateServiceRecord(yarnRegistryOperations, status); @@ -410,7 +416,10 @@ private void updateServiceRecord( } } - // Release the container , cleanup registry, hdfs dir, and record in ATS + // Called when user flexed down the container and ContainerStoppedTransition + // is not executed in this case. + // Release the container, dec running, + // cleanup registry, hdfs dir, and send record to ATS public void destroy() { LOG.info(getCompInstanceId() + ": Flexed down by user, destroying."); diagnostics.append(getCompInstanceId() + ": Flexed down by user"); @@ -420,10 +429,18 @@ public void destroy() { .releaseAssignedContainer(container.getId()); getCompSpec().removeContainer(containerSpec); } + // update metrics + if (getState() == STARTED) { + component.decRunningContainers(); + } + if (getState() == READY) { + component.decContainersReady(); + component.decRunningContainers(); + } + if (timelineServiceEnabled) { - serviceTimelinePublisher - .componentInstanceFinished(this, KILLED_BY_APPMASTER, COMPLETE, - diagnostics.toString()); + serviceTimelinePublisher.componentInstanceFinished(this, + KILLED_BY_APPMASTER, diagnostics.toString()); } scheduler.executorService.submit(this::cleanupRegistryAndCompHdfsDir); } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/conf/YarnServiceConf.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/conf/YarnServiceConf.java index 1968e95153d..a7bd58d0a00 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/conf/YarnServiceConf.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/conf/YarnServiceConf.java @@ -28,7 +28,7 @@ public class YarnServiceConf { // Retry settings for container failures public static final String CONTAINER_RETRY_MAX = "yarn.service.container-failure.retry.max"; - public static final String CONTAINER_RETRY_INTERVAL = "yarn.service.container-failure.retry-interval"; + public static final String CONTAINER_RETRY_INTERVAL = "yarn.service.container-failure.retry-interval-ms"; public static final String AM_RESTART_MAX = "yarn.service.am-restart.max-attempts"; public static final String AM_RESOURCE_MEM = "yarn.service.am-resource.memory"; @@ -40,6 +40,13 @@ public class YarnServiceConf { public static final String DEFAULT_API_SERVER_ADDRESS = "0.0.0.0:"; public static final int DEFAULT_API_SERVER_PORT = 9191; + public static final String FINAL_LOG_INCLUSION_PATTERN = "yarn.service.log.include-pattern"; + public static final String FINAL_LOG_EXCLUSION_PATTERN = "yarn.service.log.exclude-pattern"; + + public static final String ROLLING_LOG_INCLUSION_PATTERN = "yarn.service.rolling-log.include-pattern"; + public static final String ROLLING_LOG_EXCLUSION_PATTERN = "yarn.service.rolling-log.exclude-pattern"; + + /** * The yarn service base path: * Defaults to HomeDir/.yarn/ @@ -98,4 +105,9 @@ public static int getInt(String name, int defaultValue, Configuration userConf, org.apache.hadoop.conf.Configuration systemConf) { return userConf.getPropertyInt(name, systemConf.getInt(name, defaultValue)); } + + public static String get(String name, String defaultVal, + Configuration userConf, org.apache.hadoop.conf.Configuration systemConf) { + return userConf.getProperty(name, systemConf.get(name, defaultVal)); + } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/conf/YarnServiceConstants.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/conf/YarnServiceConstants.java index e5ed7030017..3973759e3c7 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/conf/YarnServiceConstants.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/conf/YarnServiceConstants.java @@ -87,4 +87,6 @@ public interface YarnServiceConstants { String OUT_FILE = "stdout.txt"; String ERR_FILE = "stderr.txt"; + + String CONTENT = "content"; } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/monitor/probe/MonitorUtils.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/monitor/probe/MonitorUtils.java index 684f655cf2d..c4f63aee7ae 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/monitor/probe/MonitorUtils.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/monitor/probe/MonitorUtils.java @@ -70,9 +70,9 @@ public static Probe getProbe(ReadinessCheck readinessCheck) { try { switch (readinessCheck.getType()) { case HTTP: - return HttpProbe.create(readinessCheck.getProps()); + return HttpProbe.create(readinessCheck.getProperties()); case PORT: - return PortProbe.create(readinessCheck.getProps()); + return PortProbe.create(readinessCheck.getProperties()); default: return null; } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/provider/AbstractClientProvider.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/provider/AbstractClientProvider.java index 0d11be202b7..75c647b1c7b 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/provider/AbstractClientProvider.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/provider/AbstractClientProvider.java @@ -27,10 +27,13 @@ import java.io.IOException; import java.nio.file.Paths; +import java.text.MessageFormat; import java.util.HashSet; import java.util.List; import java.util.Set; +import static org.apache.hadoop.yarn.service.conf.YarnServiceConstants.CONTENT; + public abstract class AbstractClientProvider { public AbstractClientProvider() { @@ -84,11 +87,15 @@ public void validateConfigFiles(List configFiles, throw new IllegalArgumentException("File type is empty"); } - if (file.getType().equals(ConfigFile.TypeEnum.TEMPLATE) && StringUtils - .isEmpty(file.getSrcFile())) { - throw new IllegalArgumentException( - "Src_file is empty for " + ConfigFile.TypeEnum.TEMPLATE); - + if (file.getType().equals(ConfigFile.TypeEnum.TEMPLATE)) { + if (StringUtils.isEmpty(file.getSrcFile()) && + !file.getProperties().containsKey(CONTENT)) { + throw new IllegalArgumentException(MessageFormat.format("For {0} " + + "format, either src_file must be specified in ConfigFile," + + " or the \"{1}\" key must be specified in " + + "the 'properties' field of ConfigFile. ", + ConfigFile.TypeEnum.TEMPLATE, CONTENT)); + } } if (!StringUtils.isEmpty(file.getSrcFile())) { Path p = new Path(file.getSrcFile()); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/provider/ProviderUtils.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/provider/ProviderUtils.java index 63fbaae04ba..582c5ac3910 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/provider/ProviderUtils.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/provider/ProviderUtils.java @@ -272,8 +272,8 @@ public static synchronized void createConfigFileAndAddLocalResource( break; } } else { - // non-template - resolveNonTemplateConfigsAndSaveOnHdfs(fs, tokensForSubstitution, + // If src_file is not specified + resolvePropsInConfigFileAndSaveOnHdfs(fs, tokensForSubstitution, instance, configFile, fileName, remoteFile); } } @@ -297,17 +297,17 @@ public static synchronized void createConfigFileAndAddLocalResource( } } - private static void resolveNonTemplateConfigsAndSaveOnHdfs(SliderFileSystem fs, + private static void resolvePropsInConfigFileAndSaveOnHdfs(SliderFileSystem fs, Map tokensForSubstitution, ComponentInstance instance, ConfigFile configFile, String fileName, Path remoteFile) throws IOException { // substitute non-template configs - substituteMapWithTokens(configFile.getProps(), tokensForSubstitution); + substituteMapWithTokens(configFile.getProperties(), tokensForSubstitution); // write configs onto hdfs PublishedConfiguration publishedConfiguration = new PublishedConfiguration(fileName, - configFile.getProps().entrySet()); + configFile.getProperties().entrySet()); if (!fs.getFileSystem().exists(remoteFile)) { PublishedConfigurationOutputter configurationOutputter = PublishedConfigurationOutputter.createOutputter( @@ -343,7 +343,7 @@ private static void resolveHadoopXmlTemplateAndSaveOnHdfs(FileSystem fs, confCopy.set(entry.getKey(), entry.getValue()); } // substitute properties - for (Map.Entry entry : configFile.getProps().entrySet()) { + for (Map.Entry entry : configFile.getProperties().entrySet()) { confCopy.set(entry.getKey(), entry.getValue()); } // substitute env variables diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceTimelineEvent.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceTimelineEvent.java index 7f7f9a10037..6c3428a748d 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceTimelineEvent.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceTimelineEvent.java @@ -30,5 +30,7 @@ public enum ServiceTimelineEvent { COMPONENT_INSTANCE_UNREGISTERED, - COMPONENT_INSTANCE_UPDATED + COMPONENT_INSTANCE_IP_HOST_UPDATE, + + COMPONENT_INSTANCE_BECOME_READY } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceTimelinePublisher.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceTimelinePublisher.java index cfe7c1b6b83..c5229868005 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceTimelinePublisher.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/main/java/org/apache/hadoop/yarn/service/timelineservice/ServiceTimelinePublisher.java @@ -20,19 +20,13 @@ import org.apache.hadoop.metrics2.AbstractMetric; import org.apache.hadoop.service.CompositeService; -import org.apache.hadoop.yarn.api.records.ContainerState; import org.apache.hadoop.yarn.api.records.FinalApplicationStatus; import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntity; import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEvent; import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetric; import org.apache.hadoop.yarn.client.api.TimelineV2Client; import org.apache.hadoop.yarn.service.ServiceContext; -import org.apache.hadoop.yarn.service.api.records.Service; -import org.apache.hadoop.yarn.service.api.records.Component; -import org.apache.hadoop.yarn.service.api.records.ConfigFile; -import org.apache.hadoop.yarn.service.api.records.Configuration; -import org.apache.hadoop.yarn.service.api.records.Container; -import org.apache.hadoop.yarn.service.api.records.ServiceState; +import org.apache.hadoop.yarn.service.api.records.*; import org.apache.hadoop.yarn.service.component.instance.ComponentInstance; import org.apache.hadoop.yarn.util.timeline.TimelineUtils; import org.slf4j.Logger; @@ -46,6 +40,8 @@ import java.util.Map.Entry; import java.util.Set; +import static org.apache.hadoop.yarn.service.api.records.ContainerState.READY; +import static org.apache.hadoop.yarn.service.api.records.ContainerState.STOPPED; import static org.apache.hadoop.yarn.service.timelineservice.ServiceTimelineMetricsConstants.DIAGNOSTICS_INFO; /** @@ -117,8 +113,8 @@ public void serviceAttemptRegistered(Service service, // publish system config - YarnConfiguration populateTimelineEntity(systemConf.iterator(), service.getId(), ServiceTimelineEntityType.SERVICE_ATTEMPT.toString()); - // publish user conf - publishUserConf(service.getConfiguration(), service.getId(), + // publish container conf + publishContainerConf(service.getConfiguration(), service.getId(), ServiceTimelineEntityType.SERVICE_ATTEMPT.toString()); // publish component as separate entity. @@ -183,7 +179,7 @@ public void componentInstanceStarted(Container container, } public void componentInstanceFinished(ComponentInstance instance, - int exitCode, ContainerState state, String diagnostics) { + int exitCode, String diagnostics) { TimelineEntity entity = createComponentInstanceEntity( instance.getContainer().getId().toString()); @@ -192,7 +188,7 @@ public void componentInstanceFinished(ComponentInstance instance, entityInfos.put(ServiceTimelineMetricsConstants.EXIT_STATUS_CODE, exitCode); entityInfos.put(DIAGNOSTICS_INFO, diagnostics); - entityInfos.put(ServiceTimelineMetricsConstants.STATE, state); + entityInfos.put(ServiceTimelineMetricsConstants.STATE, STOPPED); entity.addInfo(entityInfos); // add an event @@ -205,7 +201,7 @@ public void componentInstanceFinished(ComponentInstance instance, putEntity(entity); } - public void componentInstanceUpdated(Container container) { + public void componentInstanceIPHostUpdated(Container container) { TimelineEntity entity = createComponentInstanceEntity(container.getId()); // create info keys @@ -218,14 +214,27 @@ public void componentInstanceUpdated(Container container) { entity.addInfo(entityInfos); TimelineEvent updateEvent = new TimelineEvent(); - updateEvent - .setId(ServiceTimelineEvent.COMPONENT_INSTANCE_UPDATED.toString()); + updateEvent.setId(ServiceTimelineEvent.COMPONENT_INSTANCE_IP_HOST_UPDATE + .toString()); updateEvent.setTimestamp(System.currentTimeMillis()); entity.addEvent(updateEvent); putEntity(entity); } + public void componentInstanceBecomeReady(Container container) { + TimelineEntity entity = createComponentInstanceEntity(container.getId()); + Map entityInfo = new HashMap<>(); + entityInfo.put(ServiceTimelineMetricsConstants.STATE, READY); + entity.addInfo(entityInfo); + TimelineEvent updateEvent = new TimelineEvent(); + updateEvent.setId(ServiceTimelineEvent.COMPONENT_INSTANCE_BECOME_READY + .toString()); + updateEvent.setTimestamp(System.currentTimeMillis()); + entity.addEvent(updateEvent); + putEntity(entity); + } + private void publishComponents(List components) { long currentTimeMillis = System.currentTimeMillis(); for (Component component : components) { @@ -266,22 +275,19 @@ private void publishComponents(List components) { putEntity(entity); - // publish component specific configurations - publishUserConf(component.getConfiguration(), component.getName(), + // publish container specific configurations + publishContainerConf(component.getConfiguration(), component.getName(), ServiceTimelineEntityType.COMPONENT.toString()); } } - private void publishUserConf(Configuration configuration, + private void publishContainerConf(Configuration configuration, String entityId, String entityType) { - populateTimelineEntity(configuration.getProperties().entrySet().iterator(), - entityId, entityType); - populateTimelineEntity(configuration.getEnv().entrySet().iterator(), entityId, entityType); for (ConfigFile configFile : configuration.getFiles()) { - populateTimelineEntity(configFile.getProps().entrySet().iterator(), + populateTimelineEntity(configFile.getProperties().entrySet().iterator(), entityId, entityType); } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/java/org/apache/hadoop/yarn/service/conf/TestAppJsonResolve.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/java/org/apache/hadoop/yarn/service/conf/TestAppJsonResolve.java index 8739382603d..18318aa05eb 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/java/org/apache/hadoop/yarn/service/conf/TestAppJsonResolve.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/java/org/apache/hadoop/yarn/service/conf/TestAppJsonResolve.java @@ -96,9 +96,9 @@ public void testOverride() throws Throwable { props.put("k1", "overridden"); props.put("k2", "v2"); files.add(new ConfigFile().destFile("file1").type(ConfigFile.TypeEnum - .PROPERTIES).props(props)); + .PROPERTIES).properties(props)); files.add(new ConfigFile().destFile("file2").type(ConfigFile.TypeEnum - .XML).props(Collections.singletonMap("k3", "v3"))); + .XML).properties(Collections.singletonMap("k3", "v3"))); assertTrue(files.contains(simple.getFiles().get(0))); assertTrue(files.contains(simple.getFiles().get(1))); @@ -113,9 +113,9 @@ public void testOverride() throws Throwable { props.put("k1", "v1"); files.clear(); files.add(new ConfigFile().destFile("file1").type(ConfigFile.TypeEnum - .PROPERTIES).props(props)); + .PROPERTIES).properties(props)); files.add(new ConfigFile().destFile("file2").type(ConfigFile.TypeEnum - .XML).props(Collections.singletonMap("k3", "v3"))); + .XML).properties(Collections.singletonMap("k3", "v3"))); assertTrue(files.contains(master.getFiles().get(0))); assertTrue(files.contains(master.getFiles().get(1))); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/java/org/apache/hadoop/yarn/service/timelineservice/TestServiceTimelinePublisher.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/java/org/apache/hadoop/yarn/service/timelineservice/TestServiceTimelinePublisher.java index e7c76000c5e..80b4f5103a3 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/java/org/apache/hadoop/yarn/service/timelineservice/TestServiceTimelinePublisher.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/java/org/apache/hadoop/yarn/service/timelineservice/TestServiceTimelinePublisher.java @@ -160,7 +160,7 @@ public void testComponentInstanceEntity() { // updated container state container.setState(ContainerState.READY); - serviceTimelinePublisher.componentInstanceUpdated(container); + serviceTimelinePublisher.componentInstanceIPHostUpdated(container); lastPublishedEntities = ((DummyTimelineClient) timelineClient).getLastPublishedEntities(); assertEquals(1, lastPublishedEntities.size()); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/resources/org/apache/hadoop/yarn/service/conf/examples/app-override.json b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/resources/org/apache/hadoop/yarn/service/conf/examples/app-override.json index 753a9cdf9fe..30fa0eb8db2 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/resources/org/apache/hadoop/yarn/service/conf/examples/app-override.json +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-services/hadoop-yarn-services-core/src/test/resources/org/apache/hadoop/yarn/service/conf/examples/app-override.json @@ -10,7 +10,7 @@ { "type": "PROPERTIES", "dest_file": "file1", - "props": { + "properties": { "k1": "v1", "k2": "v2" } @@ -18,7 +18,7 @@ { "type": "XML", "dest_file": "file2", - "props": { + "properties": { "k3": "v3" } } @@ -38,7 +38,7 @@ { "type": "PROPERTIES", "dest_file": "file1", - "props": { + "properties": { "k1": "overridden" } } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Configurations.md b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Configurations.md new file mode 100644 index 00000000000..561df7f696d --- /dev/null +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Configurations.md @@ -0,0 +1,165 @@ + + +This document describes how to configure the services to be deployed on YARN + +There are mainly three types of configurations: + +* The configurations specific to the custom service running on YARN . E.g. the hbase-site.xml for a Hbase service running on YARN. + * It can be specified at both global service level(`Service#configuration`) or component level(`Component#configuration`). + * Service-level configs are considered as the default configs for all components and component-level config can override service level config. + * All config properties that uses constant variables as described below are subject to substitutions. +* The configurations specific to YARN service AM. (`Configuration#properties`). + * E.g. The `yarn.service.am-restart.max-attempts` which controls how many times +the framework AM can be retried if it fails. These configs are mainly to control the behavior of the framework AM , rather than the custom services launched by the framework. +* Some constants such as `SERVICE_NAME` for referring some system properties. + * They are substituted by the service AM before writing the config files. + +Below describes the details for each type of configurations. + +## Configuration for custom service +Below is how a configuration object typically looks like: +``` +"configuration" : { + "properties" : { + "yarn.service.am-restart.max-attempts" : 10 // config for the service AM + }, + "env" : { // The environment variables to be exported when container gets launched + "env1" : "val1" + }, + "files" : [ // The list of configuration files to be mounted for the container + { + "type": "HADOOP_XML", // The format of the config file into which the "properties" are dumped + "dest_file": "/etc/hadoop/conf/core-site.xml", // The location where the config file is mounted inside the container + "properties" : { // The list of key/value pairs to be dumped into the config file + "fs.defaultFS" : "hdfs://myhdfs" // This property will be written into core-site.xml + } + }, + { + "type": "HADOOP_XML", // The format of the config file. + "src_file" : ""hdfs://mycluster/tmp/conf/yarn-site.xml"" // The location of the source config file to be downloaded + "dest_file": "/etc/hadoop/conf/yarn-site.xml", // The location where the config file will be mounted inside the container/ + "properties" : { + "yarn.resourcemanager.webapp.address" : "${COMPONENT_INSTANCE_NAME}.${SERVICE_NAME}.${USER}.${DOMAIN}" // Merge into (or override existing property in) yarn-site.xml + } + } + ] +} +``` + +* properties: the configurations for service AM. Details below. +* env : the environment variables to be exported when container gets launched. +* files : The list of configuration files to be mounted inside the container. + * type: The format of the config file(`dest_file`) to be mounted inside the container. If `src_file` is specified, it is also the format for both `src_file` and `dest_file`. + * HADOOP_XML : the hadoop xml format. If `src_file` is specified, the file content will be read as parsed in hadoop xml format. + * XML : the standard xml format + * JSON : the standard JSON format + * YAML : the YAML format + * PROPERTIES : the java PROPERTIES format + * TEMPLATE : the plain text format. If `src_file` is specified, the content of the `src_file` will be written into `dest_file` post constant substitution. + If `src_file` is not specified, use `content` as the key in `properties` field, and the value will be the actual content to be written in the `dest_file` post constant substitution. E.g + + ``` + { + "type": "TEMPLATE" + "dest_file": "/etc/conf/hello" + "properties" : { + "content" : "Hello world" + } + } + ``` + The string `Hello world` will be written into a file located at `/etc/conf/hello` inside the container. + * src_file : [optional], the source location of the config file at a network accessible location such as hdfs. + * The format of both `src_file` and `dest_file` are defined by `type`. + * Currently, It only works with `HADOOP_XML` and `TEMPLATE` type. + * The `src_file` will be downloaded by YARN NodeManager and be mounted inside the container as in the location specified by `dest_file`. + * If any properties specified in the `properties` field, they are added into (or overwriting existing properties in) the `src_file`. + * If `src_file` is not specified, only the properties in the `properties` field will be written into the `dest_file` file. + * dest_file : the location where the config file is mounted inside the container. The file format is defined by `type`. + dest_file can be an absolute path or a relative path. If it's a relative path, the file will be located in the `$PWD/conf` directory (where `$PWD` is the container local directory which is mounted to all docker containers launched by YARN) + * properties : The list of key/value pair configs to be written into the `dest_file` in the format as defined in `type`. If `src_file` is specified, these properties will be added into (or overwriting existing properties in) the `src_file`. + +## Configuration for YARN service AM +This section describes the configurations for configuring the YARN service AM. + +These can be specified either in the cluster `yarn-site.xml` at the global level or in the `properties` field of the `Configuration` object as per service basis like below: +``` +{ + "configuration" : { + "properties" : { + "yarn.service.am-restart.max-attempts" : 10 + } + } +} +``` +Above config make the service AM to be retried at max 10 times. + +#### Available configurations: + +| Name | Description | +| ------------ | ------------- | +|yarn.service.client-am.retry.max-wait-ms | the max retry time in milliseconds for the service client to talk to the service AM. By default, it is set to 0, which means no retry | +|yarn.service.client-am.retry-interval-ms | the retry interval in milliseconds for the service client to talk to the service AM. By default, it is 2000, i.e. 2 seconds | +|yarn.service.container-failure.retry.max | the max number of retries for the container to be auto restarted if it fails. By default, it is set to -1, which means forever. +|yarn.service.container-failure.retry-interval-ms| the retry interval in milliseconds for the container to be restarted. By default, it is 30000, i.e. 30 seconds | +|yarn.service.am-restart.max-attempts| the max number of attempts for the framework AM +|yarn.service.am-resource.memory | the memory size in GB for the framework AM. By default, it is set to 1024 +|yarn.service.queue | the default queue to which the service will be submitted. By default, it is submitted to `default` queue +|yarn.service.base.path | the root location for the service artifacts on hdfs for a user. By default, it is under ${user_home_dir}/.yarn/ +|yarn.service.container-failure-per-component.threshold | the max number of container failures for a given component before the AM exits. +|yarn.service.node-blacklist.threshold | Maximum number of container failures on a node before the node is blacklisted by the AM +|yarn.service.failure-count-reset.window | The interval in seconds when the `yarn.service.container-failure-per-component.threshold` and `yarn.service.node-blacklist.threshold` gets reset. By default, it is 21600, i.e. 6 hours +|yarn.service.readiness-check-interval.seconds | The interval in seconds between readiness checks. By default, it is 30 seconds +|yarn.service.log.include-pattern| The regex expression for including log files whose file name matches it when aggregating the logs after the application completes. +|yarn.service.log.exclude-pattern| The regex expression for excluding log files whose file name matches it when aggregating the logs after the application completes. If the log file name matches both include and exclude pattern, this file will be excluded. +|yarn.service.rolling-log.include-pattern| The regex expression for including log files whose file name matches it when aggregating the logs while app is running. +|yarn.service.rolling-log.exclude-pattern| The regex expression for excluding log files whose file name matches it when aggregating the logs while app is running. If the log file name matches both include and exclude pattern, this file will be excluded. + +## Constant variables for custom service +The service framework provides some constant variables for user to configure their services. These variables are either dynamically generated by the system or are static ones such as service name defined by the user. +User can use these constants in their configurations to be dynamically substituted by the service AM.E.g. +``` +{ + "type" : "HADOOP_XML", + "dest_file" : "/etc/hadoop/hbase-site.xml", + "properties" : { + "hbase.regionserver.hostname": "${COMPONENT_INSTANCE_NAME}.${SERVICE_NAME}.${USER}.${DOMAIN}" + } +} +``` +Here, `COMPONENT_INSTANCE_NAME` and `SERVICE_NAME` are the constants to be substituted by the system. + +Suppose the `COMPONENT_INSTANCE_NAME` is `regionserver-0` and `SERVICE_NAME` is defined by user as `hbase`, user name is `devuser` and domain name is `hwxdev.site`. +Then, the config will be substituted by the service AM and written in the config file `/etc/hadoop/hbase-site.xml` inside the container as below: +``` + + hbase.regionserver.hostname + regionserver-0.hbase.devuser.hwxdev.site + +``` +where `regionserver-0` is the actual component instance name assigned by the system for this container. + +#### Available constants: +| Name | Description | +| ------------ | ------------- | +| SERVICE_NAME | name of the service defined by the user +| USER | user who submits the service | +| DOMAIN | the domain name for the cluster | +| COMPONENT_NAME | the name for a given component | +| COMPONENT_INSTANCE_NAME | the name for a given component instance (i.e. container) | +| COMPONENT_ID | the monotonically increasing integer for a given component +| CONTAINER_ID | the YARN container Id for a given container | +| ${COMPONENT_INSTANCE_NAME}_HOST | the hostname for a component instance (i.e. container), e.g. REGIONSERVER-0_HOST will be substituted by the actual hostname of the component instance. Note all characters must be uppercase. | +| ${COMPONENT_INSTANCE_NAME}_IP | the ip for a component instance (i.e. container), e.g. REGIONSERVER-0_IP will be substituted by the actual IP address of the component instance. Note all characters must be uppercase. | +| CLUSTER_FS_URI | the URI of the cluster hdfs | diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Examples.md b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Examples.md index e4881cece91..0ddd1260e1e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Examples.md +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Examples.md @@ -45,7 +45,7 @@ Note this example requires registry DNS. { "type": "ENV", "dest_file": "/var/www/html/index.html", - "props": { + "properties": { "content": "
Title
Hello from ${COMPONENT_INSTANCE_NAME}!" } } @@ -122,7 +122,7 @@ A readiness check is added for the `httpd` component: ``` "readiness_check": { "type": "HTTP", - "props": { + "properties": { "url": "http://${THIS_HOST}:8080" } }, diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Overview.md b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Overview.md index 58daee59e33..72c2d3e1dc4 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Overview.md +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/Overview.md @@ -52,9 +52,10 @@ The benefits of combining these workloads are two-fold: * [Concepts](Concepts.md): Describes the internals of the framework and some features in YARN core to support running services on YARN. * [Service REST API](YarnServiceAPI.md): The API doc for deploying/managing services on YARN. -* [Service Discovery](ServiceDiscovery.md): Descirbes the service discovery mechanism on YARN. +* [Service Discovery](ServiceDiscovery.md): Describes the service discovery mechanism on YARN. * [Registry DNS](RegistryDNS.md): Deep dives into the Registry DNS internals. * [Examples](Examples.md): List some example service definitions (`Yarnfile`). +* [Configurations](Configurations.md): Describes how to configure the custom services on YARN. diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/YarnServiceAPI.md b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/YarnServiceAPI.md index 7a2490e2116..c0e12c7e5c8 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/YarnServiceAPI.md +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site/src/site/markdown/yarn-service/YarnServiceAPI.md @@ -251,7 +251,7 @@ A config file that needs to be created and made available as a volume in a servi |type|Config file in the standard format like xml, properties, json, yaml, template.|false|enum (XML, PROPERTIES, JSON, YAML, TEMPLATE, ENV, HADOOP_XML)|| |dest_file|The path that this configuration file should be created as. If it is an absolute path, it will be mounted into the DOCKER container. Absolute paths are only allowed for DOCKER containers. If it is a relative path, only the file name should be provided, and the file will be created in the container local working directory under a folder named conf.|false|string|| |src_file|This provides the source location of the configuration file, the content of which is dumped to dest_file post property substitutions, in the format as specified in type. Typically the src_file would point to a source controlled network accessible file maintained by tools like puppet, chef, or hdfs etc. Currently, only hdfs is supported.|false|string|| -|props|A blob of key value pairs that will be dumped in the dest_file in the format as specified in type. If src_file is specified, src_file content are dumped in the dest_file and these properties will overwrite, if any, existing properties in src_file or be added as new properties in src_file.|false|object|| +|properties|A blob of key value pairs that will be dumped in the dest_file in the format as specified in type. If src_file is specified, src_file content are dumped in the dest_file and these properties will overwrite, if any, existing properties in src_file or be added as new properties in src_file.|false|object|| ### Configuration @@ -260,7 +260,7 @@ Set of configuration properties that can be injected into the service components |Name|Description|Required|Schema|Default| |----|----|----|----|----| -|properties|A blob of key-value pairs of common service properties.|false|object|| +|properties|A blob of key-value pairs for configuring YARN service AM.|false|object|| |env|A blob of key-value pairs which will be appended to the default system properties and handed off to the service at start time. All placeholder references to properties will be substituted before injection.|false|object|| |files|Array of list of files that needs to be created and made available as volumes in the service component containers.|false|ConfigFile array|| @@ -308,7 +308,7 @@ A custom command or a pluggable helper container to determine the readiness of a |Name|Description|Required|Schema|Default| |----|----|----|----|----| |type|E.g. HTTP (YARN will perform a simple REST call at a regular interval and expect a 204 No content).|true|enum (HTTP, PORT)|| -|props|A blob of key value pairs that will be used to configure the check.|false|object|| +|properties|A blob of key value pairs that will be used to configure the check.|false|object|| |artifact|Artifact of the pluggable readiness check helper container (optional). If specified, this helper container typically hosts the http uri and encapsulates the complex scripts required to perform actual container readiness check. At the end it is expected to respond a 204 No content just like the simplified use case. This pluggable framework benefits service owners who can run services without any packaging modifications. Note, artifacts of type docker only is supported for now. NOT IMPLEMENTED YET|false|Artifact|| @@ -529,14 +529,14 @@ POST URL - http://localhost:8088:/ws/v1/services/hbase-app-1 { "type": "XML", "dest_file": "/etc/hadoop/conf/core-site.xml", - "props": { + "properties": { "fs.defaultFS": "${CLUSTER_FS_URI}" } }, { "type": "XML", "dest_file": "/etc/hbase/conf/hbase-site.xml", - "props": { + "properties": { "hbase.cluster.distributed": "true", "hbase.zookeeper.quorum": "${CLUSTER_ZK_QUORUM}", "hbase.rootdir": "${SERVICE_HDFS_DIR}/hbase", @@ -569,14 +569,14 @@ POST URL - http://localhost:8088:/ws/v1/services/hbase-app-1 { "type": "XML", "dest_file": "/etc/hadoop/conf/core-site.xml", - "props": { + "properties": { "fs.defaultFS": "${CLUSTER_FS_URI}" } }, { "type": "XML", "dest_file": "/etc/hbase/conf/hbase-site.xml", - "props": { + "properties": { "hbase.cluster.distributed": "true", "hbase.zookeeper.quorum": "${CLUSTER_ZK_QUORUM}", "hbase.rootdir": "${SERVICE_HDFS_DIR}/hbase", diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/controllers/yarn-app/components.js b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/controllers/yarn-app/components.js index b4bac0d98c7..812f1e6a059 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/controllers/yarn-app/components.js +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/controllers/yarn-app/components.js @@ -49,7 +49,7 @@ export default Ember.Controller.extend({ contentPath: 'memory' }, { id: 'instances', - headerTitle: 'Number Of Instances', + headerTitle: 'Number of Instances', contentPath: 'instances', observePath: true }); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/components/metrics-table.hbs b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/components/metrics-table.hbs index 17d38cb495b..0f0156a8862 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/components/metrics-table.hbs +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/components/metrics-table.hbs @@ -20,7 +20,7 @@
-
{{type}} Metrics: Detailed Information
+
{{type}} Metrics:
diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/yarn-component-instance/info.hbs b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/yarn-component-instance/info.hbs index 5a94575aa44..ef517d024c3 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/yarn-component-instance/info.hbs +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/yarn-component-instance/info.hbs @@ -20,7 +20,7 @@ {{#if model.container}}
-
Component Information
+
Component Instance Information
@@ -30,7 +30,7 @@ - + diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/yarn-component-instances/info.hbs b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/yarn-component-instances/info.hbs index 0b642b0abbc..fb4dba27631 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/yarn-component-instances/info.hbs +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/app/templates/yarn-component-instances/info.hbs @@ -18,7 +18,7 @@
-

Active Components: {{model.componentName}}

+

Component: {{model.componentName}}

{{em-table columns=tableColumns rows=model.instances}}
{{check-availability model.container.instanceName}}
Component GroupComponent {{check-availability model.container.component}}