YARN-9016 DocumentStore as a backend for ATSv2. Contributed by Sushil Ks.
This commit is contained in:
parent
6fa229891e
commit
f235a942d5
|
@ -224,6 +224,10 @@
|
|||
<directory>hadoop-yarn/hadoop-yarn-csi/target/lib</directory>
|
||||
<outputDirectory>share/hadoop/${hadoop.component}/csi/lib</outputDirectory>
|
||||
</fileSet>
|
||||
<fileSet>
|
||||
<directory>hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore/target/lib</directory>
|
||||
<outputDirectory>share/hadoop/${hadoop.component}/timelineservice/lib</outputDirectory>
|
||||
</fileSet>
|
||||
</fileSets>
|
||||
<moduleSets>
|
||||
<moduleSet>
|
||||
|
@ -231,6 +235,7 @@
|
|||
<include>org.apache.hadoop:hadoop-yarn-server-timelineservice</include>
|
||||
<include>org.apache.hadoop:hadoop-yarn-server-timelineservice-hbase-client</include>
|
||||
<include>org.apache.hadoop:hadoop-yarn-server-timelineservice-hbase-common</include>
|
||||
<include>org.apache.hadoop:hadoop-yarn-server-timelineservice-documentstore</include>
|
||||
</includes>
|
||||
<binaries>
|
||||
<outputDirectory>share/hadoop/${hadoop.component}/timelineservice</outputDirectory>
|
||||
|
|
|
@ -0,0 +1,150 @@
|
|||
<?xml version="1.0"?>
|
||||
<!--
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License. See accompanying LICENSE file.
|
||||
-->
|
||||
<project xmlns="http://maven.apache.org/POM/4.0.0"
|
||||
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
|
||||
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0
|
||||
http://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||
<parent>
|
||||
<artifactId>hadoop-yarn-server</artifactId>
|
||||
<groupId>org.apache.hadoop</groupId>
|
||||
<version>3.3.0-SNAPSHOT</version>
|
||||
</parent>
|
||||
<modelVersion>4.0.0</modelVersion>
|
||||
<artifactId>hadoop-yarn-server-timelineservice-documentstore</artifactId>
|
||||
<name>Apache Hadoop YARN TimelineService DocumentStore</name>
|
||||
|
||||
<properties>
|
||||
<!-- Needed for generating FindBugs warnings using parent pom -->
|
||||
<yarn.basedir>${project.parent.parent.basedir}</yarn.basedir>
|
||||
<azure.documentdb.version>1.16.2</azure.documentdb.version>
|
||||
</properties>
|
||||
|
||||
<dependencies>
|
||||
<dependency>
|
||||
<groupId>org.apache.hadoop</groupId>
|
||||
<artifactId>hadoop-yarn-server-timelineservice</artifactId>
|
||||
<version>${project.version}</version>
|
||||
<scope>provided</scope>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.apache.hadoop</groupId>
|
||||
<artifactId>hadoop-common</artifactId>
|
||||
<type>test-jar</type>
|
||||
<scope>test</scope>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>com.microsoft.azure</groupId>
|
||||
<artifactId>azure-documentdb</artifactId>
|
||||
<version>${azure.documentdb.version}</version>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
<scope>test</scope>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.mockito</groupId>
|
||||
<artifactId>mockito-core</artifactId>
|
||||
<version>2.8.9</version>
|
||||
<scope>test</scope>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.powermock</groupId>
|
||||
<artifactId>powermock-api-mockito2</artifactId>
|
||||
<version>1.7.1</version>
|
||||
<scope>test</scope>
|
||||
<exclusions>
|
||||
<exclusion>
|
||||
<groupId>org.mockito</groupId>
|
||||
<artifactId>mockito-core</artifactId>
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.powermock</groupId>
|
||||
<artifactId>powermock-module-junit4</artifactId>
|
||||
<version>1.7.1</version>
|
||||
<scope>test</scope>
|
||||
<exclusions>
|
||||
<exclusion>
|
||||
<groupId>org.mockito</groupId>
|
||||
<artifactId>mockito-core</artifactId>
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
</dependencies>
|
||||
|
||||
<build>
|
||||
<plugins>
|
||||
<!-- Apache RAT -->
|
||||
<plugin>
|
||||
<groupId>org.apache.rat</groupId>
|
||||
<artifactId>apache-rat-plugin</artifactId>
|
||||
<configuration>
|
||||
<excludes>
|
||||
<exclude>src/test/resources/documents/flowactivity-doc.json</exclude>
|
||||
<exclude>src/test/resources/documents/flowrun-doc.json</exclude>
|
||||
<exclude>src/test/resources/documents/test-timeline-entities-doc.json</exclude>
|
||||
<exclude>src/test/resources/documents/timeline-app-doc.json</exclude>
|
||||
<exclude>src/test/resources/documents/timeline-entities.json</exclude>
|
||||
</excludes>
|
||||
</configuration>
|
||||
</plugin>
|
||||
|
||||
<plugin>
|
||||
<artifactId>maven-jar-plugin</artifactId>
|
||||
<executions>
|
||||
<execution>
|
||||
<goals>
|
||||
<goal>test-jar</goal>
|
||||
</goals>
|
||||
<phase>test-compile</phase>
|
||||
</execution>
|
||||
</executions>
|
||||
</plugin>
|
||||
<plugin>
|
||||
<groupId>org.apache.maven.plugins</groupId>
|
||||
<artifactId>maven-javadoc-plugin</artifactId>
|
||||
<configuration>
|
||||
<additionalDependencies>
|
||||
<additionnalDependency>
|
||||
<groupId>junit</groupId>
|
||||
<artifactId>junit</artifactId>
|
||||
<version>4.11</version>
|
||||
</additionnalDependency>
|
||||
</additionalDependencies>
|
||||
</configuration>
|
||||
</plugin>
|
||||
<plugin>
|
||||
<groupId>org.apache.maven.plugins</groupId>
|
||||
<artifactId>maven-dependency-plugin</artifactId>
|
||||
<executions>
|
||||
<execution>
|
||||
<phase>package</phase>
|
||||
<goals>
|
||||
<goal>copy-dependencies</goal>
|
||||
</goals>
|
||||
<configuration>
|
||||
<includeScope>runtime</includeScope>
|
||||
<excludeGroupIds>org.slf4j,org.apache.hadoop,com.github.stephenc.findbugs</excludeGroupIds>
|
||||
<outputDirectory>${project.build.directory}/lib</outputDirectory>
|
||||
</configuration>
|
||||
</execution>
|
||||
</executions>
|
||||
</plugin>
|
||||
</plugins>
|
||||
</build>
|
||||
</project>
|
||||
|
|
@ -0,0 +1,66 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreFactory;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.CollectionType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.DocumentStoreWriter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.storage.SchemaCreator;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* This creates the Collection for a {@link DocumentStoreVendor} backend
|
||||
* configured for storing application timeline information.
|
||||
*/
|
||||
public class DocumentStoreCollectionCreator implements SchemaCreator {
|
||||
|
||||
private static final Logger LOG = LoggerFactory
|
||||
.getLogger(DocumentStoreCollectionCreator.class);
|
||||
|
||||
|
||||
@Override
|
||||
public void createTimelineSchema(String[] args) {
|
||||
try {
|
||||
|
||||
Configuration conf = new YarnConfiguration();
|
||||
|
||||
LOG.info("Creating database and collections for DocumentStore : {}",
|
||||
DocumentStoreUtils.getStoreVendor(conf));
|
||||
|
||||
try(DocumentStoreWriter documentStoreWriter = DocumentStoreFactory
|
||||
.createDocumentStoreWriter(conf)) {
|
||||
documentStoreWriter.createDatabase();
|
||||
documentStoreWriter.createCollection(
|
||||
CollectionType.APPLICATION.getCollectionName());
|
||||
documentStoreWriter.createCollection(
|
||||
CollectionType.ENTITY.getCollectionName());
|
||||
documentStoreWriter.createCollection(
|
||||
CollectionType.FLOW_ACTIVITY.getCollectionName());
|
||||
documentStoreWriter.createCollection(
|
||||
CollectionType.FLOW_RUN.getCollectionName());
|
||||
}
|
||||
} catch (Exception e) {
|
||||
LOG.error("Error while creating Timeline Collections", e);
|
||||
}
|
||||
}
|
||||
}
|
|
@ -0,0 +1,121 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.service.AbstractService;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntityType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.reader.TimelineCollectionReader;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineDataToRetrieve;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineEntityFilters;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineReaderContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.storage.TimelineReader;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEntityDocument;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.HashSet;
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
|
||||
/**
|
||||
* This is a generic document store timeline reader for reading the timeline
|
||||
* entity information. Based on the {@link DocumentStoreVendor} that is
|
||||
* configured, the documents are read from that backend.
|
||||
*/
|
||||
public class DocumentStoreTimelineReaderImpl
|
||||
extends AbstractService implements TimelineReader {
|
||||
|
||||
private static final Logger LOG = LoggerFactory
|
||||
.getLogger(DocumentStoreTimelineReaderImpl.class);
|
||||
|
||||
private TimelineCollectionReader collectionReader;
|
||||
|
||||
public DocumentStoreTimelineReaderImpl() {
|
||||
super(DocumentStoreTimelineReaderImpl.class.getName());
|
||||
}
|
||||
|
||||
@Override
|
||||
public void serviceInit(Configuration conf) throws Exception {
|
||||
DocumentStoreVendor storeType = DocumentStoreUtils.getStoreVendor(conf);
|
||||
LOG.info("Initializing Document Store Reader for : " + storeType);
|
||||
collectionReader = new TimelineCollectionReader(conf);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void serviceStop() throws Exception {
|
||||
super.serviceStop();
|
||||
LOG.info("Stopping Document Timeline Store reader...");
|
||||
collectionReader.close();
|
||||
}
|
||||
|
||||
public TimelineEntity getEntity(TimelineReaderContext context,
|
||||
TimelineDataToRetrieve dataToRetrieve) throws IOException {
|
||||
TimelineEntityDocument timelineEntityDoc;
|
||||
switch (TimelineEntityType.valueOf(context.getEntityType())) {
|
||||
case YARN_FLOW_ACTIVITY:
|
||||
case YARN_FLOW_RUN:
|
||||
timelineEntityDoc =
|
||||
collectionReader.readDocument(context);
|
||||
return DocumentStoreUtils.createEntityToBeReturned(
|
||||
timelineEntityDoc, dataToRetrieve.getConfsToRetrieve(),
|
||||
dataToRetrieve.getMetricsToRetrieve());
|
||||
default:
|
||||
timelineEntityDoc =
|
||||
collectionReader.readDocument(context);
|
||||
}
|
||||
return DocumentStoreUtils.createEntityToBeReturned(
|
||||
timelineEntityDoc, dataToRetrieve);
|
||||
}
|
||||
|
||||
public Set<TimelineEntity> getEntities(TimelineReaderContext context,
|
||||
TimelineEntityFilters filters, TimelineDataToRetrieve dataToRetrieve)
|
||||
throws IOException {
|
||||
List<TimelineEntityDocument> entityDocs =
|
||||
collectionReader.readDocuments(context, filters.getLimit());
|
||||
|
||||
return applyFilters(filters, dataToRetrieve, entityDocs);
|
||||
}
|
||||
|
||||
public Set<String> getEntityTypes(TimelineReaderContext context) {
|
||||
return collectionReader.fetchEntityTypes(context);
|
||||
}
|
||||
|
||||
// for honoring all filters from {@link TimelineEntityFilters}
|
||||
private Set<TimelineEntity> applyFilters(TimelineEntityFilters filters,
|
||||
TimelineDataToRetrieve dataToRetrieve,
|
||||
List<TimelineEntityDocument> entityDocs) throws IOException {
|
||||
Set<TimelineEntity> timelineEntities = new HashSet<>();
|
||||
for (TimelineEntityDocument entityDoc : entityDocs) {
|
||||
final TimelineEntity timelineEntity = entityDoc.fetchTimelineEntity();
|
||||
|
||||
if (DocumentStoreUtils.isFilterNotMatching(filters, timelineEntity)) {
|
||||
continue;
|
||||
}
|
||||
|
||||
TimelineEntity entityToBeReturned = DocumentStoreUtils
|
||||
.createEntityToBeReturned(entityDoc, dataToRetrieve);
|
||||
timelineEntities.add(entityToBeReturned);
|
||||
}
|
||||
return timelineEntities;
|
||||
}
|
||||
}
|
|
@ -0,0 +1,285 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.service.AbstractService;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.*;
|
||||
import org.apache.hadoop.yarn.server.metrics.AppAttemptMetricsConstants;
|
||||
import org.apache.hadoop.yarn.server.metrics.ApplicationMetricsConstants;
|
||||
import org.apache.hadoop.yarn.server.metrics.ContainerMetricsConstants;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.collector.TimelineCollectorContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.storage.TimelineAggregationTrack;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.storage.TimelineWriter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.CollectionType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEntityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity.FlowActivityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowrun.FlowRunDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.TimelineCollectionWriter;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Set;
|
||||
|
||||
/**
|
||||
* This is a generic document store timeline writer for storing the timeline
|
||||
* entity information. Based on the {@link DocumentStoreVendor} that is
|
||||
* configured, the documents are written to that backend.
|
||||
*/
|
||||
public class DocumentStoreTimelineWriterImpl extends AbstractService
|
||||
implements TimelineWriter {
|
||||
|
||||
private static final Logger LOG = LoggerFactory
|
||||
.getLogger(DocumentStoreTimelineWriterImpl.class);
|
||||
private static final String DOC_ID_DELIMITER = "!";
|
||||
|
||||
private DocumentStoreVendor storeType;
|
||||
private TimelineCollectionWriter<TimelineEntityDocument> appCollWriter;
|
||||
private TimelineCollectionWriter<TimelineEntityDocument>
|
||||
entityCollWriter;
|
||||
private TimelineCollectionWriter<FlowActivityDocument> flowActivityCollWriter;
|
||||
private TimelineCollectionWriter<FlowRunDocument> flowRunCollWriter;
|
||||
|
||||
|
||||
public DocumentStoreTimelineWriterImpl() {
|
||||
super(DocumentStoreTimelineWriterImpl.class.getName());
|
||||
}
|
||||
|
||||
@Override
|
||||
public void serviceInit(Configuration conf) throws Exception {
|
||||
storeType = DocumentStoreUtils.getStoreVendor(conf);
|
||||
LOG.info("Initializing Document Store Writer for : " + storeType);
|
||||
super.serviceInit(conf);
|
||||
|
||||
this.appCollWriter = new TimelineCollectionWriter<>(
|
||||
CollectionType.APPLICATION, conf);
|
||||
this.entityCollWriter = new TimelineCollectionWriter<>(
|
||||
CollectionType.ENTITY, conf);
|
||||
this.flowActivityCollWriter = new TimelineCollectionWriter<>(
|
||||
CollectionType.FLOW_ACTIVITY, conf);
|
||||
this.flowRunCollWriter = new TimelineCollectionWriter<>(
|
||||
CollectionType.FLOW_RUN, conf);
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void serviceStart() throws Exception {
|
||||
super.serviceStart();
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void serviceStop() throws Exception {
|
||||
super.serviceStop();
|
||||
appCollWriter.close();
|
||||
entityCollWriter.close();
|
||||
flowActivityCollWriter.close();
|
||||
flowRunCollWriter.close();
|
||||
}
|
||||
|
||||
@Override
|
||||
public TimelineWriteResponse write(TimelineCollectorContext
|
||||
context, TimelineEntities data, UserGroupInformation callerUgi) {
|
||||
LOG.debug("Writing Timeline Entity for appID : {}", context.getAppId());
|
||||
TimelineWriteResponse putStatus = new TimelineWriteResponse();
|
||||
String subApplicationUser = callerUgi.getShortUserName();
|
||||
|
||||
//Avoiding NPE for document id
|
||||
if (DocumentStoreUtils.isNullOrEmpty(context.getFlowName(),
|
||||
context.getAppId(), context.getClusterId(), context.getUserId())) {
|
||||
LOG.warn("Found NULL for one of: flowName={} appId={} " +
|
||||
"userId={} clusterId={} . Not proceeding on writing to store : " +
|
||||
storeType);
|
||||
return putStatus;
|
||||
}
|
||||
|
||||
for (TimelineEntity timelineEntity : data.getEntities()) {
|
||||
// a set can have at most 1 null
|
||||
if(timelineEntity == null) {
|
||||
continue;
|
||||
}
|
||||
|
||||
TimelineEntityDocument entityDocument;
|
||||
//If the entity is application, it will be stored in Application
|
||||
// Collection
|
||||
if (ApplicationEntity.isApplicationEntity(timelineEntity)) {
|
||||
entityDocument = createTimelineEntityDoc(context, subApplicationUser,
|
||||
timelineEntity, true);
|
||||
// if it's an application entity, store metrics for aggregation
|
||||
FlowRunDocument flowRunDoc = createFlowRunDoc(context,
|
||||
timelineEntity.getMetrics());
|
||||
// fetch flow activity if App is created or finished
|
||||
FlowActivityDocument flowActivityDoc = getFlowActivityDoc(context,
|
||||
timelineEntity, flowRunDoc, entityDocument);
|
||||
writeApplicationDoc(entityDocument);
|
||||
writeFlowRunDoc(flowRunDoc);
|
||||
if(flowActivityDoc != null) {
|
||||
storeFlowActivityDoc(flowActivityDoc);
|
||||
}
|
||||
} else {
|
||||
entityDocument = createTimelineEntityDoc(context, subApplicationUser,
|
||||
timelineEntity, false);
|
||||
appendSubAppUserIfExists(context, subApplicationUser);
|
||||
// The entity will be stored in Entity Collection
|
||||
entityDocument.setCreatedTime(fetchEntityCreationTime(timelineEntity));
|
||||
writeEntityDoc(entityDocument);
|
||||
}
|
||||
}
|
||||
return putStatus;
|
||||
}
|
||||
|
||||
@Override
|
||||
public TimelineWriteResponse write(TimelineCollectorContext context,
|
||||
TimelineDomain domain) throws IOException {
|
||||
return null;
|
||||
}
|
||||
|
||||
private void appendSubAppUserIfExists(TimelineCollectorContext context,
|
||||
String subApplicationUser) {
|
||||
String userId = context.getUserId();
|
||||
if (!userId.equals(subApplicationUser) &&
|
||||
!userId.contains(subApplicationUser)) {
|
||||
userId = userId.concat(DOC_ID_DELIMITER).concat(subApplicationUser);
|
||||
context.setUserId(userId);
|
||||
}
|
||||
}
|
||||
|
||||
private TimelineEntityDocument createTimelineEntityDoc(
|
||||
TimelineCollectorContext context, String subApplicationUser,
|
||||
TimelineEntity timelineEntity, boolean isAppEntity) {
|
||||
TimelineEntityDocument entityDocument =
|
||||
new TimelineEntityDocument(timelineEntity);
|
||||
entityDocument.setContext(context);
|
||||
entityDocument.setFlowVersion(context.getFlowVersion());
|
||||
entityDocument.setSubApplicationUser(subApplicationUser);
|
||||
if (isAppEntity) {
|
||||
entityDocument.setId(DocumentStoreUtils.constructTimelineEntityDocId(
|
||||
context, timelineEntity.getType()));
|
||||
} else {
|
||||
entityDocument.setId(DocumentStoreUtils.constructTimelineEntityDocId(
|
||||
context, timelineEntity.getType(), timelineEntity.getId()));
|
||||
}
|
||||
return entityDocument;
|
||||
}
|
||||
|
||||
private FlowRunDocument createFlowRunDoc(TimelineCollectorContext context,
|
||||
Set<TimelineMetric> metrics) {
|
||||
FlowRunDocument flowRunDoc = new FlowRunDocument(context, metrics);
|
||||
flowRunDoc.setFlowVersion(context.getFlowVersion());
|
||||
flowRunDoc.setId(DocumentStoreUtils.constructFlowRunDocId(context));
|
||||
return flowRunDoc;
|
||||
}
|
||||
|
||||
private long fetchEntityCreationTime(TimelineEntity timelineEntity) {
|
||||
TimelineEvent event;
|
||||
switch (TimelineEntityType.valueOf(timelineEntity.getType())) {
|
||||
case YARN_CONTAINER:
|
||||
event = DocumentStoreUtils.fetchEvent(
|
||||
timelineEntity, ContainerMetricsConstants.CREATED_EVENT_TYPE);
|
||||
if (event != null) {
|
||||
return event.getTimestamp();
|
||||
}
|
||||
break;
|
||||
case YARN_APPLICATION_ATTEMPT:
|
||||
event = DocumentStoreUtils.fetchEvent(
|
||||
timelineEntity, AppAttemptMetricsConstants.REGISTERED_EVENT_TYPE);
|
||||
if (event != null) {
|
||||
return event.getTimestamp();
|
||||
}
|
||||
break;
|
||||
default:
|
||||
//NO Op
|
||||
}
|
||||
if (timelineEntity.getCreatedTime() == null) {
|
||||
return 0;
|
||||
}
|
||||
return timelineEntity.getCreatedTime();
|
||||
}
|
||||
|
||||
private FlowActivityDocument getFlowActivityDoc(
|
||||
TimelineCollectorContext context,
|
||||
TimelineEntity timelineEntity, FlowRunDocument flowRunDoc,
|
||||
TimelineEntityDocument entityDocument) {
|
||||
FlowActivityDocument flowActivityDoc = null;
|
||||
// check if the application is created
|
||||
TimelineEvent event = DocumentStoreUtils.fetchEvent(
|
||||
timelineEntity, ApplicationMetricsConstants.CREATED_EVENT_TYPE);
|
||||
if (event != null) {
|
||||
entityDocument.setCreatedTime(event.getTimestamp());
|
||||
flowRunDoc.setMinStartTime(event.getTimestamp());
|
||||
flowActivityDoc = createFlowActivityDoc(context, context.getFlowName(),
|
||||
context.getFlowVersion(), context.getFlowRunId(), event);
|
||||
}
|
||||
|
||||
// if application has finished, store it's finish time
|
||||
event = DocumentStoreUtils.fetchEvent(timelineEntity,
|
||||
ApplicationMetricsConstants.FINISHED_EVENT_TYPE);
|
||||
if (event != null) {
|
||||
flowRunDoc.setMaxEndTime(event.getTimestamp());
|
||||
|
||||
// this check is to handle in case both create and finish event exist
|
||||
// under the single list of events for an TimelineEntity
|
||||
if (flowActivityDoc == null) {
|
||||
flowActivityDoc = createFlowActivityDoc(context, context.getFlowName(),
|
||||
context.getFlowVersion(), context.getFlowRunId(), event);
|
||||
}
|
||||
}
|
||||
return flowActivityDoc;
|
||||
}
|
||||
|
||||
private FlowActivityDocument createFlowActivityDoc(
|
||||
TimelineCollectorContext context, String flowName, String flowVersion,
|
||||
long flowRunId, TimelineEvent event) {
|
||||
FlowActivityDocument flowActivityDoc = new FlowActivityDocument(flowName,
|
||||
flowVersion, flowRunId);
|
||||
flowActivityDoc.setDayTimestamp(DocumentStoreUtils.getTopOfTheDayTimestamp(
|
||||
event.getTimestamp()));
|
||||
flowActivityDoc.setFlowName(flowName);
|
||||
flowActivityDoc.setUser(context.getUserId());
|
||||
flowActivityDoc.setId(DocumentStoreUtils.constructFlowActivityDocId(
|
||||
context, event.getTimestamp()));
|
||||
return flowActivityDoc;
|
||||
}
|
||||
|
||||
private void writeFlowRunDoc(FlowRunDocument flowRunDoc) {
|
||||
flowRunCollWriter.writeDocument(flowRunDoc);
|
||||
}
|
||||
|
||||
private void storeFlowActivityDoc(FlowActivityDocument flowActivityDoc) {
|
||||
flowActivityCollWriter.writeDocument(flowActivityDoc);
|
||||
}
|
||||
|
||||
private void writeEntityDoc(TimelineEntityDocument entityDocument) {
|
||||
entityCollWriter.writeDocument(entityDocument);
|
||||
}
|
||||
|
||||
private void writeApplicationDoc(TimelineEntityDocument entityDocument) {
|
||||
appCollWriter.writeDocument(entityDocument);
|
||||
}
|
||||
|
||||
public TimelineWriteResponse aggregate(TimelineEntity data,
|
||||
TimelineAggregationTrack track) {
|
||||
return null;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void flush() {
|
||||
}
|
||||
}
|
|
@ -0,0 +1,489 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import com.microsoft.azure.documentdb.ConnectionPolicy;
|
||||
import com.microsoft.azure.documentdb.ConsistencyLevel;
|
||||
import com.microsoft.azure.documentdb.DocumentClient;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.ApplicationEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.FlowActivityEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.FlowRunEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntityType;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEvent;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetric;
|
||||
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.collector.TimelineCollectorContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineDataToRetrieve;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineEntityFilters;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.filter.TimelineFilter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.filter.TimelineFilterList;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.filter.TimelinePrefixFilter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.storage.TimelineReader;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.storage.common.TimelineStorageUtils;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEntityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEventSubDoc;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineMetricSubDoc;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity.FlowActivityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowrun.FlowRunDocument;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Collection;
|
||||
import java.util.Collections;
|
||||
import java.util.EnumSet;
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
import java.util.Map;
|
||||
import java.util.NavigableSet;
|
||||
import java.util.Set;
|
||||
import java.util.TreeSet;
|
||||
|
||||
/**
|
||||
* This class consists of all the utils required for reading or writing
|
||||
* documents for a {@link DocumentStoreVendor}.
|
||||
*/
|
||||
public final class DocumentStoreUtils {
|
||||
|
||||
private DocumentStoreUtils(){}
|
||||
|
||||
/** milliseconds in one day. */
|
||||
private static final long MILLIS_ONE_DAY = 86400000L;
|
||||
|
||||
private static final String TIMELINE_STORE_TYPE =
|
||||
YarnConfiguration.TIMELINE_SERVICE_PREFIX + "document-store-type";
|
||||
static final String TIMELINE_SERVICE_COSMOSDB_ENDPOINT =
|
||||
"yarn.timeline-service.document-store.cosmos-db.endpoint";
|
||||
static final String TIMELINE_SERVICE_COSMOSDB_MASTER_KEY =
|
||||
"yarn.timeline-service.document-store.cosmos-db.masterkey";
|
||||
static final String TIMELINE_SERVICE_DOCUMENTSTORE_DATABASE_NAME =
|
||||
"yarn.timeline-service.document-store.db-name";
|
||||
private static final String
|
||||
DEFAULT_TIMELINE_SERVICE_DOCUMENTSTORE_DATABASE_NAME = "timeline_service";
|
||||
|
||||
/**
|
||||
* Checks whether the cosmosdb conf are set properly in yarn-site.xml conf.
|
||||
* @param conf
|
||||
* related to yarn
|
||||
* @throws YarnException if required config properties are missing
|
||||
*/
|
||||
public static void validateCosmosDBConf(Configuration conf)
|
||||
throws YarnException {
|
||||
if (conf == null) {
|
||||
throw new NullPointerException("Configuration cannot be null");
|
||||
}
|
||||
if (isNullOrEmpty(conf.get(TIMELINE_SERVICE_COSMOSDB_ENDPOINT),
|
||||
conf.get(TIMELINE_SERVICE_COSMOSDB_MASTER_KEY))) {
|
||||
throw new YarnException("One or more CosmosDB configuration property is" +
|
||||
" missing in yarn-site.xml");
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Retrieves {@link DocumentStoreVendor} configured.
|
||||
* @param conf
|
||||
* related to yarn
|
||||
* @return Returns the {@link DocumentStoreVendor} that is configured, else
|
||||
* uses {@link DocumentStoreVendor#COSMOS_DB} as default
|
||||
*/
|
||||
public static DocumentStoreVendor getStoreVendor(Configuration conf) {
|
||||
return DocumentStoreVendor.getStoreType(conf.get(TIMELINE_STORE_TYPE,
|
||||
DocumentStoreVendor.COSMOS_DB.name()));
|
||||
}
|
||||
|
||||
/**
|
||||
* Retrieves a {@link TimelineEvent} from {@link TimelineEntity#events}.
|
||||
* @param timelineEntity
|
||||
* from which the set of events are examined.
|
||||
* @param eventType
|
||||
* that has to be checked.
|
||||
* @return {@link TimelineEvent} if found else null
|
||||
*/
|
||||
public static TimelineEvent fetchEvent(TimelineEntity timelineEntity,
|
||||
String eventType) {
|
||||
for (TimelineEvent event : timelineEntity.getEvents()) {
|
||||
if (event.getId().equals(eventType)) {
|
||||
return event;
|
||||
}
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
/**
|
||||
* Checks if the string is null or empty.
|
||||
* @param values
|
||||
* array of string to be checked
|
||||
* @return false if any of the string is null or empty else true
|
||||
*/
|
||||
public static boolean isNullOrEmpty(String...values) {
|
||||
for (String value : values) {
|
||||
if (value == null || value.isEmpty()) {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates CosmosDB Document Client.
|
||||
* @param conf
|
||||
* to retrieve cosmos db endpoint and key
|
||||
* @return async document client for CosmosDB
|
||||
*/
|
||||
public static DocumentClient createCosmosDBClient(Configuration conf){
|
||||
return new DocumentClient(DocumentStoreUtils.getCosmosDBEndpoint(conf),
|
||||
DocumentStoreUtils.getCosmosDBMasterKey(conf),
|
||||
ConnectionPolicy.GetDefault(), ConsistencyLevel.Session);
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns the timestamp of the day's start (which is midnight 00:00:00 AM)
|
||||
* for a given input timestamp.
|
||||
*
|
||||
* @param timeStamp Timestamp.
|
||||
* @return timestamp of that day's beginning (midnight)
|
||||
*/
|
||||
public static long getTopOfTheDayTimestamp(long timeStamp) {
|
||||
return timeStamp - (timeStamp % MILLIS_ONE_DAY);
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates a composite key for storing {@link TimelineEntityDocument}.
|
||||
* @param collectorContext
|
||||
* of the timeline writer
|
||||
* @param type
|
||||
* of the entity
|
||||
* @return composite key delimited with !
|
||||
*/
|
||||
public static String constructTimelineEntityDocId(TimelineCollectorContext
|
||||
collectorContext, String type) {
|
||||
return String.format("%s!%s!%s!%d!%s!%s",
|
||||
collectorContext.getClusterId(), collectorContext.getUserId(),
|
||||
collectorContext.getFlowName(), collectorContext.getFlowRunId(),
|
||||
collectorContext.getAppId(), type);
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates a composite key for storing {@link TimelineEntityDocument}.
|
||||
* @param collectorContext
|
||||
* of the timeline writer
|
||||
* @param type
|
||||
* of the entity
|
||||
* @param id
|
||||
* of the entity
|
||||
* @return composite key delimited with !
|
||||
*/
|
||||
public static String constructTimelineEntityDocId(TimelineCollectorContext
|
||||
collectorContext, String type, String id) {
|
||||
return String.format("%s!%s!%s!%d!%s!%s!%s",
|
||||
collectorContext.getClusterId(), collectorContext.getUserId(),
|
||||
collectorContext.getFlowName(), collectorContext.getFlowRunId(),
|
||||
collectorContext.getAppId(), type, id);
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates a composite key for storing {@link FlowRunDocument}.
|
||||
* @param collectorContext
|
||||
* of the timeline writer
|
||||
* @return composite key delimited with !
|
||||
*/
|
||||
public static String constructFlowRunDocId(TimelineCollectorContext
|
||||
collectorContext) {
|
||||
return String.format("%s!%s!%s!%s", collectorContext.getClusterId(),
|
||||
collectorContext.getUserId(), collectorContext.getFlowName(),
|
||||
collectorContext.getFlowRunId());
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates a composite key for storing {@link FlowActivityDocument}.
|
||||
* @param collectorContext
|
||||
* of the timeline writer
|
||||
* @param eventTimestamp
|
||||
* of the timeline entity
|
||||
* @return composite key delimited with !
|
||||
*/
|
||||
public static String constructFlowActivityDocId(TimelineCollectorContext
|
||||
collectorContext, long eventTimestamp) {
|
||||
return String.format("%s!%s!%s!%s", collectorContext.getClusterId(),
|
||||
getTopOfTheDayTimestamp(eventTimestamp),
|
||||
collectorContext.getUserId(), collectorContext.getFlowName());
|
||||
}
|
||||
|
||||
private static String getCosmosDBEndpoint(Configuration conf) {
|
||||
return conf.get(TIMELINE_SERVICE_COSMOSDB_ENDPOINT);
|
||||
}
|
||||
|
||||
private static String getCosmosDBMasterKey(Configuration conf) {
|
||||
return conf.get(TIMELINE_SERVICE_COSMOSDB_MASTER_KEY);
|
||||
}
|
||||
|
||||
public static String getCosmosDBDatabaseName(Configuration conf) {
|
||||
return conf.get(TIMELINE_SERVICE_DOCUMENTSTORE_DATABASE_NAME,
|
||||
getDefaultTimelineServiceDBName(conf));
|
||||
}
|
||||
|
||||
private static String getDefaultTimelineServiceDBName(
|
||||
Configuration conf) {
|
||||
return getClusterId(conf) + "_" +
|
||||
DEFAULT_TIMELINE_SERVICE_DOCUMENTSTORE_DATABASE_NAME;
|
||||
}
|
||||
|
||||
private static String getClusterId(Configuration conf) {
|
||||
return conf.get(YarnConfiguration.RM_CLUSTER_ID,
|
||||
YarnConfiguration.DEFAULT_RM_CLUSTER_ID);
|
||||
}
|
||||
|
||||
private static boolean isTimeInRange(long time, long timeBegin,
|
||||
long timeEnd) {
|
||||
return (time >= timeBegin) && (time <= timeEnd);
|
||||
}
|
||||
|
||||
/**
|
||||
* Checks if the {@link TimelineEntityFilters} are not matching for a given
|
||||
* {@link TimelineEntity}.
|
||||
* @param filters
|
||||
* that has to be checked for an entity
|
||||
* @param timelineEntity
|
||||
* for which the filters would be applied
|
||||
* @return true if any one of the filter is not matching else false
|
||||
* @throws IOException if an unsupported filter is being matched.
|
||||
*/
|
||||
static boolean isFilterNotMatching(TimelineEntityFilters filters,
|
||||
TimelineEntity timelineEntity) throws IOException {
|
||||
if (timelineEntity.getCreatedTime() != null && !isTimeInRange(timelineEntity
|
||||
.getCreatedTime(), filters.getCreatedTimeBegin(),
|
||||
filters.getCreatedTimeEnd())) {
|
||||
return true;
|
||||
}
|
||||
|
||||
if (filters.getRelatesTo() != null &&
|
||||
!filters.getRelatesTo().getFilterList().isEmpty() &&
|
||||
!TimelineStorageUtils.matchRelatesTo(timelineEntity,
|
||||
filters.getRelatesTo())) {
|
||||
return true;
|
||||
}
|
||||
|
||||
if (filters.getIsRelatedTo() != null &&
|
||||
!filters.getIsRelatedTo().getFilterList().isEmpty() &&
|
||||
!TimelineStorageUtils.matchIsRelatedTo(timelineEntity,
|
||||
filters.getIsRelatedTo())) {
|
||||
return true;
|
||||
}
|
||||
|
||||
if (filters.getInfoFilters() != null &&
|
||||
!filters.getInfoFilters().getFilterList().isEmpty() &&
|
||||
!TimelineStorageUtils.matchInfoFilters(timelineEntity,
|
||||
filters.getInfoFilters())) {
|
||||
return true;
|
||||
}
|
||||
|
||||
if (filters.getConfigFilters() != null &&
|
||||
!filters.getConfigFilters().getFilterList().isEmpty() &&
|
||||
!TimelineStorageUtils.matchConfigFilters(timelineEntity,
|
||||
filters.getConfigFilters())) {
|
||||
return true;
|
||||
}
|
||||
|
||||
if (filters.getMetricFilters() != null &&
|
||||
!filters.getMetricFilters().getFilterList().isEmpty() &&
|
||||
!TimelineStorageUtils.matchMetricFilters(timelineEntity,
|
||||
filters.getMetricFilters())) {
|
||||
return true;
|
||||
}
|
||||
|
||||
return filters.getEventFilters() != null &&
|
||||
!filters.getEventFilters().getFilterList().isEmpty() &&
|
||||
!TimelineStorageUtils.matchEventFilters(timelineEntity,
|
||||
filters.getEventFilters());
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates the final entity to be returned as the result.
|
||||
* @param timelineEntityDocument
|
||||
* which has all the information for the entity
|
||||
* @param dataToRetrieve
|
||||
* specifies filters and fields to retrieve
|
||||
* @return {@link TimelineEntity} as the result
|
||||
*/
|
||||
public static TimelineEntity createEntityToBeReturned(
|
||||
TimelineEntityDocument timelineEntityDocument,
|
||||
TimelineDataToRetrieve dataToRetrieve) {
|
||||
TimelineEntity entityToBeReturned = createTimelineEntity(
|
||||
timelineEntityDocument.getType(),
|
||||
timelineEntityDocument.fetchTimelineEntity());
|
||||
|
||||
entityToBeReturned.setIdentifier(new TimelineEntity.Identifier(
|
||||
timelineEntityDocument.getType(), timelineEntityDocument.getId()));
|
||||
entityToBeReturned.setCreatedTime(
|
||||
timelineEntityDocument.getCreatedTime());
|
||||
entityToBeReturned.setInfo(timelineEntityDocument.getInfo());
|
||||
|
||||
if (dataToRetrieve.getFieldsToRetrieve() != null) {
|
||||
fillFields(entityToBeReturned, timelineEntityDocument,
|
||||
dataToRetrieve);
|
||||
}
|
||||
return entityToBeReturned;
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates the final entity to be returned as the result.
|
||||
* @param timelineEntityDocument
|
||||
* which has all the information for the entity
|
||||
* @param confsToRetrieve
|
||||
* specifies config filters to be applied
|
||||
* @param metricsToRetrieve
|
||||
* specifies metric filters to be applied
|
||||
*
|
||||
* @return {@link TimelineEntity} as the result
|
||||
*/
|
||||
public static TimelineEntity createEntityToBeReturned(
|
||||
TimelineEntityDocument timelineEntityDocument,
|
||||
TimelineFilterList confsToRetrieve,
|
||||
TimelineFilterList metricsToRetrieve) {
|
||||
TimelineEntity timelineEntity = timelineEntityDocument
|
||||
.fetchTimelineEntity();
|
||||
if (confsToRetrieve != null) {
|
||||
timelineEntity.setConfigs(DocumentStoreUtils.applyConfigFilter(
|
||||
confsToRetrieve, timelineEntity.getConfigs()));
|
||||
}
|
||||
if (metricsToRetrieve != null) {
|
||||
timelineEntity.setMetrics(DocumentStoreUtils.transformMetrics(
|
||||
metricsToRetrieve, timelineEntityDocument.getMetrics()));
|
||||
}
|
||||
return timelineEntity;
|
||||
}
|
||||
|
||||
private static TimelineEntity createTimelineEntity(String type,
|
||||
TimelineEntity timelineEntity) {
|
||||
switch (TimelineEntityType.valueOf(type)) {
|
||||
case YARN_APPLICATION:
|
||||
return new ApplicationEntity();
|
||||
case YARN_FLOW_RUN:
|
||||
return new FlowRunEntity();
|
||||
case YARN_FLOW_ACTIVITY:
|
||||
FlowActivityEntity flowActivityEntity =
|
||||
(FlowActivityEntity) timelineEntity;
|
||||
FlowActivityEntity newFlowActivity = new FlowActivityEntity();
|
||||
newFlowActivity.addFlowRuns(flowActivityEntity.getFlowRuns());
|
||||
return newFlowActivity;
|
||||
default:
|
||||
return new TimelineEntity();
|
||||
}
|
||||
}
|
||||
|
||||
// fetch required fields for final entity to be returned
|
||||
private static void fillFields(TimelineEntity finalEntity,
|
||||
TimelineEntityDocument entityDoc,
|
||||
TimelineDataToRetrieve dataToRetrieve) {
|
||||
EnumSet<TimelineReader.Field> fieldsToRetrieve =
|
||||
dataToRetrieve.getFieldsToRetrieve();
|
||||
if (fieldsToRetrieve.contains(TimelineReader.Field.ALL)) {
|
||||
fieldsToRetrieve = EnumSet.allOf(TimelineReader.Field.class);
|
||||
}
|
||||
for (TimelineReader.Field field : fieldsToRetrieve) {
|
||||
switch(field) {
|
||||
case CONFIGS:
|
||||
finalEntity.setConfigs(applyConfigFilter(dataToRetrieve
|
||||
.getConfsToRetrieve(), entityDoc.getConfigs()));
|
||||
break;
|
||||
case METRICS:
|
||||
finalEntity.setMetrics(transformMetrics(dataToRetrieve
|
||||
.getMetricsToRetrieve(), entityDoc.getMetrics()));
|
||||
break;
|
||||
case INFO:
|
||||
finalEntity.setInfo(entityDoc.getInfo());
|
||||
break;
|
||||
case IS_RELATED_TO:
|
||||
finalEntity.setIsRelatedToEntities(entityDoc.getIsRelatedToEntities());
|
||||
break;
|
||||
case RELATES_TO:
|
||||
finalEntity.setIsRelatedToEntities(entityDoc.getIsRelatedToEntities());
|
||||
break;
|
||||
case EVENTS:
|
||||
finalEntity.setEvents(transformEvents(entityDoc.getEvents().values()));
|
||||
break;
|
||||
default:
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/* Transforms Collection<Set<TimelineEventSubDoc>> to
|
||||
NavigableSet<TimelineEvent> */
|
||||
private static NavigableSet<TimelineEvent> transformEvents(
|
||||
Collection<Set<TimelineEventSubDoc>> eventSetColl) {
|
||||
NavigableSet<TimelineEvent> timelineEvents = new TreeSet<>();
|
||||
for (Set<TimelineEventSubDoc> eventSubDocs : eventSetColl) {
|
||||
for (TimelineEventSubDoc eventSubDoc : eventSubDocs) {
|
||||
timelineEvents.add(eventSubDoc.fetchTimelineEvent());
|
||||
}
|
||||
}
|
||||
return timelineEvents;
|
||||
}
|
||||
|
||||
public static Set<TimelineMetric> transformMetrics(
|
||||
TimelineFilterList metricsToRetrieve,
|
||||
Map<String, Set<TimelineMetricSubDoc>> metrics) {
|
||||
if (metricsToRetrieve == null ||
|
||||
hasDataToBeRetrieve(metricsToRetrieve, metrics.keySet())) {
|
||||
Set<TimelineMetric> metricSet = new HashSet<>();
|
||||
for(Set<TimelineMetricSubDoc> metricSubDocs : metrics.values()) {
|
||||
for(TimelineMetricSubDoc metricSubDoc : metricSubDocs) {
|
||||
metricSet.add(metricSubDoc.fetchTimelineMetric());
|
||||
}
|
||||
}
|
||||
return metricSet;
|
||||
}
|
||||
return new HashSet<>();
|
||||
}
|
||||
|
||||
public static Map<String, String> applyConfigFilter(
|
||||
TimelineFilterList configsToRetrieve, Map<String, String> configs) {
|
||||
if (configsToRetrieve == null ||
|
||||
hasDataToBeRetrieve(configsToRetrieve, configs.keySet())) {
|
||||
return configs;
|
||||
}
|
||||
return new HashMap<>();
|
||||
}
|
||||
|
||||
private static boolean hasDataToBeRetrieve(
|
||||
TimelineFilterList timelineFilters, Set<String> dataSet) {
|
||||
Set<String> dataToBeRetrieved = new HashSet<>();
|
||||
TimelinePrefixFilter timelinePrefixFilter;
|
||||
for (TimelineFilter timelineFilter : timelineFilters.getFilterList()) {
|
||||
timelinePrefixFilter = (TimelinePrefixFilter) timelineFilter;
|
||||
dataToBeRetrieved.add(timelinePrefixFilter.getPrefix());
|
||||
}
|
||||
switch (timelineFilters.getOperator()) {
|
||||
case OR:
|
||||
if (dataToBeRetrieved.size() == 0 ||
|
||||
!Collections.disjoint(dataSet, dataToBeRetrieved)) {
|
||||
return true;
|
||||
}
|
||||
case AND:
|
||||
if (dataToBeRetrieved.size() == 0 ||
|
||||
dataSet.containsAll(dataToBeRetrieved)) {
|
||||
return true;
|
||||
}
|
||||
default:
|
||||
return false;
|
||||
}
|
||||
}
|
||||
}
|
|
@ -0,0 +1,44 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection;
|
||||
|
||||
/**
|
||||
* Specifies the different collections that are currently used for storing
|
||||
* documents.
|
||||
*/
|
||||
public enum CollectionType {
|
||||
ENTITY("EntityCollection"),
|
||||
APPLICATION("AppCollection"),
|
||||
FLOW_RUN("FlowRunCollection"),
|
||||
FLOW_ACTIVITY("FlowActivityCollection");
|
||||
|
||||
private final String collectionName;
|
||||
|
||||
CollectionType(String collectionName) {
|
||||
this.collectionName = collectionName;
|
||||
}
|
||||
|
||||
public boolean equals(String otherCollectionName) {
|
||||
return this.collectionName.equals(otherCollectionName);
|
||||
}
|
||||
|
||||
public String getCollectionName() {
|
||||
return collectionName;
|
||||
}
|
||||
}
|
|
@ -0,0 +1,39 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document;
|
||||
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
/**
|
||||
* Indicates that the document that was requested is not found from the
|
||||
* Document Store. This is a generic exception that will be thrown for all
|
||||
* the {@link DocumentStoreVendor} if there is no document while reading.
|
||||
*/
|
||||
public class NoDocumentFoundException extends IOException {
|
||||
|
||||
/**
|
||||
* Constructs exception with the specified detail message.
|
||||
* @param message detailed message.
|
||||
*/
|
||||
public NoDocumentFoundException(String message) {
|
||||
super(message);
|
||||
}
|
||||
}
|
|
@ -0,0 +1,37 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document;
|
||||
|
||||
|
||||
/**
|
||||
* This is an interface for all the Timeline Documents. Any new document that
|
||||
* has to be persisted in the document store should implement this.
|
||||
*/
|
||||
public interface TimelineDocument<Document> {
|
||||
|
||||
String getId();
|
||||
|
||||
String getType();
|
||||
|
||||
long getCreatedTime();
|
||||
|
||||
void setCreatedTime(long time);
|
||||
|
||||
void merge(Document timelineDocument);
|
||||
}
|
|
@ -0,0 +1,242 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity;
|
||||
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEvent;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetric;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.TimelineContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineReaderUtils;
|
||||
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
|
||||
/**
|
||||
* This is a generic class which contains all the meta information of some
|
||||
* conceptual entity and its related events. The timeline entity can be an
|
||||
* application, an attempt, a container or whatever the user-defined object.
|
||||
*/
|
||||
public class TimelineEntityDocument implements
|
||||
TimelineDocument<TimelineEntityDocument> {
|
||||
|
||||
private final TimelineEntity timelineEntity;
|
||||
private TimelineContext context;
|
||||
private String flowVersion;
|
||||
private String subApplicationUser;
|
||||
private final Map<String, Set<TimelineMetricSubDoc>>
|
||||
metrics = new HashMap<>();
|
||||
private final Map<String, Set<TimelineEventSubDoc>>
|
||||
events = new HashMap<>();
|
||||
|
||||
public TimelineEntityDocument() {
|
||||
timelineEntity = new TimelineEntity();
|
||||
}
|
||||
|
||||
public TimelineEntityDocument(TimelineEntity timelineEntity) {
|
||||
this.timelineEntity = timelineEntity;
|
||||
transformEvents(timelineEntity.getEvents());
|
||||
timelineMetrics(timelineEntity.getMetrics());
|
||||
}
|
||||
|
||||
// transforms TimelineMetric to TimelineMetricSubDoc
|
||||
private void timelineMetrics(Set<TimelineMetric> timelineMetrics) {
|
||||
for (TimelineMetric timelineMetric : timelineMetrics) {
|
||||
if (this.metrics.containsKey(timelineMetric.getId())) {
|
||||
this.metrics.get(timelineMetric.getId()).add(
|
||||
new TimelineMetricSubDoc(timelineMetric));
|
||||
} else {
|
||||
Set<TimelineMetricSubDoc> metricSet = new HashSet<>();
|
||||
metricSet.add(new TimelineMetricSubDoc(timelineMetric));
|
||||
this.metrics.put(timelineMetric.getId(), metricSet);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// transforms TimelineEvent to TimelineEventSubDoc
|
||||
private void transformEvents(Set<TimelineEvent> timelineEvents) {
|
||||
for (TimelineEvent timelineEvent : timelineEvents) {
|
||||
if (this.events.containsKey(timelineEvent.getId())) {
|
||||
this.events.get(timelineEvent.getId())
|
||||
.add(new TimelineEventSubDoc(timelineEvent));
|
||||
} else {
|
||||
Set<TimelineEventSubDoc> eventSet = new HashSet<>();
|
||||
eventSet.add(new TimelineEventSubDoc(timelineEvent));
|
||||
this.events.put(timelineEvent.getId(), eventSet);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Merge the TimelineEntityDocument that is passed with the current
|
||||
* document for upsert.
|
||||
*
|
||||
* @param newTimelineDocument
|
||||
* that has to be merged
|
||||
*/
|
||||
@Override
|
||||
public void merge(TimelineEntityDocument newTimelineDocument) {
|
||||
if(newTimelineDocument.getCreatedTime() > 0) {
|
||||
timelineEntity.setCreatedTime(newTimelineDocument.getCreatedTime());
|
||||
}
|
||||
setMetrics(newTimelineDocument.getMetrics());
|
||||
setEvents(newTimelineDocument.getEvents());
|
||||
timelineEntity.getInfo().putAll(newTimelineDocument.getInfo());
|
||||
timelineEntity.getConfigs().putAll(newTimelineDocument.getConfigs());
|
||||
timelineEntity.getIsRelatedToEntities().putAll(newTimelineDocument
|
||||
.getIsRelatedToEntities());
|
||||
timelineEntity.getRelatesToEntities().putAll(newTimelineDocument
|
||||
.getRelatesToEntities());
|
||||
}
|
||||
|
||||
@Override
|
||||
public String getId() {
|
||||
return timelineEntity.getId();
|
||||
}
|
||||
|
||||
public void setId(String key) {
|
||||
timelineEntity.setId(key);
|
||||
}
|
||||
|
||||
public String getType() {
|
||||
return timelineEntity.getType();
|
||||
}
|
||||
|
||||
public void setType(String type) {
|
||||
timelineEntity.setType(type);
|
||||
}
|
||||
|
||||
public Map<String, Object> getInfo() {
|
||||
timelineEntity.getInfo().put(TimelineReaderUtils.FROMID_KEY, getId());
|
||||
return timelineEntity.getInfo();
|
||||
}
|
||||
|
||||
public void setInfo(Map<String, Object> info) {
|
||||
timelineEntity.setInfo(info);
|
||||
}
|
||||
|
||||
public Map<String, Set<TimelineMetricSubDoc>> getMetrics() {
|
||||
return metrics;
|
||||
}
|
||||
|
||||
public void setMetrics(Map<String, Set<TimelineMetricSubDoc>> metrics) {
|
||||
for (String metricId : metrics.keySet()) {
|
||||
for(TimelineMetricSubDoc metricSubDoc : metrics.get(metricId)) {
|
||||
timelineEntity.addMetric(metricSubDoc.fetchTimelineMetric());
|
||||
}
|
||||
if (this.metrics.containsKey(metricId)) {
|
||||
this.metrics.get(metricId).addAll(metrics.get(metricId));
|
||||
} else {
|
||||
this.metrics.put(metricId, new HashSet<>(metrics.get(metricId)));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
public Map<String, Set<TimelineEventSubDoc>> getEvents() {
|
||||
return events;
|
||||
}
|
||||
|
||||
public void setEvents(Map<String, Set<TimelineEventSubDoc>> events) {
|
||||
for (String eventId : events.keySet()) {
|
||||
for(TimelineEventSubDoc eventSubDoc: events.get(eventId)) {
|
||||
timelineEntity.addEvent(eventSubDoc.fetchTimelineEvent());
|
||||
}
|
||||
if (this.events.containsKey(eventId)) {
|
||||
this.events.get(eventId).addAll(events.get(eventId));
|
||||
} else {
|
||||
this.events.put(eventId, new HashSet<>(events.get(eventId)));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
public Map<String, String> getConfigs() {
|
||||
return timelineEntity.getConfigs();
|
||||
}
|
||||
|
||||
public void setConfigs(Map<String, String> configs) {
|
||||
timelineEntity.setConfigs(configs);
|
||||
}
|
||||
|
||||
public Map<String, Set<String>> getIsRelatedToEntities() {
|
||||
return timelineEntity.getIsRelatedToEntities();
|
||||
}
|
||||
|
||||
public void setIsRelatedToEntities(Map<String, Set<String>>
|
||||
isRelatedToEntities) {
|
||||
timelineEntity.setIsRelatedToEntities(isRelatedToEntities);
|
||||
}
|
||||
|
||||
public Map<String, Set<String>> getRelatesToEntities() {
|
||||
return timelineEntity.getRelatesToEntities();
|
||||
}
|
||||
|
||||
public void setRelatesToEntities(Map<String, Set<String>> relatesToEntities) {
|
||||
timelineEntity.setRelatesToEntities(relatesToEntities);
|
||||
}
|
||||
|
||||
public String getFlowVersion() {
|
||||
return flowVersion;
|
||||
}
|
||||
|
||||
|
||||
public void setFlowVersion(String flowVersion) {
|
||||
this.flowVersion = flowVersion;
|
||||
}
|
||||
|
||||
public void setIdentifier(TimelineEntity.Identifier identifier) {
|
||||
timelineEntity.setIdentifier(identifier);
|
||||
}
|
||||
|
||||
public void setIdPrefix(long idPrefix) {
|
||||
timelineEntity.setIdPrefix(idPrefix);
|
||||
}
|
||||
|
||||
public String getSubApplicationUser() {
|
||||
return subApplicationUser;
|
||||
}
|
||||
|
||||
public void setSubApplicationUser(String subApplicationUser) {
|
||||
this.subApplicationUser = subApplicationUser;
|
||||
}
|
||||
|
||||
public long getCreatedTime() {
|
||||
if (timelineEntity.getCreatedTime() == null) {
|
||||
return 0;
|
||||
}
|
||||
return timelineEntity.getCreatedTime();
|
||||
}
|
||||
|
||||
public void setCreatedTime(long createdTime) {
|
||||
timelineEntity.setCreatedTime(createdTime);
|
||||
}
|
||||
|
||||
public TimelineContext getContext() {
|
||||
return context;
|
||||
}
|
||||
|
||||
public void setContext(TimelineContext context) {
|
||||
this.context = context;
|
||||
}
|
||||
|
||||
public TimelineEntity fetchTimelineEntity() {
|
||||
return timelineEntity;
|
||||
}
|
||||
}
|
|
@ -0,0 +1,96 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity;
|
||||
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEvent;
|
||||
import org.apache.hadoop.yarn.util.TimelineServiceHelper;
|
||||
|
||||
import java.util.Map;
|
||||
|
||||
/**
|
||||
* This class represents a Sub Document for {@link TimelineEvent}
|
||||
* when creating a new {@link TimelineEntityDocument}.
|
||||
*/
|
||||
public class TimelineEventSubDoc {
|
||||
|
||||
private final TimelineEvent timelineEvent;
|
||||
private boolean valid;
|
||||
|
||||
public TimelineEventSubDoc() {
|
||||
timelineEvent = new TimelineEvent();
|
||||
}
|
||||
|
||||
public TimelineEventSubDoc(TimelineEvent timelineEvent) {
|
||||
this.timelineEvent = timelineEvent;
|
||||
}
|
||||
|
||||
public String getId() {
|
||||
return timelineEvent.getId();
|
||||
}
|
||||
|
||||
public void setId(String eventId) {
|
||||
timelineEvent.setId(eventId);
|
||||
}
|
||||
|
||||
public boolean isValid() {
|
||||
return timelineEvent.isValid();
|
||||
}
|
||||
|
||||
public void setValid(boolean valid) {
|
||||
this.valid = valid;
|
||||
}
|
||||
|
||||
public long getTimestamp() {
|
||||
return timelineEvent.getTimestamp();
|
||||
}
|
||||
|
||||
public void setTimestamp(long ts) {
|
||||
timelineEvent.setTimestamp(ts);
|
||||
}
|
||||
|
||||
public Map<String, Object> getInfo() {
|
||||
return timelineEvent.getInfo();
|
||||
}
|
||||
|
||||
public void setInfo(Map<String, Object> info) {
|
||||
timelineEvent.setInfo(TimelineServiceHelper.mapCastToHashMap(info));
|
||||
}
|
||||
|
||||
@Override
|
||||
public int hashCode() {
|
||||
return 31 * timelineEvent.getId().hashCode();
|
||||
}
|
||||
|
||||
// Only check if id is equal
|
||||
@Override
|
||||
public boolean equals(Object obj) {
|
||||
if (this == obj) {
|
||||
return true;
|
||||
}
|
||||
if (!(obj instanceof TimelineEventSubDoc)) {
|
||||
return false;
|
||||
}
|
||||
TimelineEventSubDoc otherTimelineEvent = (TimelineEventSubDoc) obj;
|
||||
return this.timelineEvent.getId().equals(otherTimelineEvent.getId());
|
||||
}
|
||||
|
||||
public TimelineEvent fetchTimelineEvent() {
|
||||
return timelineEvent;
|
||||
}
|
||||
}
|
|
@ -0,0 +1,167 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity;
|
||||
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetric;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetricOperation;
|
||||
|
||||
import java.util.Map;
|
||||
import java.util.TreeMap;
|
||||
|
||||
/**
|
||||
* This class represents a Sub Document for {@link TimelineMetric} that will be
|
||||
* used when creating new {@link TimelineEntityDocument}.
|
||||
*/
|
||||
public class TimelineMetricSubDoc {
|
||||
|
||||
private final TimelineMetric timelineMetric;
|
||||
private boolean valid;
|
||||
private long singleDataTimestamp;
|
||||
private Number singleDataValue = 0;
|
||||
|
||||
public TimelineMetricSubDoc() {
|
||||
this.timelineMetric = new TimelineMetric();
|
||||
}
|
||||
|
||||
public TimelineMetricSubDoc(TimelineMetric timelineMetric) {
|
||||
this.timelineMetric = timelineMetric;
|
||||
this.valid = timelineMetric.isValid();
|
||||
if (timelineMetric.getType() == TimelineMetric.Type.SINGLE_VALUE &&
|
||||
timelineMetric.getValues().size() > 0) {
|
||||
this.singleDataTimestamp = timelineMetric.getSingleDataTimestamp();
|
||||
this.singleDataValue = timelineMetric.getSingleDataValue();
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Get the real time aggregation operation of this metric.
|
||||
*
|
||||
* @return Real time aggregation operation
|
||||
*/
|
||||
public TimelineMetricOperation getRealtimeAggregationOp() {
|
||||
return timelineMetric.getRealtimeAggregationOp();
|
||||
}
|
||||
|
||||
/**
|
||||
* Set the real time aggregation operation of this metric.
|
||||
*
|
||||
* @param op A timeline metric operation that the metric should perform on
|
||||
* real time aggregations
|
||||
*/
|
||||
public void setRealtimeAggregationOp(
|
||||
final TimelineMetricOperation op) {
|
||||
timelineMetric.setRealtimeAggregationOp(op);
|
||||
}
|
||||
|
||||
public String getId() {
|
||||
return timelineMetric.getId();
|
||||
}
|
||||
|
||||
public void setId(String metricId) {
|
||||
timelineMetric.setId(metricId);
|
||||
}
|
||||
|
||||
public void setSingleDataTimestamp(long singleDataTimestamp) {
|
||||
this.singleDataTimestamp = singleDataTimestamp;
|
||||
}
|
||||
|
||||
/**
|
||||
* Get single data timestamp of the metric.
|
||||
*
|
||||
* @return the single data timestamp
|
||||
*/
|
||||
public long getSingleDataTimestamp() {
|
||||
if (timelineMetric.getType() == TimelineMetric.Type.SINGLE_VALUE) {
|
||||
return singleDataTimestamp;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* Get single data value of the metric.
|
||||
*
|
||||
* @return the single data value
|
||||
*/
|
||||
public Number getSingleDataValue() {
|
||||
if (timelineMetric.getType() == TimelineMetric.Type.SINGLE_VALUE) {
|
||||
return singleDataValue;
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
public void setSingleDataValue(Number singleDataValue) {
|
||||
this.singleDataValue = singleDataValue;
|
||||
}
|
||||
|
||||
public Map<Long, Number> getValues() {
|
||||
return timelineMetric.getValues();
|
||||
}
|
||||
|
||||
public void setValues(Map<Long, Number> vals) {
|
||||
timelineMetric.setValues(vals);
|
||||
}
|
||||
|
||||
// required by JAXB
|
||||
public TreeMap<Long, Number> getValuesJAXB() {
|
||||
return timelineMetric.getValuesJAXB();
|
||||
}
|
||||
|
||||
public TimelineMetric.Type getType() {
|
||||
return timelineMetric.getType();
|
||||
}
|
||||
|
||||
public void setType(TimelineMetric.Type metricType) {
|
||||
timelineMetric.setType(metricType);
|
||||
}
|
||||
|
||||
public void setValid(boolean valid) {
|
||||
this.valid = valid;
|
||||
}
|
||||
|
||||
public boolean isValid() {
|
||||
return (timelineMetric.getId() != null);
|
||||
}
|
||||
|
||||
@Override
|
||||
public int hashCode() {
|
||||
int result = timelineMetric.getId().hashCode();
|
||||
result = 31 * result + timelineMetric.getType().hashCode();
|
||||
return result;
|
||||
}
|
||||
|
||||
// Only check if timestamp and id are equal
|
||||
@Override
|
||||
public boolean equals(Object obj) {
|
||||
if (this == obj) {
|
||||
return true;
|
||||
}
|
||||
if (!(obj instanceof TimelineMetricSubDoc)) {
|
||||
return false;
|
||||
}
|
||||
TimelineMetricSubDoc otherTimelineMetric = (TimelineMetricSubDoc) obj;
|
||||
if (!this.timelineMetric.getId().equals(otherTimelineMetric.getId())) {
|
||||
return false;
|
||||
}
|
||||
return this.timelineMetric.getType() == otherTimelineMetric.getType();
|
||||
}
|
||||
|
||||
public TimelineMetric fetchTimelineMetric() {
|
||||
return timelineMetric;
|
||||
}
|
||||
}
|
|
@ -0,0 +1,30 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.
|
||||
* documentstore.collection.document.entity contains
|
||||
* TimelineEntityDocument that will be common to different TimelineEntity i.e
|
||||
* Application, App Attempt, Container etc.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,131 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity;
|
||||
|
||||
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.FlowActivityEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntityType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
|
||||
import java.util.HashSet;
|
||||
import java.util.Set;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
|
||||
/**
|
||||
* This doc represents the {@link FlowActivityEntity} which is used for
|
||||
* showing all the flow runs with limited information.
|
||||
*/
|
||||
public class FlowActivityDocument implements
|
||||
TimelineDocument<FlowActivityDocument> {
|
||||
|
||||
private String id;
|
||||
private final String type = TimelineEntityType.YARN_FLOW_ACTIVITY.toString();
|
||||
private Set<FlowActivitySubDoc> flowActivities = new HashSet<>();
|
||||
private long dayTimestamp;
|
||||
private String user;
|
||||
private String flowName;
|
||||
|
||||
public FlowActivityDocument() {
|
||||
}
|
||||
|
||||
public FlowActivityDocument(String flowName, String flowVersion,
|
||||
long flowRunId) {
|
||||
flowActivities.add(new FlowActivitySubDoc(flowName,
|
||||
flowVersion, flowRunId));
|
||||
}
|
||||
|
||||
/**
|
||||
* Merge the {@link FlowActivityDocument} that is passed with the current
|
||||
* document for upsert.
|
||||
*
|
||||
* @param flowActivityDocument
|
||||
* that has to be merged
|
||||
*/
|
||||
@Override
|
||||
public void merge(FlowActivityDocument flowActivityDocument) {
|
||||
if (flowActivityDocument.getDayTimestamp() > 0) {
|
||||
this.dayTimestamp = flowActivityDocument.getDayTimestamp();
|
||||
}
|
||||
this.flowName = flowActivityDocument.getFlowName();
|
||||
this.user = flowActivityDocument.getUser();
|
||||
this.id = flowActivityDocument.getId();
|
||||
this.flowActivities.addAll(flowActivityDocument.getFlowActivities());
|
||||
}
|
||||
|
||||
@Override
|
||||
public String getId() {
|
||||
return id;
|
||||
}
|
||||
|
||||
public void setId(String id) {
|
||||
this.id = id;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String getType() {
|
||||
return type;
|
||||
}
|
||||
|
||||
public void addFlowActivity(String flowActivityName, String flowVersion,
|
||||
long flowRunId) {
|
||||
flowActivities.add(new FlowActivitySubDoc(flowActivityName,
|
||||
flowVersion, flowRunId));
|
||||
}
|
||||
|
||||
public Set<FlowActivitySubDoc> getFlowActivities() {
|
||||
return flowActivities;
|
||||
}
|
||||
|
||||
public void setFlowActivities(Set<FlowActivitySubDoc> flowActivities) {
|
||||
this.flowActivities = flowActivities;
|
||||
}
|
||||
|
||||
@Override
|
||||
public long getCreatedTime() {
|
||||
return TimeUnit.SECONDS.toMillis(dayTimestamp);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void setCreatedTime(long time) {
|
||||
}
|
||||
|
||||
public long getDayTimestamp() {
|
||||
return dayTimestamp;
|
||||
}
|
||||
|
||||
public void setDayTimestamp(long dayTimestamp) {
|
||||
this.dayTimestamp = dayTimestamp;
|
||||
}
|
||||
|
||||
public String getUser() {
|
||||
return user;
|
||||
}
|
||||
|
||||
public void setUser(String user) {
|
||||
this.user = user;
|
||||
}
|
||||
|
||||
public String getFlowName() {
|
||||
return flowName;
|
||||
}
|
||||
|
||||
public void setFlowName(String flowName) {
|
||||
this.flowName = flowName;
|
||||
}
|
||||
}
|
|
@ -0,0 +1,73 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity;
|
||||
|
||||
/**
|
||||
* This is a sub doc which represents each flow.
|
||||
*/
|
||||
public class FlowActivitySubDoc {
|
||||
private String flowName;
|
||||
private String flowVersion;
|
||||
private long flowRunId;
|
||||
|
||||
public FlowActivitySubDoc() {
|
||||
}
|
||||
|
||||
public FlowActivitySubDoc(String flowName, String flowVersion,
|
||||
long flowRunId) {
|
||||
this.flowName = flowName;
|
||||
this.flowVersion = flowVersion;
|
||||
this.flowRunId = flowRunId;
|
||||
}
|
||||
|
||||
public String getFlowName() {
|
||||
return flowName;
|
||||
}
|
||||
|
||||
public String getFlowVersion() {
|
||||
return flowVersion;
|
||||
}
|
||||
|
||||
public long getFlowRunId() {
|
||||
return flowRunId;
|
||||
}
|
||||
|
||||
@Override
|
||||
public int hashCode() {
|
||||
int result = flowVersion.hashCode();
|
||||
result = (int) (31 * result + flowRunId);
|
||||
return result;
|
||||
}
|
||||
|
||||
// Only check if type and id are equal
|
||||
@Override
|
||||
public boolean equals(Object o) {
|
||||
if (this == o) {
|
||||
return true;
|
||||
}
|
||||
if (!(o instanceof FlowActivitySubDoc)) {
|
||||
return false;
|
||||
}
|
||||
FlowActivitySubDoc m = (FlowActivitySubDoc) o;
|
||||
if (!flowVersion.equalsIgnoreCase(m.getFlowVersion())) {
|
||||
return false;
|
||||
}
|
||||
return flowRunId == m.getFlowRunId();
|
||||
}
|
||||
}
|
|
@ -0,0 +1,29 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.
|
||||
* documentstore.collection.document.flowactivity contains
|
||||
* FlowActivityDocument to audit all the flows at a day level.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,239 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowrun;
|
||||
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntityType;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetric;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetricOperation;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.collector.TimelineCollectorContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineMetricSubDoc;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
|
||||
/**
|
||||
* This doc represents the flow run information for every job.
|
||||
*/
|
||||
public class FlowRunDocument implements TimelineDocument<FlowRunDocument> {
|
||||
|
||||
private static final Logger LOG = LoggerFactory
|
||||
.getLogger(FlowRunDocument.class);
|
||||
|
||||
private String id;
|
||||
private final String type = TimelineEntityType.YARN_FLOW_RUN.toString();
|
||||
private String clusterId;
|
||||
private String username;
|
||||
private String flowName;
|
||||
private Long flowRunId;
|
||||
private String flowVersion;
|
||||
private long minStartTime;
|
||||
private long maxEndTime;
|
||||
private final Map<String, TimelineMetricSubDoc>
|
||||
metrics = new HashMap<>();
|
||||
|
||||
public FlowRunDocument() {
|
||||
}
|
||||
|
||||
public FlowRunDocument(TimelineCollectorContext collectorContext,
|
||||
Set<TimelineMetric> metrics) {
|
||||
this.clusterId = collectorContext.getClusterId();
|
||||
this.username = collectorContext.getUserId();
|
||||
this.flowName = collectorContext.getFlowName();
|
||||
this.flowRunId = collectorContext.getFlowRunId();
|
||||
transformMetrics(metrics);
|
||||
}
|
||||
|
||||
private void transformMetrics(Set<TimelineMetric> timelineMetrics) {
|
||||
for (TimelineMetric metric : timelineMetrics) {
|
||||
TimelineMetricSubDoc metricSubDoc = new TimelineMetricSubDoc(metric);
|
||||
this.metrics.put(metric.getId(), metricSubDoc);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Merge the {@link FlowRunDocument} that is passed with the current
|
||||
* document for upsert.
|
||||
*
|
||||
* @param flowRunDoc
|
||||
* that has to be merged
|
||||
*/
|
||||
@Override
|
||||
public void merge(FlowRunDocument flowRunDoc) {
|
||||
if (flowRunDoc.getMinStartTime() > 0) {
|
||||
this.minStartTime = flowRunDoc.getMinStartTime();
|
||||
}
|
||||
if (flowRunDoc.getMaxEndTime() > 0) {
|
||||
this.maxEndTime = flowRunDoc.getMaxEndTime();
|
||||
}
|
||||
this.clusterId = flowRunDoc.getClusterId();
|
||||
this.flowName = flowRunDoc.getFlowName();
|
||||
this.id = flowRunDoc.getId();
|
||||
this.username = flowRunDoc.getUsername();
|
||||
this.flowVersion = flowRunDoc.getFlowVersion();
|
||||
this.flowRunId = flowRunDoc.getFlowRunId();
|
||||
aggregateMetrics(flowRunDoc.getMetrics());
|
||||
}
|
||||
|
||||
private void aggregateMetrics(
|
||||
Map<String, TimelineMetricSubDoc> metricSubDocMap) {
|
||||
for(String metricId : metricSubDocMap.keySet()) {
|
||||
if (this.metrics.containsKey(metricId)) {
|
||||
TimelineMetric incomingMetric =
|
||||
metricSubDocMap.get(metricId).fetchTimelineMetric();
|
||||
TimelineMetric baseMetric =
|
||||
this.metrics.get(metricId).fetchTimelineMetric();
|
||||
if (incomingMetric.getValues().size() > 0) {
|
||||
baseMetric = aggregate(incomingMetric, baseMetric);
|
||||
this.metrics.put(metricId, new TimelineMetricSubDoc(baseMetric));
|
||||
} else {
|
||||
LOG.debug("No incoming metric to aggregate for : {}",
|
||||
baseMetric.getId());
|
||||
}
|
||||
} else {
|
||||
this.metrics.put(metricId, metricSubDocMap.get(metricId));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private TimelineMetric aggregate(TimelineMetric incomingMetric,
|
||||
TimelineMetric baseMetric) {
|
||||
switch (baseMetric.getRealtimeAggregationOp()) {
|
||||
case SUM:
|
||||
baseMetric = TimelineMetricOperation.SUM
|
||||
.aggregate(incomingMetric, baseMetric, null);
|
||||
break;
|
||||
case AVG:
|
||||
baseMetric = TimelineMetricOperation.AVG
|
||||
.aggregate(incomingMetric, baseMetric, null);
|
||||
break;
|
||||
case MAX:
|
||||
baseMetric = TimelineMetricOperation.MAX
|
||||
.aggregate(incomingMetric, baseMetric, null);
|
||||
break;
|
||||
case REPLACE:
|
||||
baseMetric = TimelineMetricOperation.REPLACE
|
||||
.aggregate(incomingMetric, baseMetric, null);
|
||||
default:
|
||||
//NoOP
|
||||
}
|
||||
return baseMetric;
|
||||
}
|
||||
|
||||
public String getId() {
|
||||
return id;
|
||||
}
|
||||
|
||||
public void setId(String id) {
|
||||
this.id = id;
|
||||
}
|
||||
|
||||
public String getClusterId() {
|
||||
return clusterId;
|
||||
}
|
||||
|
||||
public void setClusterId(String clusterId) {
|
||||
this.clusterId = clusterId;
|
||||
}
|
||||
|
||||
public String getUsername() {
|
||||
return username;
|
||||
}
|
||||
|
||||
public void setUsername(String username) {
|
||||
this.username = username;
|
||||
}
|
||||
|
||||
public String getFlowName() {
|
||||
return flowName;
|
||||
}
|
||||
|
||||
public void setFlowName(String flowName) {
|
||||
this.flowName = flowName;
|
||||
}
|
||||
|
||||
public Long getFlowRunId() {
|
||||
return flowRunId;
|
||||
}
|
||||
|
||||
public void setFlowRunId(Long flowRunId) {
|
||||
this.flowRunId = flowRunId;
|
||||
}
|
||||
|
||||
public Map<String, TimelineMetricSubDoc> getMetrics() {
|
||||
return metrics;
|
||||
}
|
||||
|
||||
public void setMetrics(Map<String, TimelineMetricSubDoc> metrics) {
|
||||
this.metrics.putAll(metrics);
|
||||
}
|
||||
|
||||
public Set<TimelineMetric> fetchTimelineMetrics() {
|
||||
Set<TimelineMetric> metricSet = new HashSet<>();
|
||||
for(TimelineMetricSubDoc metricSubDoc : metrics.values()) {
|
||||
metricSet.add(metricSubDoc.fetchTimelineMetric());
|
||||
}
|
||||
return metricSet;
|
||||
}
|
||||
|
||||
public long getMinStartTime() {
|
||||
return minStartTime;
|
||||
}
|
||||
|
||||
public void setMinStartTime(long minStartTime) {
|
||||
this.minStartTime = minStartTime;
|
||||
}
|
||||
|
||||
public long getMaxEndTime() {
|
||||
return maxEndTime;
|
||||
}
|
||||
|
||||
public void setMaxEndTime(long maxEndTime) {
|
||||
this.maxEndTime = maxEndTime;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String getType() {
|
||||
return type;
|
||||
}
|
||||
|
||||
@Override
|
||||
public long getCreatedTime() {
|
||||
return minStartTime;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void setCreatedTime(long createdTime) {
|
||||
if(minStartTime == 0) {
|
||||
minStartTime = createdTime;
|
||||
}
|
||||
}
|
||||
|
||||
public String getFlowVersion() {
|
||||
return flowVersion;
|
||||
}
|
||||
|
||||
public void setFlowVersion(String flowVersion) {
|
||||
this.flowVersion = flowVersion;
|
||||
}
|
||||
}
|
|
@ -0,0 +1,29 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**Package org.apache.hadoop.yarn.server.timelineservice.
|
||||
* documentstore.collection.document.flowrun contains
|
||||
* FlowRunDocument that stores the flow level information for
|
||||
* each Application and also aggregates the metrics.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowrun;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,30 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.
|
||||
* documentstore.collection.document contains interface for all the
|
||||
* Timeline Documents. Any new document that has to be persisted in
|
||||
* the document store should implement this.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,30 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.
|
||||
* documentstore.collection contains different collection types
|
||||
* for storing documents.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,96 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.lib;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.reader.DocumentStoreReader;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.reader.cosmosdb.CosmosDBDocumentStoreReader;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.cosmosdb.CosmosDBDocumentStoreWriter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.DocumentStoreUtils;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.DocumentStoreWriter;
|
||||
|
||||
import static org.apache.hadoop.yarn.server.timelineservice.documentstore.DocumentStoreUtils.getStoreVendor;
|
||||
|
||||
/**
|
||||
* Factory methods for instantiating a timeline Document Store reader or
|
||||
* writer. Based on the {@link DocumentStoreVendor} that is configured,
|
||||
* appropriate reader or writer would be instantiated.
|
||||
*/
|
||||
public final class DocumentStoreFactory {
|
||||
|
||||
// making factory class not instantiable
|
||||
private DocumentStoreFactory(){
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates a DocumentStoreWriter for a {@link DocumentStoreVendor}.
|
||||
* @param conf
|
||||
* for creating client connection
|
||||
* @param <Document> type of Document for which the writer has to be created,
|
||||
* i.e TimelineEntityDocument, FlowActivityDocument etc
|
||||
* @return document store writer
|
||||
* @throws DocumentStoreNotSupportedException if there is no implementation
|
||||
* for a configured {@link DocumentStoreVendor} or unknown
|
||||
* {@link DocumentStoreVendor} is configured.
|
||||
* @throws YarnException if the required configs for DocumentStore is missing.
|
||||
*/
|
||||
public static <Document extends TimelineDocument>
|
||||
DocumentStoreWriter <Document> createDocumentStoreWriter(
|
||||
Configuration conf) throws YarnException {
|
||||
final DocumentStoreVendor storeType = getStoreVendor(conf);
|
||||
switch (storeType) {
|
||||
case COSMOS_DB:
|
||||
DocumentStoreUtils.validateCosmosDBConf(conf);
|
||||
return new CosmosDBDocumentStoreWriter<>(conf);
|
||||
default:
|
||||
throw new DocumentStoreNotSupportedException(
|
||||
"Unable to create DocumentStoreWriter for type : "
|
||||
+ storeType);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates a DocumentStoreReader for a {@link DocumentStoreVendor}.
|
||||
* @param conf
|
||||
* for creating client connection
|
||||
* @param <Document> type of Document for which the writer has to be created,
|
||||
* i.e TimelineEntityDocument, FlowActivityDocument etc
|
||||
* @return document store reader
|
||||
* @throws DocumentStoreNotSupportedException if there is no implementation
|
||||
* for a configured {@link DocumentStoreVendor} or unknown
|
||||
* {@link DocumentStoreVendor} is configured.
|
||||
* @throws YarnException if the required configs for DocumentStore is missing.
|
||||
* */
|
||||
public static <Document extends TimelineDocument>
|
||||
DocumentStoreReader<Document> createDocumentStoreReader(
|
||||
Configuration conf) throws YarnException {
|
||||
final DocumentStoreVendor storeType = getStoreVendor(conf);
|
||||
switch (storeType) {
|
||||
case COSMOS_DB:
|
||||
DocumentStoreUtils.validateCosmosDBConf(conf);
|
||||
return new CosmosDBDocumentStoreReader<>(conf);
|
||||
default:
|
||||
throw new DocumentStoreNotSupportedException(
|
||||
"Unable to create DocumentStoreReader for type : "
|
||||
+ storeType);
|
||||
}
|
||||
}
|
||||
}
|
|
@ -0,0 +1,35 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.lib;
|
||||
|
||||
/**
|
||||
* Indicates that the document store vendor that was
|
||||
* configured does not belong to one of the {@link DocumentStoreVendor}.
|
||||
*/
|
||||
public class DocumentStoreNotSupportedException extends
|
||||
UnsupportedOperationException {
|
||||
|
||||
/**
|
||||
* Constructs exception with the specified detail message.
|
||||
* @param message detailed message.
|
||||
*/
|
||||
public DocumentStoreNotSupportedException(String message) {
|
||||
super(message);
|
||||
}
|
||||
}
|
|
@ -0,0 +1,39 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.lib;
|
||||
|
||||
/**
|
||||
* Represents the different vendors for DocumentStore.
|
||||
*/
|
||||
public enum DocumentStoreVendor {
|
||||
|
||||
COSMOS_DB,
|
||||
MONGO_DB,
|
||||
ELASTIC_SEARCH;
|
||||
|
||||
public static DocumentStoreVendor getStoreType(String storeTypeStr) {
|
||||
for (DocumentStoreVendor storeType : DocumentStoreVendor.values()) {
|
||||
if (storeType.name().equalsIgnoreCase(storeTypeStr)) {
|
||||
return DocumentStoreVendor.valueOf(storeTypeStr.toUpperCase());
|
||||
}
|
||||
}
|
||||
throw new DocumentStoreNotSupportedException(
|
||||
storeTypeStr + " is not a valid DocumentStoreVendor");
|
||||
}
|
||||
}
|
|
@ -0,0 +1,30 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.
|
||||
* documentstore.lib contains factory class for instantiating
|
||||
* different DocumentStore reader writer client based on the DocumentVendor
|
||||
* configured.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.lib;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,29 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.documentstore
|
||||
* contains DocumentStore Reader and Writer Implementation of TimelineService
|
||||
* for reading and writing documents from DocumentStore.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,45 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.reader;
|
||||
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineReaderContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.NoDocumentFoundException;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
|
||||
/**
|
||||
* Every {@link DocumentStoreVendor} have to implement this for creating
|
||||
* reader to its backend.
|
||||
*/
|
||||
public interface DocumentStoreReader<Document extends TimelineDocument>
|
||||
extends AutoCloseable {
|
||||
|
||||
Document readDocument(String collectionName, TimelineReaderContext context,
|
||||
Class<Document> documentClass) throws NoDocumentFoundException;
|
||||
|
||||
List<Document> readDocumentList(String collectionName,
|
||||
TimelineReaderContext context, Class<Document> documentClass,
|
||||
long documentsSize) throws NoDocumentFoundException;
|
||||
|
||||
Set<String> fetchEntityTypes(String collectionName,
|
||||
TimelineReaderContext context);
|
||||
}
|
|
@ -0,0 +1,220 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.reader;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.FlowActivityEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.FlowRunEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntityType;
|
||||
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineReaderContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineReaderUtils;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.DocumentStoreUtils;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.CollectionType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEntityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity.FlowActivityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity.FlowActivitySubDoc;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowrun.FlowRunDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreFactory;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
|
||||
/**
|
||||
* This is a generic Collection reader for reading documents belonging to a
|
||||
* {@link CollectionType} under a specific {@link DocumentStoreVendor} backend.
|
||||
*/
|
||||
public class TimelineCollectionReader {
|
||||
|
||||
private static final Logger LOG = LoggerFactory
|
||||
.getLogger(TimelineCollectionReader.class);
|
||||
|
||||
private final DocumentStoreReader<TimelineEntityDocument>
|
||||
genericEntityDocReader;
|
||||
private final DocumentStoreReader<FlowRunDocument>
|
||||
flowRunDocReader;
|
||||
private final DocumentStoreReader<FlowActivityDocument>
|
||||
flowActivityDocReader;
|
||||
|
||||
public TimelineCollectionReader(
|
||||
Configuration conf) throws YarnException {
|
||||
LOG.info("Initializing TimelineCollectionReader...");
|
||||
genericEntityDocReader = DocumentStoreFactory
|
||||
.createDocumentStoreReader(conf);
|
||||
flowRunDocReader = DocumentStoreFactory
|
||||
.createDocumentStoreReader(conf);
|
||||
flowActivityDocReader = DocumentStoreFactory
|
||||
.createDocumentStoreReader(conf);
|
||||
}
|
||||
|
||||
/**
|
||||
* Read a document from {@link DocumentStoreVendor} backend for
|
||||
* a {@link CollectionType}.
|
||||
* @param context
|
||||
* of the timeline reader
|
||||
* @return TimelineEntityDocument as response
|
||||
* @throws IOException on error while reading
|
||||
*/
|
||||
public TimelineEntityDocument readDocument(
|
||||
TimelineReaderContext context) throws IOException {
|
||||
LOG.debug("Fetching document for entity type {}", context.getEntityType());
|
||||
switch (TimelineEntityType.valueOf(context.getEntityType())) {
|
||||
case YARN_APPLICATION:
|
||||
return genericEntityDocReader.readDocument(
|
||||
CollectionType.APPLICATION.getCollectionName(), context,
|
||||
TimelineEntityDocument.class);
|
||||
case YARN_FLOW_RUN:
|
||||
FlowRunDocument flowRunDoc = flowRunDocReader.readDocument(
|
||||
CollectionType.FLOW_RUN.getCollectionName(), context,
|
||||
FlowRunDocument.class);
|
||||
FlowRunEntity flowRun = createFlowRunEntity(flowRunDoc);
|
||||
return new TimelineEntityDocument(flowRun);
|
||||
case YARN_FLOW_ACTIVITY:
|
||||
FlowActivityDocument flowActivityDoc = flowActivityDocReader
|
||||
.readDocument(CollectionType.FLOW_RUN.getCollectionName(),
|
||||
context, FlowActivityDocument.class);
|
||||
FlowActivityEntity flowActivity = createFlowActivityEntity(context,
|
||||
flowActivityDoc);
|
||||
return new TimelineEntityDocument(flowActivity);
|
||||
default:
|
||||
return genericEntityDocReader.readDocument(
|
||||
CollectionType.ENTITY.getCollectionName(), context,
|
||||
TimelineEntityDocument.class);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Read a list of documents from {@link DocumentStoreVendor} backend for
|
||||
* a {@link CollectionType}.
|
||||
* @param context
|
||||
* of the timeline reader
|
||||
* @param documentsSize
|
||||
* to limit
|
||||
* @return List of TimelineEntityDocument as response
|
||||
* @throws IOException on error while reading
|
||||
*/
|
||||
public List<TimelineEntityDocument> readDocuments(
|
||||
TimelineReaderContext context, long documentsSize) throws IOException {
|
||||
List<TimelineEntityDocument> entityDocs = new ArrayList<>();
|
||||
LOG.debug("Fetching documents for entity type {}", context.getEntityType());
|
||||
switch (TimelineEntityType.valueOf(context.getEntityType())) {
|
||||
case YARN_APPLICATION:
|
||||
return genericEntityDocReader.readDocumentList(
|
||||
CollectionType.APPLICATION.getCollectionName(), context,
|
||||
TimelineEntityDocument.class, documentsSize);
|
||||
case YARN_FLOW_RUN:
|
||||
List<FlowRunDocument> flowRunDocs = flowRunDocReader.readDocumentList(
|
||||
CollectionType.FLOW_RUN.getCollectionName(), context,
|
||||
FlowRunDocument.class, documentsSize);
|
||||
for (FlowRunDocument flowRunDoc : flowRunDocs) {
|
||||
entityDocs.add(new TimelineEntityDocument(createFlowRunEntity(
|
||||
flowRunDoc)));
|
||||
}
|
||||
return entityDocs;
|
||||
case YARN_FLOW_ACTIVITY:
|
||||
List<FlowActivityDocument> flowActivityDocs = flowActivityDocReader
|
||||
.readDocumentList(CollectionType.FLOW_ACTIVITY.getCollectionName(),
|
||||
context, FlowActivityDocument.class, documentsSize);
|
||||
for(FlowActivityDocument flowActivityDoc : flowActivityDocs) {
|
||||
entityDocs.add(new TimelineEntityDocument(
|
||||
createFlowActivityEntity(context, flowActivityDoc)));
|
||||
}
|
||||
return entityDocs;
|
||||
default:
|
||||
return genericEntityDocReader.readDocumentList(
|
||||
CollectionType.ENTITY.getCollectionName(), context,
|
||||
TimelineEntityDocument.class, documentsSize);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Fetches the list of Entity Types i.e (YARN_CONTAINER,
|
||||
* YARN_APPLICATION_ATTEMPT etc.) for an application Id.
|
||||
* @param context
|
||||
* of the timeline reader
|
||||
* @return List of EntityTypes as response
|
||||
*/
|
||||
public Set<String> fetchEntityTypes(
|
||||
TimelineReaderContext context) {
|
||||
LOG.debug("Fetching all entity-types for appId : {}", context.getAppId());
|
||||
return genericEntityDocReader.fetchEntityTypes(
|
||||
CollectionType.ENTITY.getCollectionName(), context);
|
||||
}
|
||||
|
||||
private FlowActivityEntity createFlowActivityEntity(
|
||||
TimelineReaderContext context, FlowActivityDocument flowActivityDoc) {
|
||||
FlowActivityEntity flowActivity = new FlowActivityEntity(
|
||||
context.getClusterId(), flowActivityDoc.getDayTimestamp(),
|
||||
flowActivityDoc.getUser(), flowActivityDoc.getFlowName());
|
||||
flowActivity.setId(flowActivityDoc.getId());
|
||||
// get the list of run ids along with the version that are associated with
|
||||
// this flow on this day
|
||||
for (FlowActivitySubDoc activity : flowActivityDoc
|
||||
.getFlowActivities()) {
|
||||
FlowRunEntity flowRunEntity = new FlowRunEntity();
|
||||
flowRunEntity.setUser(flowActivityDoc.getUser());
|
||||
flowRunEntity.setName(activity.getFlowName());
|
||||
flowRunEntity.setRunId(activity.getFlowRunId());
|
||||
flowRunEntity.setVersion(activity.getFlowVersion());
|
||||
flowRunEntity.setId(flowRunEntity.getId());
|
||||
flowActivity.addFlowRun(flowRunEntity);
|
||||
}
|
||||
flowActivity.getInfo().put(TimelineReaderUtils.FROMID_KEY,
|
||||
flowActivityDoc.getId());
|
||||
flowActivity.setCreatedTime(flowActivityDoc.getDayTimestamp());
|
||||
return flowActivity;
|
||||
}
|
||||
|
||||
private FlowRunEntity createFlowRunEntity(FlowRunDocument flowRunDoc) {
|
||||
FlowRunEntity flowRun = new FlowRunEntity();
|
||||
flowRun.setRunId(flowRunDoc.getFlowRunId());
|
||||
flowRun.setUser(flowRunDoc.getUsername());
|
||||
flowRun.setName(flowRunDoc.getFlowName());
|
||||
|
||||
// read the start time
|
||||
if (flowRunDoc.getMinStartTime() > 0) {
|
||||
flowRun.setStartTime(flowRunDoc.getMinStartTime());
|
||||
}
|
||||
|
||||
// read the end time if available
|
||||
if (flowRunDoc.getMaxEndTime() > 0) {
|
||||
flowRun.setMaxEndTime(flowRunDoc.getMaxEndTime());
|
||||
}
|
||||
|
||||
// read the flow version
|
||||
if (!DocumentStoreUtils.isNullOrEmpty(flowRunDoc.getFlowVersion())) {
|
||||
flowRun.setVersion(flowRunDoc.getFlowVersion());
|
||||
}
|
||||
flowRun.setMetrics(flowRunDoc.fetchTimelineMetrics());
|
||||
flowRun.setId(flowRunDoc.getId());
|
||||
flowRun.getInfo().put(TimelineReaderUtils.FROMID_KEY, flowRunDoc.getId());
|
||||
return flowRun;
|
||||
}
|
||||
|
||||
public void close() throws Exception {
|
||||
genericEntityDocReader.close();
|
||||
flowRunDocReader.close();
|
||||
flowActivityDocReader.close();
|
||||
}
|
||||
}
|
|
@ -0,0 +1,232 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.reader.cosmosdb;
|
||||
|
||||
import com.microsoft.azure.documentdb.Document;
|
||||
import com.microsoft.azure.documentdb.DocumentClient;
|
||||
import com.microsoft.azure.documentdb.FeedOptions;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineReaderContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.DocumentStoreUtils;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.NoDocumentFoundException;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.reader.DocumentStoreReader;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.HashSet;
|
||||
import java.util.Iterator;
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
|
||||
|
||||
/**
|
||||
* This is the Document Store Reader implementation for
|
||||
* {@link DocumentStoreVendor#COSMOS_DB}.
|
||||
*/
|
||||
public class CosmosDBDocumentStoreReader<TimelineDoc extends TimelineDocument>
|
||||
implements DocumentStoreReader<TimelineDoc> {
|
||||
|
||||
private static final Logger LOG = LoggerFactory
|
||||
.getLogger(CosmosDBDocumentStoreReader.class);
|
||||
private static final int DEFAULT_DOCUMENTS_SIZE = 1;
|
||||
|
||||
private static DocumentClient client;
|
||||
private final String databaseName;
|
||||
private final static String COLLECTION_LINK = "/dbs/%s/colls/%s";
|
||||
private final static String SELECT_TOP_FROM_COLLECTION = "SELECT TOP %d * " +
|
||||
"FROM %s c";
|
||||
private final static String SELECT_ALL_FROM_COLLECTION =
|
||||
"SELECT * FROM %s c";
|
||||
private final static String SELECT_DISTINCT_TYPES_FROM_COLLECTION =
|
||||
"SELECT distinct c.type FROM %s c";
|
||||
private static final String ENTITY_TYPE_COLUMN = "type";
|
||||
private final static String WHERE_CLAUSE = " WHERE ";
|
||||
private final static String AND_OPERATOR = " AND ";
|
||||
private final static String CONTAINS_FUNC_FOR_ID = " CONTAINS(c.id, \"%s\") ";
|
||||
private final static String CONTAINS_FUNC_FOR_TYPE = " CONTAINS(c.type, " +
|
||||
"\"%s\") ";
|
||||
private final static String ORDER_BY_CLAUSE = " ORDER BY c.createdTime";
|
||||
|
||||
public CosmosDBDocumentStoreReader(Configuration conf) {
|
||||
LOG.info("Initializing Cosmos DB DocumentStoreReader...");
|
||||
databaseName = DocumentStoreUtils.getCosmosDBDatabaseName(conf);
|
||||
// making CosmosDB Client Singleton
|
||||
if (client == null) {
|
||||
synchronized (this) {
|
||||
if (client == null) {
|
||||
LOG.info("Creating Cosmos DB Client...");
|
||||
client = DocumentStoreUtils.createCosmosDBClient(conf);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public List<TimelineDoc> readDocumentList(String collectionName,
|
||||
TimelineReaderContext context, final Class<TimelineDoc> timelineDocClass,
|
||||
long size) throws NoDocumentFoundException {
|
||||
final List<TimelineDoc> result = queryDocuments(collectionName,
|
||||
context, timelineDocClass, size);
|
||||
if (result.size() > 0) {
|
||||
return result;
|
||||
}
|
||||
throw new NoDocumentFoundException("No documents were found while " +
|
||||
"querying Collection : " + collectionName);
|
||||
}
|
||||
|
||||
@Override
|
||||
public Set<String> fetchEntityTypes(String collectionName,
|
||||
TimelineReaderContext context) {
|
||||
StringBuilder queryStrBuilder = new StringBuilder();
|
||||
queryStrBuilder.append(
|
||||
String.format(SELECT_DISTINCT_TYPES_FROM_COLLECTION, collectionName));
|
||||
String sqlQuery = addPredicates(context, collectionName, queryStrBuilder);
|
||||
|
||||
LOG.debug("Querying Collection : {} , with query {}", collectionName,
|
||||
sqlQuery);
|
||||
|
||||
Set<String> entityTypes = new HashSet<>();
|
||||
Iterator<Document> documentIterator = client.queryDocuments(
|
||||
String.format(COLLECTION_LINK, databaseName, collectionName),
|
||||
sqlQuery, null).getQueryIterator();
|
||||
while (documentIterator.hasNext()) {
|
||||
Document document = documentIterator.next();
|
||||
entityTypes.add(document.getString(ENTITY_TYPE_COLUMN));
|
||||
}
|
||||
return entityTypes;
|
||||
}
|
||||
|
||||
@Override
|
||||
public TimelineDoc readDocument(String collectionName, TimelineReaderContext
|
||||
context, final Class<TimelineDoc> timelineDocClass)
|
||||
throws NoDocumentFoundException {
|
||||
final List<TimelineDoc> result = queryDocuments(collectionName,
|
||||
context, timelineDocClass, DEFAULT_DOCUMENTS_SIZE);
|
||||
if(result.size() > 0) {
|
||||
return result.get(0);
|
||||
}
|
||||
throw new NoDocumentFoundException("No documents were found while " +
|
||||
"querying Collection : " + collectionName);
|
||||
}
|
||||
|
||||
private List<TimelineDoc> queryDocuments(String collectionName,
|
||||
TimelineReaderContext context, final Class<TimelineDoc> docClass,
|
||||
final long maxDocumentsSize) {
|
||||
final String sqlQuery = buildQueryWithPredicates(context, collectionName,
|
||||
maxDocumentsSize);
|
||||
List<TimelineDoc> timelineDocs = new ArrayList<>();
|
||||
LOG.debug("Querying Collection : {} , with query {}", collectionName,
|
||||
sqlQuery);
|
||||
|
||||
FeedOptions feedOptions = new FeedOptions();
|
||||
feedOptions.setPageSize((int) maxDocumentsSize);
|
||||
Iterator<Document> documentIterator = client.queryDocuments(
|
||||
String.format(COLLECTION_LINK, databaseName, collectionName),
|
||||
sqlQuery, feedOptions).getQueryIterator();
|
||||
while (documentIterator.hasNext()) {
|
||||
Document document = documentIterator.next();
|
||||
TimelineDoc resultDoc = document.toObject(docClass);
|
||||
if (resultDoc.getCreatedTime() == 0 &&
|
||||
document.getTimestamp() != null) {
|
||||
resultDoc.setCreatedTime(document.getTimestamp().getTime());
|
||||
}
|
||||
timelineDocs.add(resultDoc);
|
||||
}
|
||||
return timelineDocs;
|
||||
}
|
||||
|
||||
private String buildQueryWithPredicates(TimelineReaderContext context,
|
||||
String collectionName, long size) {
|
||||
StringBuilder queryStrBuilder = new StringBuilder();
|
||||
if (size == -1) {
|
||||
queryStrBuilder.append(String.format(SELECT_ALL_FROM_COLLECTION,
|
||||
collectionName));
|
||||
} else {
|
||||
queryStrBuilder.append(String.format(SELECT_TOP_FROM_COLLECTION, size,
|
||||
collectionName));
|
||||
}
|
||||
|
||||
return addPredicates(context, collectionName, queryStrBuilder);
|
||||
}
|
||||
|
||||
private String addPredicates(TimelineReaderContext context,
|
||||
String collectionName, StringBuilder queryStrBuilder) {
|
||||
boolean hasPredicate = false;
|
||||
|
||||
queryStrBuilder.append(WHERE_CLAUSE);
|
||||
|
||||
if (context.getClusterId() != null) {
|
||||
hasPredicate = true;
|
||||
queryStrBuilder.append(String.format(CONTAINS_FUNC_FOR_ID,
|
||||
context.getClusterId()));
|
||||
}
|
||||
if (context.getUserId() != null) {
|
||||
hasPredicate = true;
|
||||
queryStrBuilder.append(AND_OPERATOR)
|
||||
.append(String.format(CONTAINS_FUNC_FOR_ID, context.getUserId()));
|
||||
}
|
||||
if (context.getFlowName() != null) {
|
||||
hasPredicate = true;
|
||||
queryStrBuilder.append(AND_OPERATOR)
|
||||
.append(String.format(CONTAINS_FUNC_FOR_ID, context.getFlowName()));
|
||||
}
|
||||
if (context.getAppId() != null) {
|
||||
hasPredicate = true;
|
||||
queryStrBuilder.append(AND_OPERATOR)
|
||||
.append(String.format(CONTAINS_FUNC_FOR_ID, context.getAppId()));
|
||||
}
|
||||
if (context.getEntityId() != null) {
|
||||
hasPredicate = true;
|
||||
queryStrBuilder.append(AND_OPERATOR)
|
||||
.append(String.format(CONTAINS_FUNC_FOR_ID, context.getEntityId()));
|
||||
}
|
||||
if (context.getFlowRunId() != null) {
|
||||
hasPredicate = true;
|
||||
queryStrBuilder.append(AND_OPERATOR)
|
||||
.append(String.format(CONTAINS_FUNC_FOR_ID, context.getFlowRunId()));
|
||||
}
|
||||
if (context.getEntityType() != null){
|
||||
hasPredicate = true;
|
||||
queryStrBuilder.append(AND_OPERATOR)
|
||||
.append(String.format(CONTAINS_FUNC_FOR_TYPE,
|
||||
context.getEntityType()));
|
||||
}
|
||||
|
||||
if (hasPredicate) {
|
||||
queryStrBuilder.append(ORDER_BY_CLAUSE);
|
||||
LOG.debug("CosmosDB Sql Query with predicates : {}", queryStrBuilder);
|
||||
return queryStrBuilder.toString();
|
||||
}
|
||||
throw new IllegalArgumentException("The TimelineReaderContext does not " +
|
||||
"have enough information to query documents for Collection : " +
|
||||
collectionName);
|
||||
}
|
||||
|
||||
@Override
|
||||
public synchronized void close() {
|
||||
if (client != null) {
|
||||
LOG.info("Closing Cosmos DB Client...");
|
||||
client.close();
|
||||
client = null;
|
||||
}
|
||||
}
|
||||
}
|
|
@ -0,0 +1,28 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.documentstore
|
||||
* .reader.cosmosdb DocumentStore Reader implementation for CosmosDB.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.reader.cosmosdb;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,29 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.documentstore.reader
|
||||
* contains the implementation of different DocumentStore reader clients
|
||||
* for DocumentVendor.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.reader;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,35 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.writer;
|
||||
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.CollectionType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
|
||||
/**
|
||||
* Every {@link DocumentStoreVendor} have to implement this for creating
|
||||
* writer to its backend.
|
||||
*/
|
||||
public interface DocumentStoreWriter<Document> extends AutoCloseable {
|
||||
|
||||
void createDatabase();
|
||||
|
||||
void createCollection(String collectionName);
|
||||
|
||||
void writeDocument(Document document, CollectionType collectionType);
|
||||
}
|
|
@ -0,0 +1,146 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.writer;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.util.Time;
|
||||
import org.apache.hadoop.yarn.conf.YarnConfiguration;
|
||||
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.CollectionType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreFactory;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.metrics.PerNodeAggTimelineCollectorMetrics;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.util.HashMap;
|
||||
import java.util.Map;
|
||||
import java.util.concurrent.ExecutorService;
|
||||
import java.util.concurrent.Executors;
|
||||
import java.util.concurrent.ScheduledExecutorService;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
|
||||
/**
|
||||
* This is a generic Collection Writer that can be used for writing documents
|
||||
* belonging to different {@link CollectionType} under a specific
|
||||
* {@link DocumentStoreVendor} backend.
|
||||
*/
|
||||
public class TimelineCollectionWriter<Document extends TimelineDocument> {
|
||||
|
||||
private static final Logger LOG = LoggerFactory
|
||||
.getLogger(TimelineCollectionWriter.class);
|
||||
|
||||
private final static String DOCUMENT_BUFFER_SIZE_CONF =
|
||||
"yarn.timeline-service.document-buffer.size";
|
||||
private static final int DEFAULT_BUFFER_SIZE = 1024;
|
||||
private static final int AWAIT_TIMEOUT_SECS = 5;
|
||||
private static final PerNodeAggTimelineCollectorMetrics METRICS =
|
||||
PerNodeAggTimelineCollectorMetrics.getInstance();
|
||||
|
||||
private final CollectionType collectionType;
|
||||
private final DocumentStoreWriter<Document> documentStoreWriter;
|
||||
private final Map<String, Document> documentsBuffer;
|
||||
private final int maxBufferSize;
|
||||
private final ScheduledExecutorService scheduledDocumentsFlusher;
|
||||
private final ExecutorService documentsBufferFullFlusher;
|
||||
|
||||
public TimelineCollectionWriter(CollectionType collectionType,
|
||||
Configuration conf) throws YarnException {
|
||||
LOG.info("Initializing TimelineCollectionWriter for collection type : {}",
|
||||
collectionType);
|
||||
int flushIntervalSecs = conf.getInt(
|
||||
YarnConfiguration.TIMELINE_SERVICE_WRITER_FLUSH_INTERVAL_SECONDS,
|
||||
YarnConfiguration
|
||||
.DEFAULT_TIMELINE_SERVICE_WRITER_FLUSH_INTERVAL_SECONDS);
|
||||
maxBufferSize = conf.getInt(DOCUMENT_BUFFER_SIZE_CONF, DEFAULT_BUFFER_SIZE);
|
||||
documentsBuffer = new HashMap<>(maxBufferSize);
|
||||
this.collectionType = collectionType;
|
||||
documentStoreWriter = DocumentStoreFactory.createDocumentStoreWriter(conf);
|
||||
scheduledDocumentsFlusher = Executors.newSingleThreadScheduledExecutor();
|
||||
scheduledDocumentsFlusher.scheduleAtFixedRate(this::flush,
|
||||
flushIntervalSecs, flushIntervalSecs, TimeUnit.SECONDS);
|
||||
documentsBufferFullFlusher = Executors.newSingleThreadExecutor();
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
public void writeDocument(Document timelineDocument) {
|
||||
/*
|
||||
* The DocumentBuffer is used to buffer the most frequently used
|
||||
* documents for performing upserts on them, whenever either due to
|
||||
* buffer gets fulled or the scheduledDocumentsFlusher
|
||||
* invokes flush() periodically, all the buffered documents would be written
|
||||
* to DocumentStore in a background thread.
|
||||
*/
|
||||
long startTime = Time.monotonicNow();
|
||||
|
||||
synchronized(documentsBuffer) {
|
||||
//if buffer is full copy to flushBuffer in order to flush
|
||||
if (documentsBuffer.size() == maxBufferSize) {
|
||||
final Map<String, Document> flushedBuffer = copyToFlushBuffer();
|
||||
//flush all documents from flushBuffer in background
|
||||
documentsBufferFullFlusher.execute(() -> flush(flushedBuffer));
|
||||
}
|
||||
Document prevDocument = documentsBuffer.get(timelineDocument.getId());
|
||||
// check if Document exists inside documentsBuffer
|
||||
if (prevDocument != null) {
|
||||
prevDocument.merge(timelineDocument);
|
||||
} else { // else treat this as a new document
|
||||
prevDocument = timelineDocument;
|
||||
}
|
||||
documentsBuffer.put(prevDocument.getId(), prevDocument);
|
||||
}
|
||||
METRICS.addAsyncPutEntitiesLatency(Time.monotonicNow() - startTime,
|
||||
true);
|
||||
}
|
||||
|
||||
private Map<String, Document> copyToFlushBuffer() {
|
||||
Map<String, Document> flushBuffer = new HashMap<>();
|
||||
synchronized(documentsBuffer) {
|
||||
if (documentsBuffer.size() > 0) {
|
||||
flushBuffer.putAll(documentsBuffer);
|
||||
documentsBuffer.clear();
|
||||
}
|
||||
}
|
||||
return flushBuffer;
|
||||
}
|
||||
|
||||
private void flush(Map<String, Document> flushBuffer) {
|
||||
for (Document document : flushBuffer.values()) {
|
||||
documentStoreWriter.writeDocument(document, collectionType);
|
||||
}
|
||||
}
|
||||
|
||||
public void flush() {
|
||||
flush(copyToFlushBuffer());
|
||||
}
|
||||
|
||||
public void close() throws Exception {
|
||||
scheduledDocumentsFlusher.shutdown();
|
||||
documentsBufferFullFlusher.shutdown();
|
||||
|
||||
flush();
|
||||
|
||||
scheduledDocumentsFlusher.awaitTermination(
|
||||
AWAIT_TIMEOUT_SECS, TimeUnit.SECONDS);
|
||||
documentsBufferFullFlusher.awaitTermination(
|
||||
AWAIT_TIMEOUT_SECS, TimeUnit.SECONDS);
|
||||
documentStoreWriter.close();
|
||||
}
|
||||
}
|
|
@ -0,0 +1,235 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.cosmosdb;
|
||||
|
||||
|
||||
import com.microsoft.azure.documentdb.AccessCondition;
|
||||
import com.microsoft.azure.documentdb.AccessConditionType;
|
||||
import com.microsoft.azure.documentdb.Database;
|
||||
import com.microsoft.azure.documentdb.Document;
|
||||
import com.microsoft.azure.documentdb.DocumentClient;
|
||||
import com.microsoft.azure.documentdb.DocumentClientException;
|
||||
import com.microsoft.azure.documentdb.DocumentCollection;
|
||||
import com.microsoft.azure.documentdb.RequestOptions;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.util.Time;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.metrics.PerNodeAggTimelineCollectorMetrics;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.DocumentStoreUtils;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.CollectionType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEntityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity.FlowActivityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowrun.FlowRunDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreVendor;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.DocumentStoreWriter;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* This is the Document Store Writer implementation for
|
||||
* {@link DocumentStoreVendor#COSMOS_DB}.
|
||||
*/
|
||||
public class CosmosDBDocumentStoreWriter<TimelineDoc extends TimelineDocument>
|
||||
implements DocumentStoreWriter<TimelineDoc> {
|
||||
|
||||
private static final Logger LOG = LoggerFactory
|
||||
.getLogger(CosmosDBDocumentStoreWriter.class);
|
||||
|
||||
private static DocumentClient client;
|
||||
private final String databaseName;
|
||||
private static final PerNodeAggTimelineCollectorMetrics METRICS =
|
||||
PerNodeAggTimelineCollectorMetrics.getInstance();
|
||||
private static final String DATABASE_LINK = "/dbs/%s";
|
||||
private static final String COLLECTION_LINK = DATABASE_LINK + "/colls/%s";
|
||||
private static final String DOCUMENT_LINK = COLLECTION_LINK + "/docs/%s";
|
||||
|
||||
public CosmosDBDocumentStoreWriter(Configuration conf) {
|
||||
LOG.info("Initializing Cosmos DB DocumentStoreWriter...");
|
||||
databaseName = DocumentStoreUtils.getCosmosDBDatabaseName(conf);
|
||||
// making CosmosDB Client Singleton
|
||||
if (client == null) {
|
||||
synchronized (this) {
|
||||
if (client == null) {
|
||||
LOG.info("Creating Cosmos DB Client...");
|
||||
client = DocumentStoreUtils.createCosmosDBClient(conf);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void createDatabase() {
|
||||
try {
|
||||
client.readDatabase(String.format(
|
||||
DATABASE_LINK, databaseName), new RequestOptions());
|
||||
LOG.info("Database {} already exists.", databaseName);
|
||||
} catch (DocumentClientException docExceptionOnRead) {
|
||||
if (docExceptionOnRead.getStatusCode() == 404) {
|
||||
LOG.info("Creating new Database : {}", databaseName);
|
||||
Database databaseDefinition = new Database();
|
||||
databaseDefinition.setId(databaseName);
|
||||
try {
|
||||
client.createDatabase(databaseDefinition, new RequestOptions());
|
||||
} catch (DocumentClientException docExceptionOnCreate) {
|
||||
LOG.error("Unable to create new Database : {}", databaseName,
|
||||
docExceptionOnCreate);
|
||||
}
|
||||
} else {
|
||||
LOG.error("Error while reading Database : {}", databaseName,
|
||||
docExceptionOnRead);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void createCollection(final String collectionName) {
|
||||
LOG.info("Creating Timeline Collection : {} for Database : {}",
|
||||
collectionName, databaseName);
|
||||
try {
|
||||
client.readCollection(String.format(COLLECTION_LINK, databaseName,
|
||||
collectionName), new RequestOptions());
|
||||
LOG.info("Collection {} already exists.", collectionName);
|
||||
} catch (DocumentClientException docExceptionOnRead) {
|
||||
if (docExceptionOnRead.getStatusCode() == 404) {
|
||||
DocumentCollection collection = new DocumentCollection();
|
||||
collection.setId(collectionName);
|
||||
LOG.info("Creating collection {} under Database {}",
|
||||
collectionName, databaseName);
|
||||
try {
|
||||
client.createCollection(
|
||||
String.format(DATABASE_LINK, databaseName),
|
||||
collection, new RequestOptions());
|
||||
} catch (DocumentClientException docExceptionOnCreate) {
|
||||
LOG.error("Unable to create Collection : {} under Database : {}",
|
||||
collectionName, databaseName, docExceptionOnCreate);
|
||||
}
|
||||
} else {
|
||||
LOG.error("Error while reading Collection : {} under Database : {}",
|
||||
collectionName, databaseName, docExceptionOnRead);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void writeDocument(final TimelineDoc timelineDoc,
|
||||
final CollectionType collectionType) {
|
||||
LOG.debug("Upserting document under collection : {} with entity type : " +
|
||||
"{} under Database {}", databaseName, timelineDoc.getType(),
|
||||
collectionType.getCollectionName());
|
||||
boolean succeeded = false;
|
||||
long startTime = Time.monotonicNow();
|
||||
try {
|
||||
upsertDocument(collectionType, timelineDoc);
|
||||
succeeded = true;
|
||||
} catch (Exception e) {
|
||||
LOG.error("Unable to perform upsert for Document Id : {} under " +
|
||||
"Collection : {} under Database {}", timelineDoc.getId(),
|
||||
collectionType.getCollectionName(), databaseName, e);
|
||||
} finally {
|
||||
long latency = Time.monotonicNow() - startTime;
|
||||
METRICS.addPutEntitiesLatency(latency, succeeded);
|
||||
}
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
private void upsertDocument(final CollectionType collectionType,
|
||||
final TimelineDoc timelineDoc) {
|
||||
final String collectionLink = String.format(COLLECTION_LINK, databaseName,
|
||||
collectionType.getCollectionName());
|
||||
RequestOptions requestOptions = new RequestOptions();
|
||||
AccessCondition accessCondition = new AccessCondition();
|
||||
StringBuilder eTagStrBuilder = new StringBuilder();
|
||||
|
||||
TimelineDoc updatedTimelineDoc = applyUpdatesOnPrevDoc(collectionType,
|
||||
timelineDoc, eTagStrBuilder);
|
||||
|
||||
accessCondition.setCondition(eTagStrBuilder.toString());
|
||||
accessCondition.setType(AccessConditionType.IfMatch);
|
||||
requestOptions.setAccessCondition(accessCondition);
|
||||
|
||||
try {
|
||||
client.upsertDocument(collectionLink, updatedTimelineDoc,
|
||||
requestOptions, true);
|
||||
LOG.debug("Successfully wrote doc with id : {} and type : {} under " +
|
||||
"Database : {}", timelineDoc.getId(), timelineDoc.getType(),
|
||||
databaseName);
|
||||
} catch (DocumentClientException e) {
|
||||
if (e.getStatusCode() == 409) {
|
||||
LOG.warn("There was a conflict while upserting, hence retrying...", e);
|
||||
upsertDocument(collectionType, updatedTimelineDoc);
|
||||
}
|
||||
LOG.error("Error while upserting Collection : {} with Doc Id : {} under" +
|
||||
" Database : {}", collectionType.getCollectionName(),
|
||||
updatedTimelineDoc.getId(), databaseName, e);
|
||||
}
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
private TimelineDoc applyUpdatesOnPrevDoc(CollectionType collectionType,
|
||||
TimelineDoc timelineDoc, StringBuilder eTagStrBuilder) {
|
||||
TimelineDoc prevDocument = fetchLatestDoc(collectionType,
|
||||
timelineDoc.getId(), eTagStrBuilder);
|
||||
if (prevDocument != null) {
|
||||
prevDocument.merge(timelineDoc);
|
||||
timelineDoc = prevDocument;
|
||||
}
|
||||
return timelineDoc;
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
private TimelineDoc fetchLatestDoc(final CollectionType collectionType,
|
||||
final String documentId, StringBuilder eTagStrBuilder) {
|
||||
final String documentLink = String.format(DOCUMENT_LINK, databaseName,
|
||||
collectionType.getCollectionName(), documentId);
|
||||
try {
|
||||
Document latestDocument = client.readDocument(documentLink, new
|
||||
RequestOptions()).getResource();
|
||||
TimelineDoc timelineDoc;
|
||||
switch (collectionType) {
|
||||
case FLOW_RUN:
|
||||
timelineDoc = (TimelineDoc) latestDocument.toObject(
|
||||
FlowRunDocument.class);
|
||||
break;
|
||||
case FLOW_ACTIVITY:
|
||||
timelineDoc = (TimelineDoc) latestDocument.toObject(FlowActivityDocument
|
||||
.class);
|
||||
break;
|
||||
default:
|
||||
timelineDoc = (TimelineDoc) latestDocument.toObject(
|
||||
TimelineEntityDocument.class);
|
||||
}
|
||||
eTagStrBuilder.append(latestDocument.getETag());
|
||||
return timelineDoc;
|
||||
} catch (Exception e) {
|
||||
LOG.debug("No previous Document found with id : {} for Collection" +
|
||||
" : {} under Database : {}", documentId, collectionType
|
||||
.getCollectionName(), databaseName);
|
||||
return null;
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public synchronized void close() {
|
||||
if (client != null) {
|
||||
LOG.info("Closing Cosmos DB Client...");
|
||||
client.close();
|
||||
client = null;
|
||||
}
|
||||
}
|
||||
}
|
|
@ -0,0 +1,28 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.documentstore
|
||||
* .writer.cosmosdb DocumentStore writer implementation for CosmosDB.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.cosmosdb;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,29 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
/**
|
||||
* Package org.apache.hadoop.yarn.server.timelineservice.documentstore.writer
|
||||
* contains the implementation of different DocumentStore writer clients
|
||||
* for DocumentVendor.
|
||||
*/
|
||||
@InterfaceAudience.Private
|
||||
@InterfaceStability.Unstable
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.writer;
|
||||
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
|
@ -0,0 +1,81 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import org.apache.commons.io.IOUtils;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntity;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEntityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity.FlowActivityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowrun.FlowRunDocument;
|
||||
import org.codehaus.jackson.type.TypeReference;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.List;
|
||||
|
||||
/**
|
||||
* This is util class for baking sample TimelineEntities data for test.
|
||||
*/
|
||||
public final class DocumentStoreTestUtils {
|
||||
|
||||
private DocumentStoreTestUtils(){}
|
||||
|
||||
|
||||
public static List<TimelineEntity> bakeTimelineEntities()
|
||||
throws IOException {
|
||||
String jsonStr = IOUtils.toString(
|
||||
DocumentStoreTestUtils.class.getClassLoader().getResourceAsStream(
|
||||
"documents/timeline-entities.json"), "UTF-8");
|
||||
return JsonUtils.fromJson(jsonStr,
|
||||
new TypeReference<List<TimelineEntity>>(){});
|
||||
}
|
||||
|
||||
public static List<TimelineEntityDocument> bakeYarnAppTimelineEntities()
|
||||
throws IOException {
|
||||
String jsonStr = IOUtils.toString(
|
||||
DocumentStoreTestUtils.class.getClassLoader().getResourceAsStream(
|
||||
"documents/test-timeline-entities-doc.json"), "UTF-8");
|
||||
return JsonUtils.fromJson(jsonStr,
|
||||
new TypeReference<List<TimelineEntityDocument>>() {});
|
||||
}
|
||||
|
||||
public static TimelineEntityDocument bakeTimelineEntityDoc()
|
||||
throws IOException {
|
||||
String jsonStr = IOUtils.toString(
|
||||
DocumentStoreTestUtils.class.getClassLoader().getResourceAsStream(
|
||||
"documents/timeline-app-doc.json"), "UTF-8");
|
||||
return JsonUtils.fromJson(jsonStr,
|
||||
new TypeReference<TimelineEntityDocument>() {});
|
||||
}
|
||||
|
||||
public static FlowActivityDocument bakeFlowActivityDoc() throws IOException {
|
||||
String jsonStr = IOUtils.toString(
|
||||
DocumentStoreTestUtils.class.getClassLoader().getResourceAsStream(
|
||||
"documents/flowactivity-doc.json"), "UTF-8");
|
||||
return JsonUtils.fromJson(jsonStr,
|
||||
new TypeReference<FlowActivityDocument>() {});
|
||||
}
|
||||
|
||||
public static FlowRunDocument bakeFlowRunDoc() throws IOException {
|
||||
String jsonStr = IOUtils.toString(
|
||||
DocumentStoreTestUtils.class.getClassLoader().getResourceAsStream(
|
||||
"documents/flowrun-doc.json"), "UTF-8");
|
||||
return JsonUtils.fromJson(jsonStr,
|
||||
new TypeReference<FlowRunDocument>(){});
|
||||
}
|
||||
}
|
|
@ -0,0 +1,59 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import org.codehaus.jackson.map.DeserializationConfig;
|
||||
import org.codehaus.jackson.map.ObjectMapper;
|
||||
import org.codehaus.jackson.type.TypeReference;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
/**
|
||||
* A simple util class for Json SerDe.
|
||||
*/
|
||||
public final class JsonUtils {
|
||||
|
||||
private JsonUtils(){}
|
||||
|
||||
private static final ObjectMapper OBJECT_MAPPER = new ObjectMapper();
|
||||
|
||||
static {
|
||||
OBJECT_MAPPER.configure(
|
||||
DeserializationConfig.Feature.FAIL_ON_UNKNOWN_PROPERTIES, false);
|
||||
}
|
||||
|
||||
/**
|
||||
* Deserialize the Json String to JAVA Object.
|
||||
* @param jsonStr
|
||||
* json string that has to be deserialized
|
||||
* @param type
|
||||
* of JAVA Object
|
||||
* @return JAVA Object after deserialization
|
||||
* @throws IOException if Json String is not valid or error
|
||||
* while deserialization
|
||||
*/
|
||||
public static <T> T fromJson(final String jsonStr, final TypeReference type)
|
||||
throws IOException {
|
||||
return OBJECT_MAPPER.readValue(jsonStr, type);
|
||||
}
|
||||
|
||||
public static ObjectMapper getObjectMapper() {
|
||||
return OBJECT_MAPPER;
|
||||
}
|
||||
}
|
|
@ -0,0 +1,64 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreFactory;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.DocumentStoreWriter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.DummyDocumentStoreWriter;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.runner.RunWith;
|
||||
import org.mockito.ArgumentMatchers;
|
||||
import org.powermock.api.mockito.PowerMockito;
|
||||
import org.powermock.core.classloader.annotations.PrepareForTest;
|
||||
import org.powermock.modules.junit4.PowerMockRunner;
|
||||
|
||||
/**
|
||||
* Test case for ${@link DocumentStoreCollectionCreator}.
|
||||
*/
|
||||
@RunWith(PowerMockRunner.class)
|
||||
@PrepareForTest(DocumentStoreFactory.class)
|
||||
public class TestDocumentStoreCollectionCreator {
|
||||
|
||||
private final DocumentStoreWriter<TimelineDocument> documentStoreWriter = new
|
||||
DummyDocumentStoreWriter<>();
|
||||
private final Configuration conf = new Configuration();
|
||||
|
||||
@Before
|
||||
public void setUp() throws YarnException {
|
||||
conf.set(DocumentStoreUtils.TIMELINE_SERVICE_DOCUMENTSTORE_DATABASE_NAME,
|
||||
"TestDB");
|
||||
conf.set(DocumentStoreUtils.TIMELINE_SERVICE_COSMOSDB_ENDPOINT,
|
||||
"https://localhost:443");
|
||||
conf.set(DocumentStoreUtils.TIMELINE_SERVICE_COSMOSDB_MASTER_KEY,
|
||||
"1234567");
|
||||
PowerMockito.mockStatic(DocumentStoreFactory.class);
|
||||
PowerMockito.when(DocumentStoreFactory.createDocumentStoreWriter(
|
||||
ArgumentMatchers.any(Configuration.class)))
|
||||
.thenReturn(documentStoreWriter);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void collectionCreatorTest() {
|
||||
new DocumentStoreCollectionCreator().createTimelineSchema(new String[]{});
|
||||
}
|
||||
}
|
|
@ -0,0 +1,407 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntity;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntityType;
|
||||
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreFactory;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.reader.DocumentStoreReader;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.reader.DummyDocumentStoreReader;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineDataToRetrieve;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineEntityFilters;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineReaderContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.filter.TimelineCompareFilter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.filter.TimelineCompareOp;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.filter.TimelineExistsFilter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.filter.TimelineFilterList;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.filter.TimelineKeyValueFilter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.filter.TimelinePrefixFilter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.storage.TimelineReader;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEntityDocument;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.runner.RunWith;
|
||||
import org.mockito.ArgumentMatchers;
|
||||
import org.powermock.api.mockito.PowerMockito;
|
||||
import org.powermock.core.classloader.annotations.PrepareForTest;
|
||||
import org.powermock.modules.junit4.PowerMockRunner;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.EnumSet;
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
|
||||
|
||||
/**
|
||||
* Test case for {@link DocumentStoreTimelineReaderImpl}.
|
||||
*/
|
||||
@RunWith(PowerMockRunner.class)
|
||||
@PrepareForTest(DocumentStoreFactory.class)
|
||||
public class TestDocumentStoreTimelineReaderImpl {
|
||||
|
||||
private final DocumentStoreReader<TimelineDocument> documentStoreReader = new
|
||||
DummyDocumentStoreReader<>();
|
||||
private final List<TimelineEntity> entities = DocumentStoreTestUtils
|
||||
.bakeTimelineEntities();
|
||||
private final TimelineEntityDocument appTimelineEntity =
|
||||
DocumentStoreTestUtils.bakeTimelineEntityDoc();
|
||||
|
||||
private final Configuration conf = new Configuration();
|
||||
private final TimelineReaderContext context = new
|
||||
TimelineReaderContext(null, null, null,
|
||||
1L, null, null, null);
|
||||
private final DocumentStoreTimelineReaderImpl timelineReader = new
|
||||
DocumentStoreTimelineReaderImpl();
|
||||
|
||||
public TestDocumentStoreTimelineReaderImpl() throws IOException {
|
||||
}
|
||||
|
||||
@Before
|
||||
public void setUp() throws YarnException {
|
||||
conf.set(DocumentStoreUtils.TIMELINE_SERVICE_DOCUMENTSTORE_DATABASE_NAME,
|
||||
"TestDB");
|
||||
conf.set(DocumentStoreUtils.TIMELINE_SERVICE_COSMOSDB_ENDPOINT,
|
||||
"https://localhost:443");
|
||||
conf.set(DocumentStoreUtils.TIMELINE_SERVICE_COSMOSDB_MASTER_KEY,
|
||||
"1234567");
|
||||
PowerMockito.mockStatic(DocumentStoreFactory.class);
|
||||
PowerMockito.when(DocumentStoreFactory.createDocumentStoreReader(
|
||||
ArgumentMatchers.any(Configuration.class)))
|
||||
.thenReturn(documentStoreReader);
|
||||
}
|
||||
|
||||
@Test(expected = YarnException.class)
|
||||
public void testFailOnNoCosmosDBConfigs() throws Exception {
|
||||
DocumentStoreUtils.validateCosmosDBConf(new Configuration());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testGetEntity() throws Exception {
|
||||
context.setEntityType(TimelineEntityType.YARN_APPLICATION.toString());
|
||||
timelineReader.serviceInit(conf);
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
EnumSet<TimelineReader.Field> fieldsToRetrieve = EnumSet.noneOf(
|
||||
TimelineReader.Field.class);
|
||||
dataToRetrieve.setFieldsToRetrieve(fieldsToRetrieve);
|
||||
|
||||
TimelineEntity timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(appTimelineEntity.getCreatedTime(), timelineEntity
|
||||
.getCreatedTime().longValue());
|
||||
Assert.assertEquals(0, timelineEntity .getMetrics().size());
|
||||
Assert.assertEquals(0, timelineEntity.getEvents().size());
|
||||
Assert.assertEquals(0, timelineEntity.getConfigs().size());
|
||||
Assert.assertEquals(appTimelineEntity.getInfo().size(),
|
||||
timelineEntity.getInfo().size());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testGetEntityCustomField() throws Exception {
|
||||
context.setEntityType(TimelineEntityType.YARN_CONTAINER.toString());
|
||||
timelineReader.serviceInit(conf);
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
dataToRetrieve.getFieldsToRetrieve().add(TimelineReader.Field.METRICS);
|
||||
|
||||
TimelineEntity timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(appTimelineEntity.getCreatedTime(), timelineEntity
|
||||
.getCreatedTime().longValue());
|
||||
Assert.assertEquals(appTimelineEntity.getMetrics().size(),
|
||||
timelineEntity.getMetrics().size());
|
||||
Assert.assertEquals(0, timelineEntity.getEvents().size());
|
||||
Assert.assertEquals(0, timelineEntity.getConfigs().size());
|
||||
Assert.assertEquals(appTimelineEntity.getInfo().size(),
|
||||
timelineEntity.getInfo().size());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testGetEntityAllFields() throws Exception {
|
||||
context.setEntityType(TimelineEntityType.YARN_CONTAINER.toString());
|
||||
timelineReader.serviceInit(conf);
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
dataToRetrieve.getFieldsToRetrieve().add(TimelineReader.Field.ALL);
|
||||
|
||||
TimelineEntity timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(appTimelineEntity.getCreatedTime(), timelineEntity
|
||||
.getCreatedTime().longValue());
|
||||
Assert.assertEquals(appTimelineEntity.getMetrics().size(),
|
||||
timelineEntity .getMetrics().size());
|
||||
Assert.assertEquals(appTimelineEntity.getEvents().size(),
|
||||
timelineEntity.getEvents().size());
|
||||
Assert.assertEquals(appTimelineEntity.getConfigs().size(),
|
||||
timelineEntity.getConfigs().size());
|
||||
Assert.assertEquals(appTimelineEntity.getInfo().size(),
|
||||
timelineEntity.getInfo().size());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testGetAllEntities() throws Exception {
|
||||
context.setEntityType(TimelineEntityType.YARN_CONTAINER.toString());
|
||||
timelineReader.serviceInit(conf);
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
dataToRetrieve.getFieldsToRetrieve().add(TimelineReader.Field.ALL);
|
||||
|
||||
Set<TimelineEntity> actualEntities = timelineReader.getEntities(context,
|
||||
new TimelineEntityFilters.Builder().build(), dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(entities.size(), actualEntities.size());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testGetEntitiesWithLimit() throws Exception {
|
||||
context.setEntityType(TimelineEntityType.YARN_CONTAINER.toString());
|
||||
timelineReader.serviceInit(conf);
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
|
||||
Set<TimelineEntity> actualEntities = timelineReader.getEntities(context,
|
||||
new TimelineEntityFilters.Builder().entityLimit(2L).build(),
|
||||
dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(2, actualEntities.size());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testGetEntitiesByWindows() throws Exception {
|
||||
context.setEntityType(TimelineEntityType.YARN_CONTAINER.toString());
|
||||
timelineReader.serviceInit(conf);
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
|
||||
Set<TimelineEntity> actualEntities = timelineReader.getEntities(context,
|
||||
new TimelineEntityFilters.Builder().createdTimeBegin(1533985554927L)
|
||||
.createTimeEnd(1533985554927L).build(), dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(1, actualEntities.size());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testGetFilteredEntities() throws Exception {
|
||||
|
||||
context.setEntityType(TimelineEntityType.YARN_CONTAINER.toString());
|
||||
timelineReader.serviceInit(conf);
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
dataToRetrieve.getFieldsToRetrieve().add(TimelineReader.Field.ALL);
|
||||
|
||||
// Get entities based on info filters.
|
||||
TimelineFilterList infoFilterList = new TimelineFilterList();
|
||||
infoFilterList.addFilter(
|
||||
new TimelineKeyValueFilter(TimelineCompareOp.EQUAL,
|
||||
"YARN_APPLICATION_ATTEMPT_FINAL_STATUS", "SUCCEEDED"));
|
||||
Set<TimelineEntity> actualEntities = timelineReader.getEntities(context,
|
||||
new TimelineEntityFilters.Builder().infoFilters(infoFilterList).build(),
|
||||
dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(1, actualEntities.size());
|
||||
// Only one entity with type YARN_APPLICATION_ATTEMPT should be returned.
|
||||
for (TimelineEntity entity : actualEntities) {
|
||||
if (!entity.getType().equals("YARN_APPLICATION_ATTEMPT")) {
|
||||
Assert.fail("Incorrect filtering based on info filters");
|
||||
}
|
||||
}
|
||||
|
||||
// Get entities based on config filters.
|
||||
TimelineFilterList confFilterList = new TimelineFilterList();
|
||||
context.setEntityType(TimelineEntityType.YARN_APPLICATION.toString());
|
||||
confFilterList.addFilter(
|
||||
new TimelineKeyValueFilter(TimelineCompareOp.EQUAL,
|
||||
"YARN_AM_NODE_LABEL_EXPRESSION", "<DEFAULT_PARTITION>"));
|
||||
actualEntities = timelineReader.getEntities(context,
|
||||
new TimelineEntityFilters.Builder().configFilters(confFilterList)
|
||||
.build(), dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(1, actualEntities.size());
|
||||
// Only one entity with type YARN_APPLICATION should be returned.
|
||||
for (TimelineEntity entity : actualEntities) {
|
||||
if (!entity.getType().equals("YARN_APPLICATION")) {
|
||||
Assert.fail("Incorrect filtering based on info filters");
|
||||
}
|
||||
}
|
||||
|
||||
// Get entities based on event filters.
|
||||
context.setEntityType(TimelineEntityType.YARN_CONTAINER.toString());
|
||||
TimelineFilterList eventFilters = new TimelineFilterList();
|
||||
eventFilters.addFilter(
|
||||
new TimelineExistsFilter(TimelineCompareOp.EQUAL,
|
||||
"CONTAINER_LAUNCHED"));
|
||||
actualEntities = timelineReader.getEntities(context,
|
||||
new TimelineEntityFilters.Builder().eventFilters(eventFilters).build(),
|
||||
dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(1, actualEntities.size());
|
||||
// Only one entity with type YARN_CONTAINER should be returned.
|
||||
for (TimelineEntity entity : actualEntities) {
|
||||
if (!entity.getType().equals("YARN_CONTAINER")) {
|
||||
Assert.fail("Incorrect filtering based on info filters");
|
||||
}
|
||||
}
|
||||
|
||||
// Get entities based on metric filters.
|
||||
TimelineFilterList metricFilterList = new TimelineFilterList();
|
||||
metricFilterList.addFilter(new TimelineCompareFilter(
|
||||
TimelineCompareOp.GREATER_OR_EQUAL, "MEMORY", 150298624L));
|
||||
actualEntities = timelineReader.getEntities(context,
|
||||
new TimelineEntityFilters.Builder().metricFilters(metricFilterList)
|
||||
.build(), dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(1, actualEntities.size());
|
||||
// Only one entity with type YARN_CONTAINER should be returned.
|
||||
for (TimelineEntity entity : actualEntities) {
|
||||
if (!entity.getType().equals("YARN_CONTAINER")) {
|
||||
Assert.fail("Incorrect filtering based on info filters");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testReadingDifferentEntityTypes() throws Exception {
|
||||
|
||||
timelineReader.serviceInit(conf);
|
||||
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
|
||||
// reading YARN_FLOW_ACTIVITY
|
||||
context.setEntityType(TimelineEntityType.YARN_FLOW_ACTIVITY.toString());
|
||||
TimelineEntity timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(TimelineEntityType.YARN_FLOW_ACTIVITY.toString(),
|
||||
timelineEntity.getType());
|
||||
|
||||
// reading YARN_FLOW_RUN
|
||||
context.setEntityType(TimelineEntityType.YARN_FLOW_RUN.toString());
|
||||
timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(TimelineEntityType.YARN_FLOW_RUN.toString(),
|
||||
timelineEntity.getType());
|
||||
|
||||
// reading YARN_APPLICATION
|
||||
context.setEntityType(TimelineEntityType.YARN_APPLICATION.toString());
|
||||
timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
|
||||
Assert.assertEquals(TimelineEntityType.YARN_APPLICATION.toString(),
|
||||
timelineEntity.getType());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testReadingAllEntityTypes() throws Exception {
|
||||
|
||||
timelineReader.serviceInit(conf);
|
||||
|
||||
context.setEntityType(TimelineEntityType.YARN_CONTAINER.toString());
|
||||
Set<String> entityTypes = timelineReader.getEntityTypes(context);
|
||||
Assert.assertTrue(entityTypes.contains(TimelineEntityType.YARN_CONTAINER
|
||||
.toString()));
|
||||
Assert.assertTrue(entityTypes.contains(TimelineEntityType
|
||||
.YARN_APPLICATION_ATTEMPT.toString()));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testMetricsToRetrieve() throws Exception {
|
||||
|
||||
timelineReader.serviceInit(conf);
|
||||
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
dataToRetrieve.getFieldsToRetrieve().add(TimelineReader.Field.METRICS);
|
||||
TimelineFilterList timelineFilterList = new TimelineFilterList();
|
||||
|
||||
//testing metrics prefix for OR condition
|
||||
timelineFilterList.setOperator(TimelineFilterList.Operator.OR);
|
||||
timelineFilterList.addFilter(new TimelinePrefixFilter(
|
||||
TimelineCompareOp.EQUAL, "NOTHING"));
|
||||
dataToRetrieve.setMetricsToRetrieve(timelineFilterList);
|
||||
|
||||
context.setEntityType(TimelineEntityType.YARN_APPLICATION.toString());
|
||||
TimelineEntity timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
Assert.assertEquals(0, timelineEntity.getMetrics().size());
|
||||
|
||||
timelineFilterList.addFilter(new TimelinePrefixFilter(
|
||||
TimelineCompareOp.EQUAL,
|
||||
"YARN_APPLICATION_NON_AM_CONTAINER_PREEMPTED"));
|
||||
dataToRetrieve.setMetricsToRetrieve(timelineFilterList);
|
||||
|
||||
context.setEntityType(TimelineEntityType.YARN_APPLICATION.toString());
|
||||
timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
Assert.assertTrue(timelineEntity.getMetrics().size() > 0);
|
||||
|
||||
//testing metrics prefix for AND condition
|
||||
timelineFilterList.setOperator(TimelineFilterList.Operator.AND);
|
||||
timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
Assert.assertEquals(0, timelineEntity.getMetrics().size());
|
||||
|
||||
dataToRetrieve.getMetricsToRetrieve().getFilterList().remove(0);
|
||||
context.setEntityType(TimelineEntityType.YARN_APPLICATION.toString());
|
||||
timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
Assert.assertTrue(timelineEntity.getMetrics().size() > 0);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testConfigsToRetrieve() throws Exception {
|
||||
|
||||
timelineReader.serviceInit(conf);
|
||||
|
||||
TimelineDataToRetrieve dataToRetrieve = new TimelineDataToRetrieve();
|
||||
dataToRetrieve.getFieldsToRetrieve().add(TimelineReader.Field.CONFIGS);
|
||||
TimelineFilterList timelineFilterList = new TimelineFilterList();
|
||||
|
||||
//testing metrics prefix for OR condition
|
||||
timelineFilterList.setOperator(TimelineFilterList.Operator.OR);
|
||||
timelineFilterList.addFilter(new TimelinePrefixFilter(
|
||||
TimelineCompareOp.EQUAL, "NOTHING"));
|
||||
dataToRetrieve.setConfsToRetrieve(timelineFilterList);
|
||||
|
||||
context.setEntityType(TimelineEntityType.YARN_APPLICATION.toString());
|
||||
TimelineEntity timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
Assert.assertEquals(0, timelineEntity.getConfigs().size());
|
||||
|
||||
timelineFilterList.addFilter(new TimelinePrefixFilter(
|
||||
TimelineCompareOp.EQUAL, "YARN_AM_NODE_LABEL_EXPRESSION"));
|
||||
dataToRetrieve.setConfsToRetrieve(timelineFilterList);
|
||||
|
||||
context.setEntityType(TimelineEntityType.YARN_APPLICATION.toString());
|
||||
timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
Assert.assertTrue(timelineEntity.getConfigs().size() > 0);
|
||||
|
||||
//testing metrics prefix for AND condition
|
||||
timelineFilterList.setOperator(TimelineFilterList.Operator.AND);
|
||||
timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
Assert.assertEquals(0, timelineEntity.getConfigs().size());
|
||||
|
||||
dataToRetrieve.getConfsToRetrieve().getFilterList().remove(0);
|
||||
context.setEntityType(TimelineEntityType.YARN_APPLICATION.toString());
|
||||
timelineEntity = timelineReader.getEntity(context,
|
||||
dataToRetrieve);
|
||||
Assert.assertTrue(timelineEntity.getConfigs().size() > 0);
|
||||
}
|
||||
}
|
|
@ -0,0 +1,90 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.security.UserGroupInformation;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntities;
|
||||
import org.apache.hadoop.yarn.exceptions.YarnException;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.collector.TimelineCollectorContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.lib.DocumentStoreFactory;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.DocumentStoreWriter;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.writer.DummyDocumentStoreWriter;
|
||||
import org.junit.Before;
|
||||
import org.junit.Test;
|
||||
import org.junit.runner.RunWith;
|
||||
import org.mockito.ArgumentMatchers;
|
||||
import org.powermock.api.mockito.PowerMockito;
|
||||
import org.powermock.core.classloader.annotations.PrepareForTest;
|
||||
import org.powermock.modules.junit4.PowerMockRunner;
|
||||
|
||||
/**
|
||||
* Test case for {@link DocumentStoreTimelineWriterImpl}.
|
||||
*/
|
||||
@RunWith(PowerMockRunner.class)
|
||||
@PrepareForTest(DocumentStoreFactory.class)
|
||||
public class TestDocumentStoreTimelineWriterImpl {
|
||||
|
||||
private final DocumentStoreWriter<TimelineDocument> documentStoreWriter = new
|
||||
DummyDocumentStoreWriter<>();
|
||||
private final Configuration conf = new Configuration();
|
||||
|
||||
@Before
|
||||
public void setUp() throws YarnException {
|
||||
conf.set(DocumentStoreUtils.TIMELINE_SERVICE_DOCUMENTSTORE_DATABASE_NAME,
|
||||
"TestDB");
|
||||
conf.set(DocumentStoreUtils.TIMELINE_SERVICE_COSMOSDB_ENDPOINT,
|
||||
"https://localhost:443");
|
||||
conf.set(DocumentStoreUtils.TIMELINE_SERVICE_COSMOSDB_MASTER_KEY,
|
||||
"1234567");
|
||||
PowerMockito.mockStatic(DocumentStoreFactory.class);
|
||||
PowerMockito.when(DocumentStoreFactory.createDocumentStoreWriter(
|
||||
ArgumentMatchers.any(Configuration.class)))
|
||||
.thenReturn(documentStoreWriter);
|
||||
}
|
||||
|
||||
@Test(expected = YarnException.class)
|
||||
public void testFailOnNoCosmosDBConfigs() throws Exception {
|
||||
DocumentStoreUtils.validateCosmosDBConf(new Configuration());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testWritingToCosmosDB() throws Exception {
|
||||
DocumentStoreTimelineWriterImpl timelineWriter = new
|
||||
DocumentStoreTimelineWriterImpl();
|
||||
|
||||
timelineWriter.serviceInit(conf);
|
||||
|
||||
TimelineEntities entities = new TimelineEntities();
|
||||
entities.addEntities(DocumentStoreTestUtils.bakeTimelineEntities());
|
||||
entities.addEntity(DocumentStoreTestUtils.bakeTimelineEntityDoc()
|
||||
.fetchTimelineEntity());
|
||||
|
||||
PowerMockito.verifyStatic(DocumentStoreFactory.class);
|
||||
|
||||
TimelineCollectorContext context = new TimelineCollectorContext();
|
||||
context.setFlowName("TestFlow");
|
||||
context.setAppId("DUMMY_APP_ID");
|
||||
context.setClusterId("yarn_cluster");
|
||||
context.setUserId("test_user");
|
||||
timelineWriter.write(context, entities,
|
||||
UserGroupInformation.createRemoteUser("test_user"));
|
||||
}
|
||||
}
|
|
@ -0,0 +1,177 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.collection;
|
||||
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntityType;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetric;
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineMetricOperation;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.DocumentStoreTestUtils;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEntityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineMetricSubDoc;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity.FlowActivityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowrun.FlowRunDocument;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
/**
|
||||
* Timeline Entity Document merge and aggregation test.
|
||||
*/
|
||||
public class TestDocumentOperations {
|
||||
|
||||
private static final String MEMORY_ID = "MEMORY";
|
||||
private static final String FLOW_NAME = "DistributedShell";
|
||||
private static final String FLOW_VERSION = "1";
|
||||
|
||||
@Test
|
||||
public void testTimelineEntityDocMergeOperation() throws IOException {
|
||||
TimelineEntityDocument actualEntityDoc =
|
||||
new TimelineEntityDocument();
|
||||
TimelineEntityDocument expectedEntityDoc =
|
||||
DocumentStoreTestUtils.bakeTimelineEntityDoc();
|
||||
|
||||
Assert.assertEquals(1, actualEntityDoc.getInfo().size());
|
||||
Assert.assertEquals(0, actualEntityDoc.getMetrics().size());
|
||||
Assert.assertEquals(0, actualEntityDoc.getEvents().size());
|
||||
Assert.assertEquals(0, actualEntityDoc.getConfigs().size());
|
||||
Assert.assertEquals(0, actualEntityDoc.getIsRelatedToEntities().size());
|
||||
Assert.assertEquals(0, actualEntityDoc.getRelatesToEntities().size());
|
||||
|
||||
actualEntityDoc.merge(expectedEntityDoc);
|
||||
|
||||
Assert.assertEquals(expectedEntityDoc.getInfo().size(),
|
||||
actualEntityDoc.getInfo().size());
|
||||
Assert.assertEquals(expectedEntityDoc.getMetrics().size(),
|
||||
actualEntityDoc.getMetrics().size());
|
||||
Assert.assertEquals(expectedEntityDoc.getEvents().size(),
|
||||
actualEntityDoc.getEvents().size());
|
||||
Assert.assertEquals(expectedEntityDoc.getConfigs().size(),
|
||||
actualEntityDoc.getConfigs().size());
|
||||
Assert.assertEquals(expectedEntityDoc.getRelatesToEntities().size(),
|
||||
actualEntityDoc.getIsRelatedToEntities().size());
|
||||
Assert.assertEquals(expectedEntityDoc.getRelatesToEntities().size(),
|
||||
actualEntityDoc.getRelatesToEntities().size());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFlowActivityDocMergeOperation() throws IOException {
|
||||
FlowActivityDocument actualFlowActivityDoc = new FlowActivityDocument();
|
||||
FlowActivityDocument expectedFlowActivityDoc =
|
||||
DocumentStoreTestUtils.bakeFlowActivityDoc();
|
||||
|
||||
Assert.assertEquals(0, actualFlowActivityDoc.getDayTimestamp());
|
||||
Assert.assertEquals(0, actualFlowActivityDoc.getFlowActivities().size());
|
||||
Assert.assertNull(actualFlowActivityDoc.getFlowName());
|
||||
Assert.assertEquals(TimelineEntityType.YARN_FLOW_ACTIVITY.toString(),
|
||||
actualFlowActivityDoc.getType());
|
||||
Assert.assertNull(actualFlowActivityDoc.getUser());
|
||||
Assert.assertNull(actualFlowActivityDoc.getId());
|
||||
|
||||
actualFlowActivityDoc.merge(expectedFlowActivityDoc);
|
||||
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getDayTimestamp(),
|
||||
actualFlowActivityDoc.getDayTimestamp());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getFlowActivities().size(),
|
||||
actualFlowActivityDoc.getFlowActivities().size());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getFlowName(),
|
||||
actualFlowActivityDoc.getFlowName());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getType(),
|
||||
actualFlowActivityDoc.getType());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getUser(),
|
||||
actualFlowActivityDoc.getUser());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getId(),
|
||||
actualFlowActivityDoc.getId());
|
||||
|
||||
expectedFlowActivityDoc.addFlowActivity(FLOW_NAME,
|
||||
FLOW_VERSION, System.currentTimeMillis());
|
||||
|
||||
actualFlowActivityDoc.merge(expectedFlowActivityDoc);
|
||||
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getDayTimestamp(),
|
||||
actualFlowActivityDoc.getDayTimestamp());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getFlowActivities().size(),
|
||||
actualFlowActivityDoc.getFlowActivities().size());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getFlowName(),
|
||||
actualFlowActivityDoc.getFlowName());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getType(),
|
||||
actualFlowActivityDoc.getType());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getUser(),
|
||||
actualFlowActivityDoc.getUser());
|
||||
Assert.assertEquals(expectedFlowActivityDoc.getId(),
|
||||
actualFlowActivityDoc.getId());
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFlowRunDocMergeAndAggOperation() throws IOException {
|
||||
FlowRunDocument actualFlowRunDoc = new FlowRunDocument();
|
||||
FlowRunDocument expectedFlowRunDoc = DocumentStoreTestUtils
|
||||
.bakeFlowRunDoc();
|
||||
|
||||
final long timestamp = System.currentTimeMillis();
|
||||
final long value = 98586624;
|
||||
TimelineMetric timelineMetric = new TimelineMetric();
|
||||
timelineMetric.setId(MEMORY_ID);
|
||||
timelineMetric.setType(TimelineMetric.Type.SINGLE_VALUE);
|
||||
timelineMetric.setRealtimeAggregationOp(TimelineMetricOperation.SUM);
|
||||
timelineMetric.addValue(timestamp, value);
|
||||
TimelineMetricSubDoc metricSubDoc = new TimelineMetricSubDoc(
|
||||
timelineMetric);
|
||||
expectedFlowRunDoc.getMetrics().put(MEMORY_ID, metricSubDoc);
|
||||
|
||||
Assert.assertNull(actualFlowRunDoc.getClusterId());
|
||||
Assert.assertNull(actualFlowRunDoc.getFlowName());
|
||||
Assert.assertNull(actualFlowRunDoc.getFlowRunId());
|
||||
Assert.assertNull(actualFlowRunDoc.getFlowVersion());
|
||||
Assert.assertNull(actualFlowRunDoc.getId());
|
||||
Assert.assertNull(actualFlowRunDoc.getUsername());
|
||||
Assert.assertEquals(actualFlowRunDoc.getType(), TimelineEntityType.
|
||||
YARN_FLOW_RUN.toString());
|
||||
Assert.assertEquals(0, actualFlowRunDoc.getMinStartTime());
|
||||
Assert.assertEquals(0, actualFlowRunDoc.getMaxEndTime());
|
||||
Assert.assertEquals(0, actualFlowRunDoc.getMetrics().size());
|
||||
|
||||
actualFlowRunDoc.merge(expectedFlowRunDoc);
|
||||
|
||||
Assert.assertEquals(expectedFlowRunDoc.getClusterId(),
|
||||
actualFlowRunDoc.getClusterId());
|
||||
Assert.assertEquals(expectedFlowRunDoc.getFlowName(),
|
||||
actualFlowRunDoc.getFlowName());
|
||||
Assert.assertEquals(expectedFlowRunDoc.getFlowRunId(),
|
||||
actualFlowRunDoc.getFlowRunId());
|
||||
Assert.assertEquals(expectedFlowRunDoc.getFlowVersion(),
|
||||
actualFlowRunDoc.getFlowVersion());
|
||||
Assert.assertEquals(expectedFlowRunDoc.getId(), actualFlowRunDoc.getId());
|
||||
Assert.assertEquals(expectedFlowRunDoc.getUsername(),
|
||||
actualFlowRunDoc.getUsername());
|
||||
Assert.assertEquals(expectedFlowRunDoc.getType(),
|
||||
actualFlowRunDoc.getType());
|
||||
Assert.assertEquals(expectedFlowRunDoc.getMinStartTime(),
|
||||
actualFlowRunDoc.getMinStartTime());
|
||||
Assert.assertEquals(expectedFlowRunDoc.getMaxEndTime(),
|
||||
actualFlowRunDoc.getMaxEndTime());
|
||||
Assert.assertEquals(expectedFlowRunDoc.getMetrics().size(),
|
||||
actualFlowRunDoc.getMetrics().size());
|
||||
|
||||
actualFlowRunDoc.merge(expectedFlowRunDoc);
|
||||
|
||||
Assert.assertEquals(value + value, actualFlowRunDoc.getMetrics()
|
||||
.get(MEMORY_ID).getSingleDataValue());
|
||||
}
|
||||
}
|
|
@ -0,0 +1,118 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.reader;
|
||||
|
||||
import org.apache.hadoop.yarn.api.records.timelineservice.TimelineEntityType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.reader.TimelineReaderContext;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.DocumentStoreTestUtils;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.entity.TimelineEntityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowactivity.FlowActivityDocument;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.flowrun.FlowRunDocument;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
import java.util.stream.Collectors;
|
||||
|
||||
/**
|
||||
* Dummy Document Store Reader for mocking backend calls for unit test.
|
||||
*/
|
||||
public class DummyDocumentStoreReader<TimelineDoc extends TimelineDocument>
|
||||
implements DocumentStoreReader<TimelineDoc> {
|
||||
|
||||
private final TimelineEntityDocument entityDoc;
|
||||
private final List<TimelineEntityDocument> entityDocs;
|
||||
private final FlowRunDocument flowRunDoc;
|
||||
private final FlowActivityDocument flowActivityDoc;
|
||||
|
||||
public DummyDocumentStoreReader() {
|
||||
try {
|
||||
entityDoc = DocumentStoreTestUtils.bakeTimelineEntityDoc();
|
||||
entityDocs = DocumentStoreTestUtils.bakeYarnAppTimelineEntities();
|
||||
flowRunDoc = DocumentStoreTestUtils.bakeFlowRunDoc();
|
||||
flowActivityDoc = DocumentStoreTestUtils.bakeFlowActivityDoc();
|
||||
} catch (IOException e) {
|
||||
throw new RuntimeException("Unable to create " +
|
||||
"DummyDocumentStoreReader : ", e);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
@SuppressWarnings("unchecked")
|
||||
public TimelineDoc readDocument(String collectionName, TimelineReaderContext
|
||||
context, Class<TimelineDoc> docClass) {
|
||||
switch (TimelineEntityType.valueOf(context.getEntityType())) {
|
||||
case YARN_FLOW_ACTIVITY:
|
||||
return (TimelineDoc) flowActivityDoc;
|
||||
case YARN_FLOW_RUN:
|
||||
return (TimelineDoc) flowRunDoc;
|
||||
default:
|
||||
return (TimelineDoc) entityDoc;
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
@SuppressWarnings("unchecked")
|
||||
public List<TimelineDoc> readDocumentList(String collectionName,
|
||||
TimelineReaderContext context, Class<TimelineDoc> docClass, long size) {
|
||||
|
||||
switch (TimelineEntityType.valueOf(context.getEntityType())) {
|
||||
case YARN_FLOW_ACTIVITY:
|
||||
List<FlowActivityDocument> flowActivityDocs = new ArrayList<>();
|
||||
flowActivityDocs.add(flowActivityDoc);
|
||||
if (size > flowActivityDocs.size()) {
|
||||
size = flowActivityDocs.size();
|
||||
}
|
||||
return (List<TimelineDoc>) flowActivityDocs.subList(0, (int) size);
|
||||
case YARN_FLOW_RUN:
|
||||
List<FlowRunDocument> flowRunDocs = new ArrayList<>();
|
||||
flowRunDocs.add(flowRunDoc);
|
||||
if (size > flowRunDocs.size()) {
|
||||
size = flowRunDocs.size();
|
||||
}
|
||||
return (List<TimelineDoc>) flowRunDocs.subList(0, (int) size);
|
||||
case YARN_APPLICATION:
|
||||
List<TimelineEntityDocument> applicationEntities = new ArrayList<>();
|
||||
applicationEntities.add(entityDoc);
|
||||
if (size > applicationEntities.size()) {
|
||||
size = applicationEntities.size();
|
||||
}
|
||||
return (List<TimelineDoc>) applicationEntities.subList(0, (int) size);
|
||||
default:
|
||||
if (size > entityDocs.size() || size == -1) {
|
||||
size = entityDocs.size();
|
||||
}
|
||||
return (List<TimelineDoc>) entityDocs.subList(0, (int) size);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public Set<String> fetchEntityTypes(String collectionName,
|
||||
TimelineReaderContext context) {
|
||||
return entityDocs.stream().map(TimelineEntityDocument::getType)
|
||||
.collect(Collectors.toSet());
|
||||
}
|
||||
|
||||
@Override
|
||||
public void close() {
|
||||
}
|
||||
}
|
|
@ -0,0 +1,46 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hadoop.yarn.server.timelineservice.documentstore.writer;
|
||||
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.CollectionType;
|
||||
import org.apache.hadoop.yarn.server.timelineservice.documentstore.collection.document.TimelineDocument;
|
||||
|
||||
/**
|
||||
* Dummy Document Store Writer for mocking backend calls for unit test.
|
||||
*/
|
||||
public class DummyDocumentStoreWriter<Document extends TimelineDocument>
|
||||
implements DocumentStoreWriter<Document> {
|
||||
|
||||
@Override
|
||||
public void createDatabase() {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void createCollection(String collectionName) {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void writeDocument(Document timelineDocument,
|
||||
CollectionType collectionType) {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void close() {
|
||||
}
|
||||
}
|
|
@ -0,0 +1,20 @@
|
|||
{
|
||||
"id": "yarn_cluster!1533859200000!test_user!DistributedShell",
|
||||
"type": "YARN_FLOW_ACTIVITY",
|
||||
"flowActivities": [
|
||||
{
|
||||
"flowName": "DistributedShell",
|
||||
"flowVersion": "1",
|
||||
"flowRunId": 1533871039026
|
||||
},
|
||||
{
|
||||
"flowName": "DistributedShell",
|
||||
"flowVersion": "1",
|
||||
"flowRunId": 1533871599510
|
||||
}
|
||||
],
|
||||
"dayTimestamp": 1533859200000,
|
||||
"user": "test_user",
|
||||
"flowName": "DistributedShell",
|
||||
"createdTime": 1533859200000000
|
||||
}
|
|
@ -0,0 +1,126 @@
|
|||
{
|
||||
"id": "yarn_cluster!test_user!DistributedShell!1533871599510",
|
||||
"type": "YARN_FLOW_RUN",
|
||||
"clusterId": "yarn_cluster",
|
||||
"username": "test_user",
|
||||
"flowName": "DistributedShell",
|
||||
"flowRunId": 1533871599510,
|
||||
"flowVersion": "1",
|
||||
"minStartTime": 1533871599510,
|
||||
"maxEndTime": 1533871614645,
|
||||
"metrics": {
|
||||
"YARN_APPLICATION_NON_AM_CONTAINER_PREEMPTED": {
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_NON_AM_CONTAINER_PREEMPTED",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
},
|
||||
"YARN_APPLICATION_RESOURCE_PREEMPTED_CPU": {
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_RESOURCE_PREEMPTED_CPU",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
},
|
||||
"YARN_APPLICATION_CPU_PREEMPT_METRIC": {
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_CPU_PREEMPT_METRIC",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
},
|
||||
"YARN_APPLICATION_MEMORY": {
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 19858,
|
||||
"id": "YARN_APPLICATION_MEMORY",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 19858
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 19858
|
||||
}
|
||||
},
|
||||
"YARN_APPLICATION_MEM_PREEMPT_METRIC": {
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_MEM_PREEMPT_METRIC",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
},
|
||||
"YARN_APPLICATION_CPU": {
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 19,
|
||||
"id": "YARN_APPLICATION_CPU",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 19
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 19
|
||||
}
|
||||
},
|
||||
"YARN_APPLICATION_RESOURCE_PREEMPTED_MEMORY": {
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_RESOURCE_PREEMPTED_MEMORY",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
},
|
||||
"YARN_APPLICATION_AM_CONTAINER_PREEMPTED": {
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_AM_CONTAINER_PREEMPTED",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
}
|
||||
},
|
||||
"createdTime": 1533871599510
|
||||
}
|
|
@ -0,0 +1,185 @@
|
|||
[
|
||||
{
|
||||
"context": {
|
||||
"clusterId": "yarn_cluster",
|
||||
"userId": "test_user",
|
||||
"flowName": "DistributedShell",
|
||||
"flowRunId": 1533985547564,
|
||||
"appId": "application_1533985489663_0001"
|
||||
},
|
||||
"flowVersion": "1",
|
||||
"subApplicationUser": "test_user",
|
||||
"metrics": {},
|
||||
"events": {
|
||||
"YARN_RM_CONTAINER_REQUESTED_TYPE": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_RM_CONTAINER_REQUESTED_TYPE",
|
||||
"timestamp": 1533985547824,
|
||||
"info": {
|
||||
"YARN_RM_CONTAINER_ALLOCATION_REQUEST_ID": 0
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_ATTEMPT_FINISHED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_APPLICATION_ATTEMPT_FINISHED",
|
||||
"timestamp": 1533985561254,
|
||||
"info": {}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_ATTEMPT_REGISTERED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_APPLICATION_ATTEMPT_REGISTERED",
|
||||
"timestamp": 1533985554927,
|
||||
"info": {}
|
||||
}
|
||||
]
|
||||
},
|
||||
"id": "yarn_cluster!test_user!DistributedShell!1533985547564!application_1533985489663_0001!YARN_APPLICATION_ATTEMPT!appattempt_1533985489663_0001_000001",
|
||||
"type": "YARN_APPLICATION_ATTEMPT",
|
||||
"configs": {},
|
||||
"info": {
|
||||
"SYSTEM_INFO_PARENT_ENTITY": {
|
||||
"type": "YARN_APPLICATION",
|
||||
"id": "application_1533985489663_0001"
|
||||
},
|
||||
"YARN_APPLICATION_ATTEMPT_HOST": "test_user/10.171.19.25",
|
||||
"YARN_APPLICATION_ATTEMPT_MASTER_CONTAINER": "container_1533985489663_0001_01_000001",
|
||||
"YARN_APPLICATION_ATTEMPT_ORIGINAL_TRACKING_URL": "N/A",
|
||||
"YARN_APPLICATION_ATTEMPT_RPC_PORT": -1,
|
||||
"YARN_APPLICATION_ATTEMPT_TRACKING_URL": "http://test_user:8088/proxy/application_1533985489663_0001/",
|
||||
"YARN_APPLICATION_ATTEMPT_DIAGNOSTICS_INFO": "",
|
||||
"YARN_APPLICATION_ATTEMPT_STATE": "FINISHED",
|
||||
"YARN_APPLICATION_ATTEMPT_FINAL_STATUS": "SUCCEEDED"
|
||||
},
|
||||
"createdTime": 1533985554927,
|
||||
"relatesToEntities": {},
|
||||
"isRelatedToEntities": {}
|
||||
},
|
||||
{
|
||||
"context": {
|
||||
"clusterId": "yarn_cluster",
|
||||
"userId": "test_user",
|
||||
"flowName": "DistributedShell",
|
||||
"flowRunId": 1533985547564,
|
||||
"appId": "application_1533985489663_0001"
|
||||
},
|
||||
"flowVersion": "1",
|
||||
"subApplicationUser": "test_user",
|
||||
"metrics": {
|
||||
"MEMORY": [
|
||||
{
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533985556335,
|
||||
"singleDataValue": 150298624,
|
||||
"id": "MEMORY",
|
||||
"type": "SINGLE_VALUE",
|
||||
"realtimeAggregationOp": "SUM",
|
||||
"values": {
|
||||
"1533985556335": 150298624
|
||||
},
|
||||
"valuesJAXB": {
|
||||
"1533985556335": 150298624
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
"events": {
|
||||
"YARN_RM_CONTAINER_CREATED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_RM_CONTAINER_CREATED",
|
||||
"timestamp": 1533985548047,
|
||||
"info": {
|
||||
"YARN_RM_CONTAINER_ALLOCATION_REQUEST_ID": 0
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_CONTAINER_CREATED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_CONTAINER_CREATED",
|
||||
"timestamp": 1533985549474,
|
||||
"info": {
|
||||
"YARN_NM_EVENT_SOURCE": "CONTAINER_EVENT",
|
||||
"YARN_CONTAINER_STATE": "NEW",
|
||||
"YARN_APPLICATION_STATE": "RUNNING"
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_CONTAINER_FINISHED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_CONTAINER_FINISHED",
|
||||
"timestamp": 1533985560616,
|
||||
"info": {
|
||||
"YARN_NM_EVENT_SOURCE": "CONTAINER_EVENT"
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
"id": "yarn_cluster!test_user!DistributedShell!1533985547564!application_1533985489663_0001!YARN_CONTAINER!container_1533985489663_0001_01_000001",
|
||||
"type": "YARN_CONTAINER",
|
||||
"configs": {},
|
||||
"info": {
|
||||
"YARN_CONTAINER_ALLOCATED_PORT": 13076,
|
||||
"YARN_CONTAINER_ALLOCATED_MEMORY": 1024,
|
||||
"YARN_CONTAINER_ALLOCATED_PRIORITY": "0",
|
||||
"YARN_CONTAINER_ALLOCATED_HOST": "test_user",
|
||||
"YARN_CONTAINER_ALLOCATED_HOST_HTTP_ADDRESS": "test_user:8042",
|
||||
"YARN_CONTAINER_ALLOCATED_VCORE": 1,
|
||||
"SYSTEM_INFO_PARENT_ENTITY": {
|
||||
"type": "YARN_APPLICATION_ATTEMPT",
|
||||
"id": "appattempt_1533985489663_0001_000001"
|
||||
},
|
||||
"YARN_CONTAINER_STATE": "COMPLETE",
|
||||
"YARN_CONTAINER_EXIT_STATUS": 0,
|
||||
"YARN_CONTAINER_DIAGNOSTICS_INFO": "",
|
||||
"YARN_CONTAINER_FINISHED_TIME": 1533985560616
|
||||
},
|
||||
"createdTime": 1533985549474,
|
||||
"relatesToEntities": {},
|
||||
"isRelatedToEntities": {}
|
||||
},
|
||||
{
|
||||
"context": {
|
||||
"clusterId": "yarn_cluster",
|
||||
"userId": "test_user",
|
||||
"flowName": "DistributedShell",
|
||||
"flowRunId": 1533985547564,
|
||||
"appId": "application_1533985489663_0001"
|
||||
},
|
||||
"flowVersion": "1",
|
||||
"subApplicationUser": "test_user",
|
||||
"metrics": {},
|
||||
"events": {
|
||||
"CONTAINER_LAUNCHED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "CONTAINER_LAUNCHED",
|
||||
"timestamp": 1533985557747,
|
||||
"info": {
|
||||
"YARN_NM_EVENT_SOURCE": "CONTAINER_EVENT",
|
||||
"YARN_CONTAINER_STATE": "LOCALIZED",
|
||||
"YARN_APPLICATION_STATE": "RUNNING"
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
"id": "yarn_cluster!test_user!DistributedShell!1533985547564!application_1533985489663_0001!YARN_CONTAINER!container_1533985489663_0001_01_000002",
|
||||
"type": "YARN_CONTAINER",
|
||||
"configs": {},
|
||||
"info": {
|
||||
"SYSTEM_INFO_PARENT_ENTITY": {
|
||||
"type": "YARN_APPLICATION_ATTEMPT",
|
||||
"id": "appattempt_1533985489663_0001_000001"
|
||||
}
|
||||
},
|
||||
"createdTime": 0,
|
||||
"relatesToEntities": {},
|
||||
"isRelatedToEntities": {}
|
||||
}
|
||||
]
|
|
@ -0,0 +1,203 @@
|
|||
{
|
||||
"context": {
|
||||
"clusterId": "yarn_cluster",
|
||||
"userId": "test_user",
|
||||
"flowName": "DistributedShell",
|
||||
"flowRunId": 1533871599510,
|
||||
"appId": "application_1533871545292_0001"
|
||||
},
|
||||
"flowVersion": "1",
|
||||
"subApplicationUser": "test_user",
|
||||
"metrics": {
|
||||
"YARN_APPLICATION_NON_AM_CONTAINER_PREEMPTED": [
|
||||
{
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_NON_AM_CONTAINER_PREEMPTED",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_RESOURCE_PREEMPTED_CPU": [
|
||||
{
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_RESOURCE_PREEMPTED_CPU",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_CPU_PREEMPT_METRIC": [
|
||||
{
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_CPU_PREEMPT_METRIC",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_MEMORY": [
|
||||
{
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 19858,
|
||||
"id": "YARN_APPLICATION_MEMORY",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 19858
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 19858
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_MEM_PREEMPT_METRIC": [
|
||||
{
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_MEM_PREEMPT_METRIC",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_CPU": [
|
||||
{
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 19,
|
||||
"id": "YARN_APPLICATION_CPU",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 19
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 19
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_RESOURCE_PREEMPTED_MEMORY": [
|
||||
{
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_RESOURCE_PREEMPTED_MEMORY",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_AM_CONTAINER_PREEMPTED": [
|
||||
{
|
||||
"valid": true,
|
||||
"singleDataTimestamp": 1533871614645,
|
||||
"singleDataValue": 0,
|
||||
"id": "YARN_APPLICATION_AM_CONTAINER_PREEMPTED",
|
||||
"type": "SINGLE_VALUE",
|
||||
"values": {
|
||||
"1533871614645": 0
|
||||
},
|
||||
"realtimeAggregationOp": "NOP",
|
||||
"valuesJAXB": {
|
||||
"1533871614645": 0
|
||||
}
|
||||
}
|
||||
]
|
||||
},
|
||||
"events": {
|
||||
"YARN_APPLICATION_CREATED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_APPLICATION_CREATED",
|
||||
"timestamp": 1533871599510,
|
||||
"info": {}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_ACLS_UPDATED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_APPLICATION_ACLS_UPDATED",
|
||||
"timestamp": 1533871599671,
|
||||
"info": {}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_STATE_UPDATED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_APPLICATION_STATE_UPDATED",
|
||||
"timestamp": 1533871608094,
|
||||
"info": {
|
||||
"YARN_APPLICATION_STATE": "RUNNING"
|
||||
}
|
||||
}
|
||||
],
|
||||
"YARN_APPLICATION_FINISHED": [
|
||||
{
|
||||
"valid": true,
|
||||
"id": "YARN_APPLICATION_FINISHED",
|
||||
"timestamp": 1533871614645,
|
||||
"info": {}
|
||||
}
|
||||
]
|
||||
},
|
||||
"id": "yarn_cluster!test_user!DistributedShell!1533871599510!application_1533871545292_0001!YARN_APPLICATION!application_1533871545292_0001",
|
||||
"type": "YARN_APPLICATION",
|
||||
"createdTime": 1533871599510,
|
||||
"info": {
|
||||
"YARN_APPLICATION_VIEW_ACLS": "",
|
||||
"YARN_APPLICATION_SUBMITTED_TIME": 1533871599446,
|
||||
"YARN_AM_CONTAINER_LAUNCH_COMMAND": [
|
||||
"{{JAVA_HOME}}/bin/java -Xmx10m org.apache.hadoop.yarn.applications.distributedshell.ApplicationMaster --container_memory 10 --container_vcores 1 --num_containers 1 --priority 0 1><LOG_DIR>/AppMaster.stdout 2><LOG_DIR>/AppMaster.stderr "
|
||||
],
|
||||
"YARN_APPLICATION_NAME": "DistributedShell",
|
||||
"YARN_APPLICATION_USER": "test_user",
|
||||
"YARN_APPLICATION_QUEUE": "default",
|
||||
"YARN_APPLICATION_TYPE": "YARN",
|
||||
"YARN_APPLICATION_UNMANAGED_APPLICATION": false,
|
||||
"YARN_APPLICATION_TAGS": [],
|
||||
"YARN_APPLICATION_STATE": "FINISHED",
|
||||
"YARN_APPLICATION_DIAGNOSTICS_INFO": "",
|
||||
"YARN_APPLICATION_FINAL_STATUS": "SUCCEEDED",
|
||||
"YARN_APPLICATION_LATEST_APP_ATTEMPT": "appattempt_1533871545292_0001_000001"
|
||||
},
|
||||
"configs": {
|
||||
"YARN_AM_NODE_LABEL_EXPRESSION": "<DEFAULT_PARTITION>",
|
||||
"YARN_APP_NODE_LABEL_EXPRESSION": "<Not set>"
|
||||
},
|
||||
"relatesToEntities": {},
|
||||
"isRelatedToEntities": {}
|
||||
}
|
|
@ -0,0 +1,119 @@
|
|||
[
|
||||
{
|
||||
"identifier": {
|
||||
"type": "YARN_APPLICATION",
|
||||
"id": "application_1532614298307_0002"
|
||||
},
|
||||
"info": {
|
||||
"YARN_APPLICATION_VIEW_ACLS": ""
|
||||
},
|
||||
"configs": {},
|
||||
"metrics": [],
|
||||
"events": [
|
||||
{
|
||||
"id": "YARN_APPLICATION_ACLS_UPDATED",
|
||||
"info": {},
|
||||
"timestamp": 1532614542444,
|
||||
"valid": true,
|
||||
"infoJAXB": {}
|
||||
}
|
||||
],
|
||||
"id": "application_1532614298307_0002",
|
||||
"type": "YARN_APPLICATION",
|
||||
"valid": true,
|
||||
"configsJAXB": {},
|
||||
"infoJAXB": {
|
||||
"YARN_APPLICATION_VIEW_ACLS": ""
|
||||
},
|
||||
"relatesToEntitiesJAXB": {},
|
||||
"isRelatedToEntitiesJAXB": {},
|
||||
"isrelatedto": {},
|
||||
"relatesto": {},
|
||||
"createdtime": null,
|
||||
"idprefix": 0
|
||||
},
|
||||
{
|
||||
"identifier": {
|
||||
"type": "YARN_APPLICATION_ATTEMPT",
|
||||
"id": "appattempt_1532614298307_0002_000001"
|
||||
},
|
||||
"info": {
|
||||
"YARN_APPLICATION_ATTEMPT_HOST": "test_user/10.171.19.36",
|
||||
"YARN_APPLICATION_ATTEMPT_MASTER_CONTAINER": "container_1532614298307_0002_01_000001",
|
||||
"YARN_APPLICATION_ATTEMPT_ORIGINAL_TRACKING_URL": "N/A",
|
||||
"YARN_APPLICATION_ATTEMPT_RPC_PORT": -1,
|
||||
"YARN_APPLICATION_ATTEMPT_TRACKING_URL": "http://test_user:8088/proxy/application_1532614298307_0002/"
|
||||
},
|
||||
"configs": {},
|
||||
"metrics": [],
|
||||
"events": [
|
||||
{
|
||||
"id": "YARN_APPLICATION_ATTEMPT_REGISTERED",
|
||||
"info": {},
|
||||
"timestamp": 1532614551262,
|
||||
"valid": true,
|
||||
"infoJAXB": {}
|
||||
}
|
||||
],
|
||||
"id": "appattempt_1532614298307_0002_000001",
|
||||
"type": "YARN_APPLICATION_ATTEMPT",
|
||||
"valid": true,
|
||||
"configsJAXB": {},
|
||||
"infoJAXB": {
|
||||
"YARN_APPLICATION_ATTEMPT_HOST": "test_user/10.171.19.36",
|
||||
"YARN_APPLICATION_ATTEMPT_MASTER_CONTAINER": "container_1532614298307_0002_01_000001",
|
||||
"YARN_APPLICATION_ATTEMPT_ORIGINAL_TRACKING_URL": "N/A",
|
||||
"YARN_APPLICATION_ATTEMPT_RPC_PORT": -1,
|
||||
"YARN_APPLICATION_ATTEMPT_TRACKING_URL": "http://test_user:8088/proxy/application_1532614298307_0002/"
|
||||
},
|
||||
"relatesToEntitiesJAXB": {},
|
||||
"isRelatedToEntitiesJAXB": {},
|
||||
"isrelatedto": {},
|
||||
"relatesto": {},
|
||||
"createdtime": 1532614551262,
|
||||
"idprefix": 0
|
||||
},
|
||||
{
|
||||
"identifier": {
|
||||
"type": "YARN_CONTAINER",
|
||||
"id": "container_1532614298307_0002_01_000001"
|
||||
},
|
||||
"info": {
|
||||
"YARN_CONTAINER_ALLOCATED_PORT": 2032,
|
||||
"YARN_CONTAINER_ALLOCATED_MEMORY": 1024,
|
||||
"YARN_CONTAINER_ALLOCATED_PRIORITY": 0,
|
||||
"YARN_CONTAINER_ALLOCATED_HOST": "test_user",
|
||||
"YARN_CONTAINER_ALLOCATED_HOST_HTTP_ADDRESS": "http://test_user:8042",
|
||||
"YARN_CONTAINER_ALLOCATED_VCORE": 1
|
||||
},
|
||||
"configs": {},
|
||||
"metrics": [],
|
||||
"events": [
|
||||
{
|
||||
"id": "YARN_RM_CONTAINER_CREATED",
|
||||
"info": {},
|
||||
"timestamp": 1532614543389,
|
||||
"valid": true,
|
||||
"infoJAXB": {}
|
||||
}
|
||||
],
|
||||
"id": "container_1532614298307_0002_01_000001",
|
||||
"type": "YARN_CONTAINER",
|
||||
"valid": true,
|
||||
"configsJAXB": {},
|
||||
"infoJAXB": {
|
||||
"YARN_CONTAINER_ALLOCATED_PORT": 2032,
|
||||
"YARN_CONTAINER_ALLOCATED_MEMORY": 1024,
|
||||
"YARN_CONTAINER_ALLOCATED_PRIORITY": 0,
|
||||
"YARN_CONTAINER_ALLOCATED_HOST": "test_user",
|
||||
"YARN_CONTAINER_ALLOCATED_HOST_HTTP_ADDRESS": "http://test_user:8042",
|
||||
"YARN_CONTAINER_ALLOCATED_VCORE": 1
|
||||
},
|
||||
"relatesToEntitiesJAXB": {},
|
||||
"isRelatedToEntitiesJAXB": {},
|
||||
"isrelatedto": {},
|
||||
"relatesto": {},
|
||||
"createdtime": 1532614543389,
|
||||
"idprefix": 0
|
||||
}
|
||||
]
|
|
@ -62,7 +62,7 @@ final public class PerNodeAggTimelineCollectorMetrics {
|
|||
synchronized (PerNodeAggTimelineCollectorMetrics.class) {
|
||||
if (instance == null) {
|
||||
instance =
|
||||
DefaultMetricsSystem.initialize("TimelineService").register(
|
||||
DefaultMetricsSystem.instance().register(
|
||||
METRICS_INFO.name(), METRICS_INFO.description(),
|
||||
new PerNodeAggTimelineCollectorMetrics());
|
||||
isInitialized.set(true);
|
||||
|
|
|
@ -46,5 +46,6 @@
|
|||
<module>hadoop-yarn-server-timelineservice-hbase</module>
|
||||
<module>hadoop-yarn-server-timelineservice-hbase-tests</module>
|
||||
<module>hadoop-yarn-server-router</module>
|
||||
<module>hadoop-yarn-server-timelineservice-documentstore</module>
|
||||
</modules>
|
||||
</project>
|
||||
|
|
Loading…
Reference in New Issue