From 3923a4a279565108ac4ad341c031173aa967c603 Mon Sep 17 00:00:00 2001 From: Shashikant Banerjee Date: Wed, 14 Nov 2018 15:50:46 +0530 Subject: [PATCH] HDDS-834. Datanode goes OOM based because of segment size. Contributed by Mukul Kumar Singh. --- .../apache/hadoop/hdds/scm/ScmConfigKeys.java | 2 +- .../src/main/resources/ozone-default.xml | 4 +-- .../server/ratis/ContainerStateMachine.java | 30 ++++++++++++++----- 3 files changed, 25 insertions(+), 11 deletions(-) diff --git a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java index f2cebe928b5..38c41bad606 100644 --- a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java +++ b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/ScmConfigKeys.java @@ -69,7 +69,7 @@ public final class ScmConfigKeys { public static final String DFS_CONTAINER_RATIS_SEGMENT_SIZE_KEY = "dfs.container.ratis.segment.size"; public static final int DFS_CONTAINER_RATIS_SEGMENT_SIZE_DEFAULT = - 1 * 1024 * 1024 * 1024; + 16 * 1024; public static final String DFS_CONTAINER_RATIS_SEGMENT_PREALLOCATED_SIZE_KEY = "dfs.container.ratis.segment.preallocated.size"; public static final int diff --git a/hadoop-hdds/common/src/main/resources/ozone-default.xml b/hadoop-hdds/common/src/main/resources/ozone-default.xml index 512b3ee867c..4a72d397a36 100644 --- a/hadoop-hdds/common/src/main/resources/ozone-default.xml +++ b/hadoop-hdds/common/src/main/resources/ozone-default.xml @@ -175,10 +175,10 @@ dfs.container.ratis.segment.size - 1073741824 + 16384 OZONE, RATIS, PERFORMANCE The size of the raft segment used by Apache Ratis on datanodes. - (1 GB by default) + (16 KB by default) diff --git a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java index 270e164234a..d2d2209937a 100644 --- a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java +++ b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/transport/server/ratis/ContainerStateMachine.java @@ -119,7 +119,8 @@ public class ContainerStateMachine extends BaseStateMachine { private final ConcurrentHashMap createContainerResponseMap; private ExecutorService[] executors; private final int numExecutors; - private final Map containerCommandCompletionMap; + private final Map applyTransactionCompletionMap; + private long lastIndex; /** * CSM metrics. */ @@ -137,7 +138,8 @@ public class ContainerStateMachine extends BaseStateMachine { this.executors = executors.toArray(new ExecutorService[numExecutors]); this.writeChunkFutureMap = new ConcurrentHashMap<>(); this.createContainerResponseMap = new ConcurrentHashMap<>(); - containerCommandCompletionMap = new ConcurrentHashMap<>(); + applyTransactionCompletionMap = new ConcurrentHashMap<>(); + this.lastIndex = RaftServerConstants.INVALID_LOG_INDEX; } @Override @@ -161,10 +163,12 @@ public class ContainerStateMachine extends BaseStateMachine { private long loadSnapshot(SingleFileSnapshotInfo snapshot) { if (snapshot == null) { - TermIndex empty = TermIndex.newTermIndex(0, 0); + TermIndex empty = TermIndex.newTermIndex(0, + RaftServerConstants.INVALID_LOG_INDEX); LOG.info("The snapshot info is null." + "Setting the last applied index to:" + empty); setLastAppliedTermIndex(empty); + lastIndex = RaftServerConstants.INVALID_LOG_INDEX; return RaftServerConstants.INVALID_LOG_INDEX; } @@ -173,6 +177,7 @@ public class ContainerStateMachine extends BaseStateMachine { snapshot.getFile().getPath().toFile()); LOG.info("Setting the last applied index to " + last); setLastAppliedTermIndex(last); + lastIndex = last.getIndex(); return last.getIndex(); } @@ -450,7 +455,7 @@ public class ContainerStateMachine extends BaseStateMachine { Long appliedTerm = null; long appliedIndex = -1; for(long i = getLastAppliedTermIndex().getIndex() + 1;; i++) { - final Long removed = containerCommandCompletionMap.remove(i); + final Long removed = applyTransactionCompletionMap.remove(i); if (removed == null) { break; } @@ -458,7 +463,7 @@ public class ContainerStateMachine extends BaseStateMachine { appliedIndex = i; } if (appliedTerm != null) { - updateLastAppliedTermIndex(appliedIndex, appliedTerm); + updateLastAppliedTermIndex(appliedTerm, appliedIndex); } } @@ -467,7 +472,15 @@ public class ContainerStateMachine extends BaseStateMachine { */ @Override public CompletableFuture applyTransaction(TransactionContext trx) { + // ApplyTransaction call can come with an entryIndex much greater than + // lastIndex updated because in between entries in the raft log can be + // appended because raft config persistence. Just add a dummy entry + // for those. long index = trx.getLogEntry().getIndex(); + for (long i = lastIndex + 1; i < index; i++) { + LOG.info("Gap in indexes at:{} detected, adding dummy entries ", i); + applyTransactionCompletionMap.put(i, trx.getLogEntry().getTerm()); + } try { metrics.incNumApplyTransactionsOps(); ContainerCommandRequestProto requestProto = @@ -500,8 +513,8 @@ public class ContainerStateMachine extends BaseStateMachine { getCommandExecutor(requestProto)); } else if (cmdType == Type.CreateContainer) { long containerID = requestProto.getContainerID(); - return CompletableFuture.completedFuture( - createContainerResponseMap.get(containerID)); + future = CompletableFuture.completedFuture( + createContainerResponseMap.remove(containerID)); } else { // Make sure that in write chunk, the user data is not set if (cmdType == Type.WriteChunk) { @@ -512,9 +525,10 @@ public class ContainerStateMachine extends BaseStateMachine { getCommandExecutor(requestProto)); } + lastIndex = index; future.thenAccept(m -> { final Long previous = - containerCommandCompletionMap + applyTransactionCompletionMap .put(index, trx.getLogEntry().getTerm()); Preconditions.checkState(previous == null); updateLastApplied();