HDFS-11419. Performance analysis of new DFSNetworkTopology#chooseRandom. Contributed by Chen Liang.
This commit is contained in:
parent
ca6bcc3c76
commit
d0f346af26
|
@ -0,0 +1,524 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hdfs.net;
|
||||||
|
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.StorageType;
|
||||||
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
|
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor;
|
||||||
|
import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeStorageInfo;
|
||||||
|
import org.apache.hadoop.net.NetworkTopology;
|
||||||
|
import org.apache.hadoop.net.Node;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.BeforeClass;
|
||||||
|
import org.junit.Ignore;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
import java.text.NumberFormat;
|
||||||
|
import java.util.EnumSet;
|
||||||
|
import java.util.HashSet;
|
||||||
|
import java.util.Random;
|
||||||
|
import java.util.Set;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertNotNull;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Performance test of the new DFSNetworkTopology chooseRandom.
|
||||||
|
*
|
||||||
|
* NOTE that the tests are not for correctness but for performance comparison,
|
||||||
|
* so the tests are printing and writing down values rather than doing assertion
|
||||||
|
* checks or timeout checks. Therefore, it is pointless to run these
|
||||||
|
* tests without something reading the value. So disabled the tests to for now,
|
||||||
|
* anyone interested in looking at the numbers can enable them.
|
||||||
|
*/
|
||||||
|
@Ignore
|
||||||
|
public class TestDFSNetworkTopologyPerformance {
|
||||||
|
public static final Logger LOG =
|
||||||
|
LoggerFactory.getLogger(TestDFSNetworkTopologyPerformance.class);
|
||||||
|
private static NetworkTopology cluster;
|
||||||
|
private static DFSNetworkTopology dfscluster;
|
||||||
|
private DatanodeDescriptor[] dataNodes;
|
||||||
|
|
||||||
|
private final static int NODE_NUM = 2000;
|
||||||
|
private final static int OP_NUM = 20000;
|
||||||
|
|
||||||
|
private final static int L1_NUM = 5;
|
||||||
|
private final static int L2_NUM = 10;
|
||||||
|
private final static int L3_NUM = 10;
|
||||||
|
|
||||||
|
private final static float NS_TO_MS = 1000000;
|
||||||
|
|
||||||
|
private final static Random RANDOM = new Random();
|
||||||
|
|
||||||
|
private Node node;
|
||||||
|
private long totalStart;
|
||||||
|
private long totalEnd;
|
||||||
|
private int totalTrials;
|
||||||
|
private float totalMs;
|
||||||
|
private Set<Node> excluded;
|
||||||
|
private static String[] racks;
|
||||||
|
private static String[] hosts;
|
||||||
|
private static StorageType[] types;
|
||||||
|
|
||||||
|
private static long[] records;
|
||||||
|
private long localStart;
|
||||||
|
private long localEnd;
|
||||||
|
|
||||||
|
|
||||||
|
@BeforeClass
|
||||||
|
public static void init() throws Exception {
|
||||||
|
racks = new String[NODE_NUM];
|
||||||
|
hosts = new String[NODE_NUM];
|
||||||
|
types = new StorageType[NODE_NUM];
|
||||||
|
records = new long[OP_NUM];
|
||||||
|
for (int i = 0; i < NODE_NUM; i++) {
|
||||||
|
racks[i] = getRandLocation();
|
||||||
|
hosts[i] = "host" + i;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Before
|
||||||
|
public void setup() throws Exception {
|
||||||
|
cluster = NetworkTopology.getInstance(new Configuration());
|
||||||
|
dfscluster = DFSNetworkTopology.getInstance(new Configuration());
|
||||||
|
excluded = new HashSet<>();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* In this test, all the node types are uniformly distributed. i.e. a node
|
||||||
|
* has equal of being DISK, RAM_DISK, SSD and ARCHIVE. This test did two tests
|
||||||
|
* first test runs the old chooseRandom approach, the second test runs the
|
||||||
|
* new chooseRandom approach.
|
||||||
|
* @throws Exception
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
public void testUniformStorageType() throws Exception {
|
||||||
|
EnumSet<StorageType> missingTypes = EnumSet.allOf(StorageType.class);
|
||||||
|
|
||||||
|
for (int i = 0; i < NODE_NUM; i++) {
|
||||||
|
types[i] = getRandType();
|
||||||
|
missingTypes.remove(types);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (missingTypes.size() != 0) {
|
||||||
|
// it is statistically very, very rare that missingTypes is not length 0
|
||||||
|
// at this point. But if it happened, we do the trick by inserting 1 per
|
||||||
|
// remaining types randomly into the array, such that it is guaranteed all
|
||||||
|
// types exist
|
||||||
|
Set<Integer> usedIdx = new HashSet<>();
|
||||||
|
int idx;
|
||||||
|
for (StorageType type : missingTypes) {
|
||||||
|
do {
|
||||||
|
idx = RANDOM.nextInt(NODE_NUM);
|
||||||
|
} while (usedIdx.contains(idx));
|
||||||
|
usedIdx.add(idx);
|
||||||
|
types[idx] = type;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
addNodeByTypes(types);
|
||||||
|
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("before test1");
|
||||||
|
|
||||||
|
totalStart = System.nanoTime();
|
||||||
|
totalTrials = 0;
|
||||||
|
for (int i = 0; i < OP_NUM; i++) {
|
||||||
|
StorageType type = StorageType.values()[i%StorageType.values().length];
|
||||||
|
// mimic the behaviour of current code:
|
||||||
|
// 1. chooseRandom on NetworkTopology
|
||||||
|
// 2. if it satisfies, we are good, break
|
||||||
|
// 3. if not, add to excluded, try again
|
||||||
|
// The reason of doing simulated behaviour is that in NetworkTopology,
|
||||||
|
// the code that does this is a wrapper method that is fairly complex
|
||||||
|
// and does something more than just chooseRandom. So it would be unfair
|
||||||
|
// to the current code if we compare the wrapper with the new chooseRandom
|
||||||
|
// because we should only compare the two chooseRandom methods.
|
||||||
|
// However the way current chooseRandom works involves checking storage
|
||||||
|
// type of return nodes. We still need to simulate this behaviour.
|
||||||
|
localStart = System.nanoTime();
|
||||||
|
do {
|
||||||
|
totalTrials += 1;
|
||||||
|
node = cluster.chooseRandom("", excluded);
|
||||||
|
assertNotNull(node);
|
||||||
|
if (isType(node, type)) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
excluded.add(node);
|
||||||
|
} while (true);
|
||||||
|
excluded.clear();
|
||||||
|
localEnd = System.nanoTime();
|
||||||
|
records[i] = localEnd - localStart;
|
||||||
|
}
|
||||||
|
totalEnd = System.nanoTime();
|
||||||
|
totalMs = (totalEnd - totalStart)/NS_TO_MS;
|
||||||
|
// 4 trials on average.
|
||||||
|
LOG.info("total time: {} avg time: {} avg trials: {}",
|
||||||
|
totalMs, totalMs / OP_NUM, (float)totalTrials / OP_NUM);
|
||||||
|
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("after test1 before test2");
|
||||||
|
|
||||||
|
totalStart = System.nanoTime();
|
||||||
|
for (int i = 0; i < OP_NUM; i++) {
|
||||||
|
StorageType type = StorageType.values()[i%StorageType.values().length];
|
||||||
|
localStart = System.nanoTime();
|
||||||
|
node = dfscluster.chooseRandomWithStorageType("", excluded, type);
|
||||||
|
assertNotNull(node);
|
||||||
|
// with dfs cluster, the returned is always already the required type;
|
||||||
|
// add assertion mainly to make a more fair comparison
|
||||||
|
assertTrue(isType(node, type));
|
||||||
|
localEnd = System.nanoTime();
|
||||||
|
records[i] = localEnd - localStart;
|
||||||
|
}
|
||||||
|
totalEnd = System.nanoTime();
|
||||||
|
totalMs = (totalEnd - totalStart)/NS_TO_MS;
|
||||||
|
LOG.info("total time: {} avg time: {}", totalMs, totalMs / OP_NUM);
|
||||||
|
|
||||||
|
printMemUsage("after test2");
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* There are two storage types DISK and ARCHIVE. DISK compose the majority and
|
||||||
|
* ARCHIVE only compose 1/20 nodes, uniformly distributed. This test also runs
|
||||||
|
* two tests, one with old approach and one with the new approach. Try to
|
||||||
|
* search for ARCHIVE type devices. This test show how new approach can
|
||||||
|
* outperform the old one.
|
||||||
|
* @throws Exception
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
public void testUnbalancedStorageType() throws Exception {
|
||||||
|
for (int i = 0; i < NODE_NUM; i++) {
|
||||||
|
types[i] = StorageType.DISK;
|
||||||
|
}
|
||||||
|
for (int i = 0; i < NODE_NUM/20; i++) {
|
||||||
|
types[RANDOM.nextInt(NODE_NUM)] = StorageType.ARCHIVE;
|
||||||
|
}
|
||||||
|
addNodeByTypes(types);
|
||||||
|
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("before test1");
|
||||||
|
|
||||||
|
totalStart = System.nanoTime();
|
||||||
|
totalTrials = 0;
|
||||||
|
for (int i = 0; i < OP_NUM; i++) {
|
||||||
|
// mimic the behaviour of current code:
|
||||||
|
// 1. chooseRandom on NetworkTopology
|
||||||
|
// 2. if it satisfies, we are good, break
|
||||||
|
// 3. if not, add to excluded, try again
|
||||||
|
localStart = System.nanoTime();
|
||||||
|
do {
|
||||||
|
totalTrials += 1;
|
||||||
|
node = cluster.chooseRandom("", excluded);
|
||||||
|
assertNotNull(node);
|
||||||
|
if (isType(node, StorageType.ARCHIVE)) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
excluded.add(node);
|
||||||
|
} while (true);
|
||||||
|
excluded.clear();
|
||||||
|
localEnd = System.nanoTime();
|
||||||
|
records[i] = localEnd - localStart;
|
||||||
|
}
|
||||||
|
totalEnd = System.nanoTime();
|
||||||
|
totalMs = (totalEnd - totalStart)/NS_TO_MS;
|
||||||
|
// on average it takes 20 trials
|
||||||
|
LOG.info("total time: {} avg time: {} avg trials: {}",
|
||||||
|
totalMs, totalMs / OP_NUM, (float)totalTrials / OP_NUM);
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("after test1 before test2");
|
||||||
|
|
||||||
|
totalStart = System.nanoTime();
|
||||||
|
for (int i = 0; i < OP_NUM; i++) {
|
||||||
|
localStart = System.nanoTime();
|
||||||
|
node = dfscluster.chooseRandomWithStorageType("", excluded,
|
||||||
|
StorageType.ARCHIVE);
|
||||||
|
assertNotNull(node);
|
||||||
|
// with dfs cluster, the returned is always already the required type;
|
||||||
|
// add assertion mainly to make a more fair comparison
|
||||||
|
assertTrue(isType(node, StorageType.ARCHIVE));
|
||||||
|
localEnd = System.nanoTime();
|
||||||
|
records[i] = localEnd - localStart;
|
||||||
|
}
|
||||||
|
totalEnd = System.nanoTime();
|
||||||
|
totalMs = (totalEnd - totalStart)/NS_TO_MS;
|
||||||
|
LOG.info("total time: {} avg time: {}", totalMs, totalMs/OP_NUM);
|
||||||
|
|
||||||
|
printMemUsage("after test2");
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* There is only one storage type (DISK) in the cluster. And tries to select
|
||||||
|
* a DISK devices every time also. One approach will always succeed in one
|
||||||
|
* call. This test shows how bad the new approach is compared to the optimal
|
||||||
|
* situation.
|
||||||
|
* @throws Exception
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
public void testSameStorageType() throws Exception {
|
||||||
|
for (int i = 0; i < NODE_NUM; i++) {
|
||||||
|
types[i] = StorageType.DISK;
|
||||||
|
}
|
||||||
|
addNodeByTypes(types);
|
||||||
|
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("before test1");
|
||||||
|
|
||||||
|
totalStart = System.nanoTime();
|
||||||
|
totalTrials = 0;
|
||||||
|
for (int i = 0; i < OP_NUM; i++) {
|
||||||
|
// mimic the behaviour of current code:
|
||||||
|
// 1. chooseRandom on NetworkTopology
|
||||||
|
// 2. if it satisfies, we are good, break
|
||||||
|
// 3. if not, add to excluded, try again
|
||||||
|
do {
|
||||||
|
totalTrials += 1;
|
||||||
|
node = cluster.chooseRandom("", excluded);
|
||||||
|
assertNotNull(node);
|
||||||
|
if (isType(node, StorageType.DISK)) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
excluded.add(node);
|
||||||
|
} while (true);
|
||||||
|
excluded.clear();
|
||||||
|
}
|
||||||
|
totalEnd = System.nanoTime();
|
||||||
|
totalMs = (totalEnd - totalStart)/NS_TO_MS;
|
||||||
|
// on average it takes 20 trials
|
||||||
|
LOG.info("total time: {} avg time: {} avg trials: {}",
|
||||||
|
totalMs, totalMs / OP_NUM, (float)totalTrials / OP_NUM);
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("after test1 before test2");
|
||||||
|
|
||||||
|
totalStart = System.nanoTime();
|
||||||
|
for (int i = 0; i < OP_NUM; i++) {
|
||||||
|
node = dfscluster.chooseRandomWithStorageType("", excluded,
|
||||||
|
StorageType.DISK);
|
||||||
|
assertNotNull(node);
|
||||||
|
// with dfs cluster, the returned is always already the required type;
|
||||||
|
// add assertion mainly to make a more fair comparison
|
||||||
|
assertTrue(isType(node, StorageType.DISK));
|
||||||
|
}
|
||||||
|
totalEnd = System.nanoTime();
|
||||||
|
totalMs = (totalEnd - totalStart) / NS_TO_MS;
|
||||||
|
LOG.info("total time: {} avg time: {}", totalMs, totalMs / OP_NUM);
|
||||||
|
|
||||||
|
printMemUsage("after test2");
|
||||||
|
}
|
||||||
|
|
||||||
|
private boolean coinFlip(double chance) {
|
||||||
|
return RANDOM.nextInt(NODE_NUM) <= NODE_NUM*chance;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This is a helper test, can be changed to different distribution each run.
|
||||||
|
* Changing the value percentage = X where X is between 0.0 to 1.0 will result
|
||||||
|
* in different outcomes. This is to help understanding what is the boundary
|
||||||
|
* that makes the new approach better than the old one. The lower X is, the
|
||||||
|
* less likely the old approach will succeed in one call, in which case the
|
||||||
|
* new approach is more likely to be better.
|
||||||
|
* @throws Exception
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
public void testPercentageStorageType() throws Exception {
|
||||||
|
double percentage = 0.9;
|
||||||
|
for (int i = 0; i < NODE_NUM; i++) {
|
||||||
|
if (coinFlip(percentage)) {
|
||||||
|
types[i] = StorageType.ARCHIVE;
|
||||||
|
} else {
|
||||||
|
types[i] = StorageType.DISK;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
addNodeByTypes(types);
|
||||||
|
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("before test1");
|
||||||
|
|
||||||
|
totalStart = System.nanoTime();
|
||||||
|
totalTrials = 0;
|
||||||
|
for (int i = 0; i < OP_NUM; i++) {
|
||||||
|
// mimic the behaviour of current code:
|
||||||
|
// 1. chooseRandom on NetworkTopology
|
||||||
|
// 2. if it satisfies, we are good, break
|
||||||
|
// 3. if not, add to excluded, try again
|
||||||
|
localStart = System.nanoTime();
|
||||||
|
do {
|
||||||
|
totalTrials += 1;
|
||||||
|
node = cluster.chooseRandom("", excluded);
|
||||||
|
assertNotNull(node);
|
||||||
|
if (isType(node, StorageType.ARCHIVE)) {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
excluded.add(node);
|
||||||
|
} while (true);
|
||||||
|
excluded.clear();
|
||||||
|
localEnd = System.nanoTime();
|
||||||
|
records[i] = localEnd - localStart;
|
||||||
|
}
|
||||||
|
totalEnd = System.nanoTime();
|
||||||
|
totalMs = (totalEnd - totalStart) / NS_TO_MS;
|
||||||
|
// on average it takes 20 trials
|
||||||
|
LOG.info("total time: {} avg time: {} avg trials: {}",
|
||||||
|
totalMs, totalMs / OP_NUM, (float)totalTrials / OP_NUM);
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("after test1 before test2");
|
||||||
|
|
||||||
|
totalStart = System.nanoTime();
|
||||||
|
for (int i = 0; i < OP_NUM; i++) {
|
||||||
|
localStart = System.nanoTime();
|
||||||
|
node = dfscluster.chooseRandomWithStorageType("", excluded,
|
||||||
|
StorageType.ARCHIVE);
|
||||||
|
assertNotNull(node);
|
||||||
|
// with dfs cluster, the returned is always already the required type;
|
||||||
|
// add assertion mainly to make a more fair comparison
|
||||||
|
assertTrue(isType(node, StorageType.ARCHIVE));
|
||||||
|
localEnd = System.nanoTime();
|
||||||
|
records[i] = localEnd - localStart;
|
||||||
|
}
|
||||||
|
totalEnd = System.nanoTime();
|
||||||
|
totalMs = (totalEnd - totalStart) / NS_TO_MS;
|
||||||
|
LOG.info("total time: {} avg time: {}", totalMs, totalMs / OP_NUM);
|
||||||
|
|
||||||
|
printMemUsage("after test2");
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Similar to the previous test, change the percentage value to understand
|
||||||
|
* the performance of the mixed approach. More specifically, this test takes
|
||||||
|
* the approach that, it uses old approach for the first try, only if the
|
||||||
|
* old approach failed in the first try, it makes another call with the
|
||||||
|
* new approach. There is no comparison within this test.
|
||||||
|
* @throws Exception
|
||||||
|
*/
|
||||||
|
@Test
|
||||||
|
public void testPercentageStorageTypeWithMixedTopology() throws Exception {
|
||||||
|
double percentage = 0.9;
|
||||||
|
for (int i = 0; i < NODE_NUM; i++) {
|
||||||
|
if (coinFlip(percentage)) {
|
||||||
|
types[i] = StorageType.ARCHIVE;
|
||||||
|
} else {
|
||||||
|
types[i] = StorageType.DISK;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
addNodeByTypes(types);
|
||||||
|
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("before test1");
|
||||||
|
|
||||||
|
totalStart = System.nanoTime();
|
||||||
|
totalTrials = 0;
|
||||||
|
for (int i = 0; i < OP_NUM; i++) {
|
||||||
|
// mimic the behavior of current code:
|
||||||
|
// 1. chooseRandom on NetworkTopology
|
||||||
|
// 2. if it satisfies, we are good, do the next operation
|
||||||
|
// 3. if not, chooseRandom on DFSNetworkTopology
|
||||||
|
localStart = System.nanoTime();
|
||||||
|
|
||||||
|
totalTrials += 1;
|
||||||
|
node = cluster.chooseRandom("", excluded);
|
||||||
|
assertNotNull(node);
|
||||||
|
if (!isType(node, StorageType.ARCHIVE)) {
|
||||||
|
totalTrials += 1;
|
||||||
|
excluded.add(node);
|
||||||
|
node = dfscluster.chooseRandomWithStorageType("", excluded,
|
||||||
|
StorageType.ARCHIVE);
|
||||||
|
}
|
||||||
|
assertTrue(isType(node, StorageType.ARCHIVE));
|
||||||
|
|
||||||
|
excluded.clear();
|
||||||
|
localEnd = System.nanoTime();
|
||||||
|
records[i] = localEnd - localStart;
|
||||||
|
}
|
||||||
|
totalEnd = System.nanoTime();
|
||||||
|
totalMs = (totalEnd - totalStart)/NS_TO_MS;
|
||||||
|
LOG.info("total time: {} avg time: {} avg trials: {}",
|
||||||
|
totalMs, totalMs / OP_NUM, (float)totalTrials / OP_NUM);
|
||||||
|
// wait a bit for things to become stable
|
||||||
|
Thread.sleep(1000);
|
||||||
|
printMemUsage("test StorageType with mixed topology.");
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Print out the memory usage statistics. Note that this is an estimation by
|
||||||
|
* Java Runtime. Should not take the actual value too serious but more focus
|
||||||
|
* on the relative changes.
|
||||||
|
* @param message a prefix message
|
||||||
|
* @throws Exception throws exception
|
||||||
|
*/
|
||||||
|
private void printMemUsage(String message) throws Exception {
|
||||||
|
Runtime runtime = Runtime.getRuntime();
|
||||||
|
NumberFormat format = NumberFormat.getInstance();
|
||||||
|
StringBuilder sb = new StringBuilder();
|
||||||
|
sb.append(message);
|
||||||
|
|
||||||
|
long maxMemory = runtime.maxMemory();
|
||||||
|
long allocatedMemory = runtime.totalMemory();
|
||||||
|
long freeMemory = runtime.freeMemory();
|
||||||
|
|
||||||
|
sb.append("\nfree memory: " + format.format(freeMemory / 1024));
|
||||||
|
sb.append("\nallocated memory: " + format.format(allocatedMemory / 1024));
|
||||||
|
sb.append("\nmax memory: " + format.format(maxMemory / 1024));
|
||||||
|
sb.append("\ntotal free memory: " + format.format(
|
||||||
|
(freeMemory + (maxMemory - allocatedMemory)) / 1024));
|
||||||
|
LOG.info(sb.toString());
|
||||||
|
}
|
||||||
|
|
||||||
|
private void addNodeByTypes(StorageType[] allTypes) {
|
||||||
|
DatanodeStorageInfo[] storages =
|
||||||
|
DFSTestUtil.createDatanodeStorageInfos(
|
||||||
|
NODE_NUM, racks, hosts, allTypes);
|
||||||
|
dataNodes = DFSTestUtil.toDatanodeDescriptor(storages);
|
||||||
|
for (int i = 0; i < NODE_NUM; i++) {
|
||||||
|
cluster.add(dataNodes[i]);
|
||||||
|
dfscluster.add(dataNodes[i]);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private static String getRandLocation() {
|
||||||
|
int l1 = RANDOM.nextInt(L1_NUM) + 1;
|
||||||
|
int l2 = RANDOM.nextInt(L2_NUM) + 1;
|
||||||
|
int l3 = RANDOM.nextInt(L3_NUM) + 1;
|
||||||
|
return String.format("/%d/%d/%d", l1, l2, l3);
|
||||||
|
}
|
||||||
|
|
||||||
|
private StorageType getRandType() {
|
||||||
|
int len = StorageType.values().length;
|
||||||
|
return StorageType.values()[RANDOM.nextInt(len)];
|
||||||
|
}
|
||||||
|
|
||||||
|
private boolean isType(Node n, StorageType type) {
|
||||||
|
// no need to check n == null, because it's done by the caller
|
||||||
|
if (!(n instanceof DatanodeDescriptor)) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
DatanodeDescriptor dnDescriptor = (DatanodeDescriptor)n;
|
||||||
|
return dnDescriptor.hasStorageType(type);
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue