HDFS-15115. Namenode crash caused by NPE in BlockPlacementPolicyDefault when dynamically change logger to debug. Contributed by wangzhixiang
This commit is contained in:
parent
cf5a7186e5
commit
2190fb1fcf
|
@ -777,9 +777,8 @@ public class BlockPlacementPolicyDefault extends BlockPlacementPolicy {
|
||||||
boolean avoidStaleNodes,
|
boolean avoidStaleNodes,
|
||||||
EnumMap<StorageType, Integer> storageTypes)
|
EnumMap<StorageType, Integer> storageTypes)
|
||||||
throws NotEnoughReplicasException {
|
throws NotEnoughReplicasException {
|
||||||
StringBuilder builder = null;
|
StringBuilder builder = debugLoggingBuilder.get();
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
builder = debugLoggingBuilder.get();
|
|
||||||
builder.setLength(0);
|
builder.setLength(0);
|
||||||
builder.append("[");
|
builder.append("[");
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,89 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
* <p>
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
* <p>
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hdfs.server.blockmanagement;
|
||||||
|
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.StorageType;
|
||||||
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
|
import org.apache.hadoop.net.Node;
|
||||||
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.mockito.invocation.InvocationOnMock;
|
||||||
|
import org.mockito.stubbing.Answer;
|
||||||
|
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.EnumMap;
|
||||||
|
import java.util.HashSet;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Set;
|
||||||
|
|
||||||
|
import static org.mockito.Mockito.doAnswer;
|
||||||
|
import static org.mockito.Mockito.spy;
|
||||||
|
|
||||||
|
public class TestBlockPlacementPolicyDebugLoggingBuilder extends
|
||||||
|
BaseReplicationPolicyTest {
|
||||||
|
|
||||||
|
public TestBlockPlacementPolicyDebugLoggingBuilder() {
|
||||||
|
this.blockPlacementPolicy = BlockPlacementPolicyDefault.class.getName();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
DatanodeDescriptor[] getDatanodeDescriptors(Configuration conf) {
|
||||||
|
final String[] racks = {
|
||||||
|
"/d1/r1/n1",
|
||||||
|
"/d1/r1/n2",
|
||||||
|
"/d1/r2/n3",
|
||||||
|
};
|
||||||
|
storages = DFSTestUtil.createDatanodeStorageInfos(racks);
|
||||||
|
return DFSTestUtil.toDatanodeDescriptor(storages);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testChooseRandomDynamicallyChangeLogger() throws Exception {
|
||||||
|
BlockPlacementPolicyDefault repl =
|
||||||
|
spy((BlockPlacementPolicyDefault) replicator);
|
||||||
|
|
||||||
|
GenericTestUtils.setLogLevel(BlockPlacementPolicy.LOG,
|
||||||
|
org.slf4j.event.Level.INFO);
|
||||||
|
List<DatanodeStorageInfo> results = new ArrayList<DatanodeStorageInfo>();
|
||||||
|
results.add(storages[0]);
|
||||||
|
results.add(storages[1]);
|
||||||
|
results.add(storages[2]);
|
||||||
|
Set<Node> excludeNodes = new HashSet<>();
|
||||||
|
doAnswer(new Answer() {
|
||||||
|
@Override
|
||||||
|
public Object answer(InvocationOnMock invocationOnMock) throws Throwable {
|
||||||
|
GenericTestUtils.setLogLevel(BlockPlacementPolicy.LOG,
|
||||||
|
org.slf4j.event.Level.DEBUG);
|
||||||
|
return dataNodes[0];
|
||||||
|
}
|
||||||
|
}).when(repl).chooseDataNode("/", excludeNodes);
|
||||||
|
|
||||||
|
doAnswer(new Answer() {
|
||||||
|
@Override
|
||||||
|
public Object answer(InvocationOnMock invocationOnMock) throws Throwable {
|
||||||
|
GenericTestUtils.setLogLevel(BlockPlacementPolicy.LOG,
|
||||||
|
org.slf4j.event.Level.DEBUG);
|
||||||
|
return dataNodes[0];
|
||||||
|
}
|
||||||
|
}).when(repl).chooseDataNode("/", excludeNodes, StorageType.DISK);
|
||||||
|
EnumMap<StorageType, Integer> types = new EnumMap<>(StorageType.class);
|
||||||
|
types.put(StorageType.DISK, 1);
|
||||||
|
repl.chooseRandom(1, "/", excludeNodes, 1024L, 3, results, false, types);
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue