HDFS-10609. Uncaught InvalidEncryptionKeyException during pipeline recovery may abort downstream applications. Contributed by Wei-Chiu Chuang.
(cherry picked from commit3ae652f821
) (cherry picked from commitbde787db23
)
This commit is contained in:
parent
09964a1629
commit
f0f3b6a66a
|
@ -1772,6 +1772,11 @@ public class DFSClient implements java.io.Closeable, RemotePeerFactory,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public DataEncryptionKey getEncryptionKey() {
|
||||||
|
return encryptionKey;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Get the checksum of the whole file of a range of the file. Note that the
|
* Get the checksum of the whole file of a range of the file. Note that the
|
||||||
* range always starts from the beginning of the file.
|
* range always starts from the beginning of the file.
|
||||||
|
|
|
@ -124,6 +124,89 @@ import javax.annotation.Nonnull;
|
||||||
class DataStreamer extends Daemon {
|
class DataStreamer extends Daemon {
|
||||||
static final Logger LOG = LoggerFactory.getLogger(DataStreamer.class);
|
static final Logger LOG = LoggerFactory.getLogger(DataStreamer.class);
|
||||||
|
|
||||||
|
private class RefetchEncryptionKeyPolicy {
|
||||||
|
private int fetchEncryptionKeyTimes = 0;
|
||||||
|
private InvalidEncryptionKeyException lastException;
|
||||||
|
private final DatanodeInfo src;
|
||||||
|
|
||||||
|
RefetchEncryptionKeyPolicy(DatanodeInfo src) {
|
||||||
|
this.src = src;
|
||||||
|
}
|
||||||
|
boolean continueRetryingOrThrow() throws InvalidEncryptionKeyException {
|
||||||
|
if (fetchEncryptionKeyTimes >= 2) {
|
||||||
|
// hit the same exception twice connecting to the node, so
|
||||||
|
// throw the exception and exclude the node.
|
||||||
|
throw lastException;
|
||||||
|
}
|
||||||
|
// Don't exclude this node just yet.
|
||||||
|
// Try again with a new encryption key.
|
||||||
|
LOG.info("Will fetch a new encryption key and retry, "
|
||||||
|
+ "encryption key was invalid when connecting to "
|
||||||
|
+ this.src + ": ", lastException);
|
||||||
|
// The encryption key used is invalid.
|
||||||
|
dfsClient.clearDataEncryptionKey();
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Record a connection exception.
|
||||||
|
* @param e
|
||||||
|
* @throws InvalidEncryptionKeyException
|
||||||
|
*/
|
||||||
|
void recordFailure(final InvalidEncryptionKeyException e)
|
||||||
|
throws InvalidEncryptionKeyException {
|
||||||
|
fetchEncryptionKeyTimes++;
|
||||||
|
lastException = e;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private class StreamerStreams implements java.io.Closeable {
|
||||||
|
private Socket sock = null;
|
||||||
|
private DataOutputStream out = null;
|
||||||
|
private DataInputStream in = null;
|
||||||
|
|
||||||
|
StreamerStreams(final DatanodeInfo src,
|
||||||
|
final long writeTimeout, final long readTimeout,
|
||||||
|
final Token<BlockTokenIdentifier> blockToken)
|
||||||
|
throws IOException {
|
||||||
|
sock = createSocketForPipeline(src, 2, dfsClient);
|
||||||
|
|
||||||
|
OutputStream unbufOut = NetUtils.getOutputStream(sock, writeTimeout);
|
||||||
|
InputStream unbufIn = NetUtils.getInputStream(sock, readTimeout);
|
||||||
|
IOStreamPair saslStreams = dfsClient.saslClient
|
||||||
|
.socketSend(sock, unbufOut, unbufIn, dfsClient, blockToken, src);
|
||||||
|
unbufOut = saslStreams.out;
|
||||||
|
unbufIn = saslStreams.in;
|
||||||
|
out = new DataOutputStream(new BufferedOutputStream(unbufOut,
|
||||||
|
DFSUtilClient.getSmallBufferSize(dfsClient.getConfiguration())));
|
||||||
|
in = new DataInputStream(unbufIn);
|
||||||
|
}
|
||||||
|
|
||||||
|
void sendTransferBlock(final DatanodeInfo[] targets,
|
||||||
|
final StorageType[] targetStorageTypes,
|
||||||
|
final Token<BlockTokenIdentifier> blockToken) throws IOException {
|
||||||
|
//send the TRANSFER_BLOCK request
|
||||||
|
new Sender(out)
|
||||||
|
.transferBlock(block, blockToken, dfsClient.clientName, targets,
|
||||||
|
targetStorageTypes);
|
||||||
|
out.flush();
|
||||||
|
//ack
|
||||||
|
BlockOpResponseProto transferResponse = BlockOpResponseProto
|
||||||
|
.parseFrom(PBHelperClient.vintPrefixed(in));
|
||||||
|
if (SUCCESS != transferResponse.getStatus()) {
|
||||||
|
throw new IOException("Failed to add a datanode. Response status: "
|
||||||
|
+ transferResponse.getStatus());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void close() throws IOException {
|
||||||
|
IOUtils.closeStream(in);
|
||||||
|
IOUtils.closeStream(out);
|
||||||
|
IOUtils.closeSocket(sock);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Create a socket for a write pipeline
|
* Create a socket for a write pipeline
|
||||||
*
|
*
|
||||||
|
@ -1237,50 +1320,39 @@ class DataStreamer extends Daemon {
|
||||||
new IOException("Failed to add a node");
|
new IOException("Failed to add a node");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private long computeTransferWriteTimeout() {
|
||||||
|
return dfsClient.getDatanodeWriteTimeout(2);
|
||||||
|
}
|
||||||
|
private long computeTransferReadTimeout() {
|
||||||
|
// transfer timeout multiplier based on the transfer size
|
||||||
|
// One per 200 packets = 12.8MB. Minimum is 2.
|
||||||
|
int multi = 2
|
||||||
|
+ (int) (bytesSent / dfsClient.getConf().getWritePacketSize()) / 200;
|
||||||
|
return dfsClient.getDatanodeReadTimeout(multi);
|
||||||
|
}
|
||||||
|
|
||||||
private void transfer(final DatanodeInfo src, final DatanodeInfo[] targets,
|
private void transfer(final DatanodeInfo src, final DatanodeInfo[] targets,
|
||||||
final StorageType[] targetStorageTypes,
|
final StorageType[] targetStorageTypes,
|
||||||
final Token<BlockTokenIdentifier> blockToken)
|
final Token<BlockTokenIdentifier> blockToken)
|
||||||
throws IOException {
|
throws IOException {
|
||||||
//transfer replica to the new datanode
|
//transfer replica to the new datanode
|
||||||
Socket sock = null;
|
RefetchEncryptionKeyPolicy policy = new RefetchEncryptionKeyPolicy(src);
|
||||||
DataOutputStream out = null;
|
do {
|
||||||
DataInputStream in = null;
|
StreamerStreams streams = null;
|
||||||
try {
|
try {
|
||||||
sock = createSocketForPipeline(src, 2, dfsClient);
|
final long writeTimeout = computeTransferWriteTimeout();
|
||||||
final long writeTimeout = dfsClient.getDatanodeWriteTimeout(2);
|
final long readTimeout = computeTransferReadTimeout();
|
||||||
|
|
||||||
// transfer timeout multiplier based on the transfer size
|
streams = new StreamerStreams(src, writeTimeout, readTimeout,
|
||||||
// One per 200 packets = 12.8MB. Minimum is 2.
|
blockToken);
|
||||||
int multi = 2 + (int)(bytesSent /dfsClient.getConf().getWritePacketSize())
|
streams.sendTransferBlock(targets, targetStorageTypes, blockToken);
|
||||||
/ 200;
|
return;
|
||||||
final long readTimeout = dfsClient.getDatanodeReadTimeout(multi);
|
} catch (InvalidEncryptionKeyException e) {
|
||||||
|
policy.recordFailure(e);
|
||||||
OutputStream unbufOut = NetUtils.getOutputStream(sock, writeTimeout);
|
|
||||||
InputStream unbufIn = NetUtils.getInputStream(sock, readTimeout);
|
|
||||||
IOStreamPair saslStreams = dfsClient.saslClient.socketSend(sock,
|
|
||||||
unbufOut, unbufIn, dfsClient, blockToken, src);
|
|
||||||
unbufOut = saslStreams.out;
|
|
||||||
unbufIn = saslStreams.in;
|
|
||||||
out = new DataOutputStream(new BufferedOutputStream(unbufOut,
|
|
||||||
DFSUtilClient.getSmallBufferSize(dfsClient.getConfiguration())));
|
|
||||||
in = new DataInputStream(unbufIn);
|
|
||||||
|
|
||||||
//send the TRANSFER_BLOCK request
|
|
||||||
new Sender(out).transferBlock(block, blockToken, dfsClient.clientName,
|
|
||||||
targets, targetStorageTypes);
|
|
||||||
out.flush();
|
|
||||||
|
|
||||||
//ack
|
|
||||||
BlockOpResponseProto response =
|
|
||||||
BlockOpResponseProto.parseFrom(PBHelperClient.vintPrefixed(in));
|
|
||||||
if (SUCCESS != response.getStatus()) {
|
|
||||||
throw new IOException("Failed to add a datanode");
|
|
||||||
}
|
|
||||||
} finally {
|
} finally {
|
||||||
IOUtils.closeStream(in);
|
IOUtils.closeStream(streams);
|
||||||
IOUtils.closeStream(out);
|
|
||||||
IOUtils.closeSocket(sock);
|
|
||||||
}
|
}
|
||||||
|
} while (policy.continueRetryingOrThrow());
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -49,7 +49,8 @@ public class BlockPoolTokenSecretManager extends
|
||||||
map.put(bpid, secretMgr);
|
map.put(bpid, secretMgr);
|
||||||
}
|
}
|
||||||
|
|
||||||
synchronized BlockTokenSecretManager get(String bpid) {
|
@VisibleForTesting
|
||||||
|
public synchronized BlockTokenSecretManager get(String bpid) {
|
||||||
BlockTokenSecretManager secretMgr = map.get(bpid);
|
BlockTokenSecretManager secretMgr = map.get(bpid);
|
||||||
if (secretMgr == null) {
|
if (secretMgr == null) {
|
||||||
throw new IllegalArgumentException("Block pool " + bpid
|
throw new IllegalArgumentException("Block pool " + bpid
|
||||||
|
|
|
@ -436,6 +436,12 @@ public class BlockTokenSecretManager extends
|
||||||
allKeys.clear();
|
allKeys.clear();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public synchronized boolean hasKey(int keyId) {
|
||||||
|
BlockKey key = allKeys.get(keyId);
|
||||||
|
return key != null;
|
||||||
|
}
|
||||||
|
|
||||||
@VisibleForTesting
|
@VisibleForTesting
|
||||||
public synchronized int getSerialNoForTesting() {
|
public synchronized int getSerialNoForTesting() {
|
||||||
return serialNo;
|
return serialNo;
|
||||||
|
|
|
@ -2708,6 +2708,11 @@ public class DataNode extends ReconfigurableBase
|
||||||
return directoryScanner;
|
return directoryScanner;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public BlockPoolTokenSecretManager getBlockPoolTokenSecretManager() {
|
||||||
|
return blockPoolTokenSecretManager;
|
||||||
|
}
|
||||||
|
|
||||||
public static void secureMain(String args[], SecureResources resources) {
|
public static void secureMain(String args[], SecureResources resources) {
|
||||||
int errorCode = 0;
|
int errorCode = 0;
|
||||||
try {
|
try {
|
||||||
|
|
|
@ -18,24 +18,31 @@
|
||||||
package org.apache.hadoop.hdfs;
|
package org.apache.hadoop.hdfs;
|
||||||
|
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertFalse;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
import static org.junit.Assert.fail;
|
import static org.junit.Assert.fail;
|
||||||
|
import static org.mockito.Mockito.times;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.io.OutputStream;
|
import java.io.OutputStream;
|
||||||
import java.net.InetAddress;
|
import java.net.InetAddress;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
|
import java.util.Arrays;
|
||||||
import java.util.Collection;
|
import java.util.Collection;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
import java.util.concurrent.TimeoutException;
|
||||||
|
|
||||||
|
import com.google.common.base.Supplier;
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.FSDataInputStream;
|
import org.apache.hadoop.fs.FSDataInputStream;
|
||||||
|
import org.apache.hadoop.fs.FSDataOutputStream;
|
||||||
import org.apache.hadoop.fs.FileChecksum;
|
import org.apache.hadoop.fs.FileChecksum;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
|
import org.apache.hadoop.hdfs.client.HdfsClientConfigKeys;
|
||||||
|
import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
|
||||||
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
|
import org.apache.hadoop.hdfs.protocol.LocatedBlock;
|
||||||
import org.apache.hadoop.hdfs.protocol.datatransfer.TrustedChannelResolver;
|
import org.apache.hadoop.hdfs.protocol.datatransfer.TrustedChannelResolver;
|
||||||
import org.apache.hadoop.hdfs.protocol.datatransfer.sasl.DataTransferSaslUtil;
|
import org.apache.hadoop.hdfs.protocol.datatransfer.sasl.DataTransferSaslUtil;
|
||||||
|
@ -45,9 +52,14 @@ import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
||||||
import org.apache.hadoop.test.GenericTestUtils;
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
import org.apache.hadoop.test.GenericTestUtils.LogCapturer;
|
import org.apache.hadoop.test.GenericTestUtils.LogCapturer;
|
||||||
|
import org.apache.hadoop.hdfs.security.token.block.DataEncryptionKey;
|
||||||
import org.apache.log4j.Level;
|
import org.apache.log4j.Level;
|
||||||
import org.apache.log4j.LogManager;
|
import org.apache.log4j.LogManager;
|
||||||
|
import org.junit.After;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Rule;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
import org.junit.rules.Timeout;
|
||||||
import org.junit.runner.RunWith;
|
import org.junit.runner.RunWith;
|
||||||
import org.junit.runners.Parameterized;
|
import org.junit.runners.Parameterized;
|
||||||
import org.junit.runners.Parameterized.Parameters;
|
import org.junit.runners.Parameterized.Parameters;
|
||||||
|
@ -60,6 +72,9 @@ public class TestEncryptedTransfer {
|
||||||
LogManager.getLogger(DataTransferSaslUtil.class).setLevel(Level.DEBUG);
|
LogManager.getLogger(DataTransferSaslUtil.class).setLevel(Level.DEBUG);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Rule
|
||||||
|
public Timeout timeout = new Timeout(300000);
|
||||||
|
|
||||||
@Parameters
|
@Parameters
|
||||||
public static Collection<Object[]> data() {
|
public static Collection<Object[]> data() {
|
||||||
Collection<Object[]> params = new ArrayList<Object[]>();
|
Collection<Object[]> params = new ArrayList<Object[]>();
|
||||||
|
@ -73,7 +88,11 @@ public class TestEncryptedTransfer {
|
||||||
private static final String PLAIN_TEXT = "this is very secret plain text";
|
private static final String PLAIN_TEXT = "this is very secret plain text";
|
||||||
private static final Path TEST_PATH = new Path("/non-encrypted-file");
|
private static final Path TEST_PATH = new Path("/non-encrypted-file");
|
||||||
|
|
||||||
private void setEncryptionConfigKeys(Configuration conf) {
|
private MiniDFSCluster cluster = null;
|
||||||
|
private Configuration conf = null;
|
||||||
|
private FileSystem fs = null;
|
||||||
|
|
||||||
|
private void setEncryptionConfigKeys() {
|
||||||
conf.setBoolean(DFSConfigKeys.DFS_ENCRYPT_DATA_TRANSFER_KEY, true);
|
conf.setBoolean(DFSConfigKeys.DFS_ENCRYPT_DATA_TRANSFER_KEY, true);
|
||||||
conf.setBoolean(DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_KEY, true);
|
conf.setBoolean(DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_KEY, true);
|
||||||
if (resolverClazz != null){
|
if (resolverClazz != null){
|
||||||
|
@ -96,21 +115,33 @@ public class TestEncryptedTransfer {
|
||||||
this.resolverClazz = resolverClazz;
|
this.resolverClazz = resolverClazz;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Before
|
||||||
public void testEncryptedRead() throws IOException {
|
public void setup() throws IOException {
|
||||||
MiniDFSCluster cluster = null;
|
conf = new Configuration();
|
||||||
try {
|
}
|
||||||
Configuration conf = new Configuration();
|
|
||||||
|
@After
|
||||||
|
public void teardown() throws IOException {
|
||||||
|
if (fs != null) {
|
||||||
|
fs.close();
|
||||||
|
}
|
||||||
|
if (cluster != null) {
|
||||||
|
cluster.shutdown();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private FileChecksum writeUnencryptedAndThenRestartEncryptedCluster()
|
||||||
|
throws IOException {
|
||||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
cluster = new MiniDFSCluster.Builder(conf).build();
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
fs = getFileSystem(conf);
|
||||||
writeTestDataToFile(fs);
|
writeTestDataToFile(fs);
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
FileChecksum checksum = fs.getFileChecksum(TEST_PATH);
|
FileChecksum checksum = fs.getFileChecksum(TEST_PATH);
|
||||||
fs.close();
|
fs.close();
|
||||||
cluster.shutdown();
|
cluster.shutdown();
|
||||||
|
|
||||||
setEncryptionConfigKeys(conf);
|
setEncryptionConfigKeys();
|
||||||
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf)
|
cluster = new MiniDFSCluster.Builder(conf)
|
||||||
.manageDataDfsDirs(false)
|
.manageDataDfsDirs(false)
|
||||||
|
@ -120,116 +151,20 @@ public class TestEncryptedTransfer {
|
||||||
.build();
|
.build();
|
||||||
|
|
||||||
fs = getFileSystem(conf);
|
fs = getFileSystem(conf);
|
||||||
LogCapturer logs = GenericTestUtils.LogCapturer.captureLogs(
|
return checksum;
|
||||||
LogFactory.getLog(SaslDataTransferServer.class));
|
|
||||||
LogCapturer logs1 = GenericTestUtils.LogCapturer.captureLogs(
|
|
||||||
LogFactory.getLog(DataTransferSaslUtil.class));
|
|
||||||
try {
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
|
||||||
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
|
||||||
} finally {
|
|
||||||
logs.stopCapturing();
|
|
||||||
logs1.stopCapturing();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fs.close();
|
private void testEncryptedRead(String algorithm, String cipherSuite,
|
||||||
|
boolean matchLog, boolean readAfterRestart)
|
||||||
if (resolverClazz == null) {
|
throws IOException {
|
||||||
// Test client and server negotiate cipher option
|
// set encryption algorithm and cipher suites, but don't enable transfer
|
||||||
GenericTestUtils.assertDoesNotMatch(logs.getOutput(),
|
// encryption yet.
|
||||||
"Server using cipher suite");
|
conf.set(DFSConfigKeys.DFS_DATA_ENCRYPTION_ALGORITHM_KEY, algorithm);
|
||||||
// Check the IOStreamPair
|
|
||||||
GenericTestUtils.assertDoesNotMatch(logs1.getOutput(),
|
|
||||||
"Creating IOStreamPair of CryptoInputStream and CryptoOutputStream.");
|
|
||||||
}
|
|
||||||
} finally {
|
|
||||||
if (cluster != null) {
|
|
||||||
cluster.shutdown();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
@Test
|
|
||||||
public void testEncryptedReadWithRC4() throws IOException {
|
|
||||||
MiniDFSCluster cluster = null;
|
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
|
||||||
writeTestDataToFile(fs);
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
|
||||||
FileChecksum checksum = fs.getFileChecksum(TEST_PATH);
|
|
||||||
fs.close();
|
|
||||||
cluster.shutdown();
|
|
||||||
|
|
||||||
setEncryptionConfigKeys(conf);
|
|
||||||
// It'll use 3DES by default, but we set it to rc4 here.
|
|
||||||
conf.set(DFSConfigKeys.DFS_DATA_ENCRYPTION_ALGORITHM_KEY, "rc4");
|
|
||||||
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf)
|
|
||||||
.manageDataDfsDirs(false)
|
|
||||||
.manageNameDfsDirs(false)
|
|
||||||
.format(false)
|
|
||||||
.startupOption(StartupOption.REGULAR)
|
|
||||||
.build();
|
|
||||||
|
|
||||||
fs = getFileSystem(conf);
|
|
||||||
LogCapturer logs = GenericTestUtils.LogCapturer.captureLogs(
|
|
||||||
LogFactory.getLog(SaslDataTransferServer.class));
|
|
||||||
LogCapturer logs1 = GenericTestUtils.LogCapturer.captureLogs(
|
|
||||||
LogFactory.getLog(DataTransferSaslUtil.class));
|
|
||||||
try {
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
|
||||||
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
|
||||||
} finally {
|
|
||||||
logs.stopCapturing();
|
|
||||||
logs1.stopCapturing();
|
|
||||||
}
|
|
||||||
|
|
||||||
fs.close();
|
|
||||||
|
|
||||||
if (resolverClazz == null) {
|
|
||||||
// Test client and server negotiate cipher option
|
|
||||||
GenericTestUtils.assertDoesNotMatch(logs.getOutput(),
|
|
||||||
"Server using cipher suite");
|
|
||||||
// Check the IOStreamPair
|
|
||||||
GenericTestUtils.assertDoesNotMatch(logs1.getOutput(),
|
|
||||||
"Creating IOStreamPair of CryptoInputStream and CryptoOutputStream.");
|
|
||||||
}
|
|
||||||
} finally {
|
|
||||||
if (cluster != null) {
|
|
||||||
cluster.shutdown();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
@Test
|
|
||||||
public void testEncryptedReadWithAES() throws IOException {
|
|
||||||
MiniDFSCluster cluster = null;
|
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
conf.set(HdfsClientConfigKeys.DFS_ENCRYPT_DATA_TRANSFER_CIPHER_SUITES_KEY,
|
conf.set(HdfsClientConfigKeys.DFS_ENCRYPT_DATA_TRANSFER_CIPHER_SUITES_KEY,
|
||||||
"AES/CTR/NoPadding");
|
cipherSuite);
|
||||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
FileChecksum checksum = writeUnencryptedAndThenRestartEncryptedCluster();
|
||||||
writeTestDataToFile(fs);
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
|
||||||
FileChecksum checksum = fs.getFileChecksum(TEST_PATH);
|
|
||||||
fs.close();
|
|
||||||
cluster.shutdown();
|
|
||||||
|
|
||||||
setEncryptionConfigKeys(conf);
|
|
||||||
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf)
|
|
||||||
.manageDataDfsDirs(false)
|
|
||||||
.manageNameDfsDirs(false)
|
|
||||||
.format(false)
|
|
||||||
.startupOption(StartupOption.REGULAR)
|
|
||||||
.build();
|
|
||||||
|
|
||||||
fs = getFileSystem(conf);
|
|
||||||
LogCapturer logs = GenericTestUtils.LogCapturer.captureLogs(
|
LogCapturer logs = GenericTestUtils.LogCapturer.captureLogs(
|
||||||
LogFactory.getLog(SaslDataTransferServer.class));
|
LogFactory.getLog(SaslDataTransferServer.class));
|
||||||
LogCapturer logs1 = GenericTestUtils.LogCapturer.captureLogs(
|
LogCapturer logs1 = GenericTestUtils.LogCapturer.captureLogs(
|
||||||
|
@ -242,90 +177,61 @@ public class TestEncryptedTransfer {
|
||||||
logs1.stopCapturing();
|
logs1.stopCapturing();
|
||||||
}
|
}
|
||||||
|
|
||||||
fs.close();
|
|
||||||
|
|
||||||
if (resolverClazz == null) {
|
if (resolverClazz == null) {
|
||||||
|
if (matchLog) {
|
||||||
// Test client and server negotiate cipher option
|
// Test client and server negotiate cipher option
|
||||||
GenericTestUtils.assertMatches(logs.getOutput(),
|
GenericTestUtils
|
||||||
"Server using cipher suite");
|
.assertMatches(logs.getOutput(), "Server using cipher suite");
|
||||||
// Check the IOStreamPair
|
// Check the IOStreamPair
|
||||||
GenericTestUtils.assertMatches(logs1.getOutput(),
|
GenericTestUtils.assertMatches(logs1.getOutput(),
|
||||||
"Creating IOStreamPair of CryptoInputStream and CryptoOutputStream.");
|
"Creating IOStreamPair of CryptoInputStream and CryptoOutputStream.");
|
||||||
}
|
} else {
|
||||||
} finally {
|
// Test client and server negotiate cipher option
|
||||||
if (cluster != null) {
|
GenericTestUtils
|
||||||
cluster.shutdown();
|
.assertDoesNotMatch(logs.getOutput(), "Server using cipher suite");
|
||||||
}
|
// Check the IOStreamPair
|
||||||
|
GenericTestUtils.assertDoesNotMatch(logs1.getOutput(),
|
||||||
|
"Creating IOStreamPair of CryptoInputStream and CryptoOutputStream.");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
if (readAfterRestart) {
|
||||||
public void testEncryptedReadAfterNameNodeRestart() throws IOException {
|
|
||||||
MiniDFSCluster cluster = null;
|
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
|
||||||
writeTestDataToFile(fs);
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
|
||||||
FileChecksum checksum = fs.getFileChecksum(TEST_PATH);
|
|
||||||
fs.close();
|
|
||||||
cluster.shutdown();
|
|
||||||
|
|
||||||
setEncryptionConfigKeys(conf);
|
|
||||||
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf)
|
|
||||||
.manageDataDfsDirs(false)
|
|
||||||
.manageNameDfsDirs(false)
|
|
||||||
.format(false)
|
|
||||||
.startupOption(StartupOption.REGULAR)
|
|
||||||
.build();
|
|
||||||
|
|
||||||
fs = getFileSystem(conf);
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
|
||||||
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
|
||||||
fs.close();
|
|
||||||
|
|
||||||
cluster.restartNameNode();
|
cluster.restartNameNode();
|
||||||
fs = getFileSystem(conf);
|
fs = getFileSystem(conf);
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
||||||
fs.close();
|
|
||||||
} finally {
|
|
||||||
if (cluster != null) {
|
|
||||||
cluster.shutdown();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testEncryptedReadDefaultAlgorithmCipherSuite()
|
||||||
|
throws IOException {
|
||||||
|
testEncryptedRead("", "", false, false);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testEncryptedReadWithRC4() throws IOException {
|
||||||
|
testEncryptedRead("rc4", "", false, false);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testEncryptedReadWithAES() throws IOException {
|
||||||
|
testEncryptedRead("", "AES/CTR/NoPadding", true, false);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testEncryptedReadAfterNameNodeRestart() throws IOException {
|
||||||
|
testEncryptedRead("", "", false, true);
|
||||||
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testClientThatDoesNotSupportEncryption() throws IOException {
|
public void testClientThatDoesNotSupportEncryption() throws IOException {
|
||||||
MiniDFSCluster cluster = null;
|
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
// Set short retry timeouts so this test runs faster
|
// Set short retry timeouts so this test runs faster
|
||||||
conf.setInt(HdfsClientConfigKeys.Retry.WINDOW_BASE_KEY, 10);
|
conf.setInt(HdfsClientConfigKeys.Retry.WINDOW_BASE_KEY, 10);
|
||||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
writeUnencryptedAndThenRestartEncryptedCluster();
|
||||||
writeTestDataToFile(fs);
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
|
||||||
fs.close();
|
|
||||||
cluster.shutdown();
|
|
||||||
|
|
||||||
setEncryptionConfigKeys(conf);
|
DFSClient client = DFSClientAdapter.getDFSClient((DistributedFileSystem)fs);
|
||||||
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf)
|
|
||||||
.manageDataDfsDirs(false)
|
|
||||||
.manageNameDfsDirs(false)
|
|
||||||
.format(false)
|
|
||||||
.startupOption(StartupOption.REGULAR)
|
|
||||||
.build();
|
|
||||||
|
|
||||||
|
|
||||||
fs = getFileSystem(conf);
|
|
||||||
DFSClient client = DFSClientAdapter.getDFSClient((DistributedFileSystem) fs);
|
|
||||||
DFSClient spyClient = Mockito.spy(client);
|
DFSClient spyClient = Mockito.spy(client);
|
||||||
Mockito.doReturn(false).when(spyClient).shouldEncryptData();
|
Mockito.doReturn(false).when(spyClient).shouldEncryptData();
|
||||||
DFSClientAdapter.setDFSClient((DistributedFileSystem) fs, spyClient);
|
DFSClientAdapter.setDFSClient((DistributedFileSystem) fs, spyClient);
|
||||||
|
@ -334,7 +240,8 @@ public class TestEncryptedTransfer {
|
||||||
LogFactory.getLog(DataNode.class));
|
LogFactory.getLog(DataNode.class));
|
||||||
try {
|
try {
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
if (resolverClazz != null && !resolverClazz.endsWith("TestTrustedChannelResolver")){
|
if (resolverClazz != null &&
|
||||||
|
!resolverClazz.endsWith("TestTrustedChannelResolver")){
|
||||||
fail("Should not have been able to read without encryption enabled.");
|
fail("Should not have been able to read without encryption enabled.");
|
||||||
}
|
}
|
||||||
} catch (IOException ioe) {
|
} catch (IOException ioe) {
|
||||||
|
@ -343,43 +250,17 @@ public class TestEncryptedTransfer {
|
||||||
} finally {
|
} finally {
|
||||||
logs.stopCapturing();
|
logs.stopCapturing();
|
||||||
}
|
}
|
||||||
fs.close();
|
|
||||||
|
|
||||||
if (resolverClazz == null) {
|
if (resolverClazz == null) {
|
||||||
GenericTestUtils.assertMatches(logs.getOutput(),
|
GenericTestUtils.assertMatches(logs.getOutput(),
|
||||||
"Failed to read expected encryption handshake from client at");
|
"Failed to read expected encryption handshake from client at");
|
||||||
}
|
}
|
||||||
} finally {
|
|
||||||
if (cluster != null) {
|
|
||||||
cluster.shutdown();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testLongLivedReadClientAfterRestart() throws IOException {
|
public void testLongLivedReadClientAfterRestart() throws IOException {
|
||||||
MiniDFSCluster cluster = null;
|
FileChecksum checksum = writeUnencryptedAndThenRestartEncryptedCluster();
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
|
||||||
writeTestDataToFile(fs);
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
|
||||||
FileChecksum checksum = fs.getFileChecksum(TEST_PATH);
|
|
||||||
fs.close();
|
|
||||||
cluster.shutdown();
|
|
||||||
|
|
||||||
setEncryptionConfigKeys(conf);
|
|
||||||
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf)
|
|
||||||
.manageDataDfsDirs(false)
|
|
||||||
.manageNameDfsDirs(false)
|
|
||||||
.format(false)
|
|
||||||
.startupOption(StartupOption.REGULAR)
|
|
||||||
.build();
|
|
||||||
|
|
||||||
fs = getFileSystem(conf);
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
||||||
|
|
||||||
|
@ -390,24 +271,14 @@ public class TestEncryptedTransfer {
|
||||||
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
||||||
|
|
||||||
fs.close();
|
|
||||||
} finally {
|
|
||||||
if (cluster != null) {
|
|
||||||
cluster.shutdown();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testLongLivedWriteClientAfterRestart() throws IOException {
|
public void testLongLivedWriteClientAfterRestart() throws IOException {
|
||||||
MiniDFSCluster cluster = null;
|
setEncryptionConfigKeys();
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
setEncryptionConfigKeys(conf);
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
cluster = new MiniDFSCluster.Builder(conf).build();
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
fs = getFileSystem(conf);
|
||||||
|
|
||||||
writeTestDataToFile(fs);
|
writeTestDataToFile(fs);
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
|
@ -420,37 +291,11 @@ public class TestEncryptedTransfer {
|
||||||
|
|
||||||
writeTestDataToFile(fs);
|
writeTestDataToFile(fs);
|
||||||
assertEquals(PLAIN_TEXT + PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT + PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
|
|
||||||
fs.close();
|
|
||||||
} finally {
|
|
||||||
if (cluster != null) {
|
|
||||||
cluster.shutdown();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testLongLivedClient() throws IOException, InterruptedException {
|
public void testLongLivedClient() throws IOException, InterruptedException {
|
||||||
MiniDFSCluster cluster = null;
|
FileChecksum checksum = writeUnencryptedAndThenRestartEncryptedCluster();
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf).build();
|
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
|
||||||
writeTestDataToFile(fs);
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
|
||||||
FileChecksum checksum = fs.getFileChecksum(TEST_PATH);
|
|
||||||
fs.close();
|
|
||||||
cluster.shutdown();
|
|
||||||
|
|
||||||
setEncryptionConfigKeys(conf);
|
|
||||||
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf)
|
|
||||||
.manageDataDfsDirs(false)
|
|
||||||
.manageNameDfsDirs(false)
|
|
||||||
.format(false)
|
|
||||||
.startupOption(StartupOption.REGULAR)
|
|
||||||
.build();
|
|
||||||
|
|
||||||
BlockTokenSecretManager btsm = cluster.getNamesystem().getBlockManager()
|
BlockTokenSecretManager btsm = cluster.getNamesystem().getBlockManager()
|
||||||
.getBlockTokenSecretManager();
|
.getBlockTokenSecretManager();
|
||||||
|
@ -458,7 +303,6 @@ public class TestEncryptedTransfer {
|
||||||
btsm.setTokenLifetime(2 * 1000);
|
btsm.setTokenLifetime(2 * 1000);
|
||||||
btsm.clearAllKeysForTesting();
|
btsm.clearAllKeysForTesting();
|
||||||
|
|
||||||
fs = getFileSystem(conf);
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
||||||
|
|
||||||
|
@ -472,13 +316,71 @@ public class TestEncryptedTransfer {
|
||||||
|
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
assertEquals(checksum, fs.getFileChecksum(TEST_PATH));
|
||||||
|
}
|
||||||
|
|
||||||
fs.close();
|
@Test
|
||||||
} finally {
|
public void testLongLivedClientPipelineRecovery()
|
||||||
if (cluster != null) {
|
throws IOException, InterruptedException, TimeoutException {
|
||||||
cluster.shutdown();
|
if (resolverClazz != null) {
|
||||||
|
// TestTrustedChannelResolver does not use encryption keys.
|
||||||
|
return;
|
||||||
}
|
}
|
||||||
|
// use 4 datanodes to make sure that after 1 data node is stopped,
|
||||||
|
// client only retries establishing pipeline with the 4th node.
|
||||||
|
int numDataNodes = 4;
|
||||||
|
// do not consider load factor when selecting a data node
|
||||||
|
conf.setBoolean(DFSConfigKeys.DFS_NAMENODE_REPLICATION_CONSIDERLOAD_KEY,
|
||||||
|
false);
|
||||||
|
setEncryptionConfigKeys();
|
||||||
|
|
||||||
|
cluster = new MiniDFSCluster.Builder(conf)
|
||||||
|
.numDataNodes(numDataNodes)
|
||||||
|
.build();
|
||||||
|
|
||||||
|
fs = getFileSystem(conf);
|
||||||
|
DFSClient client = DFSClientAdapter.getDFSClient((DistributedFileSystem)fs);
|
||||||
|
DFSClient spyClient = Mockito.spy(client);
|
||||||
|
DFSClientAdapter.setDFSClient((DistributedFileSystem) fs, spyClient);
|
||||||
|
writeTestDataToFile(fs);
|
||||||
|
|
||||||
|
BlockTokenSecretManager btsm = cluster.getNamesystem().getBlockManager()
|
||||||
|
.getBlockTokenSecretManager();
|
||||||
|
// Reduce key update interval and token life for testing.
|
||||||
|
btsm.setKeyUpdateIntervalForTesting(2 * 1000);
|
||||||
|
btsm.setTokenLifetime(2 * 1000);
|
||||||
|
btsm.clearAllKeysForTesting();
|
||||||
|
|
||||||
|
// Wait until the encryption key becomes invalid.
|
||||||
|
LOG.info("Wait until encryption keys become invalid...");
|
||||||
|
|
||||||
|
final DataEncryptionKey encryptionKey = spyClient.getEncryptionKey();
|
||||||
|
List<DataNode> dataNodes = cluster.getDataNodes();
|
||||||
|
for (final DataNode dn: dataNodes) {
|
||||||
|
GenericTestUtils.waitFor(
|
||||||
|
new Supplier<Boolean>() {
|
||||||
|
@Override
|
||||||
|
public Boolean get() {
|
||||||
|
return !dn.getBlockPoolTokenSecretManager().
|
||||||
|
get(encryptionKey.blockPoolId)
|
||||||
|
.hasKey(encryptionKey.keyId);
|
||||||
}
|
}
|
||||||
|
}, 100, 30*1000
|
||||||
|
);
|
||||||
|
}
|
||||||
|
LOG.info("The encryption key is invalid on all nodes now.");
|
||||||
|
try(FSDataOutputStream out = fs.append(TEST_PATH)) {
|
||||||
|
DFSOutputStream dfstream = (DFSOutputStream) out.getWrappedStream();
|
||||||
|
// shut down the first datanode in the pipeline.
|
||||||
|
DatanodeInfo[] targets = dfstream.getPipeline();
|
||||||
|
cluster.stopDataNode(targets[0].getXferAddr());
|
||||||
|
// write data to induce pipeline recovery
|
||||||
|
out.write(PLAIN_TEXT.getBytes());
|
||||||
|
out.hflush();
|
||||||
|
assertFalse("The first datanode in the pipeline was not replaced.",
|
||||||
|
Arrays.asList(dfstream.getPipeline()).contains(targets[0]));
|
||||||
|
}
|
||||||
|
// verify that InvalidEncryptionKeyException is handled properly
|
||||||
|
Mockito.verify(spyClient, times(1)).clearDataEncryptionKey();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
@ -497,14 +399,11 @@ public class TestEncryptedTransfer {
|
||||||
}
|
}
|
||||||
|
|
||||||
private void testEncryptedWrite(int numDns) throws IOException {
|
private void testEncryptedWrite(int numDns) throws IOException {
|
||||||
MiniDFSCluster cluster = null;
|
setEncryptionConfigKeys();
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
setEncryptionConfigKeys(conf);
|
|
||||||
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDns).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(numDns).build();
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
fs = getFileSystem(conf);
|
||||||
|
|
||||||
LogCapturer logs = GenericTestUtils.LogCapturer.captureLogs(
|
LogCapturer logs = GenericTestUtils.LogCapturer.captureLogs(
|
||||||
LogFactory.getLog(SaslDataTransferServer.class));
|
LogFactory.getLog(SaslDataTransferServer.class));
|
||||||
|
@ -517,7 +416,6 @@ public class TestEncryptedTransfer {
|
||||||
logs1.stopCapturing();
|
logs1.stopCapturing();
|
||||||
}
|
}
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
fs.close();
|
|
||||||
|
|
||||||
if (resolverClazz == null) {
|
if (resolverClazz == null) {
|
||||||
// Test client and server negotiate cipher option
|
// Test client and server negotiate cipher option
|
||||||
|
@ -527,49 +425,31 @@ public class TestEncryptedTransfer {
|
||||||
GenericTestUtils.assertDoesNotMatch(logs1.getOutput(),
|
GenericTestUtils.assertDoesNotMatch(logs1.getOutput(),
|
||||||
"Creating IOStreamPair of CryptoInputStream and CryptoOutputStream.");
|
"Creating IOStreamPair of CryptoInputStream and CryptoOutputStream.");
|
||||||
}
|
}
|
||||||
} finally {
|
|
||||||
if (cluster != null) {
|
|
||||||
cluster.shutdown();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testEncryptedAppend() throws IOException {
|
public void testEncryptedAppend() throws IOException {
|
||||||
MiniDFSCluster cluster = null;
|
setEncryptionConfigKeys();
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
setEncryptionConfigKeys(conf);
|
|
||||||
|
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(3).build();
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
fs = getFileSystem(conf);
|
||||||
|
|
||||||
writeTestDataToFile(fs);
|
writeTestDataToFile(fs);
|
||||||
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
|
|
||||||
writeTestDataToFile(fs);
|
writeTestDataToFile(fs);
|
||||||
assertEquals(PLAIN_TEXT + PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT + PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
|
|
||||||
fs.close();
|
|
||||||
} finally {
|
|
||||||
if (cluster != null) {
|
|
||||||
cluster.shutdown();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testEncryptedAppendRequiringBlockTransfer() throws IOException {
|
public void testEncryptedAppendRequiringBlockTransfer() throws IOException {
|
||||||
MiniDFSCluster cluster = null;
|
setEncryptionConfigKeys();
|
||||||
try {
|
|
||||||
Configuration conf = new Configuration();
|
|
||||||
setEncryptionConfigKeys(conf);
|
|
||||||
|
|
||||||
// start up 4 DNs
|
// start up 4 DNs
|
||||||
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(4).build();
|
cluster = new MiniDFSCluster.Builder(conf).numDataNodes(4).build();
|
||||||
|
|
||||||
FileSystem fs = getFileSystem(conf);
|
fs = getFileSystem(conf);
|
||||||
|
|
||||||
// Create a file with replication 3, so its block is on 3 / 4 DNs.
|
// Create a file with replication 3, so its block is on 3 / 4 DNs.
|
||||||
writeTestDataToFile(fs);
|
writeTestDataToFile(fs);
|
||||||
|
@ -581,20 +461,14 @@ public class TestEncryptedTransfer {
|
||||||
in.close();
|
in.close();
|
||||||
assertEquals(1, locatedBlocks.size());
|
assertEquals(1, locatedBlocks.size());
|
||||||
assertEquals(3, locatedBlocks.get(0).getLocations().length);
|
assertEquals(3, locatedBlocks.get(0).getLocations().length);
|
||||||
DataNode dn = cluster.getDataNode(locatedBlocks.get(0).getLocations()[0].getIpcPort());
|
DataNode dn = cluster.getDataNode(
|
||||||
|
locatedBlocks.get(0).getLocations()[0].getIpcPort());
|
||||||
dn.shutdown();
|
dn.shutdown();
|
||||||
|
|
||||||
// Reopen the file for append, which will need to add another DN to the
|
// Reopen the file for append, which will need to add another DN to the
|
||||||
// pipeline and in doing so trigger a block transfer.
|
// pipeline and in doing so trigger a block transfer.
|
||||||
writeTestDataToFile(fs);
|
writeTestDataToFile(fs);
|
||||||
assertEquals(PLAIN_TEXT + PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
assertEquals(PLAIN_TEXT + PLAIN_TEXT, DFSTestUtil.readFile(fs, TEST_PATH));
|
||||||
|
|
||||||
fs.close();
|
|
||||||
} finally {
|
|
||||||
if (cluster != null) {
|
|
||||||
cluster.shutdown();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private static void writeTestDataToFile(FileSystem fs) throws IOException {
|
private static void writeTestDataToFile(FileSystem fs) throws IOException {
|
||||||
|
|
Loading…
Reference in New Issue