HDFS-11361. Block Storage: add cache interface. Contributed by Chen Liang.
This commit is contained in:
parent
d251a55ff9
commit
23044c1db8
|
@ -24,6 +24,7 @@ import org.apache.hadoop.hdfs.protocol.DatanodeID;
|
||||||
import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos;
|
import org.apache.hadoop.hdfs.protocol.proto.HdfsProtos;
|
||||||
|
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
|
import java.util.Arrays;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
import java.util.TreeMap;
|
import java.util.TreeMap;
|
||||||
|
@ -35,6 +36,12 @@ public class Pipeline {
|
||||||
private String containerName;
|
private String containerName;
|
||||||
private String leaderID;
|
private String leaderID;
|
||||||
private Map<String, DatanodeID> datanodes;
|
private Map<String, DatanodeID> datanodes;
|
||||||
|
/**
|
||||||
|
* Allows you to maintain private data on pipelines.
|
||||||
|
* This is not serialized via protobuf, just allows us to maintain some
|
||||||
|
* private data.
|
||||||
|
*/
|
||||||
|
private byte[] data;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Constructs a new pipeline data structure.
|
* Constructs a new pipeline data structure.
|
||||||
|
@ -44,6 +51,7 @@ public class Pipeline {
|
||||||
public Pipeline(String leaderID) {
|
public Pipeline(String leaderID) {
|
||||||
this.leaderID = leaderID;
|
this.leaderID = leaderID;
|
||||||
datanodes = new TreeMap<>();
|
datanodes = new TreeMap<>();
|
||||||
|
data = null;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -124,4 +132,27 @@ public class Pipeline {
|
||||||
public void setContainerName(String containerName) {
|
public void setContainerName(String containerName) {
|
||||||
this.containerName = containerName;
|
this.containerName = containerName;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Set private data on pipeline.
|
||||||
|
* @param data -- private data.
|
||||||
|
*/
|
||||||
|
public void setData(byte[] data) {
|
||||||
|
if (data != null) {
|
||||||
|
this.data = Arrays.copyOf(data, data.length);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns private data that is set on this pipeline.
|
||||||
|
*
|
||||||
|
* @return blob, the user can interpret it any way they like.
|
||||||
|
*/
|
||||||
|
public byte[] getData() {
|
||||||
|
if (this.data != null) {
|
||||||
|
return Arrays.copyOf(this.data, this.data.length);
|
||||||
|
} else {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -17,12 +17,12 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.cblock;
|
package org.apache.hadoop.cblock;
|
||||||
|
|
||||||
|
import static java.lang.Thread.NORM_PRIORITY;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This class contains constants for configuration keys used in CBlock.
|
* This class contains constants for configuration keys used in CBlock.
|
||||||
*/
|
*/
|
||||||
public final class CBlockConfigKeys {
|
public final class CBlockConfigKeys {
|
||||||
public static final String DFS_CBLOCK_ENABLED_KEY =
|
|
||||||
"dfs.cblock.enabled";
|
|
||||||
public static final String DFS_CBLOCK_SERVICERPC_ADDRESS_KEY =
|
public static final String DFS_CBLOCK_SERVICERPC_ADDRESS_KEY =
|
||||||
"dfs.cblock.servicerpc-address";
|
"dfs.cblock.servicerpc-address";
|
||||||
public static final String DFS_CBLOCK_SERVICERPC_PORT_KEY =
|
public static final String DFS_CBLOCK_SERVICERPC_PORT_KEY =
|
||||||
|
@ -33,8 +33,6 @@ public final class CBlockConfigKeys {
|
||||||
"dfs.cblock.servicerpc.hostname";
|
"dfs.cblock.servicerpc.hostname";
|
||||||
public static final String DFS_CBLOCK_SERVICERPC_HOSTNAME_DEFAULT =
|
public static final String DFS_CBLOCK_SERVICERPC_HOSTNAME_DEFAULT =
|
||||||
"0.0.0.0";
|
"0.0.0.0";
|
||||||
public static final String DFS_CBLOCK_RPCSERVICE_IP_DEFAULT =
|
|
||||||
"0.0.0.0";
|
|
||||||
public static final String DFS_CBLOCK_SERVICERPC_ADDRESS_DEFAULT =
|
public static final String DFS_CBLOCK_SERVICERPC_ADDRESS_DEFAULT =
|
||||||
DFS_CBLOCK_SERVICERPC_HOSTNAME_DEFAULT
|
DFS_CBLOCK_SERVICERPC_HOSTNAME_DEFAULT
|
||||||
+ ":" + DFS_CBLOCK_SERVICERPC_PORT_DEFAULT;
|
+ ":" + DFS_CBLOCK_SERVICERPC_PORT_DEFAULT;
|
||||||
|
@ -46,7 +44,7 @@ public final class CBlockConfigKeys {
|
||||||
public static final int DFS_CBLOCK_JSCSI_PORT_DEFAULT =
|
public static final int DFS_CBLOCK_JSCSI_PORT_DEFAULT =
|
||||||
9811;
|
9811;
|
||||||
public static final String DFS_CBLOCK_JSCSIRPC_ADDRESS_DEFAULT =
|
public static final String DFS_CBLOCK_JSCSIRPC_ADDRESS_DEFAULT =
|
||||||
DFS_CBLOCK_RPCSERVICE_IP_DEFAULT
|
DFS_CBLOCK_SERVICERPC_HOSTNAME_DEFAULT
|
||||||
+ ":" + DFS_CBLOCK_JSCSI_PORT_DEFAULT;
|
+ ":" + DFS_CBLOCK_JSCSI_PORT_DEFAULT;
|
||||||
|
|
||||||
|
|
||||||
|
@ -69,6 +67,66 @@ public final class CBlockConfigKeys {
|
||||||
public static final String DFS_CBLOCK_SERVICE_LEVELDB_PATH_DEFAULT =
|
public static final String DFS_CBLOCK_SERVICE_LEVELDB_PATH_DEFAULT =
|
||||||
"/tmp/cblock_levelDB.dat";
|
"/tmp/cblock_levelDB.dat";
|
||||||
|
|
||||||
|
|
||||||
|
public static final String DFS_CBLOCK_DISK_CACHE_PATH_KEY =
|
||||||
|
"dfs.cblock.disk.cache.path";
|
||||||
|
public static final String DFS_CBLOCK_DISK_CACHE_PATH_DEFAULT =
|
||||||
|
"/tmp/cblockCacheDB";
|
||||||
|
/**
|
||||||
|
* Setting this flag to true makes the block layer compute a sha256 hash of
|
||||||
|
* the data and log that information along with block ID. This is very
|
||||||
|
* useful for doing trace based simulation of various workloads. Since it is
|
||||||
|
* computing a hash for each block this could be expensive, hence default
|
||||||
|
* is false.
|
||||||
|
*/
|
||||||
|
public static final String DFS_CBLOCK_TRACE_IO = "dfs.cblock.trace.io";
|
||||||
|
public static final boolean DFS_CBLOCK_TRACE_IO_DEFAULT = false;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Cache size in 1000s of entries. 256 indicates 256 * 1024.
|
||||||
|
*/
|
||||||
|
public static final String DFS_CBLOCK_CACHE_QUEUE_SIZE_KB =
|
||||||
|
"dfs.cblock.cache.cache.size.in.kb";
|
||||||
|
public static final int DFS_CBLOCK_CACHE_QUEUE_SIZE_KB_DEFAULT = 256;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Minimum Number of threads that cache pool will use for background I/O.
|
||||||
|
*/
|
||||||
|
public static final String DFS_CBLOCK_CACHE_CORE_POOL_SIZE =
|
||||||
|
"dfs.cblock.cache.core.pool.size";
|
||||||
|
public static final int DFS_CBLOCK_CACHE_CORE_POOL_SIZE_DEFAULT = 16;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Maximum Number of threads that cache pool will use for background I/O.
|
||||||
|
*/
|
||||||
|
|
||||||
|
public static final String DFS_CBLOCK_CACHE_MAX_POOL_SIZE =
|
||||||
|
"dfs.cblock.cache.max.pool.size";
|
||||||
|
public static final int DFS_CBLOCK_CACHE_MAX_POOL_SIZE_DEFAULT = 256;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Number of seconds to keep the Thread alive when it is idle.
|
||||||
|
*/
|
||||||
|
public static final String DFS_CBLOCK_CACHE_KEEP_ALIVE_SECONDS =
|
||||||
|
"dfs.cblock.cache.keep.alive.seconds";
|
||||||
|
public static final long DFS_CBLOCK_CACHE_KEEP_ALIVE_SECONDS_DEFAULT = 60;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Priority of cache flusher thread, affecting the relative performance of
|
||||||
|
* write and read.
|
||||||
|
*/
|
||||||
|
public static final String DFS_CBLOCK_CACHE_THREAD_PRIORITY =
|
||||||
|
"dfs.cblock.cache.thread.priority";
|
||||||
|
public static final int DFS_CBLOCK_CACHE_THREAD_PRIORITY_DEFAULT =
|
||||||
|
NORM_PRIORITY;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Block Buffer size in 1024 entries, 128 means 128 * 1024 blockIDs.
|
||||||
|
*/
|
||||||
|
public static final String DFS_CBLOCK_CACHE_BLOCK_BUFFER_SIZE =
|
||||||
|
"dfs.cblock.cache.block.buffer.size";
|
||||||
|
public static final int DFS_CBLOCK_CACHE_BLOCK_BUFFER_SIZE_DEFAULT = 128;
|
||||||
|
|
||||||
private CBlockConfigKeys() {
|
private CBlockConfigKeys() {
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,155 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
* <p>
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
* <p>
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper;
|
||||||
|
|
||||||
|
import com.google.common.base.Preconditions;
|
||||||
|
import com.google.common.primitives.Longs;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.cache.LogicalBlock;
|
||||||
|
import org.apache.hadoop.scm.XceiverClientSpi;
|
||||||
|
import org.apache.hadoop.scm.container.common.helpers.Pipeline;
|
||||||
|
import org.apache.hadoop.scm.storage.ContainerProtocolCalls;
|
||||||
|
import org.apache.hadoop.util.Time;
|
||||||
|
|
||||||
|
import java.io.File;
|
||||||
|
import java.io.FileOutputStream;
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
import java.nio.channels.FileChannel;
|
||||||
|
import java.nio.file.Paths;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The blockWriter task.
|
||||||
|
*/
|
||||||
|
public class BlockWriterTask implements Runnable {
|
||||||
|
private final LogicalBlock block;
|
||||||
|
private int tryCount;
|
||||||
|
private final ContainerCacheFlusher flusher;
|
||||||
|
private final String dbPath;
|
||||||
|
private final String fileName;
|
||||||
|
private static final String RETRY_LOG_PREFIX = "RetryLog";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructs a BlockWriterTask.
|
||||||
|
*
|
||||||
|
* @param block - Block Information.
|
||||||
|
* @param flusher - ContainerCacheFlusher.
|
||||||
|
*/
|
||||||
|
public BlockWriterTask(LogicalBlock block, ContainerCacheFlusher flusher,
|
||||||
|
String dbPath, String fileName) {
|
||||||
|
this.block = block;
|
||||||
|
this.flusher = flusher;
|
||||||
|
this.dbPath = dbPath;
|
||||||
|
tryCount = 0;
|
||||||
|
this.fileName = fileName;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* When an object implementing interface <code>Runnable</code> is used
|
||||||
|
* to create a thread, starting the thread causes the object's
|
||||||
|
* <code>run</code> method to be called in that separately executing
|
||||||
|
* thread.
|
||||||
|
* <p>
|
||||||
|
* The general contract of the method <code>run</code> is that it may
|
||||||
|
* take any action whatsoever.
|
||||||
|
*
|
||||||
|
* @see Thread#run()
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
public void run() {
|
||||||
|
String containerName = null;
|
||||||
|
XceiverClientSpi client = null;
|
||||||
|
flusher.getLOG().debug(
|
||||||
|
"Writing block to remote. block ID: {}", block.getBlockID());
|
||||||
|
try {
|
||||||
|
incTryCount();
|
||||||
|
Pipeline pipeline = flusher.getPipeline(this.dbPath, block.getBlockID());
|
||||||
|
client = flusher.getXceiverClientManager().acquireClient(pipeline);
|
||||||
|
byte[] keybuf = Longs.toByteArray(block.getBlockID());
|
||||||
|
byte[] data;
|
||||||
|
long startTime = Time.monotonicNow();
|
||||||
|
data = flusher.getCacheDB(this.dbPath).get(keybuf);
|
||||||
|
long endTime = Time.monotonicNow();
|
||||||
|
Preconditions.checkState(data.length > 0, "Block data is zero length");
|
||||||
|
startTime = Time.monotonicNow();
|
||||||
|
// BUG: fix the trace ID.
|
||||||
|
ContainerProtocolCalls.writeSmallFile(client, containerName,
|
||||||
|
Long.toString(block.getBlockID()), data, "");
|
||||||
|
endTime = Time.monotonicNow();
|
||||||
|
flusher.getTargetMetrics().updateContainerWriteLatency(
|
||||||
|
endTime - startTime);
|
||||||
|
flusher.getLOG().debug("Time taken for Write Small File : {} ms",
|
||||||
|
endTime - startTime);
|
||||||
|
|
||||||
|
flusher.incrementremoteIO();
|
||||||
|
|
||||||
|
} catch (IOException ex) {
|
||||||
|
flusher.getLOG().error("Writing of block failed, We have attempted " +
|
||||||
|
"to write this block {} times to the container {}.Trace ID:{}",
|
||||||
|
this.getTryCount(), containerName, "", ex);
|
||||||
|
writeRetryBlock(block);
|
||||||
|
} finally {
|
||||||
|
flusher.incFinishCount(fileName);
|
||||||
|
if(client != null) {
|
||||||
|
flusher.getXceiverClientManager().releaseClient(client);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
private void writeRetryBlock(LogicalBlock currentBlock) {
|
||||||
|
boolean append = false;
|
||||||
|
String retryFileName =
|
||||||
|
String.format("%s.%d.%s", RETRY_LOG_PREFIX, currentBlock.getBlockID(),
|
||||||
|
Time.monotonicNow());
|
||||||
|
File logDir = new File(this.dbPath);
|
||||||
|
if (!logDir.exists() && !logDir.mkdirs()) {
|
||||||
|
flusher.getLOG().error(
|
||||||
|
"Unable to create the log directory, Crticial error cannot continue");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
String log = Paths.get(this.dbPath, retryFileName).toString();
|
||||||
|
ByteBuffer buffer = ByteBuffer.allocate(Long.SIZE / Byte.SIZE);
|
||||||
|
buffer.putLong(currentBlock.getBlockID());
|
||||||
|
try {
|
||||||
|
FileChannel channel = new FileOutputStream(log, append).getChannel();
|
||||||
|
channel.write(buffer);
|
||||||
|
channel.close();
|
||||||
|
flusher.processDirtyBlocks(this.dbPath, retryFileName);
|
||||||
|
} catch (IOException e) {
|
||||||
|
flusher.getLOG().error("Unable to write the retry block. Block ID: {}",
|
||||||
|
currentBlock.getBlockID(), e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Increments the try count. This is done each time we try this block
|
||||||
|
* write to the container.
|
||||||
|
*/
|
||||||
|
private void incTryCount() {
|
||||||
|
tryCount++;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the retry count.
|
||||||
|
*
|
||||||
|
* @return int
|
||||||
|
*/
|
||||||
|
public int getTryCount() {
|
||||||
|
return tryCount;
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,440 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
* <p>
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
* <p>
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper;
|
||||||
|
|
||||||
|
import org.apache.commons.codec.digest.DigestUtils;
|
||||||
|
import org.apache.commons.lang.StringUtils;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.cache.CacheModule;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.cache.LogicalBlock;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.cache.impl.CBlockLocalCache;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.scm.XceiverClientManager;
|
||||||
|
import org.apache.hadoop.scm.container.common.helpers.Pipeline;
|
||||||
|
import org.jscsi.target.storage.IStorageModule;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys.DFS_CBLOCK_TRACE_IO;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_TRACE_IO_DEFAULT;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The SCSI Target class for CBlockSCSIServer.
|
||||||
|
*/
|
||||||
|
final public class CBlockIStorageImpl implements IStorageModule {
|
||||||
|
private static final Logger LOGGER =
|
||||||
|
LoggerFactory.getLogger(CBlockIStorageImpl.class);
|
||||||
|
private static final Logger TRACER =
|
||||||
|
LoggerFactory.getLogger("TraceIO");
|
||||||
|
|
||||||
|
private CacheModule cache;
|
||||||
|
private final long volumeSize;
|
||||||
|
private final int blockSize;
|
||||||
|
private final String userName;
|
||||||
|
private final String volumeName;
|
||||||
|
private final boolean traceEnabled;
|
||||||
|
private final Configuration conf;
|
||||||
|
private final ContainerCacheFlusher flusher;
|
||||||
|
private List<Pipeline> fullContainerList;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* private: constructs a SCSI Target.
|
||||||
|
*
|
||||||
|
* @param config - config
|
||||||
|
* @param userName - Username
|
||||||
|
* @param volumeName - Name of the volume
|
||||||
|
* @param volumeSize - Size of the volume
|
||||||
|
* @param blockSize - Size of the block
|
||||||
|
* @param fullContainerList - Ordered list of containers that make up this
|
||||||
|
* volume.
|
||||||
|
* @param flusher - flusher which is used to flush data from
|
||||||
|
* level db cache to containers
|
||||||
|
* @throws IOException - Throws IOException.
|
||||||
|
*/
|
||||||
|
private CBlockIStorageImpl(Configuration config, String userName,
|
||||||
|
String volumeName, long volumeSize, int blockSize,
|
||||||
|
List<Pipeline> fullContainerList, ContainerCacheFlusher flusher) {
|
||||||
|
this.conf = config;
|
||||||
|
this.userName = userName;
|
||||||
|
this.volumeName = volumeName;
|
||||||
|
this.volumeSize = volumeSize;
|
||||||
|
this.blockSize = blockSize;
|
||||||
|
this.fullContainerList = new ArrayList<>(fullContainerList);
|
||||||
|
this.flusher = flusher;
|
||||||
|
this.traceEnabled = conf.getBoolean(DFS_CBLOCK_TRACE_IO,
|
||||||
|
DFS_CBLOCK_TRACE_IO_DEFAULT);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* private: initialize the cache.
|
||||||
|
*
|
||||||
|
* @param xceiverClientManager - client manager that is used for creating new
|
||||||
|
* connections to containers.
|
||||||
|
* @param metrics - target metrics to maintain metrics for target server
|
||||||
|
* @throws IOException - Throws IOException.
|
||||||
|
*/
|
||||||
|
private void initCache(XceiverClientManager xceiverClientManager,
|
||||||
|
CBlockTargetMetrics metrics) throws IOException {
|
||||||
|
this.cache = CBlockLocalCache.newBuilder()
|
||||||
|
.setConfiguration(conf)
|
||||||
|
.setVolumeName(this.volumeName)
|
||||||
|
.setUserName(this.userName)
|
||||||
|
.setPipelines(this.fullContainerList)
|
||||||
|
.setClientManager(xceiverClientManager)
|
||||||
|
.setBlockSize(blockSize)
|
||||||
|
.setVolumeSize(volumeSize)
|
||||||
|
.setFlusher(flusher)
|
||||||
|
.setCBlockTargetMetrics(metrics)
|
||||||
|
.build();
|
||||||
|
this.cache.start();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets a new builder for CBlockStorageImpl.
|
||||||
|
*
|
||||||
|
* @return builder
|
||||||
|
*/
|
||||||
|
public static Builder newBuilder() {
|
||||||
|
return new Builder();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get Cache.
|
||||||
|
*
|
||||||
|
* @return - Cache
|
||||||
|
*/
|
||||||
|
public CacheModule getCache() {
|
||||||
|
return cache;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns block size of this volume.
|
||||||
|
*
|
||||||
|
* @return int size of block for this volume.
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
public int getBlockSize() {
|
||||||
|
return blockSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Checks the index boundary of a block address.
|
||||||
|
*
|
||||||
|
* @param logicalBlockAddress the index of the first block of data to be read
|
||||||
|
* or written
|
||||||
|
* @param transferLengthInBlocks the total number of consecutive blocks about
|
||||||
|
* to be read or written
|
||||||
|
* @return 0 == Success, 1 indicates the LBA address is out of bounds and 2
|
||||||
|
* indicates that LBA + transfer size is out of bounds.
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
public int checkBounds(long logicalBlockAddress, int transferLengthInBlocks) {
|
||||||
|
long sizeInBlocks = volumeSize / blockSize;
|
||||||
|
int res = 0;
|
||||||
|
if (logicalBlockAddress < 0 || logicalBlockAddress >= sizeInBlocks) {
|
||||||
|
res = 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (transferLengthInBlocks < 0 ||
|
||||||
|
logicalBlockAddress + transferLengthInBlocks > sizeInBlocks) {
|
||||||
|
if (res == 0) {
|
||||||
|
res = 2;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return res;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Number of blocks that make up this volume.
|
||||||
|
*
|
||||||
|
* @return long - count of blocks.
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
public long getSizeInBlocks() {
|
||||||
|
return volumeSize / blockSize;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Reads the number of bytes that can be read into the bytes buffer from the
|
||||||
|
* location indicated.
|
||||||
|
*
|
||||||
|
* @param bytes the array into which the data will be copied will be filled
|
||||||
|
* with data from storage
|
||||||
|
* @param storageIndex the position of the first byte to be copied
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
public void read(byte[] bytes, long storageIndex) throws IOException {
|
||||||
|
int startingIdxInBlock = (int) storageIndex % blockSize;
|
||||||
|
int idxInBytes = 0;
|
||||||
|
if (this.traceEnabled) {
|
||||||
|
TRACER.info("Task=ReadStart,length={},location={}",
|
||||||
|
bytes.length, storageIndex);
|
||||||
|
}
|
||||||
|
while (idxInBytes < bytes.length - 1) {
|
||||||
|
long blockId = (storageIndex + idxInBytes) / blockSize;
|
||||||
|
byte[] dataBytes;
|
||||||
|
|
||||||
|
try {
|
||||||
|
LogicalBlock block = this.cache.get(blockId);
|
||||||
|
dataBytes = block.getData().array();
|
||||||
|
|
||||||
|
if (this.traceEnabled) {
|
||||||
|
TRACER.info("Task=ReadBlock,BlockID={},length={},SHA={}",
|
||||||
|
blockId,
|
||||||
|
dataBytes.length,
|
||||||
|
dataBytes.length > 0 ? DigestUtils.sha256Hex(dataBytes) : null);
|
||||||
|
}
|
||||||
|
} catch (IOException e) {
|
||||||
|
// For an non-existing block cache.get will return a block with zero
|
||||||
|
// bytes filled. So any error here is a real error.
|
||||||
|
LOGGER.error("getting errors when reading data:" + e);
|
||||||
|
throw e;
|
||||||
|
}
|
||||||
|
|
||||||
|
int length = blockSize - startingIdxInBlock;
|
||||||
|
if (length > bytes.length - idxInBytes) {
|
||||||
|
length = bytes.length - idxInBytes;
|
||||||
|
}
|
||||||
|
if (dataBytes.length >= length) {
|
||||||
|
System.arraycopy(dataBytes, startingIdxInBlock, bytes, idxInBytes,
|
||||||
|
length);
|
||||||
|
}
|
||||||
|
startingIdxInBlock = 0;
|
||||||
|
idxInBytes += length;
|
||||||
|
}
|
||||||
|
if (this.traceEnabled) {
|
||||||
|
TRACER.info("Task=ReadEnd,length={},location={},SHA={}",
|
||||||
|
bytes.length, storageIndex, DigestUtils.sha256Hex(bytes));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void write(byte[] bytes, long storageIndex) throws IOException {
|
||||||
|
int startingIdxInBlock = (int) storageIndex % blockSize;
|
||||||
|
int idxInBytes = 0;
|
||||||
|
if (this.traceEnabled) {
|
||||||
|
TRACER.info("Task=WriteStart,length={},location={},SHA={}",
|
||||||
|
bytes.length, storageIndex,
|
||||||
|
bytes.length > 0 ? DigestUtils.sha256Hex(bytes) : null);
|
||||||
|
}
|
||||||
|
|
||||||
|
ByteBuffer dataByte = ByteBuffer.allocate(blockSize);
|
||||||
|
while (idxInBytes < bytes.length - 1) {
|
||||||
|
long blockId = (storageIndex + idxInBytes) / blockSize;
|
||||||
|
int length = blockSize - startingIdxInBlock;
|
||||||
|
if (length > bytes.length - idxInBytes) {
|
||||||
|
length = bytes.length - idxInBytes;
|
||||||
|
}
|
||||||
|
System.arraycopy(bytes, idxInBytes, dataByte.array(), startingIdxInBlock,
|
||||||
|
length);
|
||||||
|
this.cache.put(blockId, dataByte.array());
|
||||||
|
|
||||||
|
if (this.traceEnabled) {
|
||||||
|
TRACER.info("Task=WriteBlock,BlockID={},length={},SHA={}",
|
||||||
|
blockId, dataByte.array().length,
|
||||||
|
dataByte.array().length > 0 ?
|
||||||
|
DigestUtils.sha256Hex(dataByte.array()) : null);
|
||||||
|
}
|
||||||
|
dataByte.clear();
|
||||||
|
startingIdxInBlock = 0;
|
||||||
|
idxInBytes += length;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (this.traceEnabled) {
|
||||||
|
TRACER.info("Task=WriteEnd,length={},location={} ",
|
||||||
|
bytes.length, storageIndex);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void close() throws IOException {
|
||||||
|
try {
|
||||||
|
cache.close();
|
||||||
|
} catch (IllegalStateException ise) {
|
||||||
|
LOGGER.error("Can not close the storage {}", ise);
|
||||||
|
throw ise;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Builder class for CBlocklocalCache.
|
||||||
|
*/
|
||||||
|
public static class Builder {
|
||||||
|
private String userName;
|
||||||
|
private String volumeName;
|
||||||
|
private long volumeSize;
|
||||||
|
private int blockSize;
|
||||||
|
private List<Pipeline> containerList;
|
||||||
|
private Configuration conf;
|
||||||
|
private XceiverClientManager clientManager;
|
||||||
|
private ContainerCacheFlusher flusher;
|
||||||
|
private CBlockTargetMetrics metrics;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructs a builder.
|
||||||
|
*/
|
||||||
|
Builder() {
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
public Builder setFlusher(ContainerCacheFlusher cacheFlusher) {
|
||||||
|
this.flusher = cacheFlusher;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* set config.
|
||||||
|
*
|
||||||
|
* @param config - config
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setConf(Configuration config) {
|
||||||
|
this.conf = config;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* set user name.
|
||||||
|
*
|
||||||
|
* @param cblockUserName - user name
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setUserName(String cblockUserName) {
|
||||||
|
this.userName = cblockUserName;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* set volume name.
|
||||||
|
*
|
||||||
|
* @param cblockVolumeName -- volume name
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setVolumeName(String cblockVolumeName) {
|
||||||
|
this.volumeName = cblockVolumeName;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* set volume size.
|
||||||
|
*
|
||||||
|
* @param cblockVolumeSize -- set volume size.
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setVolumeSize(long cblockVolumeSize) {
|
||||||
|
this.volumeSize = cblockVolumeSize;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* set block size.
|
||||||
|
*
|
||||||
|
* @param cblockBlockSize -- block size
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setBlockSize(int cblockBlockSize) {
|
||||||
|
this.blockSize = cblockBlockSize;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Set contianer list.
|
||||||
|
*
|
||||||
|
* @param cblockContainerList - set the pipeline list
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setContainerList(List<Pipeline> cblockContainerList) {
|
||||||
|
this.containerList = cblockContainerList;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Set client manager.
|
||||||
|
*
|
||||||
|
* @param xceiverClientManager -- sets the client manager.
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setClientManager(XceiverClientManager xceiverClientManager) {
|
||||||
|
this.clientManager = xceiverClientManager;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Set Cblock Target Metrics.
|
||||||
|
*
|
||||||
|
* @param targetMetrics -- sets the cblock target metrics
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setCBlockTargetMetrics(CBlockTargetMetrics targetMetrics) {
|
||||||
|
this.metrics = targetMetrics;
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Builds the CBlockStorageImpl.
|
||||||
|
*
|
||||||
|
* @return builds the CBlock Scsi Target.
|
||||||
|
*/
|
||||||
|
public CBlockIStorageImpl build() throws IOException {
|
||||||
|
if (StringUtils.isBlank(userName)) {
|
||||||
|
throw new IllegalArgumentException("User name cannot be null or empty" +
|
||||||
|
".");
|
||||||
|
}
|
||||||
|
if (StringUtils.isBlank(volumeName)) {
|
||||||
|
throw new IllegalArgumentException("Volume name cannot be null or " +
|
||||||
|
"empty");
|
||||||
|
}
|
||||||
|
|
||||||
|
if (volumeSize < 1) {
|
||||||
|
throw new IllegalArgumentException("Volume size cannot be negative or" +
|
||||||
|
" zero.");
|
||||||
|
}
|
||||||
|
|
||||||
|
if (blockSize < 1) {
|
||||||
|
throw new IllegalArgumentException("Block size cannot be negative or " +
|
||||||
|
"zero.");
|
||||||
|
}
|
||||||
|
|
||||||
|
if (containerList == null || containerList.size() == 0) {
|
||||||
|
throw new IllegalArgumentException("Container list cannot be null or " +
|
||||||
|
"empty");
|
||||||
|
}
|
||||||
|
if (clientManager == null) {
|
||||||
|
throw new IllegalArgumentException("Client manager cannot be null");
|
||||||
|
}
|
||||||
|
if (conf == null) {
|
||||||
|
throw new IllegalArgumentException("Configuration cannot be null");
|
||||||
|
}
|
||||||
|
|
||||||
|
if (flusher == null) {
|
||||||
|
throw new IllegalArgumentException("Flusher Cannot be null.");
|
||||||
|
}
|
||||||
|
CBlockIStorageImpl impl = new CBlockIStorageImpl(this.conf, this.userName,
|
||||||
|
this.volumeName, this.volumeSize, this.blockSize, this.containerList,
|
||||||
|
this.flusher);
|
||||||
|
impl.initCache(this.clientManager, this.metrics);
|
||||||
|
return impl;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,119 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper;
|
||||||
|
|
||||||
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
|
import org.apache.hadoop.metrics2.MetricsSystem;
|
||||||
|
import org.apache.hadoop.metrics2.annotation.Metric;
|
||||||
|
import org.apache.hadoop.metrics2.lib.DefaultMetricsSystem;
|
||||||
|
import org.apache.hadoop.metrics2.lib.MutableCounterLong;
|
||||||
|
import org.apache.hadoop.metrics2.lib.MutableRate;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This class is for maintaining the various Cblock Target statistics
|
||||||
|
* and publishing them through the metrics interfaces.
|
||||||
|
* This also registers the JMX MBean for RPC.
|
||||||
|
*
|
||||||
|
* This class maintains stats like cache hit and miss ratio
|
||||||
|
* as well as the latency time of read and write ops.
|
||||||
|
*/
|
||||||
|
public class CBlockTargetMetrics {
|
||||||
|
@Metric private MutableCounterLong numReadOps;
|
||||||
|
@Metric private MutableCounterLong numWriteOps;
|
||||||
|
@Metric private MutableCounterLong numReadCacheHits;
|
||||||
|
@Metric private MutableCounterLong numReadCacheMiss;
|
||||||
|
@Metric private MutableCounterLong numReadLostBlocks;
|
||||||
|
|
||||||
|
@Metric private MutableRate dbReadLatency;
|
||||||
|
@Metric private MutableRate containerReadLatency;
|
||||||
|
|
||||||
|
@Metric private MutableRate dbWriteLatency;
|
||||||
|
@Metric private MutableRate containerWriteLatency;
|
||||||
|
|
||||||
|
public CBlockTargetMetrics() {
|
||||||
|
}
|
||||||
|
|
||||||
|
public static CBlockTargetMetrics create() {
|
||||||
|
MetricsSystem ms = DefaultMetricsSystem.instance();
|
||||||
|
return ms.register("CBlockTargetMetrics",
|
||||||
|
"CBlock Target Metrics",
|
||||||
|
new CBlockTargetMetrics());
|
||||||
|
}
|
||||||
|
|
||||||
|
public void incNumReadOps() {
|
||||||
|
numReadOps.incr();
|
||||||
|
}
|
||||||
|
|
||||||
|
public void incNumWriteOps() {
|
||||||
|
numWriteOps.incr();
|
||||||
|
}
|
||||||
|
|
||||||
|
public void incNumReadCacheHits() {
|
||||||
|
numReadCacheHits.incr();
|
||||||
|
}
|
||||||
|
|
||||||
|
public void incNumReadCacheMiss() {
|
||||||
|
numReadCacheMiss.incr();
|
||||||
|
}
|
||||||
|
|
||||||
|
public void incNumReadLostBlocks() {
|
||||||
|
numReadLostBlocks.incr();
|
||||||
|
}
|
||||||
|
|
||||||
|
public void updateDBReadLatency(long latency) {
|
||||||
|
dbReadLatency.add(latency);
|
||||||
|
}
|
||||||
|
|
||||||
|
public void updateContainerReadLatency(long latency) {
|
||||||
|
containerReadLatency.add(latency);
|
||||||
|
}
|
||||||
|
|
||||||
|
public void updateDBWriteLatency(long latency) {
|
||||||
|
dbWriteLatency.add(latency);
|
||||||
|
}
|
||||||
|
|
||||||
|
public void updateContainerWriteLatency(long latency) {
|
||||||
|
containerWriteLatency.add(latency);
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public long getNumReadOps() {
|
||||||
|
return numReadOps.value();
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public long getNumWriteOps() {
|
||||||
|
return numWriteOps.value();
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public long getNumReadCacheHits() {
|
||||||
|
return numReadCacheHits.value();
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public long getNumReadCacheMiss() {
|
||||||
|
return numReadCacheMiss.value();
|
||||||
|
}
|
||||||
|
|
||||||
|
@VisibleForTesting
|
||||||
|
public long getNumReadLostBlocks() {
|
||||||
|
return numReadLostBlocks.value();
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,512 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
* <p>
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
* <p>
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper;
|
||||||
|
|
||||||
|
import com.google.common.base.Preconditions;
|
||||||
|
import com.google.common.primitives.Longs;
|
||||||
|
import com.google.common.util.concurrent.ThreadFactoryBuilder;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.cache.LogicalBlock;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.cache.impl.DiskBlock;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.scm.XceiverClientManager;
|
||||||
|
import org.apache.hadoop.scm.container.common.helpers.Pipeline;
|
||||||
|
import org.apache.hadoop.utils.LevelDBStore;
|
||||||
|
import org.iq80.leveldb.Options;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
import java.io.File;
|
||||||
|
import java.io.FileInputStream;
|
||||||
|
import java.io.FileNotFoundException;
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
import java.nio.channels.ReadableByteChannel;
|
||||||
|
import java.nio.file.Files;
|
||||||
|
import java.nio.file.Path;
|
||||||
|
import java.nio.file.Paths;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.HashMap;
|
||||||
|
import java.util.Map;
|
||||||
|
import java.util.concurrent.ArrayBlockingQueue;
|
||||||
|
import java.util.concurrent.ConcurrentHashMap;
|
||||||
|
import java.util.concurrent.ConcurrentMap;
|
||||||
|
import java.util.concurrent.LinkedBlockingQueue;
|
||||||
|
import java.util.concurrent.ThreadFactory;
|
||||||
|
import java.util.concurrent.ThreadPoolExecutor;
|
||||||
|
import java.util.concurrent.TimeUnit;
|
||||||
|
import java.util.concurrent.atomic.AtomicBoolean;
|
||||||
|
import java.util.concurrent.atomic.AtomicInteger;
|
||||||
|
import java.util.concurrent.atomic.AtomicLong;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_BLOCK_BUFFER_SIZE;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_BLOCK_BUFFER_SIZE_DEFAULT;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_CORE_POOL_SIZE;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_CORE_POOL_SIZE_DEFAULT;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_KEEP_ALIVE_SECONDS;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_KEEP_ALIVE_SECONDS_DEFAULT;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_MAX_POOL_SIZE;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_MAX_POOL_SIZE_DEFAULT;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_QUEUE_SIZE_KB;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_QUEUE_SIZE_KB_DEFAULT;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_THREAD_PRIORITY;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys
|
||||||
|
.DFS_CBLOCK_CACHE_THREAD_PRIORITY_DEFAULT;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys.DFS_CBLOCK_DISK_CACHE_PATH_DEFAULT;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys.DFS_CBLOCK_DISK_CACHE_PATH_KEY;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Class that writes to remote containers.
|
||||||
|
*/
|
||||||
|
public class ContainerCacheFlusher implements Runnable {
|
||||||
|
private static final Logger LOG =
|
||||||
|
LoggerFactory.getLogger(ContainerCacheFlusher.class);
|
||||||
|
private final LinkedBlockingQueue<Message> messageQueue;
|
||||||
|
private final ThreadPoolExecutor threadPoolExecutor;
|
||||||
|
private final ArrayBlockingQueue<Runnable> workQueue;
|
||||||
|
private final ConcurrentMap<String, RefCountedDB> dbMap;
|
||||||
|
private final ByteBuffer blockIDBuffer;
|
||||||
|
private final ConcurrentMap<String, Pipeline[]> pipelineMap;
|
||||||
|
private final AtomicLong remoteIO;
|
||||||
|
private final XceiverClientManager xceiverClientManager;
|
||||||
|
private final CBlockTargetMetrics metrics;
|
||||||
|
private AtomicBoolean shutdown;
|
||||||
|
|
||||||
|
private final ConcurrentMap<String, FinishCounter> finishCountMap;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructs the writers to remote queue.
|
||||||
|
*/
|
||||||
|
public ContainerCacheFlusher(Configuration config,
|
||||||
|
XceiverClientManager xceiverClientManager,
|
||||||
|
CBlockTargetMetrics metrics) {
|
||||||
|
int queueSize = config.getInt(DFS_CBLOCK_CACHE_QUEUE_SIZE_KB,
|
||||||
|
DFS_CBLOCK_CACHE_QUEUE_SIZE_KB_DEFAULT) * 1024;
|
||||||
|
int corePoolSize = config.getInt(DFS_CBLOCK_CACHE_CORE_POOL_SIZE,
|
||||||
|
DFS_CBLOCK_CACHE_CORE_POOL_SIZE_DEFAULT);
|
||||||
|
int maxPoolSize = config.getInt(DFS_CBLOCK_CACHE_MAX_POOL_SIZE,
|
||||||
|
DFS_CBLOCK_CACHE_MAX_POOL_SIZE_DEFAULT);
|
||||||
|
long keepAlive = config.getLong(DFS_CBLOCK_CACHE_KEEP_ALIVE_SECONDS,
|
||||||
|
DFS_CBLOCK_CACHE_KEEP_ALIVE_SECONDS_DEFAULT);
|
||||||
|
int threadPri = config.getInt(DFS_CBLOCK_CACHE_THREAD_PRIORITY,
|
||||||
|
DFS_CBLOCK_CACHE_THREAD_PRIORITY_DEFAULT);
|
||||||
|
int blockBufferSize = config.getInt(DFS_CBLOCK_CACHE_BLOCK_BUFFER_SIZE,
|
||||||
|
DFS_CBLOCK_CACHE_BLOCK_BUFFER_SIZE_DEFAULT) * 1024;
|
||||||
|
|
||||||
|
LOG.info("Cache: Core Pool Size: {}", corePoolSize);
|
||||||
|
LOG.info("Cache: Keep Alive: {}", keepAlive);
|
||||||
|
LOG.info("Cache: Max Pool Size: {}", maxPoolSize);
|
||||||
|
LOG.info("Cache: Thread Pri: {}", threadPri);
|
||||||
|
LOG.info("Cache: BlockBuffer Size: {}", blockBufferSize);
|
||||||
|
|
||||||
|
shutdown = new AtomicBoolean(false);
|
||||||
|
messageQueue = new LinkedBlockingQueue<>();
|
||||||
|
workQueue = new ArrayBlockingQueue<>(queueSize, true);
|
||||||
|
|
||||||
|
ThreadFactory workerThreadFactory = new ThreadFactoryBuilder()
|
||||||
|
.setNameFormat("Cache Block Writer Thread #%d")
|
||||||
|
.setDaemon(true)
|
||||||
|
.setPriority(threadPri)
|
||||||
|
.build();
|
||||||
|
threadPoolExecutor = new ThreadPoolExecutor(corePoolSize, maxPoolSize,
|
||||||
|
keepAlive, TimeUnit.SECONDS, workQueue, workerThreadFactory,
|
||||||
|
new ThreadPoolExecutor.AbortPolicy());
|
||||||
|
threadPoolExecutor.prestartAllCoreThreads();
|
||||||
|
|
||||||
|
dbMap = new ConcurrentHashMap<>();
|
||||||
|
pipelineMap = new ConcurrentHashMap<>();
|
||||||
|
blockIDBuffer = ByteBuffer.allocateDirect(blockBufferSize);
|
||||||
|
this.xceiverClientManager = xceiverClientManager;
|
||||||
|
this.metrics = metrics;
|
||||||
|
this.remoteIO = new AtomicLong();
|
||||||
|
|
||||||
|
this.finishCountMap = new ConcurrentHashMap<>();
|
||||||
|
checkExisitingDirtyLog(config);
|
||||||
|
}
|
||||||
|
|
||||||
|
private void checkExisitingDirtyLog(Configuration config) {
|
||||||
|
File dbPath = Paths.get(config.get(DFS_CBLOCK_DISK_CACHE_PATH_KEY,
|
||||||
|
DFS_CBLOCK_DISK_CACHE_PATH_DEFAULT)).toFile();
|
||||||
|
if (!dbPath.exists()) {
|
||||||
|
LOG.info("No existing dirty log found at {}", dbPath);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
LOG.info("Need to check and requeue existing dirty log {}", dbPath);
|
||||||
|
HashMap<String, ArrayList<String>> allFiles = new HashMap<>();
|
||||||
|
traverse(dbPath, allFiles);
|
||||||
|
for (Map.Entry<String, ArrayList<String>> entry : allFiles.entrySet()) {
|
||||||
|
String parentPath = entry.getKey();
|
||||||
|
for (String fileName : entry.getValue()) {
|
||||||
|
LOG.info("found this {} with {}", parentPath, fileName);
|
||||||
|
processDirtyBlocks(parentPath, fileName);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void traverse(File path, HashMap<String, ArrayList<String>> files) {
|
||||||
|
if (path.isFile()) {
|
||||||
|
if (path.getName().startsWith("DirtyLog")) {
|
||||||
|
LOG.debug("found this {} with {}", path.getParent(), path.getName());
|
||||||
|
if (!files.containsKey(path.getParent())) {
|
||||||
|
files.put(path.getParent(), new ArrayList<>());
|
||||||
|
}
|
||||||
|
files.get(path.getParent()).add(path.getName());
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
File[] listFiles = path.listFiles();
|
||||||
|
if (listFiles != null) {
|
||||||
|
for (File subPath : listFiles) {
|
||||||
|
traverse(subPath, files);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets the CBlockTargetMetrics.
|
||||||
|
*
|
||||||
|
* @return CBlockTargetMetrics
|
||||||
|
*/
|
||||||
|
public CBlockTargetMetrics getTargetMetrics() {
|
||||||
|
return metrics;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets the getXceiverClientManager.
|
||||||
|
*
|
||||||
|
* @return XceiverClientManager
|
||||||
|
*/
|
||||||
|
public XceiverClientManager getXceiverClientManager() {
|
||||||
|
return xceiverClientManager;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Shutdown this instance.
|
||||||
|
*/
|
||||||
|
public void shutdown() {
|
||||||
|
this.shutdown.set(true);
|
||||||
|
threadPoolExecutor.shutdown();
|
||||||
|
}
|
||||||
|
|
||||||
|
public long incrementremoteIO() {
|
||||||
|
return remoteIO.incrementAndGet();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Processes a block cache file and queues those blocks for the remote I/O.
|
||||||
|
*
|
||||||
|
* @param dbPath - Location where the DB can be found.
|
||||||
|
* @param fileName - Block Cache File Name
|
||||||
|
*/
|
||||||
|
public void processDirtyBlocks(String dbPath, String fileName) {
|
||||||
|
LOG.info("Adding {}/{} to queue. Queue Length: {}", dbPath, fileName,
|
||||||
|
messageQueue.size());
|
||||||
|
this.messageQueue.add(new Message(dbPath, fileName));
|
||||||
|
}
|
||||||
|
|
||||||
|
public Logger getLOG() {
|
||||||
|
return LOG;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Opens a DB if needed or returns a handle to an already open DB.
|
||||||
|
*
|
||||||
|
* @param dbPath -- dbPath
|
||||||
|
* @param cacheSize - cacheSize
|
||||||
|
* @return the levelDB on the given path.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
public synchronized LevelDBStore openDB(String dbPath, int cacheSize)
|
||||||
|
throws IOException {
|
||||||
|
if (dbMap.containsKey(dbPath)) {
|
||||||
|
RefCountedDB refDB = dbMap.get(dbPath);
|
||||||
|
refDB.open();
|
||||||
|
return refDB.db;
|
||||||
|
} else {
|
||||||
|
Options options = new Options();
|
||||||
|
options.cacheSize(cacheSize * (1024L * 1024L));
|
||||||
|
options.createIfMissing(true);
|
||||||
|
LevelDBStore cacheDB = new LevelDBStore(
|
||||||
|
new File(getDBFileName(dbPath)), options);
|
||||||
|
RefCountedDB refDB = new RefCountedDB(dbPath, cacheDB);
|
||||||
|
dbMap.put(dbPath, refDB);
|
||||||
|
return cacheDB;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Updates the contianer map. This data never changes so we will update this
|
||||||
|
* during restarts and it should not hurt us.
|
||||||
|
*
|
||||||
|
* @param dbPath - DbPath
|
||||||
|
* @param containerList - Contianer List.
|
||||||
|
*/
|
||||||
|
public void register(String dbPath, Pipeline[] containerList) {
|
||||||
|
pipelineMap.put(dbPath, containerList);
|
||||||
|
}
|
||||||
|
|
||||||
|
private String getDBFileName(String dbPath) {
|
||||||
|
return dbPath + ".db";
|
||||||
|
}
|
||||||
|
|
||||||
|
LevelDBStore getCacheDB(String dbPath) {
|
||||||
|
return dbMap.get(dbPath).db;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Close the DB if we don't have any outstanding refrences.
|
||||||
|
*
|
||||||
|
* @param dbPath - dbPath
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
public synchronized void closeDB(String dbPath) throws IOException {
|
||||||
|
if (dbMap.containsKey(dbPath)) {
|
||||||
|
RefCountedDB refDB = dbMap.get(dbPath);
|
||||||
|
int count = refDB.close();
|
||||||
|
if (count == 0) {
|
||||||
|
dbMap.remove(dbPath);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Pipeline getPipeline(String dbPath, long blockId) {
|
||||||
|
Pipeline[] containerList = pipelineMap.get(dbPath);
|
||||||
|
Preconditions.checkNotNull(containerList);
|
||||||
|
int containerIdx = (int) blockId % containerList.length;
|
||||||
|
long cBlockIndex =
|
||||||
|
Longs.fromByteArray(containerList[containerIdx].getData());
|
||||||
|
if (cBlockIndex > 0) {
|
||||||
|
// This catches the case when we get a wrong container in the ordering
|
||||||
|
// of the containers.
|
||||||
|
Preconditions.checkState(containerIdx % cBlockIndex == 0,
|
||||||
|
"The container ID computed should match with the container index " +
|
||||||
|
"returned from cBlock Server.");
|
||||||
|
}
|
||||||
|
return containerList[containerIdx];
|
||||||
|
}
|
||||||
|
|
||||||
|
public void incFinishCount(String fileName) {
|
||||||
|
if (!finishCountMap.containsKey(fileName)) {
|
||||||
|
LOG.error("No record for such file:" + fileName);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
finishCountMap.get(fileName).incCount();
|
||||||
|
if (finishCountMap.get(fileName).isFileDeleted()) {
|
||||||
|
finishCountMap.remove(fileName);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* When an object implementing interface <code>Runnable</code> is used
|
||||||
|
* to create a thread, starting the thread causes the object's
|
||||||
|
* <code>run</code> method to be called in that separately executing
|
||||||
|
* thread.
|
||||||
|
* <p>
|
||||||
|
* The general contract of the method <code>run</code> is that it may
|
||||||
|
* take any action whatsoever.
|
||||||
|
*
|
||||||
|
* @see Thread#run()
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
public void run() {
|
||||||
|
while (!this.shutdown.get()) {
|
||||||
|
try {
|
||||||
|
Message message = messageQueue.take();
|
||||||
|
LOG.debug("Got message to process -- DB Path : {} , FileName; {}",
|
||||||
|
message.getDbPath(), message.getFileName());
|
||||||
|
String fullPath = Paths.get(message.getDbPath(),
|
||||||
|
message.getFileName()).toString();
|
||||||
|
ReadableByteChannel fileChannel = new FileInputStream(fullPath)
|
||||||
|
.getChannel();
|
||||||
|
// TODO: We can batch and unique the IOs here. First getting the code
|
||||||
|
// to work, we will add those later.
|
||||||
|
int bytesRead = fileChannel.read(blockIDBuffer);
|
||||||
|
LOG.debug("Read blockID log of size: {} position {} remaining {}",
|
||||||
|
bytesRead, blockIDBuffer.position(), blockIDBuffer.remaining());
|
||||||
|
// current position of in the buffer in bytes, divided by number of
|
||||||
|
// bytes per long (which is calculated by number of bits per long
|
||||||
|
// divided by number of bits per byte) gives the number of blocks
|
||||||
|
int blockCount = blockIDBuffer.position()/(Long.SIZE / Byte.SIZE);
|
||||||
|
if (finishCountMap.containsKey(message.getFileName())) {
|
||||||
|
// In theory this should never happen. But if it happened,
|
||||||
|
// we need to know it...
|
||||||
|
LOG.error("Adding DirtyLog file again {} current count {} new {}",
|
||||||
|
message.getFileName(),
|
||||||
|
finishCountMap.get(message.getFileName()).expectedCount,
|
||||||
|
blockCount);
|
||||||
|
}
|
||||||
|
finishCountMap.put(message.getFileName(),
|
||||||
|
new FinishCounter(blockCount, message.getDbPath(),
|
||||||
|
message.getFileName()));
|
||||||
|
// should be flip instead of rewind, because we also need to make sure
|
||||||
|
// the end position is correct.
|
||||||
|
blockIDBuffer.flip();
|
||||||
|
LOG.debug("Remaining blocks count {} and {}", blockIDBuffer.remaining(),
|
||||||
|
blockCount);
|
||||||
|
while (blockIDBuffer.remaining() >= (Long.SIZE / Byte.SIZE)) {
|
||||||
|
long blockID = blockIDBuffer.getLong();
|
||||||
|
LogicalBlock block = new DiskBlock(blockID, null, false);
|
||||||
|
BlockWriterTask blockWriterTask = new BlockWriterTask(block, this,
|
||||||
|
message.getDbPath(), message.getFileName());
|
||||||
|
threadPoolExecutor.submit(blockWriterTask);
|
||||||
|
}
|
||||||
|
blockIDBuffer.clear();
|
||||||
|
} catch (InterruptedException e) {
|
||||||
|
LOG.info("ContainerCacheFlusher is interrupted.", e);
|
||||||
|
} catch (FileNotFoundException e) {
|
||||||
|
LOG.error("Unable to find the dirty blocks file. This will cause " +
|
||||||
|
"data errors. Please stop using this volume.", e);
|
||||||
|
} catch (IOException e) {
|
||||||
|
LOG.error("Unable to read the dirty blocks file. This will cause " +
|
||||||
|
"data errors. Please stop using this volume.", e);
|
||||||
|
} catch (Exception e) {
|
||||||
|
LOG.error("Generic exception.", e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
LOG.info("Exiting flusher");
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Keeps a Reference counted DB that we close only when the total Reference
|
||||||
|
* has gone to zero.
|
||||||
|
*/
|
||||||
|
private static class RefCountedDB {
|
||||||
|
private LevelDBStore db;
|
||||||
|
private AtomicInteger refcount;
|
||||||
|
private String dbPath;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* RefCountedDB DB ctor.
|
||||||
|
*
|
||||||
|
* @param dbPath - DB path.
|
||||||
|
* @param db - LevelDBStore db
|
||||||
|
*/
|
||||||
|
RefCountedDB(String dbPath, LevelDBStore db) {
|
||||||
|
this.db = db;
|
||||||
|
this.refcount = new AtomicInteger(1);
|
||||||
|
this.dbPath = dbPath;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* close the DB if possible.
|
||||||
|
*/
|
||||||
|
public int close() throws IOException {
|
||||||
|
int count = this.refcount.decrementAndGet();
|
||||||
|
if (count == 0) {
|
||||||
|
LOG.info("Closing the LevelDB. {} ", this.dbPath);
|
||||||
|
db.close();
|
||||||
|
}
|
||||||
|
return count;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void open() {
|
||||||
|
this.refcount.incrementAndGet();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The message held in processing queue.
|
||||||
|
*/
|
||||||
|
private static class Message {
|
||||||
|
private String dbPath;
|
||||||
|
private String fileName;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A message that holds the info about which path dirty blocks log and
|
||||||
|
* which path contains db.
|
||||||
|
*
|
||||||
|
* @param dbPath
|
||||||
|
* @param fileName
|
||||||
|
*/
|
||||||
|
Message(String dbPath, String fileName) {
|
||||||
|
this.dbPath = dbPath;
|
||||||
|
this.fileName = fileName;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getDbPath() {
|
||||||
|
return dbPath;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setDbPath(String dbPath) {
|
||||||
|
this.dbPath = dbPath;
|
||||||
|
}
|
||||||
|
|
||||||
|
public String getFileName() {
|
||||||
|
return fileName;
|
||||||
|
}
|
||||||
|
|
||||||
|
public void setFileName(String fileName) {
|
||||||
|
this.fileName = fileName;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private static class FinishCounter {
|
||||||
|
private final long expectedCount;
|
||||||
|
private final String dbPath;
|
||||||
|
private final String dirtyLogPath;
|
||||||
|
private final AtomicLong currentCount;
|
||||||
|
private AtomicBoolean fileDeleted;
|
||||||
|
|
||||||
|
FinishCounter(long expectedCount, String dbPath,
|
||||||
|
String dirtyLogPath) {
|
||||||
|
this.expectedCount = expectedCount;
|
||||||
|
this.dbPath = dbPath;
|
||||||
|
this.dirtyLogPath = dirtyLogPath;
|
||||||
|
this.currentCount = new AtomicLong(0);
|
||||||
|
this.fileDeleted = new AtomicBoolean(false);
|
||||||
|
}
|
||||||
|
|
||||||
|
public boolean isFileDeleted() {
|
||||||
|
return fileDeleted.get();
|
||||||
|
}
|
||||||
|
|
||||||
|
public void incCount() {
|
||||||
|
long count = this.currentCount.incrementAndGet();
|
||||||
|
if (count >= expectedCount) {
|
||||||
|
String filePath = String.format("%s/%s", dbPath, dirtyLogPath);
|
||||||
|
LOG.debug(
|
||||||
|
"Deleting {} with count {} {}", filePath, count, expectedCount);
|
||||||
|
try {
|
||||||
|
Path path = Paths.get(filePath);
|
||||||
|
Files.delete(path);
|
||||||
|
// the following part tries to remove the directory if it is empty
|
||||||
|
// but not sufficient, because the .db directory still exists....
|
||||||
|
// TODO how to handle the .db directory?
|
||||||
|
/*Path parent = path.getParent();
|
||||||
|
if (parent.toFile().listFiles().length == 0) {
|
||||||
|
Files.delete(parent);
|
||||||
|
}*/
|
||||||
|
fileDeleted.set(true);
|
||||||
|
} catch (IOException e) {
|
||||||
|
LOG.error(
|
||||||
|
"Error deleting dirty log file {} {}", filePath, e.toString());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,52 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper.cache;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Defines the interface for cache implementations. The cache will be called
|
||||||
|
* by cblock storage module when it performs IO operations.
|
||||||
|
*/
|
||||||
|
public interface CacheModule {
|
||||||
|
/**
|
||||||
|
* check if the key is cached, if yes, returned the cached object.
|
||||||
|
* otherwise, load from data source. Then put it into cache.
|
||||||
|
*
|
||||||
|
* @param blockID
|
||||||
|
* @return the target block.
|
||||||
|
*/
|
||||||
|
LogicalBlock get(long blockID) throws IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* put the value of the key into cache.
|
||||||
|
* @param blockID
|
||||||
|
* @param value
|
||||||
|
*/
|
||||||
|
void put(long blockID, byte[] value) throws IOException;
|
||||||
|
|
||||||
|
void flush() throws IOException;
|
||||||
|
|
||||||
|
void start() throws IOException;
|
||||||
|
|
||||||
|
void stop() throws IOException;
|
||||||
|
|
||||||
|
void close() throws IOException;
|
||||||
|
|
||||||
|
boolean isDirtyCache();
|
||||||
|
}
|
|
@ -0,0 +1,50 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper.cache;
|
||||||
|
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Logical Block is the data structure that we write to the cache,
|
||||||
|
* the key and data gets written to remote contianers. Rest is used for
|
||||||
|
* book keeping for the cache.
|
||||||
|
*/
|
||||||
|
public interface LogicalBlock {
|
||||||
|
/**
|
||||||
|
* Returns the data stream of this block.
|
||||||
|
* @return - ByteBuffer
|
||||||
|
*/
|
||||||
|
ByteBuffer getData();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Frees the byte buffer since we don't need it any more.
|
||||||
|
*/
|
||||||
|
void clearData();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns the Block ID for this Block.
|
||||||
|
* @return long - BlockID
|
||||||
|
*/
|
||||||
|
long getBlockID();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Flag that tells us if this block has been persisted to container.
|
||||||
|
* @return whether this block is now persistent
|
||||||
|
*/
|
||||||
|
boolean isPersisted();
|
||||||
|
}
|
|
@ -0,0 +1,180 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper.cache.impl;
|
||||||
|
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.CBlockTargetMetrics;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.ContainerCacheFlusher;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.cache.CacheModule;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.cache.LogicalBlock;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.scm.XceiverClientManager;
|
||||||
|
import org.apache.hadoop.scm.container.common.helpers.Pipeline;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A local cache used by the CBlock ISCSI server. This class is enabled or
|
||||||
|
* disabled via config settings.
|
||||||
|
*
|
||||||
|
* TODO : currently, this class is a just a place holder.
|
||||||
|
*/
|
||||||
|
final public class CBlockLocalCache implements CacheModule {
|
||||||
|
|
||||||
|
private CBlockLocalCache() {
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public LogicalBlock get(long blockID) throws IOException {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void put(long blockID, byte[] data) throws IOException {
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void flush() throws IOException {
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void start() throws IOException {
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void stop() throws IOException {
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void close() throws IOException {
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean isDirtyCache() {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
public static Builder newBuilder() {
|
||||||
|
return new Builder();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Builder class for CBlocklocalCache.
|
||||||
|
*/
|
||||||
|
public static class Builder {
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the Config to be used by this cache.
|
||||||
|
*
|
||||||
|
* @param configuration - Config
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setConfiguration(Configuration configuration) {
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the user name who is the owner of this volume.
|
||||||
|
*
|
||||||
|
* @param userName - name of the owner, please note this is not the current
|
||||||
|
* user name.
|
||||||
|
* @return - Builder
|
||||||
|
*/
|
||||||
|
public Builder setUserName(String userName) {
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the VolumeName.
|
||||||
|
*
|
||||||
|
* @param volumeName - Name of the volume
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setVolumeName(String volumeName) {
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the Pipelines that form this volume.
|
||||||
|
*
|
||||||
|
* @param pipelines - list of pipelines
|
||||||
|
* @return Builder
|
||||||
|
*/
|
||||||
|
public Builder setPipelines(List<Pipeline> pipelines) {
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the Client Manager that manages the communication with containers.
|
||||||
|
*
|
||||||
|
* @param clientManager - clientManager.
|
||||||
|
* @return - Builder
|
||||||
|
*/
|
||||||
|
public Builder setClientManager(XceiverClientManager clientManager) {
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the block size -- Typical sizes are 4KB, 8KB etc.
|
||||||
|
*
|
||||||
|
* @param blockSize - BlockSize.
|
||||||
|
* @return - Builder
|
||||||
|
*/
|
||||||
|
public Builder setBlockSize(int blockSize) {
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the volumeSize.
|
||||||
|
*
|
||||||
|
* @param volumeSize - VolumeSize
|
||||||
|
* @return - Builder
|
||||||
|
*/
|
||||||
|
public Builder setVolumeSize(long volumeSize) {
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Set flusher.
|
||||||
|
* @param flusher - cache Flusher
|
||||||
|
* @return Builder.
|
||||||
|
*/
|
||||||
|
public Builder setFlusher(ContainerCacheFlusher flusher) {
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the cblock Metrics.
|
||||||
|
*
|
||||||
|
* @param targetMetrics - CBlock Target Metrics
|
||||||
|
* @return - Builder
|
||||||
|
*/
|
||||||
|
public Builder setCBlockTargetMetrics(CBlockTargetMetrics targetMetrics) {
|
||||||
|
return this;
|
||||||
|
}
|
||||||
|
|
||||||
|
public CBlockLocalCache build() throws IOException {
|
||||||
|
return new CBlockLocalCache();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,77 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper.cache.impl;
|
||||||
|
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.cache.LogicalBlock;
|
||||||
|
import java.nio.ByteBuffer;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Impl class for LogicalBlock.
|
||||||
|
*/
|
||||||
|
public class DiskBlock implements LogicalBlock {
|
||||||
|
private ByteBuffer data;
|
||||||
|
private long blockID;
|
||||||
|
private boolean persisted;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructs a DiskBlock Class from the following params.
|
||||||
|
* @param blockID - 64-bit block ID
|
||||||
|
* @param data - Byte Array
|
||||||
|
* @param persisted - Flag which tells us if this is persisted to remote
|
||||||
|
*/
|
||||||
|
public DiskBlock(long blockID, byte[] data, boolean persisted) {
|
||||||
|
if (data !=null) {
|
||||||
|
this.data = ByteBuffer.wrap(data);
|
||||||
|
}
|
||||||
|
this.blockID = blockID;
|
||||||
|
this.persisted = persisted;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ByteBuffer getData() {
|
||||||
|
return data;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Frees the byte buffer since we don't need it any more.
|
||||||
|
*/
|
||||||
|
@Override
|
||||||
|
public void clearData() {
|
||||||
|
data.clear();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public long getBlockID() {
|
||||||
|
return blockID;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean isPersisted() {
|
||||||
|
return persisted;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Sets the value of persisted.
|
||||||
|
* @param value - True if this has been persisted to container, false
|
||||||
|
* otherwise.
|
||||||
|
*/
|
||||||
|
public void setPersisted(boolean value) {
|
||||||
|
persisted = value;
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
|
@ -0,0 +1,18 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper.cache.impl;
|
|
@ -0,0 +1,18 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper.cache;
|
|
@ -0,0 +1,18 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock.jscsiHelper;
|
|
@ -0,0 +1,157 @@
|
||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.cblock;
|
||||||
|
|
||||||
|
import com.google.common.primitives.Longs;
|
||||||
|
import org.apache.commons.codec.digest.DigestUtils;
|
||||||
|
import org.apache.commons.lang.RandomStringUtils;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.CBlockIStorageImpl;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.CBlockTargetMetrics;
|
||||||
|
import org.apache.hadoop.cblock.jscsiHelper.ContainerCacheFlusher;
|
||||||
|
import org.apache.hadoop.io.IOUtils;
|
||||||
|
import org.apache.hadoop.ozone.MiniOzoneCluster;
|
||||||
|
import org.apache.hadoop.ozone.OzoneConfiguration;
|
||||||
|
import org.apache.hadoop.ozone.container.ozoneimpl.TestOzoneContainer;
|
||||||
|
import org.apache.hadoop.scm.XceiverClientManager;
|
||||||
|
import org.apache.hadoop.scm.XceiverClientSpi;
|
||||||
|
import org.apache.hadoop.scm.container.common.helpers.Pipeline;
|
||||||
|
import org.apache.hadoop.scm.protocolPB.StorageContainerLocationProtocolClientSideTranslatorPB;
|
||||||
|
import org.apache.hadoop.scm.storage.ContainerProtocolCalls;
|
||||||
|
import org.apache.hadoop.test.GenericTestUtils;
|
||||||
|
import org.junit.AfterClass;
|
||||||
|
import org.junit.BeforeClass;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.net.URL;
|
||||||
|
import java.nio.charset.StandardCharsets;
|
||||||
|
import java.util.LinkedList;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys.DFS_CBLOCK_DISK_CACHE_PATH_KEY;
|
||||||
|
import static org.apache.hadoop.cblock.CBlockConfigKeys.DFS_CBLOCK_TRACE_IO;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This class tests the cblock storage layer.
|
||||||
|
*/
|
||||||
|
public class TestStorageImpl {
|
||||||
|
|
||||||
|
private static final Logger LOG =
|
||||||
|
LoggerFactory.getLogger(TestStorageImpl.class);
|
||||||
|
private final static long GB = 1024 * 1024 * 1024;
|
||||||
|
private final static int KB = 1024;
|
||||||
|
private static MiniOzoneCluster cluster;
|
||||||
|
private static OzoneConfiguration config;
|
||||||
|
private static StorageContainerLocationProtocolClientSideTranslatorPB
|
||||||
|
storageContainerLocationClient;
|
||||||
|
private static XceiverClientManager xceiverClientManager;
|
||||||
|
|
||||||
|
@BeforeClass
|
||||||
|
public static void init() throws IOException {
|
||||||
|
config = new OzoneConfiguration();
|
||||||
|
URL p = config.getClass().getResource("");
|
||||||
|
String path = p.getPath().concat(
|
||||||
|
TestOzoneContainer.class.getSimpleName());
|
||||||
|
config.set(DFS_CBLOCK_DISK_CACHE_PATH_KEY, path);
|
||||||
|
config.setBoolean(DFS_CBLOCK_TRACE_IO, true);
|
||||||
|
cluster = new MiniOzoneCluster.Builder(config)
|
||||||
|
.numDataNodes(1).setHandlerType("distributed").build();
|
||||||
|
storageContainerLocationClient = cluster
|
||||||
|
.createStorageContainerLocationClient();
|
||||||
|
xceiverClientManager = new XceiverClientManager(config);
|
||||||
|
}
|
||||||
|
|
||||||
|
@AfterClass
|
||||||
|
public static void shutdown() throws InterruptedException {
|
||||||
|
if (cluster != null) {
|
||||||
|
cluster.shutdown();
|
||||||
|
}
|
||||||
|
IOUtils.cleanup(null, storageContainerLocationClient, cluster);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* getContainerPipelines creates a set of containers and returns the
|
||||||
|
* Pipelines that define those containers.
|
||||||
|
*
|
||||||
|
* @param count - Number of containers to create.
|
||||||
|
* @return - List of Pipelines.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
private List<Pipeline> getContainerPipeline(int count) throws IOException {
|
||||||
|
List<Pipeline> containerPipelines = new LinkedList<>();
|
||||||
|
for (int x = 0; x < count; x++) {
|
||||||
|
String traceID = "trace" + RandomStringUtils.randomNumeric(4);
|
||||||
|
String containerName = "container" + RandomStringUtils.randomNumeric(10);
|
||||||
|
Pipeline pipeline =
|
||||||
|
storageContainerLocationClient.allocateContainer(containerName);
|
||||||
|
XceiverClientSpi client = xceiverClientManager.acquireClient(pipeline);
|
||||||
|
ContainerProtocolCalls.createContainer(client, traceID);
|
||||||
|
// This step is needed since we set private data on pipelines, when we
|
||||||
|
// read the list from CBlockServer. So we mimic that action here.
|
||||||
|
pipeline.setData(Longs.toByteArray(x));
|
||||||
|
containerPipelines.add(pipeline);
|
||||||
|
}
|
||||||
|
return containerPipelines;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testStorageImplBasicReadWrite() throws Exception {
|
||||||
|
OzoneConfiguration oConfig = new OzoneConfiguration();
|
||||||
|
String userName = "user" + RandomStringUtils.randomNumeric(5);
|
||||||
|
String volumeName = "volume" + RandomStringUtils.randomNumeric(5);
|
||||||
|
long volumeSize = 50L * (1024L * 1024L * 1024L);
|
||||||
|
int blockSize = 4096;
|
||||||
|
byte[] data =
|
||||||
|
RandomStringUtils.randomAlphanumeric(10 * (1024 * 1024))
|
||||||
|
.getBytes(StandardCharsets.UTF_8);
|
||||||
|
String hash = DigestUtils.sha256Hex(data);
|
||||||
|
CBlockTargetMetrics metrics = CBlockTargetMetrics.create();
|
||||||
|
ContainerCacheFlusher flusher = new ContainerCacheFlusher(oConfig,
|
||||||
|
xceiverClientManager, metrics);
|
||||||
|
CBlockIStorageImpl ozoneStore = CBlockIStorageImpl.newBuilder()
|
||||||
|
.setUserName(userName)
|
||||||
|
.setVolumeName(volumeName)
|
||||||
|
.setVolumeSize(volumeSize)
|
||||||
|
.setBlockSize(blockSize)
|
||||||
|
.setContainerList(getContainerPipeline(10))
|
||||||
|
.setClientManager(xceiverClientManager)
|
||||||
|
.setConf(oConfig)
|
||||||
|
.setFlusher(flusher)
|
||||||
|
.setCBlockTargetMetrics(metrics)
|
||||||
|
.build();
|
||||||
|
ozoneStore.write(data, 0);
|
||||||
|
|
||||||
|
// Currently, local cache is a placeholder and does not actually handle
|
||||||
|
// read and write. So the below write is guaranteed to fail. After
|
||||||
|
// CBlockLocalCache is properly implemented, we should uncomment the
|
||||||
|
// following lines
|
||||||
|
// TODO uncomment the following.
|
||||||
|
|
||||||
|
//byte[] newData = new byte[10 * 1024 * 1024];
|
||||||
|
//ozoneStore.read(newData, 0);
|
||||||
|
//String newHash = DigestUtils.sha256Hex(newData);
|
||||||
|
//Assert.assertEquals("hashes don't match.", hash, newHash);
|
||||||
|
GenericTestUtils.waitFor(() -> !ozoneStore.getCache().isDirtyCache(),
|
||||||
|
100, 20 *
|
||||||
|
1000);
|
||||||
|
|
||||||
|
ozoneStore.close();
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue