HDFS-11299. Support multiple Datanode File IO hooks. Contributed by Hanisha Koneru.

This commit is contained in:
Arpit Agarwal 2017-01-10 10:43:02 -08:00
parent c18590fce2
commit 4046794a53
11 changed files with 284 additions and 465 deletions

View File

@ -326,7 +326,7 @@ Each metrics record contains tags such as SessionId and Hostname as additional i
FsVolume FsVolume
-------- --------
Per-volume metrics contain Datanode Volume IO related statistics. Per-volume metrics are off by default. They can be enbabled by setting `dfs.datanode.fileio.events.class` to **org.apache.hadoop.hdfs.server.datanode.ProfilingFileIoEvents**, but enabling per-volume metrics may have a performance impact. Each metrics record contains tags such as Hostname as additional information along with metrics. Per-volume metrics contain Datanode Volume IO related statistics. Per-volume metrics are off by default. They can be enbabled by setting `dfs.datanode.enable.fileio.profiling` to **true**, but enabling per-volume metrics may have a performance impact. Each metrics record contains tags such as Hostname as additional information along with metrics.
| Name | Description | | Name | Description |
|:---- |:---- | |:---- |:---- |

View File

@ -698,8 +698,14 @@ public class DFSConfigKeys extends CommonConfigurationKeys {
public static final String DFS_DATANODE_PLUGINS_KEY = "dfs.datanode.plugins"; public static final String DFS_DATANODE_PLUGINS_KEY = "dfs.datanode.plugins";
public static final String DFS_DATANODE_FSDATASET_FACTORY_KEY = "dfs.datanode.fsdataset.factory"; public static final String DFS_DATANODE_FSDATASET_FACTORY_KEY = "dfs.datanode.fsdataset.factory";
public static final String DFS_DATANODE_FSDATASET_VOLUME_CHOOSING_POLICY_KEY = "dfs.datanode.fsdataset.volume.choosing.policy"; public static final String DFS_DATANODE_FSDATASET_VOLUME_CHOOSING_POLICY_KEY = "dfs.datanode.fsdataset.volume.choosing.policy";
public static final String DFS_DATANODE_FILE_IO_EVENTS_CLASS_KEY = public static final String DFS_DATANODE_ENABLE_FILEIO_PROFILING_KEY =
"dfs.datanode.fileio.events.class"; "dfs.datanode.enable.fileio.profiling";
public static final boolean DFS_DATANODE_ENABLE_FILEIO_PROFILING_DEFAULT =
false;
public static final String DFS_DATANODE_ENABLE_FILEIO_FAULT_INJECTION_KEY =
"dfs.datanode.enable.fileio.fault.injection";
public static final boolean
DFS_DATANODE_ENABLE_FILEIO_FAULT_INJECTION_DEFAULT = false;
public static final String DFS_DATANODE_AVAILABLE_SPACE_VOLUME_CHOOSING_POLICY_BALANCED_SPACE_THRESHOLD_KEY = "dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold"; public static final String DFS_DATANODE_AVAILABLE_SPACE_VOLUME_CHOOSING_POLICY_BALANCED_SPACE_THRESHOLD_KEY = "dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold";
public static final long DFS_DATANODE_AVAILABLE_SPACE_VOLUME_CHOOSING_POLICY_BALANCED_SPACE_THRESHOLD_DEFAULT = 1024L * 1024L * 1024L * 10L; // 10 GB public static final long DFS_DATANODE_AVAILABLE_SPACE_VOLUME_CHOOSING_POLICY_BALANCED_SPACE_THRESHOLD_DEFAULT = 1024L * 1024L * 1024L * 10L; // 10 GB
public static final String DFS_DATANODE_AVAILABLE_SPACE_VOLUME_CHOOSING_POLICY_BALANCED_SPACE_PREFERENCE_FRACTION_KEY = "dfs.datanode.available-space-volume-choosing-policy.balanced-space-preference-fraction"; public static final String DFS_DATANODE_AVAILABLE_SPACE_VOLUME_CHOOSING_POLICY_BALANCED_SPACE_PREFERENCE_FRACTION_KEY = "dfs.datanode.available-space-volume-choosing-policy.balanced-space-preference-fraction";

View File

@ -1,106 +0,0 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
* <p>
* http://www.apache.org/licenses/LICENSE-2.0
* <p>
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hdfs.server.datanode;
import com.fasterxml.jackson.annotation.JsonProperty;
import com.fasterxml.jackson.core.JsonProcessingException;
import com.fasterxml.jackson.databind.ObjectMapper;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.hdfs.server.datanode.FileIoProvider.OPERATION;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi;
import javax.annotation.Nullable;
import java.util.HashMap;
import java.util.Map;
import java.util.concurrent.atomic.AtomicLong;
/**
* {@link FileIoEvents} that simply counts the number of operations.
* Not meant to be used outside of testing.
*/
@InterfaceAudience.Private
@InterfaceStability.Unstable
public class CountingFileIoEvents extends FileIoEvents {
private final Map<OPERATION, Counts> counts;
private static class Counts {
private final AtomicLong successes = new AtomicLong(0);
private final AtomicLong failures = new AtomicLong(0);
@JsonProperty("Successes")
public long getSuccesses() {
return successes.get();
}
@JsonProperty("Failures")
public long getFailures() {
return failures.get();
}
}
public CountingFileIoEvents() {
counts = new HashMap<>();
for (OPERATION op : OPERATION.values()) {
counts.put(op, new Counts());
}
}
@Override
public long beforeMetadataOp(
@Nullable FsVolumeSpi volume, OPERATION op) {
return 0;
}
@Override
public void afterMetadataOp(
@Nullable FsVolumeSpi volume, OPERATION op, long begin) {
counts.get(op).successes.incrementAndGet();
}
@Override
public long beforeFileIo(
@Nullable FsVolumeSpi volume, OPERATION op, long len) {
return 0;
}
@Override
public void afterFileIo(
@Nullable FsVolumeSpi volume, OPERATION op, long begin, long len) {
counts.get(op).successes.incrementAndGet();
}
@Override
public void onFailure(
@Nullable FsVolumeSpi volume, OPERATION op, Exception e, long begin) {
counts.get(op).failures.incrementAndGet();
}
@Override
public String getStatistics() {
ObjectMapper objectMapper = new ObjectMapper();
try {
return objectMapper.writeValueAsString(counts);
} catch (JsonProcessingException e) {
// Failed to serialize. Don't log the exception call stack.
FileIoProvider.LOG.error("Failed to serialize statistics" + e);
return null;
}
}
}

View File

@ -3024,11 +3024,6 @@ public class DataNode extends ReconfigurableBase
} }
} }
@Override // DataNodeMXBean
public String getFileIoProviderStatistics() {
return fileIoProvider.getStatistics();
}
public void refreshNamenodes(Configuration conf) throws IOException { public void refreshNamenodes(Configuration conf) throws IOException {
blockPoolManager.refreshNamenodes(conf); blockPoolManager.refreshNamenodes(conf);
} }

View File

@ -121,11 +121,6 @@ public interface DataNodeMXBean {
*/ */
String getDiskBalancerStatus(); String getDiskBalancerStatus();
/**
* Gets the {@link FileIoProvider} statistics.
*/
String getFileIoProviderStatistics();
/** /**
* Gets the average info (e.g. time) of SendPacketDownstream when the DataNode * Gets the average info (e.g. time) of SendPacketDownstream when the DataNode
* acts as the penultimate (2nd to the last) node in pipeline. * acts as the penultimate (2nd to the last) node in pipeline.

View File

@ -20,48 +20,36 @@ package org.apache.hadoop.hdfs.server.datanode;
import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hdfs.server.datanode.FileIoProvider.OPERATION; import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi;
import javax.annotation.Nullable; import javax.annotation.Nullable;
/** /**
* The default implementation of {@link FileIoEvents} that do nothing. * Injects faults in the metadata and data related operations on datanode
* volumes.
*/ */
@InterfaceAudience.Private @InterfaceAudience.Private
@InterfaceStability.Unstable public class FaultInjectorFileIoEvents {
public final class DefaultFileIoEvents extends FileIoEvents {
@Override private final boolean isEnabled;
public long beforeMetadataOp(
@Nullable FsVolumeSpi volume, OPERATION op) { public FaultInjectorFileIoEvents(@Nullable Configuration conf) {
return 0; if (conf != null) {
isEnabled = conf.getBoolean(DFSConfigKeys
.DFS_DATANODE_ENABLE_FILEIO_FAULT_INJECTION_KEY, DFSConfigKeys
.DFS_DATANODE_ENABLE_FILEIO_FAULT_INJECTION_DEFAULT);
} else {
isEnabled = false;
}
} }
@Override public void beforeMetadataOp(
public void afterMetadataOp( @Nullable FsVolumeSpi volume, FileIoProvider.OPERATION op) {
@Nullable FsVolumeSpi volume, OPERATION op, long begin) {
} }
@Override public void beforeFileIo(
public long beforeFileIo( @Nullable FsVolumeSpi volume, FileIoProvider.OPERATION op, long len) {
@Nullable FsVolumeSpi volume, OPERATION op, long len) {
return 0;
}
@Override
public void afterFileIo(
@Nullable FsVolumeSpi volume, OPERATION op, long begin, long len) {
}
@Override
public void onFailure(
@Nullable FsVolumeSpi volume, OPERATION op, Exception e, long begin) {
}
@Override
public @Nullable String getStatistics() {
// null is valid JSON.
return null;
} }
} }

View File

@ -1,115 +0,0 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
* <p>
* http://www.apache.org/licenses/LICENSE-2.0
* <p>
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hdfs.server.datanode;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.hdfs.server.datanode.FileIoProvider.OPERATION;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi;
import javax.annotation.Nullable;
/**
* The following hooks can be implemented for instrumentation/fault
* injection.
*/
@InterfaceAudience.Private
@InterfaceStability.Unstable
public abstract class FileIoEvents {
/**
* Invoked before a filesystem metadata operation.
*
* @param volume target volume for the operation. Null if unavailable.
* @param op type of operation.
* @return timestamp at which the operation was started. 0 if
* unavailable.
*/
abstract long beforeMetadataOp(@Nullable FsVolumeSpi volume, OPERATION op);
/**
* Invoked after a filesystem metadata operation has completed.
*
* @param volume target volume for the operation. Null if unavailable.
* @param op type of operation.
* @param begin timestamp at which the operation was started. 0
* if unavailable.
*/
abstract void afterMetadataOp(@Nullable FsVolumeSpi volume, OPERATION op,
long begin);
/**
* Invoked before a read/write/flush/channel transfer operation.
*
* @param volume target volume for the operation. Null if unavailable.
* @param op type of operation.
* @param len length of the file IO. 0 for flush.
* @return timestamp at which the operation was started. 0 if
* unavailable.
*/
abstract long beforeFileIo(@Nullable FsVolumeSpi volume, OPERATION op,
long len);
/**
* Invoked after a read/write/flush/channel transfer operation
* has completed.
*
* @param volume target volume for the operation. Null if unavailable.
* @param op type of operation.
* @param len of the file IO. 0 for flush.
* @return timestamp at which the operation was started. 0 if
* unavailable.
*/
abstract void afterFileIo(@Nullable FsVolumeSpi volume, OPERATION op,
long begin, long len);
/**
* Invoked if an operation fails with an exception.
* @param volume target volume for the operation. Null if unavailable.
* @param op type of operation.
* @param e Exception encountered during the operation.
* @param begin time at which the operation was started.
*/
abstract void onFailure(
@Nullable FsVolumeSpi volume, OPERATION op, Exception e, long begin);
/**
* Invoked by FileIoProvider if an operation fails with an exception.
* @param datanode datanode that runs volume check upon volume io failure
* @param volume target volume for the operation. Null if unavailable.
* @param op type of operation.
* @param e Exception encountered during the operation.
* @param begin time at which the operation was started.
*/
void onFailure(DataNode datanode,
@Nullable FsVolumeSpi volume, OPERATION op, Exception e, long begin) {
onFailure(volume, op, e, begin);
if (datanode != null && volume != null) {
datanode.checkDiskErrorAsync(volume);
}
}
/**
* Return statistics as a JSON string.
* @return
*/
@Nullable abstract String getStatistics();
}

View File

@ -18,7 +18,6 @@
package org.apache.hadoop.hdfs.server.datanode; package org.apache.hadoop.hdfs.server.datanode;
import org.apache.commons.io.FileUtils; import org.apache.commons.io.FileUtils;
import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.classification.InterfaceStability;
@ -34,7 +33,6 @@ import org.apache.hadoop.io.LongWritable;
import org.apache.hadoop.io.nativeio.NativeIO; import org.apache.hadoop.io.nativeio.NativeIO;
import org.apache.hadoop.io.nativeio.NativeIOException; import org.apache.hadoop.io.nativeio.NativeIOException;
import org.apache.hadoop.net.SocketOutputStream; import org.apache.hadoop.net.SocketOutputStream;
import org.apache.hadoop.util.ReflectionUtils;
import org.slf4j.Logger; import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;
@ -59,11 +57,14 @@ import static org.apache.hadoop.hdfs.server.datanode.FileIoProvider.OPERATION.*;
/** /**
* This class abstracts out various file IO operations performed by the * This class abstracts out various file IO operations performed by the
* DataNode and invokes event hooks before and after each file IO. * DataNode and invokes profiling (for collecting stats) and fault injection
* (for testing) event hooks before and after each file IO.
* *
* Behavior can be injected into these events by implementing * Behavior can be injected into these events by enabling the
* {@link FileIoEvents} and replacing the default implementation * profiling and/or fault injection event hooks through
* with {@link DFSConfigKeys#DFS_DATANODE_FILE_IO_EVENTS_CLASS_KEY}. * {@link DFSConfigKeys#DFS_DATANODE_ENABLE_FILEIO_PROFILING_KEY} and
* {@link DFSConfigKeys#DFS_DATANODE_ENABLE_FILEIO_FAULT_INJECTION_KEY}.
* These event hooks are disabled by default.
* *
* Most functions accept an optional {@link FsVolumeSpi} parameter for * Most functions accept an optional {@link FsVolumeSpi} parameter for
* instrumentation/logging. * instrumentation/logging.
@ -78,9 +79,12 @@ public class FileIoProvider {
public static final Logger LOG = LoggerFactory.getLogger( public static final Logger LOG = LoggerFactory.getLogger(
FileIoProvider.class); FileIoProvider.class);
private final FileIoEvents eventHooks; private final ProfilingFileIoEvents profilingEventHook;
private final FaultInjectorFileIoEvents faultInjectorEventHook;
private final DataNode datanode; private final DataNode datanode;
private static final int LEN_INT = 4;
/** /**
* @param conf Configuration object. May be null. When null, * @param conf Configuration object. May be null. When null,
* the event handlers are no-ops. * the event handlers are no-ops.
@ -89,15 +93,8 @@ public class FileIoProvider {
*/ */
public FileIoProvider(@Nullable Configuration conf, public FileIoProvider(@Nullable Configuration conf,
final DataNode datanode) { final DataNode datanode) {
if (conf != null) { profilingEventHook = new ProfilingFileIoEvents(conf);
final Class<? extends FileIoEvents> clazz = conf.getClass( faultInjectorEventHook = new FaultInjectorFileIoEvents(conf);
DFSConfigKeys.DFS_DATANODE_FILE_IO_EVENTS_CLASS_KEY,
DefaultFileIoEvents.class,
FileIoEvents.class);
eventHooks = ReflectionUtils.newInstance(clazz, conf);
} else {
eventHooks = new DefaultFileIoEvents();
}
this.datanode = datanode; this.datanode = datanode;
} }
@ -122,15 +119,6 @@ public class FileIoProvider {
NATIVE_COPY NATIVE_COPY
} }
/**
* Retrieve statistics from the underlying {@link FileIoEvents}
* implementation as a JSON string, if it maintains them.
* @return statistics as a JSON string. May be null.
*/
public @Nullable String getStatistics() {
return eventHooks.getStatistics();
}
/** /**
* See {@link Flushable#flush()}. * See {@link Flushable#flush()}.
* *
@ -139,12 +127,13 @@ public class FileIoProvider {
*/ */
public void flush( public void flush(
@Nullable FsVolumeSpi volume, Flushable f) throws IOException { @Nullable FsVolumeSpi volume, Flushable f) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, FLUSH, 0); final long begin = profilingEventHook.beforeFileIo(volume, FLUSH, 0);
try { try {
faultInjectorEventHook.beforeFileIo(volume, FLUSH, 0);
f.flush(); f.flush();
eventHooks.afterFileIo(volume, FLUSH, begin, 0); profilingEventHook.afterFileIo(volume, FLUSH, begin, 0);
} catch (Exception e) { } catch (Exception e) {
eventHooks.onFailure(datanode, volume, FLUSH, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -157,12 +146,13 @@ public class FileIoProvider {
*/ */
public void sync( public void sync(
@Nullable FsVolumeSpi volume, FileOutputStream fos) throws IOException { @Nullable FsVolumeSpi volume, FileOutputStream fos) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, SYNC, 0); final long begin = profilingEventHook.beforeFileIo(volume, SYNC, 0);
try { try {
faultInjectorEventHook.beforeFileIo(volume, SYNC, 0);
fos.getChannel().force(true); fos.getChannel().force(true);
eventHooks.afterFileIo(volume, SYNC, begin, 0); profilingEventHook.afterFileIo(volume, SYNC, begin, 0);
} catch (Exception e) { } catch (Exception e) {
eventHooks.onFailure(datanode, volume, SYNC, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -176,12 +166,13 @@ public class FileIoProvider {
public void syncFileRange( public void syncFileRange(
@Nullable FsVolumeSpi volume, FileDescriptor outFd, @Nullable FsVolumeSpi volume, FileDescriptor outFd,
long offset, long numBytes, int flags) throws NativeIOException { long offset, long numBytes, int flags) throws NativeIOException {
final long begin = eventHooks.beforeFileIo(volume, SYNC, 0); final long begin = profilingEventHook.beforeFileIo(volume, SYNC, 0);
try { try {
faultInjectorEventHook.beforeFileIo(volume, SYNC, 0);
NativeIO.POSIX.syncFileRangeIfPossible(outFd, offset, numBytes, flags); NativeIO.POSIX.syncFileRangeIfPossible(outFd, offset, numBytes, flags);
eventHooks.afterFileIo(volume, SYNC, begin, 0); profilingEventHook.afterFileIo(volume, SYNC, begin, 0);
} catch (Exception e) { } catch (Exception e) {
eventHooks.onFailure(datanode, volume, SYNC, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -195,13 +186,14 @@ public class FileIoProvider {
public void posixFadvise( public void posixFadvise(
@Nullable FsVolumeSpi volume, String identifier, FileDescriptor outFd, @Nullable FsVolumeSpi volume, String identifier, FileDescriptor outFd,
long offset, long length, int flags) throws NativeIOException { long offset, long length, int flags) throws NativeIOException {
final long begin = eventHooks.beforeMetadataOp(volume, FADVISE); final long begin = profilingEventHook.beforeMetadataOp(volume, FADVISE);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, FADVISE);
NativeIO.POSIX.getCacheManipulator().posixFadviseIfPossible( NativeIO.POSIX.getCacheManipulator().posixFadviseIfPossible(
identifier, outFd, offset, length, flags); identifier, outFd, offset, length, flags);
eventHooks.afterMetadataOp(volume, FADVISE, begin); profilingEventHook.afterMetadataOp(volume, FADVISE, begin);
} catch (Exception e) { } catch (Exception e) {
eventHooks.onFailure(datanode, volume, FADVISE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -213,13 +205,14 @@ public class FileIoProvider {
* @return true if the file was successfully deleted. * @return true if the file was successfully deleted.
*/ */
public boolean delete(@Nullable FsVolumeSpi volume, File f) { public boolean delete(@Nullable FsVolumeSpi volume, File f) {
final long begin = eventHooks.beforeMetadataOp(volume, DELETE); final long begin = profilingEventHook.beforeMetadataOp(volume, DELETE);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, DELETE);
boolean deleted = f.delete(); boolean deleted = f.delete();
eventHooks.afterMetadataOp(volume, DELETE, begin); profilingEventHook.afterMetadataOp(volume, DELETE, begin);
return deleted; return deleted;
} catch (Exception e) { } catch (Exception e) {
eventHooks.onFailure(datanode, volume, DELETE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -232,16 +225,17 @@ public class FileIoProvider {
* existed. * existed.
*/ */
public boolean deleteWithExistsCheck(@Nullable FsVolumeSpi volume, File f) { public boolean deleteWithExistsCheck(@Nullable FsVolumeSpi volume, File f) {
final long begin = eventHooks.beforeMetadataOp(volume, DELETE); final long begin = profilingEventHook.beforeMetadataOp(volume, DELETE);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, DELETE);
boolean deleted = !f.exists() || f.delete(); boolean deleted = !f.exists() || f.delete();
eventHooks.afterMetadataOp(volume, DELETE, begin); profilingEventHook.afterMetadataOp(volume, DELETE, begin);
if (!deleted) { if (!deleted) {
LOG.warn("Failed to delete file {}", f); LOG.warn("Failed to delete file {}", f);
} }
return deleted; return deleted;
} catch (Exception e) { } catch (Exception e) {
eventHooks.onFailure(datanode, volume, DELETE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -263,13 +257,14 @@ public class FileIoProvider {
@Nullable FsVolumeSpi volume, SocketOutputStream sockOut, @Nullable FsVolumeSpi volume, SocketOutputStream sockOut,
FileChannel fileCh, long position, int count, FileChannel fileCh, long position, int count,
LongWritable waitTime, LongWritable transferTime) throws IOException { LongWritable waitTime, LongWritable transferTime) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, TRANSFER, count); final long begin = profilingEventHook.beforeFileIo(volume, TRANSFER, count);
try { try {
faultInjectorEventHook.beforeFileIo(volume, TRANSFER, count);
sockOut.transferToFully(fileCh, position, count, sockOut.transferToFully(fileCh, position, count,
waitTime, transferTime); waitTime, transferTime);
eventHooks.afterFileIo(volume, TRANSFER, begin, count); profilingEventHook.afterFileIo(volume, TRANSFER, begin, count);
} catch (Exception e) { } catch (Exception e) {
eventHooks.onFailure(datanode, volume, TRANSFER, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -284,13 +279,14 @@ public class FileIoProvider {
*/ */
public boolean createFile( public boolean createFile(
@Nullable FsVolumeSpi volume, File f) throws IOException { @Nullable FsVolumeSpi volume, File f) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, OPEN); final long begin = profilingEventHook.beforeMetadataOp(volume, OPEN);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, OPEN);
boolean created = f.createNewFile(); boolean created = f.createNewFile();
eventHooks.afterMetadataOp(volume, OPEN, begin); profilingEventHook.afterMetadataOp(volume, OPEN, begin);
return created; return created;
} catch (Exception e) { } catch (Exception e) {
eventHooks.onFailure(datanode, volume, OPEN, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -309,15 +305,16 @@ public class FileIoProvider {
*/ */
public FileInputStream getFileInputStream( public FileInputStream getFileInputStream(
@Nullable FsVolumeSpi volume, File f) throws FileNotFoundException { @Nullable FsVolumeSpi volume, File f) throws FileNotFoundException {
final long begin = eventHooks.beforeMetadataOp(volume, OPEN); final long begin = profilingEventHook.beforeMetadataOp(volume, OPEN);
FileInputStream fis = null; FileInputStream fis = null;
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, OPEN);
fis = new WrappedFileInputStream(volume, f); fis = new WrappedFileInputStream(volume, f);
eventHooks.afterMetadataOp(volume, OPEN, begin); profilingEventHook.afterMetadataOp(volume, OPEN, begin);
return fis; return fis;
} catch(Exception e) { } catch(Exception e) {
org.apache.commons.io.IOUtils.closeQuietly(fis); org.apache.commons.io.IOUtils.closeQuietly(fis);
eventHooks.onFailure(datanode, volume, OPEN, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -339,15 +336,16 @@ public class FileIoProvider {
public FileOutputStream getFileOutputStream( public FileOutputStream getFileOutputStream(
@Nullable FsVolumeSpi volume, File f, @Nullable FsVolumeSpi volume, File f,
boolean append) throws FileNotFoundException { boolean append) throws FileNotFoundException {
final long begin = eventHooks.beforeMetadataOp(volume, OPEN); final long begin = profilingEventHook.beforeMetadataOp(volume, OPEN);
FileOutputStream fos = null; FileOutputStream fos = null;
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, OPEN);
fos = new WrappedFileOutputStream(volume, f, append); fos = new WrappedFileOutputStream(volume, f, append);
eventHooks.afterMetadataOp(volume, OPEN, begin); profilingEventHook.afterMetadataOp(volume, OPEN, begin);
return fos; return fos;
} catch(Exception e) { } catch(Exception e) {
org.apache.commons.io.IOUtils.closeQuietly(fos); org.apache.commons.io.IOUtils.closeQuietly(fos);
eventHooks.onFailure(datanode, volume, OPEN, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -403,16 +401,17 @@ public class FileIoProvider {
public FileInputStream getShareDeleteFileInputStream( public FileInputStream getShareDeleteFileInputStream(
@Nullable FsVolumeSpi volume, File f, @Nullable FsVolumeSpi volume, File f,
long offset) throws IOException { long offset) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, OPEN); final long begin = profilingEventHook.beforeMetadataOp(volume, OPEN);
FileInputStream fis = null; FileInputStream fis = null;
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, OPEN);
fis = new WrappedFileInputStream(volume, fis = new WrappedFileInputStream(volume,
NativeIO.getShareDeleteFileDescriptor(f, offset)); NativeIO.getShareDeleteFileDescriptor(f, offset));
eventHooks.afterMetadataOp(volume, OPEN, begin); profilingEventHook.afterMetadataOp(volume, OPEN, begin);
return fis; return fis;
} catch(Exception e) { } catch(Exception e) {
org.apache.commons.io.IOUtils.closeQuietly(fis); org.apache.commons.io.IOUtils.closeQuietly(fis);
eventHooks.onFailure(datanode, volume, OPEN, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -434,16 +433,17 @@ public class FileIoProvider {
*/ */
public FileInputStream openAndSeek( public FileInputStream openAndSeek(
@Nullable FsVolumeSpi volume, File f, long offset) throws IOException { @Nullable FsVolumeSpi volume, File f, long offset) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, OPEN); final long begin = profilingEventHook.beforeMetadataOp(volume, OPEN);
FileInputStream fis = null; FileInputStream fis = null;
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, OPEN);
fis = new WrappedFileInputStream(volume, fis = new WrappedFileInputStream(volume,
FsDatasetUtil.openAndSeek(f, offset)); FsDatasetUtil.openAndSeek(f, offset));
eventHooks.afterMetadataOp(volume, OPEN, begin); profilingEventHook.afterMetadataOp(volume, OPEN, begin);
return fis; return fis;
} catch(Exception e) { } catch(Exception e) {
org.apache.commons.io.IOUtils.closeQuietly(fis); org.apache.commons.io.IOUtils.closeQuietly(fis);
eventHooks.onFailure(datanode, volume, OPEN, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -465,15 +465,16 @@ public class FileIoProvider {
public RandomAccessFile getRandomAccessFile( public RandomAccessFile getRandomAccessFile(
@Nullable FsVolumeSpi volume, File f, @Nullable FsVolumeSpi volume, File f,
String mode) throws FileNotFoundException { String mode) throws FileNotFoundException {
final long begin = eventHooks.beforeMetadataOp(volume, OPEN); final long begin = profilingEventHook.beforeMetadataOp(volume, OPEN);
RandomAccessFile raf = null; RandomAccessFile raf = null;
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, OPEN);
raf = new WrappedRandomAccessFile(volume, f, mode); raf = new WrappedRandomAccessFile(volume, f, mode);
eventHooks.afterMetadataOp(volume, OPEN, begin); profilingEventHook.afterMetadataOp(volume, OPEN, begin);
return raf; return raf;
} catch(Exception e) { } catch(Exception e) {
org.apache.commons.io.IOUtils.closeQuietly(raf); org.apache.commons.io.IOUtils.closeQuietly(raf);
eventHooks.onFailure(datanode, volume, OPEN, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -486,13 +487,14 @@ public class FileIoProvider {
* @return true on success false on failure. * @return true on success false on failure.
*/ */
public boolean fullyDelete(@Nullable FsVolumeSpi volume, File dir) { public boolean fullyDelete(@Nullable FsVolumeSpi volume, File dir) {
final long begin = eventHooks.beforeMetadataOp(volume, DELETE); final long begin = profilingEventHook.beforeMetadataOp(volume, DELETE);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, DELETE);
boolean deleted = FileUtil.fullyDelete(dir); boolean deleted = FileUtil.fullyDelete(dir);
eventHooks.afterMetadataOp(volume, DELETE, begin); profilingEventHook.afterMetadataOp(volume, DELETE, begin);
return deleted; return deleted;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, DELETE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -508,12 +510,13 @@ public class FileIoProvider {
*/ */
public void replaceFile( public void replaceFile(
@Nullable FsVolumeSpi volume, File src, File target) throws IOException { @Nullable FsVolumeSpi volume, File src, File target) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, MOVE); final long begin = profilingEventHook.beforeMetadataOp(volume, MOVE);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, MOVE);
FileUtil.replaceFile(src, target); FileUtil.replaceFile(src, target);
eventHooks.afterMetadataOp(volume, MOVE, begin); profilingEventHook.afterMetadataOp(volume, MOVE, begin);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, MOVE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -530,12 +533,13 @@ public class FileIoProvider {
public void rename( public void rename(
@Nullable FsVolumeSpi volume, File src, File target) @Nullable FsVolumeSpi volume, File src, File target)
throws IOException { throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, MOVE); final long begin = profilingEventHook.beforeMetadataOp(volume, MOVE);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, MOVE);
Storage.rename(src, target); Storage.rename(src, target);
eventHooks.afterMetadataOp(volume, MOVE, begin); profilingEventHook.afterMetadataOp(volume, MOVE, begin);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, MOVE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -552,12 +556,13 @@ public class FileIoProvider {
public void moveFile( public void moveFile(
@Nullable FsVolumeSpi volume, File src, File target) @Nullable FsVolumeSpi volume, File src, File target)
throws IOException { throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, MOVE); final long begin = profilingEventHook.beforeMetadataOp(volume, MOVE);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, MOVE);
FileUtils.moveFile(src, target); FileUtils.moveFile(src, target);
eventHooks.afterMetadataOp(volume, MOVE, begin); profilingEventHook.afterMetadataOp(volume, MOVE, begin);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, MOVE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -576,12 +581,13 @@ public class FileIoProvider {
public void move( public void move(
@Nullable FsVolumeSpi volume, Path src, Path target, @Nullable FsVolumeSpi volume, Path src, Path target,
CopyOption... options) throws IOException { CopyOption... options) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, MOVE); final long begin = profilingEventHook.beforeMetadataOp(volume, MOVE);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, MOVE);
Files.move(src, target, options); Files.move(src, target, options);
eventHooks.afterMetadataOp(volume, MOVE, begin); profilingEventHook.afterMetadataOp(volume, MOVE, begin);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, MOVE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -600,12 +606,14 @@ public class FileIoProvider {
@Nullable FsVolumeSpi volume, File src, File target, @Nullable FsVolumeSpi volume, File src, File target,
boolean preserveFileDate) throws IOException { boolean preserveFileDate) throws IOException {
final long length = src.length(); final long length = src.length();
final long begin = eventHooks.beforeFileIo(volume, NATIVE_COPY, length); final long begin = profilingEventHook.beforeFileIo(volume, NATIVE_COPY,
length);
try { try {
faultInjectorEventHook.beforeFileIo(volume, NATIVE_COPY, length);
Storage.nativeCopyFileUnbuffered(src, target, preserveFileDate); Storage.nativeCopyFileUnbuffered(src, target, preserveFileDate);
eventHooks.afterFileIo(volume, NATIVE_COPY, begin, length); profilingEventHook.afterFileIo(volume, NATIVE_COPY, begin, length);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, NATIVE_COPY, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -622,15 +630,16 @@ public class FileIoProvider {
*/ */
public boolean mkdirs( public boolean mkdirs(
@Nullable FsVolumeSpi volume, File dir) throws IOException { @Nullable FsVolumeSpi volume, File dir) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, MKDIRS); final long begin = profilingEventHook.beforeMetadataOp(volume, MKDIRS);
boolean created = false; boolean created = false;
boolean isDirectory; boolean isDirectory;
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, MKDIRS);
created = dir.mkdirs(); created = dir.mkdirs();
isDirectory = !created && dir.isDirectory(); isDirectory = !created && dir.isDirectory();
eventHooks.afterMetadataOp(volume, MKDIRS, begin); profilingEventHook.afterMetadataOp(volume, MKDIRS, begin);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, MKDIRS, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
@ -650,13 +659,14 @@ public class FileIoProvider {
*/ */
public void mkdirsWithExistsCheck( public void mkdirsWithExistsCheck(
@Nullable FsVolumeSpi volume, File dir) throws IOException { @Nullable FsVolumeSpi volume, File dir) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, MKDIRS); final long begin = profilingEventHook.beforeMetadataOp(volume, MKDIRS);
boolean succeeded = false; boolean succeeded = false;
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, MKDIRS);
succeeded = dir.isDirectory() || dir.mkdirs(); succeeded = dir.isDirectory() || dir.mkdirs();
eventHooks.afterMetadataOp(volume, MKDIRS, begin); profilingEventHook.afterMetadataOp(volume, MKDIRS, begin);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, MKDIRS, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
@ -676,13 +686,14 @@ public class FileIoProvider {
*/ */
public File[] listFiles( public File[] listFiles(
@Nullable FsVolumeSpi volume, File dir) throws IOException { @Nullable FsVolumeSpi volume, File dir) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, LIST); final long begin = profilingEventHook.beforeMetadataOp(volume, LIST);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, LIST);
File[] children = FileUtil.listFiles(dir); File[] children = FileUtil.listFiles(dir);
eventHooks.afterMetadataOp(volume, LIST, begin); profilingEventHook.afterMetadataOp(volume, LIST, begin);
return children; return children;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, LIST, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -698,13 +709,14 @@ public class FileIoProvider {
*/ */
public String[] list( public String[] list(
@Nullable FsVolumeSpi volume, File dir) throws IOException { @Nullable FsVolumeSpi volume, File dir) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, LIST); final long begin = profilingEventHook.beforeMetadataOp(volume, LIST);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, LIST);
String[] children = FileUtil.list(dir); String[] children = FileUtil.list(dir);
eventHooks.afterMetadataOp(volume, LIST, begin); profilingEventHook.afterMetadataOp(volume, LIST, begin);
return children; return children;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, LIST, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -721,13 +733,14 @@ public class FileIoProvider {
public List<String> listDirectory( public List<String> listDirectory(
@Nullable FsVolumeSpi volume, File dir, @Nullable FsVolumeSpi volume, File dir,
FilenameFilter filter) throws IOException { FilenameFilter filter) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, LIST); final long begin = profilingEventHook.beforeMetadataOp(volume, LIST);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, LIST);
List<String> children = IOUtils.listDirectory(dir, filter); List<String> children = IOUtils.listDirectory(dir, filter);
eventHooks.afterMetadataOp(volume, LIST, begin); profilingEventHook.afterMetadataOp(volume, LIST, begin);
return children; return children;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, LIST, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -743,13 +756,14 @@ public class FileIoProvider {
*/ */
public int getHardLinkCount( public int getHardLinkCount(
@Nullable FsVolumeSpi volume, File f) throws IOException { @Nullable FsVolumeSpi volume, File f) throws IOException {
final long begin = eventHooks.beforeMetadataOp(volume, LIST); final long begin = profilingEventHook.beforeMetadataOp(volume, LIST);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, LIST);
int count = HardLink.getLinkCount(f); int count = HardLink.getLinkCount(f);
eventHooks.afterMetadataOp(volume, LIST, begin); profilingEventHook.afterMetadataOp(volume, LIST, begin);
return count; return count;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, LIST, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -762,13 +776,14 @@ public class FileIoProvider {
* @return true if the file exists. * @return true if the file exists.
*/ */
public boolean exists(@Nullable FsVolumeSpi volume, File f) { public boolean exists(@Nullable FsVolumeSpi volume, File f) {
final long begin = eventHooks.beforeMetadataOp(volume, EXISTS); final long begin = profilingEventHook.beforeMetadataOp(volume, EXISTS);
try { try {
faultInjectorEventHook.beforeMetadataOp(volume, EXISTS);
boolean exists = f.exists(); boolean exists = f.exists();
eventHooks.afterMetadataOp(volume, EXISTS, begin); profilingEventHook.afterMetadataOp(volume, EXISTS, begin);
return exists; return exists;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, EXISTS, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -803,13 +818,14 @@ public class FileIoProvider {
*/ */
@Override @Override
public int read() throws IOException { public int read() throws IOException {
final long begin = eventHooks.beforeFileIo(volume, READ, 1); final long begin = profilingEventHook.beforeFileIo(volume, READ, LEN_INT);
try { try {
faultInjectorEventHook.beforeFileIo(volume, READ, LEN_INT);
int b = super.read(); int b = super.read();
eventHooks.afterFileIo(volume, READ, begin, 1); profilingEventHook.afterFileIo(volume, READ, begin, LEN_INT);
return b; return b;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, READ, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -819,13 +835,15 @@ public class FileIoProvider {
*/ */
@Override @Override
public int read(@Nonnull byte[] b) throws IOException { public int read(@Nonnull byte[] b) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, READ, b.length); final long begin = profilingEventHook.beforeFileIo(volume, READ, b
.length);
try { try {
faultInjectorEventHook.beforeFileIo(volume, READ, b.length);
int numBytesRead = super.read(b); int numBytesRead = super.read(b);
eventHooks.afterFileIo(volume, READ, begin, numBytesRead); profilingEventHook.afterFileIo(volume, READ, begin, numBytesRead);
return numBytesRead; return numBytesRead;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, READ, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -835,13 +853,14 @@ public class FileIoProvider {
*/ */
@Override @Override
public int read(@Nonnull byte[] b, int off, int len) throws IOException { public int read(@Nonnull byte[] b, int off, int len) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, READ, len); final long begin = profilingEventHook.beforeFileIo(volume, READ, len);
try { try {
faultInjectorEventHook.beforeFileIo(volume, READ, len);
int numBytesRead = super.read(b, off, len); int numBytesRead = super.read(b, off, len);
eventHooks.afterFileIo(volume, READ, begin, numBytesRead); profilingEventHook.afterFileIo(volume, READ, begin, numBytesRead);
return numBytesRead; return numBytesRead;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, READ, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -878,12 +897,14 @@ public class FileIoProvider {
*/ */
@Override @Override
public void write(int b) throws IOException { public void write(int b) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, WRITE, 1); final long begin = profilingEventHook.beforeFileIo(volume, WRITE,
LEN_INT);
try { try {
faultInjectorEventHook.beforeFileIo(volume, WRITE, LEN_INT);
super.write(b); super.write(b);
eventHooks.afterFileIo(volume, WRITE, begin, 1); profilingEventHook.afterFileIo(volume, WRITE, begin, LEN_INT);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, WRITE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -893,12 +914,14 @@ public class FileIoProvider {
*/ */
@Override @Override
public void write(@Nonnull byte[] b) throws IOException { public void write(@Nonnull byte[] b) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, WRITE, b.length); final long begin = profilingEventHook.beforeFileIo(volume, WRITE, b
.length);
try { try {
faultInjectorEventHook.beforeFileIo(volume, WRITE, b.length);
super.write(b); super.write(b);
eventHooks.afterFileIo(volume, WRITE, begin, b.length); profilingEventHook.afterFileIo(volume, WRITE, begin, b.length);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, WRITE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -908,12 +931,13 @@ public class FileIoProvider {
*/ */
@Override @Override
public void write(@Nonnull byte[] b, int off, int len) throws IOException { public void write(@Nonnull byte[] b, int off, int len) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, WRITE, len); final long begin = profilingEventHook.beforeFileIo(volume, WRITE, len);
try { try {
faultInjectorEventHook.beforeFileIo(volume, WRITE, len);
super.write(b, off, len); super.write(b, off, len);
eventHooks.afterFileIo(volume, WRITE, begin, len); profilingEventHook.afterFileIo(volume, WRITE, begin, len);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, WRITE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@ -935,77 +959,93 @@ public class FileIoProvider {
@Override @Override
public int read() throws IOException { public int read() throws IOException {
final long begin = eventHooks.beforeFileIo(volume, READ, 1); final long begin = profilingEventHook.beforeFileIo(volume, READ, LEN_INT);
try { try {
faultInjectorEventHook.beforeFileIo(volume, READ, LEN_INT);
int b = super.read(); int b = super.read();
eventHooks.afterFileIo(volume, READ, begin, 1); profilingEventHook.afterFileIo(volume, READ, begin, LEN_INT);
return b; return b;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, READ, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@Override @Override
public int read(byte[] b, int off, int len) throws IOException { public int read(byte[] b, int off, int len) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, READ, len); final long begin = profilingEventHook.beforeFileIo(volume, READ, len);
try { try {
faultInjectorEventHook.beforeFileIo(volume, READ, len);
int numBytesRead = super.read(b, off, len); int numBytesRead = super.read(b, off, len);
eventHooks.afterFileIo(volume, READ, begin, numBytesRead); profilingEventHook.afterFileIo(volume, READ, begin, numBytesRead);
return numBytesRead; return numBytesRead;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, READ, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@Override @Override
public int read(byte[] b) throws IOException { public int read(byte[] b) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, READ, b.length); final long begin = profilingEventHook.beforeFileIo(volume, READ, b
.length);
try { try {
faultInjectorEventHook.beforeFileIo(volume, READ, b.length);
int numBytesRead = super.read(b); int numBytesRead = super.read(b);
eventHooks.afterFileIo(volume, READ, begin, numBytesRead); profilingEventHook.afterFileIo(volume, READ, begin, numBytesRead);
return numBytesRead; return numBytesRead;
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, READ, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@Override @Override
public void write(int b) throws IOException { public void write(int b) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, WRITE, 1); final long begin = profilingEventHook.beforeFileIo(volume, WRITE,
LEN_INT);
try { try {
faultInjectorEventHook.beforeFileIo(volume, WRITE, LEN_INT);
super.write(b); super.write(b);
eventHooks.afterFileIo(volume, WRITE, begin, 1); profilingEventHook.afterFileIo(volume, WRITE, begin, LEN_INT);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, WRITE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@Override @Override
public void write(@Nonnull byte[] b) throws IOException { public void write(@Nonnull byte[] b) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, WRITE, b.length); final long begin = profilingEventHook.beforeFileIo(volume, WRITE, b
.length);
try { try {
faultInjectorEventHook.beforeFileIo(volume, WRITE, b.length);
super.write(b); super.write(b);
eventHooks.afterFileIo(volume, WRITE, begin, b.length); profilingEventHook.afterFileIo(volume, WRITE, begin, b.length);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, WRITE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
@Override @Override
public void write(byte[] b, int off, int len) throws IOException { public void write(byte[] b, int off, int len) throws IOException {
final long begin = eventHooks.beforeFileIo(volume, WRITE, len); final long begin = profilingEventHook.beforeFileIo(volume, WRITE, len);
try { try {
faultInjectorEventHook.beforeFileIo(volume, WRITE, len);
super.write(b, off, len); super.write(b, off, len);
eventHooks.afterFileIo(volume, WRITE, begin, len); profilingEventHook.afterFileIo(volume, WRITE, begin, len);
} catch(Exception e) { } catch(Exception e) {
eventHooks.onFailure(datanode, volume, WRITE, e, begin); onFailure(volume, begin);
throw e; throw e;
} }
} }
} }
private void onFailure(@Nullable FsVolumeSpi volume, long begin) {
if (datanode != null && volume != null) {
datanode.checkDiskErrorAsync(volume);
}
profilingEventHook.onFailure(volume, begin);
}
} }

View File

@ -19,6 +19,8 @@
package org.apache.hadoop.hdfs.server.datanode; package org.apache.hadoop.hdfs.server.datanode;
import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.DataNodeVolumeMetrics; import org.apache.hadoop.hdfs.server.datanode.fsdataset.DataNodeVolumeMetrics;
import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi;
import org.apache.hadoop.util.Time; import org.apache.hadoop.util.Time;
@ -26,84 +28,96 @@ import org.apache.hadoop.util.Time;
import javax.annotation.Nullable; import javax.annotation.Nullable;
/** /**
* {@link FileIoEvents} that profiles the performance of the metadata and data * Profiles the performance of the metadata and data related operations on
* related operations on datanode volumes. * datanode volumes.
*/ */
@InterfaceAudience.Private @InterfaceAudience.Private
class ProfilingFileIoEvents extends FileIoEvents { class ProfilingFileIoEvents {
private final boolean isEnabled;
public ProfilingFileIoEvents(@Nullable Configuration conf) {
if (conf != null) {
isEnabled = conf.getBoolean(DFSConfigKeys
.DFS_DATANODE_ENABLE_FILEIO_PROFILING_KEY, DFSConfigKeys
.DFS_DATANODE_ENABLE_FILEIO_PROFILING_DEFAULT);
} else {
isEnabled = false;
}
}
@Override
public long beforeMetadataOp(@Nullable FsVolumeSpi volume, public long beforeMetadataOp(@Nullable FsVolumeSpi volume,
FileIoProvider.OPERATION op) { FileIoProvider.OPERATION op) {
DataNodeVolumeMetrics metrics = getVolumeMetrics(volume); if (isEnabled) {
if (metrics != null) { DataNodeVolumeMetrics metrics = getVolumeMetrics(volume);
return Time.monotonicNow(); if (metrics != null) {
return Time.monotonicNow();
}
} }
return 0; return 0;
} }
@Override
public void afterMetadataOp(@Nullable FsVolumeSpi volume, public void afterMetadataOp(@Nullable FsVolumeSpi volume,
FileIoProvider.OPERATION op, long begin) { FileIoProvider.OPERATION op, long begin) {
DataNodeVolumeMetrics metrics = getVolumeMetrics(volume); if (isEnabled) {
if (metrics != null) { DataNodeVolumeMetrics metrics = getVolumeMetrics(volume);
metrics.addMetadastaOperationLatency(Time.monotonicNow() - begin); if (metrics != null) {
} metrics.addMetadastaOperationLatency(Time.monotonicNow() - begin);
}
@Override
public long beforeFileIo(@Nullable FsVolumeSpi volume,
FileIoProvider.OPERATION op, long len) {
DataNodeVolumeMetrics metrics = getVolumeMetrics(volume);
if (metrics != null) {
return Time.monotonicNow();
}
return 0;
}
@Override
public void afterFileIo(@Nullable FsVolumeSpi volume,
FileIoProvider.OPERATION op, long begin, long len) {
DataNodeVolumeMetrics metrics = getVolumeMetrics(volume);
if (metrics != null) {
long latency = Time.monotonicNow() - begin;
metrics.addDataFileIoLatency(latency);
switch (op) {
case SYNC:
metrics.addSyncIoLatency(latency);
break;
case FLUSH:
metrics.addFlushIoLatency(latency);
break;
case READ:
metrics.addReadIoLatency(latency);
break;
case WRITE:
metrics.addWriteIoLatency(latency);
break;
default:
} }
} }
} }
@Override public long beforeFileIo(@Nullable FsVolumeSpi volume,
public void onFailure(@Nullable FsVolumeSpi volume, FileIoProvider.OPERATION op, long len) {
FileIoProvider.OPERATION op, Exception e, long begin) { if (isEnabled) {
DataNodeVolumeMetrics metrics = getVolumeMetrics(volume); DataNodeVolumeMetrics metrics = getVolumeMetrics(volume);
if (metrics != null) { if (metrics != null) {
metrics.addFileIoError(Time.monotonicNow() - begin); return Time.monotonicNow();
}
}
return 0;
}
public void afterFileIo(@Nullable FsVolumeSpi volume,
FileIoProvider.OPERATION op, long begin, long len) {
if (isEnabled) {
DataNodeVolumeMetrics metrics = getVolumeMetrics(volume);
if (metrics != null) {
long latency = Time.monotonicNow() - begin;
metrics.addDataFileIoLatency(latency);
switch (op) {
case SYNC:
metrics.addSyncIoLatency(latency);
break;
case FLUSH:
metrics.addFlushIoLatency(latency);
break;
case READ:
metrics.addReadIoLatency(latency);
break;
case WRITE:
metrics.addWriteIoLatency(latency);
break;
default:
}
}
} }
} }
@Nullable public void onFailure(@Nullable FsVolumeSpi volume, long begin) {
@Override if (isEnabled) {
public String getStatistics() { DataNodeVolumeMetrics metrics = getVolumeMetrics(volume);
return null; if (metrics != null) {
metrics.addFileIoError(Time.monotonicNow() - begin);
}
}
} }
private DataNodeVolumeMetrics getVolumeMetrics(final FsVolumeSpi volume) { private DataNodeVolumeMetrics getVolumeMetrics(final FsVolumeSpi volume) {
if (volume != null) { if (isEnabled) {
return volume.getMetrics(); if (volume != null) {
return volume.getMetrics();
}
} }
return null; return null;
} }

View File

@ -121,8 +121,8 @@ public class TestDataNodeVolumeMetrics {
private MiniDFSCluster setupClusterForVolumeMetrics() throws IOException { private MiniDFSCluster setupClusterForVolumeMetrics() throws IOException {
Configuration conf = new HdfsConfiguration(); Configuration conf = new HdfsConfiguration();
conf.set(DFSConfigKeys.DFS_DATANODE_FILE_IO_EVENTS_CLASS_KEY, conf.setBoolean(DFSConfigKeys
"org.apache.hadoop.hdfs.server.datanode.ProfilingFileIoEvents"); .DFS_DATANODE_ENABLE_FILEIO_PROFILING_KEY, true);
SimulatedFSDataset.setFactory(conf); SimulatedFSDataset.setFactory(conf);
return new MiniDFSCluster.Builder(conf) return new MiniDFSCluster.Builder(conf)
.numDataNodes(NUM_DATANODES) .numDataNodes(NUM_DATANODES)

View File

@ -100,7 +100,9 @@ public class TestHdfsConfigFields extends TestConfigurationFieldsBase {
configurationPropsToSkipCompare configurationPropsToSkipCompare
.add(DFSConfigKeys.DFS_NAMENODE_STARTUP_KEY); .add(DFSConfigKeys.DFS_NAMENODE_STARTUP_KEY);
configurationPropsToSkipCompare configurationPropsToSkipCompare
.add(DFSConfigKeys.DFS_DATANODE_FILE_IO_EVENTS_CLASS_KEY); .add(DFSConfigKeys.DFS_DATANODE_ENABLE_FILEIO_PROFILING_KEY);
configurationPropsToSkipCompare.add(DFSConfigKeys
.DFS_DATANODE_ENABLE_FILEIO_FAULT_INJECTION_KEY);
// Allocate // Allocate
xmlPropsToSkipCompare = new HashSet<String>(); xmlPropsToSkipCompare = new HashSet<String>();