HBASE-23686 Revert binary incompatible change in ByteRangeUtils and removed reflections in CommonFSUtils

Signed-off-by: Sean Busbey <busbey@apache.org>
This commit is contained in:
Jan Hentschel 2020-01-19 18:06:31 +01:00 committed by GitHub
parent 75b85016eb
commit 50e264485d
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
4 changed files with 44 additions and 121 deletions

View File

@ -51,4 +51,8 @@
<suppress checks="EmptyBlock" files="org.apache.hadoop.hbase.TestTimeout"/> <suppress checks="EmptyBlock" files="org.apache.hadoop.hbase.TestTimeout"/>
<suppress checks="InnerAssignment" files="org.apache.hadoop.hbase.rest.PerformanceEvaluation"/> <suppress checks="InnerAssignment" files="org.apache.hadoop.hbase.rest.PerformanceEvaluation"/>
<suppress checks="EmptyBlock" files="org.apache.hadoop.hbase.rest.PerformanceEvaluation"/> <suppress checks="EmptyBlock" files="org.apache.hadoop.hbase.rest.PerformanceEvaluation"/>
<!-- Will not have a private constructor, because it is InterfaceAudience.Public -->
<suppress checks="HideUtilityClassConstructor" files="org.apache.hadoop.hbase.util.ByteRangeUtils"/>
<!-- Will not be final, because it is InterfaceAudience.Public -->
<suppress checks="FinalClass" files="org.apache.hadoop.hbase.net.Address"/>
</suppressions> </suppressions>

View File

@ -31,7 +31,7 @@ import org.apache.hbase.thirdparty.com.google.common.net.HostAndPort;
* We cannot have Guava classes in our API hence this Type. * We cannot have Guava classes in our API hence this Type.
*/ */
@InterfaceAudience.Public @InterfaceAudience.Public
public final class Address implements Comparable<Address> { public class Address implements Comparable<Address> {
private HostAndPort hostAndPort; private HostAndPort hostAndPort;
private Address(HostAndPort hostAndPort) { private Address(HostAndPort hostAndPort) {

View File

@ -30,10 +30,7 @@ import org.apache.hbase.thirdparty.com.google.common.collect.Lists;
* Utility methods for working with {@link ByteRange}. * Utility methods for working with {@link ByteRange}.
*/ */
@InterfaceAudience.Public @InterfaceAudience.Public
public final class ByteRangeUtils { public class ByteRangeUtils {
private ByteRangeUtils() {
}
public static int numEqualPrefixBytes(ByteRange left, ByteRange right, int rightInnerOffset) { public static int numEqualPrefixBytes(ByteRange left, ByteRange right, int rightInnerOffset) {
int maxCompares = Math.min(left.getLength(), right.getLength() - rightInnerOffset); int maxCompares = Math.min(left.getLength(), right.getLength() - rightInnerOffset);
final byte[] lbytes = left.getBytes(); final byte[] lbytes = left.getBytes();

View File

@ -149,69 +149,22 @@ public abstract class CommonFSUtils {
* Return the number of bytes that large input files should be optimally * Return the number of bytes that large input files should be optimally
* be split into to minimize i/o time. * be split into to minimize i/o time.
* *
* use reflection to search for getDefaultBlockSize(Path f)
* if the method doesn't exist, fall back to using getDefaultBlockSize()
*
* @param fs filesystem object * @param fs filesystem object
* @return the default block size for the path's filesystem * @return the default block size for the path's filesystem
* @throws IOException e
*/ */
public static long getDefaultBlockSize(final FileSystem fs, final Path path) throws IOException { public static long getDefaultBlockSize(final FileSystem fs, final Path path) {
Method m = null;
Class<? extends FileSystem> cls = fs.getClass();
try {
m = cls.getMethod("getDefaultBlockSize", Path.class);
} catch (NoSuchMethodException e) {
LOG.info("FileSystem doesn't support getDefaultBlockSize");
} catch (SecurityException e) {
LOG.info("Doesn't have access to getDefaultBlockSize on FileSystems", e);
m = null; // could happen on setAccessible()
}
if (m == null) {
return fs.getDefaultBlockSize(path); return fs.getDefaultBlockSize(path);
} else {
try {
Object ret = m.invoke(fs, path);
return ((Long)ret).longValue();
} catch (Exception e) {
throw new IOException(e);
}
}
} }
/* /*
* Get the default replication. * Get the default replication.
* *
* use reflection to search for getDefaultReplication(Path f)
* if the method doesn't exist, fall back to using getDefaultReplication()
*
* @param fs filesystem object * @param fs filesystem object
* @param f path of file * @param f path of file
* @return default replication for the path's filesystem * @return default replication for the path's filesystem
* @throws IOException e
*/ */
public static short getDefaultReplication(final FileSystem fs, final Path path) public static short getDefaultReplication(final FileSystem fs, final Path path) {
throws IOException {
Method m = null;
Class<? extends FileSystem> cls = fs.getClass();
try {
m = cls.getMethod("getDefaultReplication", Path.class);
} catch (NoSuchMethodException e) {
LOG.info("FileSystem doesn't support getDefaultReplication");
} catch (SecurityException e) {
LOG.info("Doesn't have access to getDefaultReplication on FileSystems", e);
m = null; // could happen on setAccessible()
}
if (m == null) {
return fs.getDefaultReplication(path); return fs.getDefaultReplication(path);
} else {
try {
Object ret = m.invoke(fs, path);
return ((Number)ret).shortValue();
} catch (Exception e) {
throw new IOException(e);
}
}
} }
/** /**
@ -568,40 +521,11 @@ public abstract class CommonFSUtils {
*/ */
private static void invokeSetStoragePolicy(final FileSystem fs, final Path path, private static void invokeSetStoragePolicy(final FileSystem fs, final Path path,
final String storagePolicy) throws IOException { final String storagePolicy) throws IOException {
Method m = null;
Exception toThrow = null; Exception toThrow = null;
try { try {
m = fs.getClass().getDeclaredMethod("setStoragePolicy", Path.class, String.class); fs.setStoragePolicy(path, storagePolicy);
m.setAccessible(true);
} catch (NoSuchMethodException e) {
toThrow = e;
final String msg = "FileSystem doesn't support setStoragePolicy; HDFS-6584, HDFS-9345 " +
"not available. This is normal and expected on earlier Hadoop versions.";
if (!warningMap.containsKey(fs)) {
warningMap.put(fs, true);
LOG.warn(msg, e);
} else if (LOG.isDebugEnabled()) {
LOG.debug(msg, e);
}
m = null;
} catch (SecurityException e) {
toThrow = e;
final String msg = "No access to setStoragePolicy on FileSystem from the SecurityManager; " +
"HDFS-6584, HDFS-9345 not available. This is unusual and probably warrants an email " +
"to the user@hbase mailing list. Please be sure to include a link to your configs, and " +
"logs that include this message and period of time before it. Logs around service " +
"start up will probably be useful as well.";
if (!warningMap.containsKey(fs)) {
warningMap.put(fs, true);
LOG.warn(msg, e);
} else if (LOG.isDebugEnabled()) {
LOG.debug(msg, e);
}
m = null; // could happen on setAccessible() or getDeclaredMethod()
}
if (m != null) {
try {
m.invoke(fs, path, storagePolicy);
if (LOG.isDebugEnabled()) { if (LOG.isDebugEnabled()) {
LOG.debug("Set storagePolicy={} for path={}", storagePolicy, path); LOG.debug("Set storagePolicy={} for path={}", storagePolicy, path);
} }
@ -616,12 +540,11 @@ public abstract class CommonFSUtils {
} else if (LOG.isDebugEnabled()) { } else if (LOG.isDebugEnabled()) {
LOG.debug("Unable to set storagePolicy=" + storagePolicy + " for path=" + path, e); LOG.debug("Unable to set storagePolicy=" + storagePolicy + " for path=" + path, e);
} }
// check for lack of HDFS-7228 // check for lack of HDFS-7228
if (e instanceof InvocationTargetException) { if (e instanceof RemoteException &&
final Throwable exception = e.getCause();
if (exception instanceof RemoteException &&
HadoopIllegalArgumentException.class.getName().equals( HadoopIllegalArgumentException.class.getName().equals(
((RemoteException)exception).getClassName())) { ((RemoteException)e).getClassName())) {
if (LOG.isDebugEnabled()) { if (LOG.isDebugEnabled()) {
LOG.debug("Given storage policy, '" +storagePolicy +"', was rejected and probably " + LOG.debug("Given storage policy, '" +storagePolicy +"', was rejected and probably " +
"isn't a valid policy for the version of Hadoop you're running. I.e. if you're " + "isn't a valid policy for the version of Hadoop you're running. I.e. if you're " +
@ -630,7 +553,7 @@ public abstract class CommonFSUtils {
} }
// Hadoop 2.8+, 3.0-a1+ added FileSystem.setStoragePolicy with a default implementation // Hadoop 2.8+, 3.0-a1+ added FileSystem.setStoragePolicy with a default implementation
// that throws UnsupportedOperationException // that throws UnsupportedOperationException
} else if (exception instanceof UnsupportedOperationException) { } else if (e instanceof UnsupportedOperationException) {
if (LOG.isDebugEnabled()) { if (LOG.isDebugEnabled()) {
LOG.debug("The underlying FileSystem implementation doesn't support " + LOG.debug("The underlying FileSystem implementation doesn't support " +
"setStoragePolicy. This is probably intentional on their part, since HDFS-9345 " + "setStoragePolicy. This is probably intentional on their part, since HDFS-9345 " +
@ -639,12 +562,11 @@ public abstract class CommonFSUtils {
"specification docs from HADOOP-11981, and/or related documentation from the " + "specification docs from HADOOP-11981, and/or related documentation from the " +
"provider of the underlying FileSystem (its name should appear in the " + "provider of the underlying FileSystem (its name should appear in the " +
"stacktrace that accompanies this message). Note in particular that Hadoop's " + "stacktrace that accompanies this message). Note in particular that Hadoop's " +
"local filesystem implementation doesn't support storage policies.", exception); "local filesystem implementation doesn't support storage policies.", e);
}
}
} }
} }
} }
if (toThrow != null) { if (toThrow != null) {
throw new IOException(toThrow); throw new IOException(toThrow);
} }