HBASE-13200 Improper configuration can leads to endless lease recovery during failover (He Liangliang)
This commit is contained in:
parent
f9a17edc25
commit
6709feeb65
|
@ -180,9 +180,11 @@ public class FSHDFSUtils extends FSUtils {
|
|||
long firstPause = conf.getInt("hbase.lease.recovery.first.pause", 4000);
|
||||
// This should be set to how long it'll take for us to timeout against primary datanode if it
|
||||
// is dead. We set it to 61 seconds, 1 second than the default READ_TIMEOUT in HDFS, the
|
||||
// default value for DFS_CLIENT_SOCKET_TIMEOUT_KEY.
|
||||
long subsequentPause = conf.getInt("hbase.lease.recovery.dfs.timeout", 61 * 1000);
|
||||
|
||||
// default value for DFS_CLIENT_SOCKET_TIMEOUT_KEY. If recovery is still failing after this
|
||||
// timeout, then further recovery will take liner backoff with this base, to avoid endless
|
||||
// preemptions when this value is not properly configured.
|
||||
long subsequentPauseBase = conf.getLong("hbase.lease.recovery.dfs.timeout", 61 * 1000);
|
||||
|
||||
Method isFileClosedMeth = null;
|
||||
// whether we need to look for isFileClosed method
|
||||
boolean findIsFileClosedMeth = true;
|
||||
|
@ -198,11 +200,11 @@ public class FSHDFSUtils extends FSUtils {
|
|||
if (nbAttempt == 0) {
|
||||
Thread.sleep(firstPause);
|
||||
} else {
|
||||
// Cycle here until subsequentPause elapses. While spinning, check isFileClosed if
|
||||
// available (should be in hadoop 2.0.5... not in hadoop 1 though.
|
||||
// Cycle here until (subsequentPause * nbAttempt) elapses. While spinning, check
|
||||
// isFileClosed if available (should be in hadoop 2.0.5... not in hadoop 1 though.
|
||||
long localStartWaiting = EnvironmentEdgeManager.currentTime();
|
||||
while ((EnvironmentEdgeManager.currentTime() - localStartWaiting) <
|
||||
subsequentPause) {
|
||||
subsequentPauseBase * nbAttempt) {
|
||||
Thread.sleep(conf.getInt("hbase.lease.recovery.pause", 1000));
|
||||
if (findIsFileClosedMeth) {
|
||||
try {
|
||||
|
|
Loading…
Reference in New Issue