HBASE-930 RegionServer stuck: HLog: Could not append. Requesting close of log java.io.IOException: Could not get block locations. Aborting...
git-svn-id: https://svn.apache.org/repos/asf/hadoop/hbase/branches/0.18@705064 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
8b7d4da0f9
commit
06d1a00f83
|
@ -13,6 +13,8 @@ Release 0.18.1 - Unreleased
|
||||||
HBASE-928 NPE throwing RetriesExhaustedException
|
HBASE-928 NPE throwing RetriesExhaustedException
|
||||||
HBASE-576 Investigate IPC performance; partial.
|
HBASE-576 Investigate IPC performance; partial.
|
||||||
HBASE-924 Update hadoop in lib on 0.18 hbase branch to 0.18.1
|
HBASE-924 Update hadoop in lib on 0.18 hbase branch to 0.18.1
|
||||||
|
HBASE-930 RegionServer stuck: HLog: Could not append. Requesting close of
|
||||||
|
log java.io.IOException: Could not get block locations. Aborting...
|
||||||
|
|
||||||
IMPROVEMENTS
|
IMPROVEMENTS
|
||||||
HBASE-920 Make region balancing sloppier
|
HBASE-920 Make region balancing sloppier
|
||||||
|
|
|
@ -178,7 +178,7 @@ class Flusher extends Thread implements FlushRequester {
|
||||||
// is required. Currently the only way to do this is a restart of
|
// is required. Currently the only way to do this is a restart of
|
||||||
// the server. Abort because hdfs is probably bad (HBASE-644 is a case
|
// the server. Abort because hdfs is probably bad (HBASE-644 is a case
|
||||||
// where hdfs was bad but passed the hdfs check).
|
// where hdfs was bad but passed the hdfs check).
|
||||||
LOG.fatal("Replay of hlog required. Forcing server restart", ex);
|
LOG.fatal("Replay of hlog required. Forcing server shutdown", ex);
|
||||||
server.abort();
|
server.abort();
|
||||||
return false;
|
return false;
|
||||||
} catch (IOException ex) {
|
} catch (IOException ex) {
|
||||||
|
|
|
@ -226,9 +226,10 @@ public class HLog implements HConstants {
|
||||||
* cacheFlushLock and then completeCacheFlush could be called which would wait
|
* cacheFlushLock and then completeCacheFlush could be called which would wait
|
||||||
* for the lock on this and consequently never release the cacheFlushLock
|
* for the lock on this and consequently never release the cacheFlushLock
|
||||||
*
|
*
|
||||||
|
* @throws FailedLogCloseException
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
*/
|
*/
|
||||||
public void rollWriter() throws IOException {
|
public void rollWriter() throws FailedLogCloseException, IOException {
|
||||||
this.cacheFlushLock.lock();
|
this.cacheFlushLock.lock();
|
||||||
try {
|
try {
|
||||||
if (closed) {
|
if (closed) {
|
||||||
|
@ -237,7 +238,14 @@ public class HLog implements HConstants {
|
||||||
synchronized (updateLock) {
|
synchronized (updateLock) {
|
||||||
if (this.writer != null) {
|
if (this.writer != null) {
|
||||||
// Close the current writer, get a new one.
|
// Close the current writer, get a new one.
|
||||||
this.writer.close();
|
try {
|
||||||
|
this.writer.close();
|
||||||
|
} catch (IOException e) {
|
||||||
|
// Failed close of log file. Means we're losing edits. For now,
|
||||||
|
// shut ourselves down to minimize loss. Alternative is to try and
|
||||||
|
// keep going. See HBASE-930.
|
||||||
|
throw new FailedLogCloseException("#" + this.filenum, e);
|
||||||
|
}
|
||||||
Path p = computeFilename(old_filenum);
|
Path p = computeFilename(old_filenum);
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
LOG.debug("Closing current log writer " + FSUtils.getPath(p));
|
LOG.debug("Closing current log writer " + FSUtils.getPath(p));
|
||||||
|
|
|
@ -77,8 +77,11 @@ class LogRoller extends Thread implements LogRollListener {
|
||||||
try {
|
try {
|
||||||
LOG.info("Rolling hlog. Number of entries: " + server.getLog().getNumEntries());
|
LOG.info("Rolling hlog. Number of entries: " + server.getLog().getNumEntries());
|
||||||
server.getLog().rollWriter();
|
server.getLog().rollWriter();
|
||||||
|
} catch (FailedLogCloseException e) {
|
||||||
|
LOG.fatal("Forcing server shutdown", e);
|
||||||
|
server.abort();
|
||||||
} catch (IOException ex) {
|
} catch (IOException ex) {
|
||||||
LOG.error("Log rolling failed",
|
LOG.error("Log rolling failed with ioe: ",
|
||||||
RemoteExceptionHandler.checkIOException(ex));
|
RemoteExceptionHandler.checkIOException(ex));
|
||||||
server.checkFileSystem();
|
server.checkFileSystem();
|
||||||
} catch (Exception ex) {
|
} catch (Exception ex) {
|
||||||
|
|
|
@ -72,10 +72,8 @@ public class FSUtils {
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
exception = RemoteExceptionHandler.checkIOException(e);
|
exception = RemoteExceptionHandler.checkIOException(e);
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
fs.close();
|
fs.close();
|
||||||
|
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
LOG.error("file system close failed: ", e);
|
LOG.error("file system close failed: ", e);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue