HBASE-17224 Fix lots of spelling errors in HBase logging and exception messages (Grant Sohn)

This commit is contained in:
Jonathan M Hsieh 2016-12-01 08:57:16 -08:00
parent 540ede376b
commit eeaea4aea3
29 changed files with 38 additions and 38 deletions

View File

@ -1249,7 +1249,7 @@ public class HColumnDescriptor implements Comparable<HColumnDescriptor> {
public HColumnDescriptor setDFSReplication(short replication) { public HColumnDescriptor setDFSReplication(short replication) {
if (replication < 1 && replication != DEFAULT_DFS_REPLICATION) { if (replication < 1 && replication != DEFAULT_DFS_REPLICATION) {
throw new IllegalArgumentException( throw new IllegalArgumentException(
"DFS replication factor cannot be less than 1 if explictly set."); "DFS replication factor cannot be less than 1 if explicitly set.");
} }
setValue(DFS_REPLICATION, Short.toString(replication)); setValue(DFS_REPLICATION, Short.toString(replication));
return this; return this;

View File

@ -551,7 +551,7 @@ public class ReplicationAdmin implements Closeable {
} else if (!peerHtd.equals(htd)) { } else if (!peerHtd.equals(htd)) {
throw new IllegalArgumentException("Table " + tableName.getNameAsString() throw new IllegalArgumentException("Table " + tableName.getNameAsString()
+ " exists in peer cluster " + repPeer.getId() + " exists in peer cluster " + repPeer.getId()
+ ", but the table descriptors are not same when comapred with source cluster." + ", but the table descriptors are not same when compared with source cluster."
+ " Thus can not enable the table's replication switch."); + " Thus can not enable the table's replication switch.");
} }
} }

View File

@ -182,7 +182,7 @@ public class ReplicationPeerZKImpl extends ReplicationStateZKBase
@Override @Override
public void abort(String why, Throwable e) { public void abort(String why, Throwable e) {
LOG.fatal("The ReplicationPeer coresponding to peer " + peerConfig LOG.fatal("The ReplicationPeer corresponding to peer " + peerConfig
+ " was aborted for the following reason(s):" + why, e); + " was aborted for the following reason(s):" + why, e);
} }

View File

@ -138,7 +138,7 @@ public class BoundedByteBufferPool {
if (LOG.isTraceEnabled()) { if (LOG.isTraceEnabled()) {
long allocations = allocationsRef.incrementAndGet(); long allocations = allocationsRef.incrementAndGet();
LOG.trace("runningAverage=" + runningAverage + ", alloctions=" + allocations); LOG.trace("runningAverage=" + runningAverage + ", allocations=" + allocations);
} }
return bb; return bb;
} }

View File

@ -113,7 +113,7 @@ public class HDFSBlocksDistribution {
*/ */
@Override @Override
public synchronized String toString() { public synchronized String toString() {
return "number of unique hosts in the disribution=" + return "number of unique hosts in the distribution=" +
this.hostAndWeights.size(); this.hostAndWeights.size();
} }

View File

@ -138,7 +138,7 @@ public class ZNodeClearer {
String[] rsZnodeParts = rsZnodePath.split("/"); String[] rsZnodeParts = rsZnodePath.split("/");
masterServerName = rsZnodeParts[rsZnodeParts.length -1]; masterServerName = rsZnodeParts[rsZnodeParts.length -1];
} catch (IndexOutOfBoundsException e) { } catch (IndexOutOfBoundsException e) {
LOG.warn("String " + rsZnodePath + " has wrong fromat", e); LOG.warn("String " + rsZnodePath + " has wrong format", e);
} }
return masterServerName; return masterServerName;
} }

View File

@ -1598,7 +1598,7 @@ public class HFileBlock implements Cacheable {
blk = readBlockDataInternal(is, offset, onDiskSizeWithHeaderL, pread, blk = readBlockDataInternal(is, offset, onDiskSizeWithHeaderL, pread,
doVerificationThruHBaseChecksum); doVerificationThruHBaseChecksum);
if (blk != null) { if (blk != null) {
HFile.LOG.warn("HDFS checksum verification suceeded for file " + HFile.LOG.warn("HDFS checksum verification succeeded for file " +
pathName + " at offset " + pathName + " at offset " +
offset + " filesize " + fileSize); offset + " filesize " + fileSize);
} }

View File

@ -1024,7 +1024,7 @@ public class BucketCache implements BlockCache, HeapSize {
if (this.ioErrorStartTime > 0) { if (this.ioErrorStartTime > 0) {
if (cacheEnabled && (now - ioErrorStartTime) > this.ioErrorsTolerationDuration) { if (cacheEnabled && (now - ioErrorStartTime) > this.ioErrorsTolerationDuration) {
LOG.error("IO errors duration time has exceeded " + ioErrorsTolerationDuration + LOG.error("IO errors duration time has exceeded " + ioErrorsTolerationDuration +
"ms, disabing cache, please check your IOEngine"); "ms, disabling cache, please check your IOEngine");
disableCache(); disableCache();
} }
} else { } else {

View File

@ -473,7 +473,7 @@ public class LoadIncrementalHFiles extends Configured implements Tool {
// need to reload split keys each iteration. // need to reload split keys each iteration.
final Pair<byte[][], byte[][]> startEndKeys = regionLocator.getStartEndKeys(); final Pair<byte[][], byte[][]> startEndKeys = regionLocator.getStartEndKeys();
if (count != 0) { if (count != 0) {
LOG.info("Split occured while grouping HFiles, retry attempt " + LOG.info("Split occurred while grouping HFiles, retry attempt " +
+ count + " with " + queue.size() + " files remaining to group or split"); + count + " with " + queue.size() + " files remaining to group or split");
} }
@ -903,12 +903,12 @@ public class LoadIncrementalHFiles extends Configured implements Tool {
if (indexForCallable < 0) { if (indexForCallable < 0) {
throw new IOException("The first region info for table " throw new IOException("The first region info for table "
+ table.getName() + table.getName()
+ " cann't be found in hbase:meta.Please use hbck tool to fix it first."); + " can't be found in hbase:meta.Please use hbck tool to fix it first.");
} else if ((indexForCallable == startEndKeys.getFirst().length - 1) } else if ((indexForCallable == startEndKeys.getFirst().length - 1)
&& !Bytes.equals(startEndKeys.getSecond()[indexForCallable], HConstants.EMPTY_BYTE_ARRAY)) { && !Bytes.equals(startEndKeys.getSecond()[indexForCallable], HConstants.EMPTY_BYTE_ARRAY)) {
throw new IOException("The last region info for table " throw new IOException("The last region info for table "
+ table.getName() + table.getName()
+ " cann't be found in hbase:meta.Please use hbck tool to fix it first."); + " can't be found in hbase:meta.Please use hbck tool to fix it first.");
} else if (indexForCallable + 1 < startEndKeys.getFirst().length } else if (indexForCallable + 1 < startEndKeys.getFirst().length
&& !(Bytes.compareTo(startEndKeys.getSecond()[indexForCallable], && !(Bytes.compareTo(startEndKeys.getSecond()[indexForCallable],
startEndKeys.getFirst()[indexForCallable + 1]) == 0)) { startEndKeys.getFirst()[indexForCallable + 1]) == 0)) {

View File

@ -298,7 +298,7 @@ public class VerifyReplication extends Configured implements Tool {
return pair; return pair;
} catch (ReplicationException e) { } catch (ReplicationException e) {
throw new IOException( throw new IOException(
"An error occured while trying to connect to the remove peer cluster", e); "An error occurred while trying to connect to the remove peer cluster", e);
} finally { } finally {
if (peer != null) { if (peer != null) {
peer.close(); peer.close();

View File

@ -460,7 +460,7 @@ public class AssignmentVerificationReport {
public void print(boolean isDetailMode) { public void print(boolean isDetailMode) {
if (!isFilledUp) { if (!isFilledUp) {
System.err.println("[Error] Region assignment verfication report" + System.err.println("[Error] Region assignment verification report" +
"hasn't been filled up"); "hasn't been filled up");
} }
DecimalFormat df = new java.text.DecimalFormat( "#.##"); DecimalFormat df = new java.text.DecimalFormat( "#.##");

View File

@ -89,7 +89,7 @@ public class MasterMobCompactionThread {
try { try {
master.reportMobCompactionEnd(tableName); master.reportMobCompactionEnd(tableName);
} catch (IOException e1) { } catch (IOException e1) {
LOG.error("Failed to mark end of mob compation", e1); LOG.error("Failed to mark end of mob compaction", e1);
} }
throw e; throw e;
} }
@ -131,7 +131,7 @@ public class MasterMobCompactionThread {
try { try {
master.reportMobCompactionEnd(tableName); master.reportMobCompactionEnd(tableName);
} catch (IOException e) { } catch (IOException e) {
LOG.error("Failed to mark end of mob compation", e); LOG.error("Failed to mark end of mob compaction", e);
} }
} }
} }

View File

@ -976,8 +976,8 @@ public class RegionPlacementMaintainer {
opt.addOption("tables", true, opt.addOption("tables", true,
"The list of table names splitted by ',' ;" + "The list of table names splitted by ',' ;" +
"For example: -tables: t1,t2,...,tn"); "For example: -tables: t1,t2,...,tn");
opt.addOption("l", "locality", true, "enforce the maxium locality"); opt.addOption("l", "locality", true, "enforce the maximum locality");
opt.addOption("m", "min-move", true, "enforce minium assignment move"); opt.addOption("m", "min-move", true, "enforce minimum assignment move");
opt.addOption("diff", false, "calculate difference between assignment plans"); opt.addOption("diff", false, "calculate difference between assignment plans");
opt.addOption("munkres", false, opt.addOption("munkres", false,
"use munkres to place secondaries and tertiaries"); "use munkres to place secondaries and tertiaries");

View File

@ -1091,7 +1091,7 @@ public class RegionStates {
} }
return hri; return hri;
} catch (IOException e) { } catch (IOException e) {
server.abort("Aborting because error occoured while reading " server.abort("Aborting because error occurred while reading "
+ Bytes.toStringBinary(regionName) + " from hbase:meta", e); + Bytes.toStringBinary(regionName) + " from hbase:meta", e);
return null; return null;
} }

View File

@ -460,7 +460,7 @@ public class FavoredNodeAssignmentHelper {
getOneRandomServer(primaryRack, serverSkipSet); getOneRandomServer(primaryRack, serverSkipSet);
if (secondaryRS == null || tertiaryRS == null) { if (secondaryRS == null || tertiaryRS == null) {
LOG.error("Cannot place the secondary and terinary" + LOG.error("Cannot place the secondary and ternary" +
"region server for region " + "region server for region " +
regionInfo.getRegionNameAsString()); regionInfo.getRegionNameAsString());
} }
@ -498,7 +498,7 @@ public class FavoredNodeAssignmentHelper {
ServerName tertiaryRS = getOneRandomServer(secondaryRack, skipServerSet); ServerName tertiaryRS = getOneRandomServer(secondaryRack, skipServerSet);
if (secondaryRS == null || tertiaryRS == null) { if (secondaryRS == null || tertiaryRS == null) {
LOG.error("Cannot place the secondary and terinary" + LOG.error("Cannot place the secondary and ternary" +
"region server for region " + "region server for region " +
regionInfo.getRegionNameAsString()); regionInfo.getRegionNameAsString());
} }
@ -603,4 +603,4 @@ public class FavoredNodeAssignmentHelper {
} }
return strBuf.toString(); return strBuf.toString();
} }
} }

View File

@ -349,7 +349,7 @@ public class DeleteTableProcedure
} }
} }
if (!deletes.isEmpty()) { if (!deletes.isEmpty()) {
LOG.warn("Deleting some vestigal " + deletes.size() + " rows of " + tableName + LOG.warn("Deleting some vestigial " + deletes.size() + " rows of " + tableName +
" from " + TableName.META_TABLE_NAME); " from " + TableName.META_TABLE_NAME);
metaTable.delete(deletes); metaTable.delete(deletes);
} }

View File

@ -416,14 +416,14 @@ public class DispatchMergingRegionsProcedure
forcible, forcible,
getUser()); getUser());
LOG.info("Sent merge to server " + getServerName(env) + " for region " + LOG.info("Sent merge to server " + getServerName(env) + " for region " +
getRegionsToMergeListEncodedNameString() + ", focible=" + forcible); getRegionsToMergeListEncodedNameString() + ", forcible=" + forcible);
return; return;
} catch (RegionOpeningException roe) { } catch (RegionOpeningException roe) {
// Do a retry since region should be online on RS immediately // Do a retry since region should be online on RS immediately
LOG.warn("Failed mergering regions in " + getServerName(env) + ", retrying...", roe); LOG.warn("Failed mergering regions in " + getServerName(env) + ", retrying...", roe);
} catch (Exception ie) { } catch (Exception ie) {
LOG.warn("Failed sending merge to " + getServerName(env) + " for regions " + LOG.warn("Failed sending merge to " + getServerName(env) + " for regions " +
getRegionsToMergeListEncodedNameString() + ", focible=" + forcible, ie); getRegionsToMergeListEncodedNameString() + ", forcible=" + forcible, ie);
return; return;
} }
} while ((duration = EnvironmentEdgeManager.currentTime() - startTime) <= getTimeout(env)); } while ((duration = EnvironmentEdgeManager.currentTime() - startTime) <= getTimeout(env));

View File

@ -697,7 +697,7 @@ public class SplitTableRegionProcedure
HRegionInfo.parseRegionName(p.getRow()); HRegionInfo.parseRegionName(p.getRow());
} }
} catch (IOException e) { } catch (IOException e) {
LOG.error("Row key of mutation from coprossor is not parsable as region name." LOG.error("Row key of mutation from coprocessor is not parsable as region name."
+ "Mutations from coprocessor should only for hbase:meta table."); + "Mutations from coprocessor should only for hbase:meta table.");
throw e; throw e;
} }

View File

@ -199,7 +199,7 @@ public class ZKProcedureMemberRpcs implements ProcedureMemberRpcs {
try { try {
byte[] data = ZKUtil.getData(zkController.getWatcher(), path); byte[] data = ZKUtil.getData(zkController.getWatcher(), path);
if (!ProtobufUtil.isPBMagicPrefix(data)) { if (!ProtobufUtil.isPBMagicPrefix(data)) {
String msg = "Data in for starting procuedure " + opName + String msg = "Data in for starting procedure " + opName +
" is illegally formatted (no pb magic). " + " is illegally formatted (no pb magic). " +
"Killing the procedure: " + Bytes.toString(data); "Killing the procedure: " + Bytes.toString(data);
LOG.error(msg); LOG.error(msg);

View File

@ -100,7 +100,7 @@ public class CompactedHFilesDischarger extends ScheduledChore {
+ region.getRegionInfo() + " under the store " + store.getColumnFamilyName()); + region.getRegionInfo() + " under the store " + store.getColumnFamilyName());
} }
} catch (Exception e) { } catch (Exception e) {
LOG.error("Exception while trying to close and archive the comapcted store " LOG.error("Exception while trying to close and archive the compacted store "
+ "files of the store " + store.getColumnFamilyName() + " in the" + " region " + "files of the store " + store.getColumnFamilyName() + " in the" + " region "
+ region.getRegionInfo(), e); + region.getRegionInfo(), e);
} }

View File

@ -206,7 +206,7 @@ public class HeapMemoryManager {
public void stop() { public void stop() {
// The thread is Daemon. Just interrupting the ongoing process. // The thread is Daemon. Just interrupting the ongoing process.
LOG.info("Stoping HeapMemoryTuner chore."); LOG.info("Stopping HeapMemoryTuner chore.");
this.heapMemTunerChore.cancel(true); this.heapMemTunerChore.cancel(true);
} }

View File

@ -257,7 +257,7 @@ public class SplitTransactionImpl implements SplitTransaction {
HRegionInfo.parseRegionName(p.getRow()); HRegionInfo.parseRegionName(p.getRow());
} }
} catch (IOException e) { } catch (IOException e) {
LOG.error("Row key of mutation from coprossor is not parsable as region name." LOG.error("Row key of mutation from coprocessor is not parsable as region name."
+ "Mutations from coprocessor should only for hbase:meta table."); + "Mutations from coprocessor should only for hbase:meta table.");
throw e; throw e;
} }

View File

@ -77,7 +77,7 @@ public class WALSplitterHandler extends EventHandler {
break; break;
case PREEMPTED: case PREEMPTED:
SplitLogCounters.tot_wkr_preempt_task.incrementAndGet(); SplitLogCounters.tot_wkr_preempt_task.incrementAndGet();
LOG.warn("task execution prempted " + splitTaskDetails.getWALFile()); LOG.warn("task execution preempted " + splitTaskDetails.getWALFile());
break; break;
case ERR: case ERR:
if (server != null && !server.isStopped()) { if (server != null && !server.isStopped()) {

View File

@ -177,7 +177,7 @@ public class HFileReplicator {
// need to reload split keys each iteration. // need to reload split keys each iteration.
startEndKeys = locator.getStartEndKeys(); startEndKeys = locator.getStartEndKeys();
if (count != 0) { if (count != 0) {
LOG.warn("Error occured while replicating HFiles, retry attempt " + count + " with " LOG.warn("Error occurred while replicating HFiles, retry attempt " + count + " with "
+ queue.size() + " files still remaining to replicate."); + queue.size() + " files still remaining to replicate.");
} }

View File

@ -225,7 +225,7 @@ public class TableAuthManager implements Closeable {
mtime.incrementAndGet(); mtime.incrementAndGet();
} catch (IOException e) { } catch (IOException e) {
// Never happens // Never happens
LOG.error("Error occured while updating the global cache", e); LOG.error("Error occurred while updating the global cache", e);
} }
} }

View File

@ -152,7 +152,7 @@ public class AuthenticationTokenSecretManager
synchronized (this) { synchronized (this) {
if (!leaderElector.isAlive() || leaderElector.isStopped()) { if (!leaderElector.isAlive() || leaderElector.isStopped()) {
LOG.warn("Thread leaderElector[" + leaderElector.getName() + ":" LOG.warn("Thread leaderElector[" + leaderElector.getName() + ":"
+ leaderElector.getId() + "] is stoped or not alive"); + leaderElector.getId() + "] is stopped or not alive");
leaderElector.start(); leaderElector.start();
LOG.info("Thread leaderElector [" + leaderElector.getName() + ":" LOG.info("Thread leaderElector [" + leaderElector.getName() + ":"
+ leaderElector.getId() + "] is started"); + leaderElector.getId() + "] is started");

View File

@ -899,7 +899,7 @@ public class HBaseFsck extends Configured implements Closeable {
if (!valid) { if (!valid) {
errors.reportError(ERROR_CODE.BOUNDARIES_ERROR, "Found issues with regions boundaries", errors.reportError(ERROR_CODE.BOUNDARIES_ERROR, "Found issues with regions boundaries",
tablesInfo.get(regionInfo.getTable())); tablesInfo.get(regionInfo.getTable()));
LOG.warn("Region's boundaries not alligned between stores and META for:"); LOG.warn("Region's boundaries not aligned between stores and META for:");
LOG.warn(currentRegionBoundariesInformation); LOG.warn(currentRegionBoundariesInformation);
} }
} }
@ -933,7 +933,7 @@ public class HBaseFsck extends Configured implements Closeable {
FileSystem fs = p.getFileSystem(getConf()); FileSystem fs = p.getFileSystem(getConf());
FileStatus[] dirs = fs.listStatus(p); FileStatus[] dirs = fs.listStatus(p);
if (dirs == null) { if (dirs == null) {
LOG.warn("Attempt to adopt ophan hdfs region skipped becuase no files present in " + LOG.warn("Attempt to adopt orphan hdfs region skipped because no files present in " +
p + ". This dir could probably be deleted."); p + ". This dir could probably be deleted.");
return ; return ;
} }
@ -1103,7 +1103,7 @@ public class HBaseFsck extends Configured implements Closeable {
Path rootDir = getSidelineDir(); Path rootDir = getSidelineDir();
Path dst = new Path(rootDir, pathStr.substring(index + 1)); Path dst = new Path(rootDir, pathStr.substring(index + 1));
fs.mkdirs(dst.getParent()); fs.mkdirs(dst.getParent());
LOG.info("Trying to sildeline reference file " LOG.info("Trying to sideline reference file "
+ path + " to " + dst); + path + " to " + dst);
setShouldRerun(); setShouldRerun();

View File

@ -368,7 +368,7 @@ public class HFileCorruptionChecker {
try { try {
f.get(); f.get();
} catch (ExecutionException e) { } catch (ExecutionException e) {
LOG.warn("Failed to quaratine an HFile in regiondir " LOG.warn("Failed to quarantine an HFile in regiondir "
+ rdcs.get(i).regionDir, e.getCause()); + rdcs.get(i).regionDir, e.getCause());
// rethrow IOExceptions // rethrow IOExceptions
if (e.getCause() instanceof IOException) { if (e.getCause() instanceof IOException) {

View File

@ -1001,7 +1001,7 @@ public class WALSplitter {
try { try {
controller.dataAvailable.wait(2000); controller.dataAvailable.wait(2000);
} catch (InterruptedException e) { } catch (InterruptedException e) {
LOG.warn("Got intrerrupted while waiting for EntryBuffers is drained"); LOG.warn("Got interrupted while waiting for EntryBuffers is drained");
Thread.interrupted(); Thread.interrupted();
break; break;
} }