HBASE-3695 Some improvements to Hbck to test the entire region chain in Meta and provide better error reporting
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1099566 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
4e4ded4cde
commit
3d6b1485bf
|
@ -285,6 +285,8 @@ Release 0.90.3 - Unreleased
|
||||||
HBASE-3580 Remove RS from DeadServer when new instance checks in
|
HBASE-3580 Remove RS from DeadServer when new instance checks in
|
||||||
HBASE-2470 Add Scan.setTimeRange() support in Shell (Harsh J Chouraria)
|
HBASE-2470 Add Scan.setTimeRange() support in Shell (Harsh J Chouraria)
|
||||||
HBASE-3805 Log RegionState that are processed too late in the master
|
HBASE-3805 Log RegionState that are processed too late in the master
|
||||||
|
HBASE-3695 Some improvements to Hbck to test the entire region chain in
|
||||||
|
Meta and provide better error reporting (Marc Limotte)
|
||||||
|
|
||||||
TASKS
|
TASKS
|
||||||
HBASE-3748 Add rolling of thrift/rest daemons to graceful_stop.sh script
|
HBASE-3748 Add rolling of thrift/rest daemons to graceful_stop.sh script
|
||||||
|
|
|
@ -20,14 +20,7 @@
|
||||||
package org.apache.hadoop.hbase.util;
|
package org.apache.hadoop.hbase.util;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.Collection;
|
import java.util.*;
|
||||||
import java.util.Collections;
|
|
||||||
import java.util.Comparator;
|
|
||||||
import java.util.HashSet;
|
|
||||||
import java.util.List;
|
|
||||||
import java.util.Set;
|
|
||||||
import java.util.TreeMap;
|
|
||||||
import java.util.TreeSet;
|
|
||||||
import java.util.concurrent.atomic.AtomicInteger;
|
import java.util.concurrent.atomic.AtomicInteger;
|
||||||
import java.util.concurrent.ThreadPoolExecutor;
|
import java.util.concurrent.ThreadPoolExecutor;
|
||||||
import java.util.concurrent.TimeUnit;
|
import java.util.concurrent.TimeUnit;
|
||||||
|
@ -69,6 +62,8 @@ import org.apache.zookeeper.KeeperException;
|
||||||
import com.google.common.base.Joiner;
|
import com.google.common.base.Joiner;
|
||||||
import com.google.common.collect.Lists;
|
import com.google.common.collect.Lists;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Check consistency among the in-memory states of the master and the
|
* Check consistency among the in-memory states of the master and the
|
||||||
* region server(s) and the state of data in HDFS.
|
* region server(s) and the state of data in HDFS.
|
||||||
|
@ -101,7 +96,7 @@ public class HBaseFsck {
|
||||||
private Set<Result> emptyRegionInfoQualifiers = new HashSet<Result>();
|
private Set<Result> emptyRegionInfoQualifiers = new HashSet<Result>();
|
||||||
private int numThreads = MAX_NUM_THREADS;
|
private int numThreads = MAX_NUM_THREADS;
|
||||||
|
|
||||||
ThreadPoolExecutor executor; // threads to retrieve data from regionservers
|
ThreadPoolExecutor executor; // threads to retrieve data from regionservers
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Constructor
|
* Constructor
|
||||||
|
@ -140,20 +135,21 @@ public class HBaseFsck {
|
||||||
tablesInfo.clear();
|
tablesInfo.clear();
|
||||||
emptyRegionInfoQualifiers.clear();
|
emptyRegionInfoQualifiers.clear();
|
||||||
disabledTables.clear();
|
disabledTables.clear();
|
||||||
|
errors.clear();
|
||||||
|
|
||||||
// get a list of all regions from the master. This involves
|
// get a list of all regions from the master. This involves
|
||||||
// scanning the META table
|
// scanning the META table
|
||||||
if (!recordRootRegion()) {
|
if (!recordRootRegion()) {
|
||||||
// Will remove later if we can fix it
|
// Will remove later if we can fix it
|
||||||
errors.reportError("Encountered fatal error. Exitting...");
|
errors.reportError("Encountered fatal error. Exiting...");
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
getMetaEntries();
|
getMetaEntries();
|
||||||
|
|
||||||
// Check if .META. is found only once and on the right place
|
// Check if .META. is found only once and in the right place
|
||||||
if (!checkMetaEntries()) {
|
if (!checkMetaEntries()) {
|
||||||
// Will remove later if we can fix it
|
// Will remove later if we can fix it
|
||||||
errors.reportError("Encountered fatal error. Exitting...");
|
errors.reportError("Encountered fatal error. Exiting...");
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -225,6 +221,10 @@ public class HBaseFsck {
|
||||||
return errors.summarize();
|
return errors.summarize();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public ErrorReporter getErrors() {
|
||||||
|
return errors;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Load the list of disabled tables in ZK into local set.
|
* Load the list of disabled tables in ZK into local set.
|
||||||
* @throws ZooKeeperConnectionException
|
* @throws ZooKeeperConnectionException
|
||||||
|
@ -281,7 +281,8 @@ public class HBaseFsck {
|
||||||
|
|
||||||
// verify that version file exists
|
// verify that version file exists
|
||||||
if (!foundVersionFile) {
|
if (!foundVersionFile) {
|
||||||
errors.reportError("Version file does not exist in root dir " + rootDir);
|
errors.reportError(ERROR_CODE.NO_VERSION_FILE,
|
||||||
|
"Version file does not exist in root dir " + rootDir);
|
||||||
}
|
}
|
||||||
|
|
||||||
// level 1: <HBASE_DIR>/*
|
// level 1: <HBASE_DIR>/*
|
||||||
|
@ -315,7 +316,8 @@ public class HBaseFsck {
|
||||||
// Check if Root region is valid and existing
|
// Check if Root region is valid and existing
|
||||||
if (rootLocation == null || rootLocation.getRegionInfo() == null ||
|
if (rootLocation == null || rootLocation.getRegionInfo() == null ||
|
||||||
rootLocation.getHostname() == null) {
|
rootLocation.getHostname() == null) {
|
||||||
errors.reportError("Root Region or some of its attributes is null.");
|
errors.reportError(ERROR_CODE.NULL_ROOT_REGION,
|
||||||
|
"Root Region or some of its attributes are null.");
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
ServerName sn;
|
ServerName sn;
|
||||||
|
@ -428,7 +430,7 @@ public class HBaseFsck {
|
||||||
LOG.debug("Region " + descriptiveName + " offline, ignoring.");
|
LOG.debug("Region " + descriptiveName + " offline, ignoring.");
|
||||||
return;
|
return;
|
||||||
} else if (recentlyModified) {
|
} else if (recentlyModified) {
|
||||||
LOG.info("Region " + descriptiveName + " was recently modified -- skipping");
|
LOG.warn("Region " + descriptiveName + " was recently modified -- skipping");
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
// ========== Cases where the region is not in META =============
|
// ========== Cases where the region is not in META =============
|
||||||
|
@ -436,24 +438,29 @@ public class HBaseFsck {
|
||||||
// We shouldn't have record of this region at all then!
|
// We shouldn't have record of this region at all then!
|
||||||
assert false : "Entry for region with no data";
|
assert false : "Entry for region with no data";
|
||||||
} else if (!inMeta && !inHdfs && isDeployed) {
|
} else if (!inMeta && !inHdfs && isDeployed) {
|
||||||
errors.reportError("Region " + descriptiveName + ", key=" + key + ", not on HDFS or in META but " +
|
errors.reportError(ERROR_CODE.NOT_IN_META_HDFS, "Region "
|
||||||
"deployed on " + Joiner.on(", ").join(hbi.deployedOn));
|
+ descriptiveName + ", key=" + key + ", not on HDFS or in META but " +
|
||||||
|
"deployed on " + Joiner.on(", ").join(hbi.deployedOn));
|
||||||
} else if (!inMeta && inHdfs && !isDeployed) {
|
} else if (!inMeta && inHdfs && !isDeployed) {
|
||||||
errors.reportError("Region " + descriptiveName + " on HDFS, but not listed in META " +
|
errors.reportError(ERROR_CODE.NOT_IN_META_OR_DEPLOYED, "Region "
|
||||||
"or deployed on any region server.");
|
+ descriptiveName + " on HDFS, but not listed in META " +
|
||||||
|
"or deployed on any region server.");
|
||||||
} else if (!inMeta && inHdfs && isDeployed) {
|
} else if (!inMeta && inHdfs && isDeployed) {
|
||||||
errors.reportError("Region " + descriptiveName + " not in META, but deployed on " +
|
errors.reportError(ERROR_CODE.NOT_IN_META, "Region " + descriptiveName
|
||||||
Joiner.on(", ").join(hbi.deployedOn));
|
+ " not in META, but deployed on " + Joiner.on(", ").join(hbi.deployedOn));
|
||||||
|
|
||||||
// ========== Cases where the region is in META =============
|
// ========== Cases where the region is in META =============
|
||||||
} else if (inMeta && !inHdfs && !isDeployed) {
|
} else if (inMeta && !inHdfs && !isDeployed) {
|
||||||
errors.reportError("Region " + descriptiveName + " found in META, but not in HDFS " +
|
errors.reportError(ERROR_CODE.NOT_IN_HDFS_OR_DEPLOYED, "Region "
|
||||||
"or deployed on any region server.");
|
+ descriptiveName + " found in META, but not in HDFS "
|
||||||
|
+ "or deployed on any region server.");
|
||||||
} else if (inMeta && !inHdfs && isDeployed) {
|
} else if (inMeta && !inHdfs && isDeployed) {
|
||||||
errors.reportError("Region " + descriptiveName + " found in META, but not in HDFS, " +
|
errors.reportError(ERROR_CODE.NOT_IN_HDFS, "Region " + descriptiveName
|
||||||
"and deployed on " + Joiner.on(", ").join(hbi.deployedOn));
|
+ " found in META, but not in HDFS, " +
|
||||||
|
"and deployed on " + Joiner.on(", ").join(hbi.deployedOn));
|
||||||
} else if (inMeta && inHdfs && !isDeployed && shouldBeDeployed) {
|
} else if (inMeta && inHdfs && !isDeployed && shouldBeDeployed) {
|
||||||
errors.reportError("Region " + descriptiveName + " not deployed on any region server.");
|
errors.reportError(ERROR_CODE.NOT_DEPLOYED, "Region " + descriptiveName
|
||||||
|
+ " not deployed on any region server.");
|
||||||
// If we are trying to fix the errors
|
// If we are trying to fix the errors
|
||||||
if (shouldFix()) {
|
if (shouldFix()) {
|
||||||
errors.print("Trying to fix unassigned region...");
|
errors.print("Trying to fix unassigned region...");
|
||||||
|
@ -461,12 +468,14 @@ public class HBaseFsck {
|
||||||
HBaseFsckRepair.fixUnassigned(this.conf, hbi.metaEntry);
|
HBaseFsckRepair.fixUnassigned(this.conf, hbi.metaEntry);
|
||||||
}
|
}
|
||||||
} else if (inMeta && inHdfs && isDeployed && !shouldBeDeployed) {
|
} else if (inMeta && inHdfs && isDeployed && !shouldBeDeployed) {
|
||||||
errors.reportError("Region " + descriptiveName + " should not be deployed according " +
|
errors.reportError(ERROR_CODE.SHOULD_NOT_BE_DEPLOYED, "Region "
|
||||||
"to META, but is deployed on " + Joiner.on(", ").join(hbi.deployedOn));
|
+ descriptiveName + " should not be deployed according " +
|
||||||
|
"to META, but is deployed on " + Joiner.on(", ").join(hbi.deployedOn));
|
||||||
} else if (inMeta && inHdfs && isMultiplyDeployed) {
|
} else if (inMeta && inHdfs && isMultiplyDeployed) {
|
||||||
errors.reportError("Region " + descriptiveName + " is listed in META on region server " +
|
errors.reportError(ERROR_CODE.MULTI_DEPLOYED, "Region " + descriptiveName
|
||||||
hbi.metaEntry.regionServer + " but is multiply assigned to region servers " +
|
+ " is listed in META on region server " + hbi.metaEntry.regionServer
|
||||||
Joiner.on(", ").join(hbi.deployedOn));
|
+ " but is multiply assigned to region servers " +
|
||||||
|
Joiner.on(", ").join(hbi.deployedOn));
|
||||||
// If we are trying to fix the errors
|
// If we are trying to fix the errors
|
||||||
if (shouldFix()) {
|
if (shouldFix()) {
|
||||||
errors.print("Trying to fix assignment error...");
|
errors.print("Trying to fix assignment error...");
|
||||||
|
@ -474,9 +483,10 @@ public class HBaseFsck {
|
||||||
HBaseFsckRepair.fixDupeAssignment(this.conf, hbi.metaEntry, hbi.deployedOn);
|
HBaseFsckRepair.fixDupeAssignment(this.conf, hbi.metaEntry, hbi.deployedOn);
|
||||||
}
|
}
|
||||||
} else if (inMeta && inHdfs && isDeployed && !deploymentMatchesMeta) {
|
} else if (inMeta && inHdfs && isDeployed && !deploymentMatchesMeta) {
|
||||||
errors.reportError("Region " + descriptiveName + " listed in META on region server " +
|
errors.reportError(ERROR_CODE.SERVER_DOES_NOT_MATCH_META, "Region "
|
||||||
hbi.metaEntry.regionServer + " but found on region server " +
|
+ descriptiveName + " listed in META on region server " +
|
||||||
hbi.deployedOn.get(0));
|
hbi.metaEntry.regionServer + " but found on region server " +
|
||||||
|
hbi.deployedOn.get(0));
|
||||||
// If we are trying to fix the errors
|
// If we are trying to fix the errors
|
||||||
if (shouldFix()) {
|
if (shouldFix()) {
|
||||||
errors.print("Trying to fix assignment error...");
|
errors.print("Trying to fix assignment error...");
|
||||||
|
@ -484,13 +494,14 @@ public class HBaseFsck {
|
||||||
HBaseFsckRepair.fixDupeAssignment(this.conf, hbi.metaEntry, hbi.deployedOn);
|
HBaseFsckRepair.fixDupeAssignment(this.conf, hbi.metaEntry, hbi.deployedOn);
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
errors.reportError("Region " + descriptiveName + " is in an unforeseen state:" +
|
errors.reportError(ERROR_CODE.UNKNOWN, "Region " + descriptiveName +
|
||||||
" inMeta=" + inMeta +
|
" is in an unforeseen state:" +
|
||||||
" inHdfs=" + inHdfs +
|
" inMeta=" + inMeta +
|
||||||
" isDeployed=" + isDeployed +
|
" inHdfs=" + inHdfs +
|
||||||
" isMultiplyDeployed=" + isMultiplyDeployed +
|
" isDeployed=" + isDeployed +
|
||||||
" deploymentMatchesMeta=" + deploymentMatchesMeta +
|
" isMultiplyDeployed=" + isMultiplyDeployed +
|
||||||
" shouldBeDeployed=" + shouldBeDeployed);
|
" deploymentMatchesMeta=" + deploymentMatchesMeta +
|
||||||
|
" shouldBeDeployed=" + shouldBeDeployed);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -504,10 +515,15 @@ public class HBaseFsck {
|
||||||
// Check only valid, working regions
|
// Check only valid, working regions
|
||||||
if (hbi.metaEntry == null) continue;
|
if (hbi.metaEntry == null) continue;
|
||||||
if (hbi.metaEntry.regionServer == null) continue;
|
if (hbi.metaEntry.regionServer == null) continue;
|
||||||
if (hbi.foundRegionDir == null) continue;
|
|
||||||
if (hbi.deployedOn.size() != 1) continue;
|
|
||||||
if (hbi.onlyEdits) continue;
|
if (hbi.onlyEdits) continue;
|
||||||
|
|
||||||
|
// Missing regionDir or over-deployment is checked elsewhere. Include
|
||||||
|
// these cases in modTInfo, so we can evaluate those regions as part of
|
||||||
|
// the region chain in META
|
||||||
|
//if (hbi.foundRegionDir == null) continue;
|
||||||
|
//if (hbi.deployedOn.size() != 1) continue;
|
||||||
|
if (hbi.deployedOn.size() == 0) continue;
|
||||||
|
|
||||||
// We should be safe here
|
// We should be safe here
|
||||||
String tableName = hbi.metaEntry.getTableDesc().getNameAsString();
|
String tableName = hbi.metaEntry.getTableDesc().getNameAsString();
|
||||||
TInfo modTInfo = tablesInfo.get(tableName);
|
TInfo modTInfo = tablesInfo.get(tableName);
|
||||||
|
@ -517,13 +533,16 @@ public class HBaseFsck {
|
||||||
for (ServerName server : hbi.deployedOn) {
|
for (ServerName server : hbi.deployedOn) {
|
||||||
modTInfo.addServer(server);
|
modTInfo.addServer(server);
|
||||||
}
|
}
|
||||||
modTInfo.addEdge(hbi.metaEntry.getStartKey(), hbi.metaEntry.getEndKey());
|
|
||||||
|
//modTInfo.addEdge(hbi.metaEntry.getStartKey(), hbi.metaEntry.getEndKey());
|
||||||
|
modTInfo.addRegionInfo(hbi);
|
||||||
|
|
||||||
tablesInfo.put(tableName, modTInfo);
|
tablesInfo.put(tableName, modTInfo);
|
||||||
}
|
}
|
||||||
|
|
||||||
for (TInfo tInfo : tablesInfo.values()) {
|
for (TInfo tInfo : tablesInfo.values()) {
|
||||||
if (!tInfo.check()) {
|
if (!tInfo.checkRegionChain()) {
|
||||||
errors.reportError("Found inconsistency in table " + tInfo.getName());
|
errors.report("Found inconsistency in table " + tInfo.getName());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -533,17 +552,16 @@ public class HBaseFsck {
|
||||||
*/
|
*/
|
||||||
private class TInfo {
|
private class TInfo {
|
||||||
String tableName;
|
String tableName;
|
||||||
TreeMap <byte[], byte[]> edges;
|
|
||||||
TreeSet <ServerName> deployedOn;
|
TreeSet <ServerName> deployedOn;
|
||||||
|
List<HbckInfo> regions = new ArrayList<HbckInfo>();
|
||||||
|
|
||||||
TInfo(String name) {
|
TInfo(String name) {
|
||||||
this.tableName = name;
|
this.tableName = name;
|
||||||
edges = new TreeMap <byte[], byte[]> (Bytes.BYTES_COMPARATOR);
|
|
||||||
deployedOn = new TreeSet <ServerName>();
|
deployedOn = new TreeSet <ServerName>();
|
||||||
}
|
}
|
||||||
|
|
||||||
public void addEdge(byte[] fromNode, byte[] toNode) {
|
public void addRegionInfo (HbckInfo r) {
|
||||||
this.edges.put(fromNode, toNode);
|
regions.add(r);
|
||||||
}
|
}
|
||||||
|
|
||||||
public void addServer(ServerName server) {
|
public void addServer(ServerName server) {
|
||||||
|
@ -555,46 +573,77 @@ public class HBaseFsck {
|
||||||
}
|
}
|
||||||
|
|
||||||
public int getNumRegions() {
|
public int getNumRegions() {
|
||||||
return edges.size();
|
return regions.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
public boolean check() {
|
/**
|
||||||
byte[] last = new byte[0];
|
* Check the region chain (from META) of this table. We are looking for
|
||||||
byte[] next = new byte[0];
|
* holes, overlaps, and cycles.
|
||||||
TreeSet <byte[]> visited = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR);
|
* @return false if there are errors
|
||||||
// Each table should start with a zero-length byte[] and end at a
|
*/
|
||||||
// zero-length byte[]. Just follow the edges to see if this is true
|
public boolean checkRegionChain() {
|
||||||
while (true) {
|
Collections.sort(regions);
|
||||||
// Check if chain is broken
|
HbckInfo last = null;
|
||||||
if (!edges.containsKey(last)) {
|
|
||||||
errors.detail("Chain of regions in table " + tableName +
|
for (HbckInfo r : regions) {
|
||||||
" is broken; edges does not contain " + Bytes.toString(last));
|
if (last == null) {
|
||||||
return false;
|
// This is the first region, check that the start key is empty
|
||||||
}
|
if (! Bytes.equals(r.metaEntry.getStartKey(), HConstants.EMPTY_BYTE_ARRAY)) {
|
||||||
next = edges.get(last);
|
errors.reportError(ERROR_CODE.FIRST_REGION_STARTKEY_NOT_EMPTY,
|
||||||
// Found a cycle
|
"First region should start with an empty key.",
|
||||||
if (visited.contains(next)) {
|
this, r);
|
||||||
errors.detail("Chain of regions in table " + tableName +
|
|
||||||
" has a cycle around " + Bytes.toString(next));
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
// Mark next node as visited
|
|
||||||
visited.add(next);
|
|
||||||
// If next is zero-length byte[] we are possibly at the end of the chain
|
|
||||||
if (next.length == 0) {
|
|
||||||
// If we have visited all elements we are fine
|
|
||||||
if (edges.size() != visited.size()) {
|
|
||||||
errors.detail("Chain of regions in table " + tableName +
|
|
||||||
" contains less elements than are listed in META; visited=" + visited.size() +
|
|
||||||
", edges=" + edges.size());
|
|
||||||
return false;
|
|
||||||
}
|
}
|
||||||
return true;
|
} else {
|
||||||
|
|
||||||
|
// Check if endKey < startKey
|
||||||
|
// Previous implementation of this code checked for a cycle in the
|
||||||
|
// region chain. A cycle would imply that the endKey comes before
|
||||||
|
// the startKey (i.e. endKey < startKey).
|
||||||
|
if (! Bytes.equals(r.metaEntry.getEndKey(), HConstants.EMPTY_BYTE_ARRAY)) {
|
||||||
|
// continue with this check if this is not the last region
|
||||||
|
int cmpRegionKeys = Bytes.compareTo(r.metaEntry.getStartKey(),
|
||||||
|
r.metaEntry.getEndKey());
|
||||||
|
if (cmpRegionKeys > 0) {
|
||||||
|
errors.reportError(ERROR_CODE.REGION_CYCLE,
|
||||||
|
String.format("The endkey for this region comes before the "
|
||||||
|
+ "startkey, startkey=%s, endkey=%s",
|
||||||
|
Bytes.toString(r.metaEntry.getStartKey()),
|
||||||
|
Bytes.toString(r.metaEntry.getEndKey())),
|
||||||
|
this, r, last);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if the startkeys are different
|
||||||
|
if (Bytes.equals(r.metaEntry.getStartKey(), last.metaEntry.getStartKey())) {
|
||||||
|
errors.reportError(ERROR_CODE.DUPE_STARTKEYS,
|
||||||
|
"Two regions have the same startkey: "
|
||||||
|
+ Bytes.toString(r.metaEntry.getStartKey()),
|
||||||
|
this, r, last);
|
||||||
|
} else {
|
||||||
|
// Check that the startkey is the same as the previous end key
|
||||||
|
int cmp = Bytes.compareTo(r.metaEntry.getStartKey(),
|
||||||
|
last.metaEntry.getEndKey());
|
||||||
|
if (cmp > 0) {
|
||||||
|
// hole
|
||||||
|
errors.reportError(ERROR_CODE.HOLE_IN_REGION_CHAIN,
|
||||||
|
"There is a hole in the region chain.",
|
||||||
|
this, r, last);
|
||||||
|
} else if (cmp < 0) {
|
||||||
|
// overlap
|
||||||
|
errors.reportError(ERROR_CODE.OVERLAP_IN_REGION_CHAIN,
|
||||||
|
"There is an overlap in the region chain.",
|
||||||
|
this, r, last);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
last = next;
|
|
||||||
|
last = r;
|
||||||
}
|
}
|
||||||
// How did we get here?
|
|
||||||
|
return errors.getErrorList().size() == 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -603,7 +652,6 @@ public class HBaseFsck {
|
||||||
* if any of the REGIONINFO_QUALIFIER, SERVER_QUALIFIER, STARTCODE_QUALIFIER,
|
* if any of the REGIONINFO_QUALIFIER, SERVER_QUALIFIER, STARTCODE_QUALIFIER,
|
||||||
* SPLITA_QUALIFIER, SPLITB_QUALIFIER have not changed in the last
|
* SPLITA_QUALIFIER, SPLITB_QUALIFIER have not changed in the last
|
||||||
* milliseconds specified by timelag, then the table is a candidate to be returned.
|
* milliseconds specified by timelag, then the table is a candidate to be returned.
|
||||||
* @param regionList - all entries found in .META
|
|
||||||
* @return tables that have not been modified recently
|
* @return tables that have not been modified recently
|
||||||
* @throws IOException if an error is encountered
|
* @throws IOException if an error is encountered
|
||||||
*/
|
*/
|
||||||
|
@ -668,7 +716,7 @@ public class HBaseFsck {
|
||||||
|
|
||||||
// If there is no region holding .META.
|
// If there is no region holding .META.
|
||||||
if (metaRegions.size() == 0) {
|
if (metaRegions.size() == 0) {
|
||||||
errors.reportError(".META. is not found on any region.");
|
errors.reportError(ERROR_CODE.NO_META_REGION, ".META. is not found on any region.");
|
||||||
if (shouldFix()) {
|
if (shouldFix()) {
|
||||||
errors.print("Trying to fix a problem with .META...");
|
errors.print("Trying to fix a problem with .META...");
|
||||||
setShouldRerun();
|
setShouldRerun();
|
||||||
|
@ -678,7 +726,7 @@ public class HBaseFsck {
|
||||||
}
|
}
|
||||||
// If there are more than one regions pretending to hold the .META.
|
// If there are more than one regions pretending to hold the .META.
|
||||||
else if (metaRegions.size() > 1) {
|
else if (metaRegions.size() > 1) {
|
||||||
errors.reportError(".META. is found on more than one region.");
|
errors.reportError(ERROR_CODE.MULTI_META_REGION, ".META. is found on more than one region.");
|
||||||
if (shouldFix()) {
|
if (shouldFix()) {
|
||||||
errors.print("Trying to fix a problem with .META...");
|
errors.print("Trying to fix a problem with .META...");
|
||||||
setShouldRerun();
|
setShouldRerun();
|
||||||
|
@ -773,7 +821,7 @@ public class HBaseFsck {
|
||||||
/**
|
/**
|
||||||
* Maintain information about a particular region.
|
* Maintain information about a particular region.
|
||||||
*/
|
*/
|
||||||
static class HbckInfo {
|
static class HbckInfo implements Comparable {
|
||||||
boolean onlyEdits = false;
|
boolean onlyEdits = false;
|
||||||
MetaEntry metaEntry = null;
|
MetaEntry metaEntry = null;
|
||||||
FileStatus foundRegionDir = null;
|
FileStatus foundRegionDir = null;
|
||||||
|
@ -796,6 +844,16 @@ public class HBaseFsck {
|
||||||
return "UNKNOWN_REGION on " + Joiner.on(", ").join(deployedOn);
|
return "UNKNOWN_REGION on " + Joiner.on(", ").join(deployedOn);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int compareTo(Object o) {
|
||||||
|
HbckInfo other = (HbckInfo) o;
|
||||||
|
int startComparison = Bytes.compareTo(this.metaEntry.getStartKey(), other.metaEntry.getStartKey());
|
||||||
|
if (startComparison != 0)
|
||||||
|
return startComparison;
|
||||||
|
else
|
||||||
|
return Bytes.compareTo(this.metaEntry.getEndKey(), other.metaEntry.getEndKey());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -804,10 +862,10 @@ public class HBaseFsck {
|
||||||
private void printTableSummary() {
|
private void printTableSummary() {
|
||||||
System.out.println("Summary:");
|
System.out.println("Summary:");
|
||||||
for (TInfo tInfo : tablesInfo.values()) {
|
for (TInfo tInfo : tablesInfo.values()) {
|
||||||
if (tInfo.check()) {
|
if (errors.tableHasErrors(tInfo)) {
|
||||||
System.out.println(" " + tInfo.getName() + " is okay.");
|
|
||||||
} else {
|
|
||||||
System.out.println("Table " + tInfo.getName() + " is inconsistent.");
|
System.out.println("Table " + tInfo.getName() + " is inconsistent.");
|
||||||
|
} else {
|
||||||
|
System.out.println(" " + tInfo.getName() + " is okay.");
|
||||||
}
|
}
|
||||||
System.out.println(" Number of regions: " + tInfo.getNumRegions());
|
System.out.println(" Number of regions: " + tInfo.getNumRegions());
|
||||||
System.out.print(" Deployed on: ");
|
System.out.print(" Deployed on: ");
|
||||||
|
@ -819,19 +877,45 @@ public class HBaseFsck {
|
||||||
}
|
}
|
||||||
|
|
||||||
interface ErrorReporter {
|
interface ErrorReporter {
|
||||||
|
public static enum ERROR_CODE {
|
||||||
|
UNKNOWN, NO_META_REGION, NULL_ROOT_REGION, NO_VERSION_FILE, NOT_IN_META_HDFS, NOT_IN_META,
|
||||||
|
NOT_IN_META_OR_DEPLOYED, NOT_IN_HDFS_OR_DEPLOYED, NOT_IN_HDFS, SERVER_DOES_NOT_MATCH_META, NOT_DEPLOYED,
|
||||||
|
MULTI_DEPLOYED, SHOULD_NOT_BE_DEPLOYED, MULTI_META_REGION, RS_CONNECT_FAILURE,
|
||||||
|
FIRST_REGION_STARTKEY_NOT_EMPTY, DUPE_STARTKEYS,
|
||||||
|
HOLE_IN_REGION_CHAIN, OVERLAP_IN_REGION_CHAIN, REGION_CYCLE
|
||||||
|
}
|
||||||
|
public void clear();
|
||||||
|
public void report(String message);
|
||||||
public void reportError(String message);
|
public void reportError(String message);
|
||||||
|
public void reportError(ERROR_CODE errorCode, String message);
|
||||||
|
public void reportError(ERROR_CODE errorCode, String message, TInfo table, HbckInfo info);
|
||||||
|
public void reportError(ERROR_CODE errorCode, String message, TInfo table, HbckInfo info1, HbckInfo info2);
|
||||||
public int summarize();
|
public int summarize();
|
||||||
public void detail(String details);
|
public void detail(String details);
|
||||||
|
public ArrayList<ERROR_CODE> getErrorList();
|
||||||
public void progress();
|
public void progress();
|
||||||
public void print(String message);
|
public void print(String message);
|
||||||
public void resetErrors();
|
public void resetErrors();
|
||||||
|
public boolean tableHasErrors(TInfo table);
|
||||||
}
|
}
|
||||||
|
|
||||||
private static class PrintingErrorReporter implements ErrorReporter {
|
private static class PrintingErrorReporter implements ErrorReporter {
|
||||||
public int errorCount = 0;
|
public int errorCount = 0;
|
||||||
private int showProgress;
|
private int showProgress;
|
||||||
|
|
||||||
public synchronized void reportError(String message) {
|
Set<TInfo> errorTables = new HashSet<TInfo>();
|
||||||
|
|
||||||
|
// for use by unit tests to verify which errors were discovered
|
||||||
|
private ArrayList<ERROR_CODE> errorList = new ArrayList<ERROR_CODE>();
|
||||||
|
|
||||||
|
public void clear() {
|
||||||
|
errorTables.clear();
|
||||||
|
errorList.clear();
|
||||||
|
errorCount = 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void reportError(ERROR_CODE errorCode, String message) {
|
||||||
|
errorList.add(errorCode);
|
||||||
if (!summary) {
|
if (!summary) {
|
||||||
System.out.println("ERROR: " + message);
|
System.out.println("ERROR: " + message);
|
||||||
}
|
}
|
||||||
|
@ -839,6 +923,37 @@ public class HBaseFsck {
|
||||||
showProgress = 0;
|
showProgress = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public synchronized void reportError(ERROR_CODE errorCode, String message, TInfo table,
|
||||||
|
HbckInfo info) {
|
||||||
|
errorTables.add(table);
|
||||||
|
String reference = "(region " + info.metaEntry.getRegionNameAsString() + ")";
|
||||||
|
reportError(errorCode, reference + " " + message);
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void reportError(ERROR_CODE errorCode, String message, TInfo table,
|
||||||
|
HbckInfo info1, HbckInfo info2) {
|
||||||
|
errorTables.add(table);
|
||||||
|
String reference = "(regions " + info1.metaEntry.getRegionNameAsString()
|
||||||
|
+ " and " + info2.metaEntry.getRegionNameAsString() + ")";
|
||||||
|
reportError(errorCode, reference + " " + message);
|
||||||
|
}
|
||||||
|
|
||||||
|
public synchronized void reportError(String message) {
|
||||||
|
reportError(ERROR_CODE.UNKNOWN, message);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Report error information, but do not increment the error count. Intended for cases
|
||||||
|
* where the actual error would have been reported previously.
|
||||||
|
* @param message
|
||||||
|
*/
|
||||||
|
public synchronized void report(String message) {
|
||||||
|
if (! summary) {
|
||||||
|
System.out.println("ERROR: " + message);
|
||||||
|
}
|
||||||
|
showProgress = 0;
|
||||||
|
}
|
||||||
|
|
||||||
public synchronized int summarize() {
|
public synchronized int summarize() {
|
||||||
System.out.println(Integer.toString(errorCount) +
|
System.out.println(Integer.toString(errorCount) +
|
||||||
" inconsistencies detected.");
|
" inconsistencies detected.");
|
||||||
|
@ -851,12 +966,21 @@ public class HBaseFsck {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public ArrayList<ERROR_CODE> getErrorList() {
|
||||||
|
return errorList;
|
||||||
|
}
|
||||||
|
|
||||||
public synchronized void print(String message) {
|
public synchronized void print(String message) {
|
||||||
if (!summary) {
|
if (!summary) {
|
||||||
System.out.println(message);
|
System.out.println(message);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean tableHasErrors(TInfo table) {
|
||||||
|
return errorTables.contains(table);
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void resetErrors() {
|
public void resetErrors() {
|
||||||
errorCount = 0;
|
errorCount = 0;
|
||||||
|
@ -923,14 +1047,14 @@ public class HBaseFsck {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// check to see if the existance of this region matches the region in META
|
// check to see if the existence of this region matches the region in META
|
||||||
for (HRegionInfo r:regions) {
|
for (HRegionInfo r:regions) {
|
||||||
HbckInfo hbi = hbck.getOrCreateInfo(r.getEncodedName());
|
HbckInfo hbi = hbck.getOrCreateInfo(r.getEncodedName());
|
||||||
hbi.addServer(rsinfo);
|
hbi.addServer(rsinfo);
|
||||||
}
|
}
|
||||||
} catch (IOException e) { // unable to connect to the region server.
|
} catch (IOException e) { // unable to connect to the region server.
|
||||||
errors.reportError("RegionServer: " + rsinfo.getServerName() +
|
errors.reportError(ERROR_CODE.RS_CONNECT_FAILURE, "RegionServer: " + rsinfo.getServerName() +
|
||||||
" Unable to fetch region information. " + e);
|
" Unable to fetch region information. " + e);
|
||||||
} finally {
|
} finally {
|
||||||
done = true;
|
done = true;
|
||||||
notifyAll(); // wakeup anybody waiting for this item to be done
|
notifyAll(); // wakeup anybody waiting for this item to be done
|
||||||
|
@ -1000,7 +1124,7 @@ public class HBaseFsck {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} catch (IOException e) { // unable to connect to the region server.
|
} catch (IOException e) { // unable to connect to the region server.
|
||||||
errors.reportError("Table Directory: " + tableDir.getPath().getName() +
|
errors.reportError(ERROR_CODE.RS_CONNECT_FAILURE, "Table Directory: " + tableDir.getPath().getName() +
|
||||||
" Unable to fetch region information. " + e);
|
" Unable to fetch region information. " + e);
|
||||||
} finally {
|
} finally {
|
||||||
done = true;
|
done = true;
|
||||||
|
|
|
@ -89,6 +89,7 @@ public class HBaseTestingUtility {
|
||||||
*/
|
*/
|
||||||
private boolean passedZkCluster = false;
|
private boolean passedZkCluster = false;
|
||||||
private MiniDFSCluster dfsCluster = null;
|
private MiniDFSCluster dfsCluster = null;
|
||||||
|
|
||||||
private MiniHBaseCluster hbaseCluster = null;
|
private MiniHBaseCluster hbaseCluster = null;
|
||||||
private MiniMRCluster mrCluster = null;
|
private MiniMRCluster mrCluster = null;
|
||||||
// If non-null, then already a cluster running.
|
// If non-null, then already a cluster running.
|
||||||
|
@ -113,6 +114,10 @@ public class HBaseTestingUtility {
|
||||||
this.conf = conf;
|
this.conf = conf;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public MiniHBaseCluster getHbaseCluster() {
|
||||||
|
return hbaseCluster;
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Returns this classes's instance of {@link Configuration}. Be careful how
|
* Returns this classes's instance of {@link Configuration}. Be careful how
|
||||||
* you use the returned Configuration since {@link HConnection} instances
|
* you use the returned Configuration since {@link HConnection} instances
|
||||||
|
|
|
@ -19,23 +19,28 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.util;
|
package org.apache.hadoop.hbase.util;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.hbase.util.HBaseFsck.ErrorReporter.ERROR_CODE;
|
||||||
|
import static org.junit.Assert.assertArrayEquals;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.ServerName;
|
import org.apache.hadoop.hbase.*;
|
||||||
import org.apache.hadoop.hbase.client.HTable;
|
import org.apache.hadoop.hbase.client.*;
|
||||||
import org.apache.hadoop.hbase.client.Put;
|
import org.apache.hadoop.hbase.ipc.HRegionInterface;
|
||||||
import org.apache.hadoop.hbase.client.Result;
|
|
||||||
import org.apache.hadoop.hbase.client.ResultScanner;
|
|
||||||
import org.apache.hadoop.hbase.client.Scan;
|
|
||||||
import org.junit.BeforeClass;
|
import org.junit.BeforeClass;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.Arrays;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Random;
|
||||||
|
|
||||||
public class TestHBaseFsck {
|
public class TestHBaseFsck {
|
||||||
|
|
||||||
final Log LOG = LogFactory.getLog(getClass());
|
final Log LOG = LogFactory.getLog(getClass());
|
||||||
|
@ -50,32 +55,39 @@ public class TestHBaseFsck {
|
||||||
TEST_UTIL.startMiniCluster(3);
|
TEST_UTIL.startMiniCluster(3);
|
||||||
}
|
}
|
||||||
|
|
||||||
private int doFsck(boolean fix) throws Exception {
|
private List doFsck(boolean fix) throws Exception {
|
||||||
HBaseFsck fsck = new HBaseFsck(conf);
|
HBaseFsck fsck = new HBaseFsck(conf);
|
||||||
fsck.displayFullReport();
|
fsck.displayFullReport(); // i.e. -details
|
||||||
fsck.setTimeLag(0);
|
fsck.setTimeLag(0);
|
||||||
fsck.setFixErrors(fix);
|
fsck.setFixErrors(fix);
|
||||||
// Most basic check ever, 0 tables
|
fsck.doWork();
|
||||||
return fsck.doWork();
|
return fsck.getErrors().getErrorList();
|
||||||
|
}
|
||||||
|
|
||||||
|
private void assertNoErrors(List errs) throws Exception {
|
||||||
|
assertEquals(0, errs.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
private void assertErrors(List errs, ERROR_CODE[] expectedErrors) {
|
||||||
|
assertEquals(Arrays.asList(expectedErrors), errs);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testHBaseFsck() throws Exception {
|
public void testHBaseFsck() throws Exception {
|
||||||
int result = doFsck(false);
|
assertNoErrors(doFsck(false));
|
||||||
assertEquals(0, result);
|
|
||||||
|
|
||||||
TEST_UTIL.createTable(TABLE, FAM);
|
TEST_UTIL.createTable(TABLE, FAM);
|
||||||
|
|
||||||
// We created 1 table, should be fine
|
// We created 1 table, should be fine
|
||||||
result = doFsck(false);
|
assertNoErrors(doFsck(false));
|
||||||
assertEquals(0, result);
|
|
||||||
|
|
||||||
// Now let's mess it up and change the assignment in .META. to
|
// Now let's mess it up and change the assignment in .META. to
|
||||||
// point to a different region server
|
// point to a different region server
|
||||||
HTable meta = new HTable(conf, HTableDescriptor.META_TABLEDESC.getName());
|
HTable meta = new HTable(conf, HTableDescriptor.META_TABLEDESC.getName());
|
||||||
ResultScanner scanner = meta.getScanner(new Scan());
|
ResultScanner scanner = meta.getScanner(new Scan());
|
||||||
|
|
||||||
resforloop : for (Result res : scanner) {
|
resforloop:
|
||||||
|
for (Result res : scanner) {
|
||||||
long startCode = Bytes.toLong(res.getValue(HConstants.CATALOG_FAMILY,
|
long startCode = Bytes.toLong(res.getValue(HConstants.CATALOG_FAMILY,
|
||||||
HConstants.STARTCODE_QUALIFIER));
|
HConstants.STARTCODE_QUALIFIER));
|
||||||
|
|
||||||
|
@ -98,13 +110,90 @@ public class TestHBaseFsck {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Try to fix the data
|
// Try to fix the data
|
||||||
result = doFsck(true);
|
assertErrors(doFsck(true), new ERROR_CODE[]{
|
||||||
assertEquals(-1, result);
|
ERROR_CODE.SERVER_DOES_NOT_MATCH_META});
|
||||||
|
|
||||||
Thread.sleep(15000);
|
Thread.sleep(15000);
|
||||||
result = doFsck(false);
|
|
||||||
// Should have fixed
|
// Should be fixed now
|
||||||
assertEquals(0, result);
|
assertNoErrors(doFsck(false));
|
||||||
|
|
||||||
|
// comment needed - what is the purpose of this line
|
||||||
new HTable(conf, TABLE).getScanner(new Scan());
|
new HTable(conf, TABLE).getScanner(new Scan());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private HRegionInfo createRegion(Configuration conf, final HTableDescriptor
|
||||||
|
htd, byte[] startKey, byte[] endKey)
|
||||||
|
throws IOException {
|
||||||
|
HTable meta = new HTable(conf, HConstants.META_TABLE_NAME);
|
||||||
|
HRegionInfo hri = new HRegionInfo(htd, startKey, endKey);
|
||||||
|
Put put = new Put(hri.getRegionName());
|
||||||
|
put.add(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER,
|
||||||
|
Writables.getBytes(hri));
|
||||||
|
meta.put(put);
|
||||||
|
return hri;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
/**
|
||||||
|
* Tests for inconsistencies in the META data (duplicate start keys, or holes)
|
||||||
|
*/
|
||||||
|
public void testHBaseFsckMeta() throws Exception {
|
||||||
|
assertNoErrors(doFsck(false));
|
||||||
|
|
||||||
|
HTable tbl = TEST_UTIL.createTable(Bytes.toBytes("table2"), FAM);
|
||||||
|
|
||||||
|
HRegionInfo hriOrig = tbl.getRegionsInfo().keySet().iterator().next();
|
||||||
|
HServerAddress rsAddressOrig = tbl.getRegionsInfo().get(hriOrig);
|
||||||
|
|
||||||
|
byte[][] startKeys = new byte[][]{
|
||||||
|
HConstants.EMPTY_BYTE_ARRAY,
|
||||||
|
Bytes.toBytes("A"),
|
||||||
|
Bytes.toBytes("B"),
|
||||||
|
Bytes.toBytes("C")
|
||||||
|
};
|
||||||
|
TEST_UTIL.createMultiRegions(conf, tbl, FAM, startKeys);
|
||||||
|
Path rootDir = new Path(conf.get(HConstants.HBASE_DIR));
|
||||||
|
FileSystem fs = rootDir.getFileSystem(conf);
|
||||||
|
Path p = new Path(rootDir + "/table2", hriOrig.getEncodedName());
|
||||||
|
fs.delete(p, true);
|
||||||
|
|
||||||
|
Thread.sleep(1 * 1000);
|
||||||
|
ArrayList servers = new ArrayList();
|
||||||
|
servers.add(rsAddressOrig);
|
||||||
|
HBaseFsckRepair.fixDupeAssignment(conf, hriOrig, servers);
|
||||||
|
|
||||||
|
// We created 1 table, should be fine
|
||||||
|
assertNoErrors(doFsck(false));
|
||||||
|
|
||||||
|
// Now let's mess it up, by adding a region with a duplicate startkey
|
||||||
|
HRegionInfo hriDupe = createRegion(conf, tbl.getTableDescriptor(),
|
||||||
|
Bytes.toBytes("A"), Bytes.toBytes("A2"));
|
||||||
|
TEST_UTIL.getHBaseCluster().getMaster().assignRegion(hriDupe);
|
||||||
|
TEST_UTIL.getHBaseCluster().getMaster().getAssignmentManager()
|
||||||
|
.waitForAssignment(hriDupe);
|
||||||
|
assertErrors(doFsck(false), new ERROR_CODE[]{ERROR_CODE.DUPE_STARTKEYS});
|
||||||
|
|
||||||
|
// Mess it up by creating an overlap in the metadata
|
||||||
|
HRegionInfo hriOverlap = createRegion(conf, tbl.getTableDescriptor(),
|
||||||
|
Bytes.toBytes("A2"), Bytes.toBytes("B2"));
|
||||||
|
TEST_UTIL.getHBaseCluster().getMaster().assignRegion(hriOverlap);
|
||||||
|
TEST_UTIL.getHBaseCluster().getMaster().getAssignmentManager()
|
||||||
|
.waitForAssignment(hriOverlap);
|
||||||
|
assertErrors(doFsck(false), new ERROR_CODE[]{
|
||||||
|
ERROR_CODE.DUPE_STARTKEYS, ERROR_CODE.OVERLAP_IN_REGION_CHAIN,
|
||||||
|
ERROR_CODE.OVERLAP_IN_REGION_CHAIN});
|
||||||
|
|
||||||
|
// Mess it up by leaving a hole in the meta data
|
||||||
|
HRegionInfo hriHole = createRegion(conf, tbl.getTableDescriptor(),
|
||||||
|
Bytes.toBytes("D"), Bytes.toBytes("E"));
|
||||||
|
TEST_UTIL.getHBaseCluster().getMaster().assignRegion(hriHole);
|
||||||
|
TEST_UTIL.getHBaseCluster().getMaster().getAssignmentManager()
|
||||||
|
.waitForAssignment(hriHole);
|
||||||
|
// assertError(doFsck(false), ERROR_CODE.OVERLAP_IN_REGION_CHAIN);
|
||||||
|
assertErrors(doFsck(false), new ERROR_CODE[]{ ERROR_CODE.DUPE_STARTKEYS,
|
||||||
|
ERROR_CODE.OVERLAP_IN_REGION_CHAIN, ERROR_CODE.OVERLAP_IN_REGION_CHAIN,
|
||||||
|
ERROR_CODE.HOLE_IN_REGION_CHAIN });
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue