HBASE-12122 Try not to assign user regions to master all the time

This commit is contained in:
Jimmy Xiang 2014-09-30 10:59:15 -07:00
parent 546f436a41
commit a463aef8bc
15 changed files with 185 additions and 214 deletions

View File

@ -262,6 +262,7 @@ public final class ProtobufUtil {
* @return True if passed <code>bytes</code> has {@link #PB_MAGIC} for a prefix. * @return True if passed <code>bytes</code> has {@link #PB_MAGIC} for a prefix.
*/ */
public static boolean isPBMagicPrefix(final byte [] bytes) { public static boolean isPBMagicPrefix(final byte [] bytes) {
if (bytes == null) return false;
return isPBMagicPrefix(bytes, 0, bytes.length); return isPBMagicPrefix(bytes, 0, bytes.length);
} }

View File

@ -1699,15 +1699,15 @@ public class AssignmentManager {
ServerName serverName = regionState.getServerName(); ServerName serverName = regionState.getServerName();
ReentrantLock lock = locker.acquireLock(hri.getEncodedName()); ReentrantLock lock = locker.acquireLock(hri.getEncodedName());
try { try {
if (!regionState.equals(regionStates.getRegionState(hri))) {
return; // Region is not in the expected state any more
}
for (int i = 1; i <= maximumAttempts; i++) { for (int i = 1; i <= maximumAttempts; i++) {
if (!serverManager.isServerOnline(serverName) if (!serverManager.isServerOnline(serverName)
|| server.isStopped() || server.isAborted()) { || server.isStopped() || server.isAborted()) {
return; // No need any more return; // No need any more
} }
try { try {
if (!regionState.equals(regionStates.getRegionState(hri))) {
return; // Region is not in the expected state any more
}
List<ServerName> favoredNodes = ServerName.EMPTY_SERVER_LIST; List<ServerName> favoredNodes = ServerName.EMPTY_SERVER_LIST;
if (shouldAssignRegionsWithFavoredNodes) { if (shouldAssignRegionsWithFavoredNodes) {
favoredNodes = ((FavoredNodeLoadBalancer)balancer).getFavoredNodes(hri); favoredNodes = ((FavoredNodeLoadBalancer)balancer).getFavoredNodes(hri);
@ -1771,15 +1771,15 @@ public class AssignmentManager {
ServerName serverName = regionState.getServerName(); ServerName serverName = regionState.getServerName();
ReentrantLock lock = locker.acquireLock(hri.getEncodedName()); ReentrantLock lock = locker.acquireLock(hri.getEncodedName());
try { try {
if (!regionState.equals(regionStates.getRegionState(hri))) {
return; // Region is not in the expected state any more
}
for (int i = 1; i <= maximumAttempts; i++) { for (int i = 1; i <= maximumAttempts; i++) {
if (!serverManager.isServerOnline(serverName) if (!serverManager.isServerOnline(serverName)
|| server.isStopped() || server.isAborted()) { || server.isStopped() || server.isAborted()) {
return; // No need any more return; // No need any more
} }
try { try {
if (!regionState.equals(regionStates.getRegionState(hri))) {
return; // Region is not in the expected state any more
}
serverManager.sendRegionClose(serverName, hri, null); serverManager.sendRegionClose(serverName, hri, null);
return; // Done. return; // Done.
} catch (Throwable t) { } catch (Throwable t) {

View File

@ -1270,7 +1270,9 @@ public class MasterRpcServices extends RSRpcServices
RegionStateTransition rt = req.getTransition(0); RegionStateTransition rt = req.getTransition(0);
TableName tableName = ProtobufUtil.toTableName( TableName tableName = ProtobufUtil.toTableName(
rt.getRegionInfo(0).getTableName()); rt.getRegionInfo(0).getTableName());
if (!TableName.META_TABLE_NAME.equals(tableName) RegionStates regionStates = master.assignmentManager.getRegionStates();
if (!(TableName.META_TABLE_NAME.equals(tableName)
&& regionStates.getRegionState(HRegionInfo.FIRST_META_REGIONINFO) != null)
&& !master.assignmentManager.isFailoverCleanupDone()) { && !master.assignmentManager.isFailoverCleanupDone()) {
// Meta region is assigned before master finishes the // Meta region is assigned before master finishes the
// failover cleanup. So no need this check for it // failover cleanup. So no need this check for it

View File

@ -444,7 +444,7 @@ public class ServerManager {
} }
/** @return the count of active regionservers */ /** @return the count of active regionservers */
private int countOfRegionServers() { public int countOfRegionServers() {
// Presumes onlineServers is a concurrent map // Presumes onlineServers is a concurrent map
return this.onlineServers.size(); return this.onlineServers.size();
} }

View File

@ -91,8 +91,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
* topology in terms of server names, hostnames and racks. * topology in terms of server names, hostnames and racks.
*/ */
protected static class Cluster { protected static class Cluster {
ServerName masterServerName;
Set<String> tablesOnMaster;
ServerName[] servers; ServerName[] servers;
String[] hosts; // ServerName uniquely identifies a region server. multiple RS can run on the same host String[] hosts; // ServerName uniquely identifies a region server. multiple RS can run on the same host
String[] racks; String[] racks;
@ -101,7 +99,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
ArrayList<String> tables; ArrayList<String> tables;
HRegionInfo[] regions; HRegionInfo[] regions;
Deque<RegionLoad>[] regionLoads; Deque<RegionLoad>[] regionLoads;
int activeMasterIndex = -1;
int[][] regionLocations; //regionIndex -> list of serverIndex sorted by locality int[][] regionLocations; //regionIndex -> list of serverIndex sorted by locality
@ -122,7 +119,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
int[] regionIndexToTableIndex; //regionIndex -> tableIndex int[] regionIndexToTableIndex; //regionIndex -> tableIndex
int[][] numRegionsPerServerPerTable; //serverIndex -> tableIndex -> # regions int[][] numRegionsPerServerPerTable; //serverIndex -> tableIndex -> # regions
int[] numMaxRegionsPerTable; //tableIndex -> max number of regions in a single RS int[] numMaxRegionsPerTable; //tableIndex -> max number of regions in a single RS
int numUserRegionsOnMaster; //number of user regions on the active master
int[] regionIndexToPrimaryIndex; //regionIndex -> regionIndex of the primary int[] regionIndexToPrimaryIndex; //regionIndex -> regionIndex of the primary
boolean hasRegionReplicas = false; //whether there is regions with replicas boolean hasRegionReplicas = false; //whether there is regions with replicas
@ -142,40 +138,32 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
int numMovedRegions = 0; //num moved regions from the initial configuration int numMovedRegions = 0; //num moved regions from the initial configuration
// num of moved regions away from master that should be on the master // num of moved regions away from master that should be on the master
int numMovedMasterHostedRegions = 0;
int numMovedMetaRegions = 0; //num of moved regions that are META int numMovedMetaRegions = 0; //num of moved regions that are META
Map<ServerName, List<HRegionInfo>> clusterState; Map<ServerName, List<HRegionInfo>> clusterState;
protected final RackManager rackManager; protected final RackManager rackManager;
protected Cluster( protected Cluster(
ServerName masterServerName,
Map<ServerName, List<HRegionInfo>> clusterState, Map<ServerName, List<HRegionInfo>> clusterState,
Map<String, Deque<RegionLoad>> loads, Map<String, Deque<RegionLoad>> loads,
RegionLocationFinder regionFinder, RegionLocationFinder regionFinder,
Set<String> tablesOnMaster,
RackManager rackManager) { RackManager rackManager) {
this(masterServerName, null, clusterState, loads, regionFinder, this(null, clusterState, loads, regionFinder,
tablesOnMaster, rackManager); rackManager);
} }
@SuppressWarnings("unchecked") @SuppressWarnings("unchecked")
protected Cluster( protected Cluster(
ServerName masterServerName,
Collection<HRegionInfo> unassignedRegions, Collection<HRegionInfo> unassignedRegions,
Map<ServerName, List<HRegionInfo>> clusterState, Map<ServerName, List<HRegionInfo>> clusterState,
Map<String, Deque<RegionLoad>> loads, Map<String, Deque<RegionLoad>> loads,
RegionLocationFinder regionFinder, RegionLocationFinder regionFinder,
Set<String> tablesOnMaster,
RackManager rackManager) { RackManager rackManager) {
if (unassignedRegions == null) { if (unassignedRegions == null) {
unassignedRegions = EMPTY_REGION_LIST; unassignedRegions = EMPTY_REGION_LIST;
} }
this.masterServerName = masterServerName;
this.tablesOnMaster = tablesOnMaster;
serversToIndex = new HashMap<String, Integer>(); serversToIndex = new HashMap<String, Integer>();
hostsToIndex = new HashMap<String, Integer>(); hostsToIndex = new HashMap<String, Integer>();
racksToIndex = new HashMap<String, Integer>(); racksToIndex = new HashMap<String, Integer>();
@ -264,10 +252,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
} }
primariesOfRegionsPerServer[serverIndex] = new int[regionsPerServer[serverIndex].length]; primariesOfRegionsPerServer[serverIndex] = new int[regionsPerServer[serverIndex].length];
serverIndicesSortedByRegionCount[serverIndex] = serverIndex; serverIndicesSortedByRegionCount[serverIndex] = serverIndex;
if (servers[serverIndex].equals(masterServerName)) {
activeMasterIndex = serverIndex;
}
} }
hosts = new String[numHosts]; hosts = new String[numHosts];
@ -642,16 +626,8 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
regionIndexToServerIndex[region] = newServer; regionIndexToServerIndex[region] = newServer;
if (initialRegionIndexToServerIndex[region] == newServer) { if (initialRegionIndexToServerIndex[region] == newServer) {
numMovedRegions--; //region moved back to original location numMovedRegions--; //region moved back to original location
if (shouldBeOnMaster(regions[region]) && isActiveMaster(newServer)) {
//Master hosted region moved back to the active master
numMovedMasterHostedRegions--;
}
} else if (oldServer >= 0 && initialRegionIndexToServerIndex[region] == oldServer) { } else if (oldServer >= 0 && initialRegionIndexToServerIndex[region] == oldServer) {
numMovedRegions++; //region moved from original location numMovedRegions++; //region moved from original location
if (shouldBeOnMaster(regions[region]) && isActiveMaster(oldServer)) {
// Master hosted region moved away from active the master
numMovedMasterHostedRegions++;
}
} }
int tableIndex = regionIndexToTableIndex[region]; int tableIndex = regionIndexToTableIndex[region];
if (oldServer >= 0) { if (oldServer >= 0) {
@ -767,15 +743,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
return regionsPerServer[server].length; return regionsPerServer[server].length;
} }
boolean isActiveMaster(int server) {
return activeMasterIndex == server;
}
boolean shouldBeOnMaster(HRegionInfo region) {
return tablesOnMaster != null && tablesOnMaster.contains(
region.getTable().getNameAsString());
}
boolean contains(int[] arr, int val) { boolean contains(int[] arr, int val) {
return Arrays.binarySearch(arr, val) >= 0; return Arrays.binarySearch(arr, val) >= 0;
} }
@ -815,8 +782,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
numTables + numTables +
", numMovedRegions=" + ", numMovedRegions=" +
numMovedRegions + numMovedRegions +
", numMovedMasterHostedRegions=" +
numMovedMasterHostedRegions +
'}'; '}';
return desc; return desc;
} }
@ -902,7 +867,8 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
*/ */
protected List<RegionPlan> balanceMasterRegions( protected List<RegionPlan> balanceMasterRegions(
Map<ServerName, List<HRegionInfo>> clusterMap) { Map<ServerName, List<HRegionInfo>> clusterMap) {
if (services == null || clusterMap.size() <= 1) return null; if (masterServerName == null
|| clusterMap == null || clusterMap.size() <= 1) return null;
List<RegionPlan> plans = null; List<RegionPlan> plans = null;
List<HRegionInfo> regions = clusterMap.get(masterServerName); List<HRegionInfo> regions = clusterMap.get(masterServerName);
if (regions != null) { if (regions != null) {
@ -946,6 +912,27 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
return plans; return plans;
} }
/**
* Assign the regions that should be on master regionserver.
*/
protected Map<ServerName, List<HRegionInfo>> assignMasterRegions(
Collection<HRegionInfo> regions, List<ServerName> servers) {
if (servers == null || regions == null || regions.isEmpty()) {
return null;
}
Map<ServerName, List<HRegionInfo>> assignments
= new TreeMap<ServerName, List<HRegionInfo>>();
if (masterServerName != null && servers.contains(masterServerName)) {
assignments.put(masterServerName, new ArrayList<HRegionInfo>());
for (HRegionInfo region: regions) {
if (shouldBeOnMaster(region)) {
assignments.get(masterServerName).add(region);
}
}
}
return assignments;
}
@Override @Override
public Configuration getConf() { public Configuration getConf() {
return this.config; return this.config;
@ -969,8 +956,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
} }
protected boolean needsBalance(Cluster c) { protected boolean needsBalance(Cluster c) {
ClusterLoadState cs = new ClusterLoadState( ClusterLoadState cs = new ClusterLoadState(c.clusterState);
masterServerName, c.clusterState);
if (cs.getNumServers() < MIN_SERVER_BALANCE) { if (cs.getNumServers() < MIN_SERVER_BALANCE) {
if (LOG.isDebugEnabled()) { if (LOG.isDebugEnabled()) {
LOG.debug("Not running balancer because only " + cs.getNumServers() LOG.debug("Not running balancer because only " + cs.getNumServers()
@ -1031,8 +1017,21 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
public Map<ServerName, List<HRegionInfo>> roundRobinAssignment(List<HRegionInfo> regions, public Map<ServerName, List<HRegionInfo>> roundRobinAssignment(List<HRegionInfo> regions,
List<ServerName> servers) { List<ServerName> servers) {
metricsBalancer.incrMiscInvocations(); metricsBalancer.incrMiscInvocations();
Map<ServerName, List<HRegionInfo>> assignments = assignMasterRegions(regions, servers);
if (assignments != null && !assignments.isEmpty()) {
servers = new ArrayList<ServerName>(servers);
// Guarantee not to put other regions on master
servers.remove(masterServerName);
List<HRegionInfo> masterRegions = assignments.get(masterServerName);
if (!masterRegions.isEmpty()) {
regions = new ArrayList<HRegionInfo>(regions);
for (HRegionInfo region: masterRegions) {
regions.remove(region);
}
}
}
if (regions == null || regions.isEmpty()) { if (regions == null || regions.isEmpty()) {
return null; return assignments;
} }
int numServers = servers == null ? 0 : servers.size(); int numServers = servers == null ? 0 : servers.size();
@ -1046,29 +1045,18 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
// generator for AssignRegionAction. The LB will ensure the regions are mostly local // generator for AssignRegionAction. The LB will ensure the regions are mostly local
// and balanced. This should also run fast with fewer number of iterations. // and balanced. This should also run fast with fewer number of iterations.
Map<ServerName, List<HRegionInfo>> assignments = new TreeMap<ServerName, List<HRegionInfo>>();
if (numServers == 1) { // Only one server, nothing fancy we can do here if (numServers == 1) { // Only one server, nothing fancy we can do here
ServerName server = servers.get(0); ServerName server = servers.get(0);
assignments.put(server, new ArrayList<HRegionInfo>(regions)); assignments.put(server, new ArrayList<HRegionInfo>(regions));
return assignments; return assignments;
} }
List<HRegionInfo> masterRegions = null;
if (servers.contains(masterServerName)) {
masterRegions = new ArrayList<HRegionInfo>();
}
Cluster cluster = createCluster(servers, regions, tablesOnMaster); Cluster cluster = createCluster(servers, regions);
List<HRegionInfo> unassignedRegions = new ArrayList<HRegionInfo>(); List<HRegionInfo> unassignedRegions = new ArrayList<HRegionInfo>();
roundRobinAssignment(cluster, regions, unassignedRegions, roundRobinAssignment(cluster, regions, unassignedRegions,
servers, masterRegions, assignments); servers, assignments);
if (masterRegions != null && !masterRegions.isEmpty()) {
assignments.put(masterServerName, masterRegions);
for (HRegionInfo r : masterRegions) {
cluster.doAssignRegion(r, masterServerName);
}
}
List<HRegionInfo> lastFewRegions = new ArrayList<HRegionInfo>(); List<HRegionInfo> lastFewRegions = new ArrayList<HRegionInfo>();
// assign the remaining by going through the list and try to assign to servers one-by-one // assign the remaining by going through the list and try to assign to servers one-by-one
int serverIdx = RANDOM.nextInt(numServers); int serverIdx = RANDOM.nextInt(numServers);
@ -1076,9 +1064,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
boolean assigned = false; boolean assigned = false;
for (int j = 0; j < numServers; j++) { // try all servers one by one for (int j = 0; j < numServers; j++) { // try all servers one by one
ServerName serverName = servers.get((j + serverIdx) % numServers); ServerName serverName = servers.get((j + serverIdx) % numServers);
if (serverName.equals(masterServerName)) {
continue;
}
if (!cluster.wouldLowerAvailability(region, serverName)) { if (!cluster.wouldLowerAvailability(region, serverName)) {
List<HRegionInfo> serverRegions = assignments.get(serverName); List<HRegionInfo> serverRegions = assignments.get(serverName);
if (serverRegions == null) { if (serverRegions == null) {
@ -1101,11 +1086,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
for (HRegionInfo region : lastFewRegions) { for (HRegionInfo region : lastFewRegions) {
int i = RANDOM.nextInt(numServers); int i = RANDOM.nextInt(numServers);
ServerName server = servers.get(i); ServerName server = servers.get(i);
if (server.equals(masterServerName)) {
// Try to avoid master for a user region
i = (i == 0 ? 1 : i - 1);
server = servers.get(i);
}
List<HRegionInfo> serverRegions = assignments.get(server); List<HRegionInfo> serverRegions = assignments.get(server);
if (serverRegions == null) { if (serverRegions == null) {
serverRegions = new ArrayList<HRegionInfo>(); serverRegions = new ArrayList<HRegionInfo>();
@ -1118,7 +1098,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
} }
protected Cluster createCluster(List<ServerName> servers, protected Cluster createCluster(List<ServerName> servers,
Collection<HRegionInfo> regions, Set<String> tablesOnMaster) { Collection<HRegionInfo> regions) {
// Get the snapshot of the current assignments for the regions in question, and then create // Get the snapshot of the current assignments for the regions in question, and then create
// a cluster out of it. Note that we might have replicas already assigned to some servers // a cluster out of it. Note that we might have replicas already assigned to some servers
// earlier. So we want to get the snapshot to see those assignments, but this will only contain // earlier. So we want to get the snapshot to see those assignments, but this will only contain
@ -1130,8 +1110,8 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
clusterState.put(server, EMPTY_REGION_LIST); clusterState.put(server, EMPTY_REGION_LIST);
} }
} }
return new Cluster(masterServerName, regions, clusterState, null, this.regionFinder, return new Cluster(regions, clusterState, null, this.regionFinder,
tablesOnMaster, rackManager); rackManager);
} }
/** /**
@ -1173,6 +1153,15 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
@Override @Override
public ServerName randomAssignment(HRegionInfo regionInfo, List<ServerName> servers) { public ServerName randomAssignment(HRegionInfo regionInfo, List<ServerName> servers) {
metricsBalancer.incrMiscInvocations(); metricsBalancer.incrMiscInvocations();
if (servers != null && servers.contains(masterServerName)) {
if (shouldBeOnMaster(regionInfo)) {
return masterServerName;
}
servers = new ArrayList<ServerName>(servers);
// Guarantee not to put other regions on master
servers.remove(masterServerName);
}
int numServers = servers == null ? 0 : servers.size(); int numServers = servers == null ? 0 : servers.size();
if (numServers == 0) { if (numServers == 0) {
LOG.warn("Wanted to do retain assignment but no servers to assign to"); LOG.warn("Wanted to do retain assignment but no servers to assign to");
@ -1181,13 +1170,9 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
if (numServers == 1) { // Only one server, nothing fancy we can do here if (numServers == 1) { // Only one server, nothing fancy we can do here
return servers.get(0); return servers.get(0);
} }
if (shouldBeOnMaster(regionInfo)
&& servers.contains(masterServerName)) {
return masterServerName;
}
List<HRegionInfo> regions = Lists.newArrayList(regionInfo); List<HRegionInfo> regions = Lists.newArrayList(regionInfo);
Cluster cluster = createCluster(servers, regions, tablesOnMaster); Cluster cluster = createCluster(servers, regions);
return randomAssignment(cluster, regionInfo, servers); return randomAssignment(cluster, regionInfo, servers);
} }
@ -1213,8 +1198,22 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
List<ServerName> servers) { List<ServerName> servers) {
// Update metrics // Update metrics
metricsBalancer.incrMiscInvocations(); metricsBalancer.incrMiscInvocations();
Map<ServerName, List<HRegionInfo>> assignments
= assignMasterRegions(regions.keySet(), servers);
if (assignments != null && !assignments.isEmpty()) {
servers = new ArrayList<ServerName>(servers);
// Guarantee not to put other regions on master
servers.remove(masterServerName);
List<HRegionInfo> masterRegions = assignments.get(masterServerName);
if (!masterRegions.isEmpty()) {
regions = new HashMap<HRegionInfo, ServerName>(regions);
for (HRegionInfo region: masterRegions) {
regions.remove(region);
}
}
}
if (regions == null || regions.isEmpty()) { if (regions == null || regions.isEmpty()) {
return null; return assignments;
} }
int numServers = servers == null ? 0 : servers.size(); int numServers = servers == null ? 0 : servers.size();
@ -1222,7 +1221,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
LOG.warn("Wanted to do retain assignment but no servers to assign to"); LOG.warn("Wanted to do retain assignment but no servers to assign to");
return null; return null;
} }
Map<ServerName, List<HRegionInfo>> assignments = new TreeMap<ServerName, List<HRegionInfo>>();
if (numServers == 1) { // Only one server, nothing fancy we can do here if (numServers == 1) { // Only one server, nothing fancy we can do here
ServerName server = servers.get(0); ServerName server = servers.get(0);
assignments.put(server, new ArrayList<HRegionInfo>(regions.keySet())); assignments.put(server, new ArrayList<HRegionInfo>(regions.keySet()));
@ -1238,23 +1236,18 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
ArrayListMultimap<String, ServerName> serversByHostname = ArrayListMultimap.create(); ArrayListMultimap<String, ServerName> serversByHostname = ArrayListMultimap.create();
for (ServerName server : servers) { for (ServerName server : servers) {
assignments.put(server, new ArrayList<HRegionInfo>()); assignments.put(server, new ArrayList<HRegionInfo>());
if (!server.equals(masterServerName)) {
serversByHostname.put(server.getHostname(), server); serversByHostname.put(server.getHostname(), server);
} }
}
// Collection of the hostnames that used to have regions // Collection of the hostnames that used to have regions
// assigned, but for which we no longer have any RS running // assigned, but for which we no longer have any RS running
// after the cluster restart. // after the cluster restart.
Set<String> oldHostsNoLongerPresent = Sets.newTreeSet(); Set<String> oldHostsNoLongerPresent = Sets.newTreeSet();
// Master regionserver is in the server list.
boolean masterIncluded = servers.contains(masterServerName);
int numRandomAssignments = 0; int numRandomAssignments = 0;
int numRetainedAssigments = 0; int numRetainedAssigments = 0;
Cluster cluster = createCluster(servers, regions.keySet(), tablesOnMaster); Cluster cluster = createCluster(servers, regions.keySet());
for (Map.Entry<HRegionInfo, ServerName> entry : regions.entrySet()) { for (Map.Entry<HRegionInfo, ServerName> entry : regions.entrySet()) {
HRegionInfo region = entry.getKey(); HRegionInfo region = entry.getKey();
@ -1263,14 +1256,7 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
if (oldServerName != null) { if (oldServerName != null) {
localServers = serversByHostname.get(oldServerName.getHostname()); localServers = serversByHostname.get(oldServerName.getHostname());
} }
if (masterIncluded && shouldBeOnMaster(region)) { if (localServers.isEmpty()) {
assignments.get(masterServerName).add(region);
if (localServers.contains(masterServerName)) {
numRetainedAssigments++;
} else {
numRandomAssignments++;
}
} else if (localServers.isEmpty()) {
// No servers on the new cluster match up with this hostname, // No servers on the new cluster match up with this hostname,
// assign randomly. // assign randomly.
ServerName randomServer = randomAssignment(cluster, region, servers); ServerName randomServer = randomAssignment(cluster, region, servers);
@ -1355,11 +1341,6 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
do { do {
int i = RANDOM.nextInt(numServers); int i = RANDOM.nextInt(numServers);
sn = servers.get(i); sn = servers.get(i);
if (sn.equals(masterServerName)) {
// Try to avoid master for a user region
i = (i == 0 ? 1 : i - 1);
sn = servers.get(i);
}
} while (cluster.wouldLowerAvailability(regionInfo, sn) } while (cluster.wouldLowerAvailability(regionInfo, sn)
&& iterations++ < maxIterations); && iterations++ < maxIterations);
cluster.doAssignRegion(regionInfo, sn); cluster.doAssignRegion(regionInfo, sn);
@ -1371,16 +1352,11 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
*/ */
private void roundRobinAssignment(Cluster cluster, List<HRegionInfo> regions, private void roundRobinAssignment(Cluster cluster, List<HRegionInfo> regions,
List<HRegionInfo> unassignedRegions, List<ServerName> servers, List<HRegionInfo> unassignedRegions, List<ServerName> servers,
List<HRegionInfo> masterRegions, Map<ServerName, List<HRegionInfo>> assignments) { Map<ServerName, List<HRegionInfo>> assignments) {
boolean masterIncluded = servers.contains(masterServerName);
int numServers = servers.size(); int numServers = servers.size();
int skipServers = numServers;
if (masterIncluded) {
skipServers--;
}
int numRegions = regions.size(); int numRegions = regions.size();
int max = (int) Math.ceil((float) numRegions / skipServers); int max = (int) Math.ceil((float) numRegions / numServers);
int serverIdx = 0; int serverIdx = 0;
if (numServers > 1) { if (numServers > 1) {
serverIdx = RANDOM.nextInt(numServers); serverIdx = RANDOM.nextInt(numServers);
@ -1389,25 +1365,15 @@ public abstract class BaseLoadBalancer implements LoadBalancer {
for (int j = 0; j < numServers; j++) { for (int j = 0; j < numServers; j++) {
ServerName server = servers.get((j + serverIdx) % numServers); ServerName server = servers.get((j + serverIdx) % numServers);
if (masterIncluded && server.equals(masterServerName)) {
// Don't put non-special region on the master regionserver,
// So that it is not overloaded.
continue;
}
List<HRegionInfo> serverRegions = new ArrayList<HRegionInfo>(max); List<HRegionInfo> serverRegions = new ArrayList<HRegionInfo>(max);
for (int i = regionIdx; i < numRegions; i += skipServers) { for (int i = regionIdx; i < numRegions; i += numServers) {
HRegionInfo region = regions.get(i % numRegions); HRegionInfo region = regions.get(i % numRegions);
if (masterRegions == null || !shouldBeOnMaster(region)) {
if (cluster.wouldLowerAvailability(region, server)) { if (cluster.wouldLowerAvailability(region, server)) {
unassignedRegions.add(region); unassignedRegions.add(region);
} else { } else {
serverRegions.add(region); serverRegions.add(region);
cluster.doAssignRegion(region, server); cluster.doAssignRegion(region, server);
} }
continue;
}
// Master is in the list and this is a special region
masterRegions.add(region);
} }
assignments.put(server, serverRegions); assignments.put(server, serverRegions);
regionIdx++; regionIdx++;

View File

@ -35,19 +35,13 @@ public class ClusterLoadState {
private int numRegions = 0; private int numRegions = 0;
private int numServers = 0; private int numServers = 0;
public ClusterLoadState(ServerName master, public ClusterLoadState(Map<ServerName, List<HRegionInfo>> clusterState) {
Map<ServerName, List<HRegionInfo>> clusterState) {
this.numRegions = 0; this.numRegions = 0;
this.numServers = clusterState.size(); this.numServers = clusterState.size();
this.clusterState = clusterState; this.clusterState = clusterState;
serversByLoad = new TreeMap<ServerAndLoad, List<HRegionInfo>>(); serversByLoad = new TreeMap<ServerAndLoad, List<HRegionInfo>>();
// Iterate so we can count regions as we build the map // Iterate so we can count regions as we build the map
for (Map.Entry<ServerName, List<HRegionInfo>> server : clusterState.entrySet()) { for (Map.Entry<ServerName, List<HRegionInfo>> server : clusterState.entrySet()) {
if (master != null && numServers > 1 && master.equals(server.getKey())) {
// Don't count the master since its load is meant to be low.
numServers--;
continue;
}
List<HRegionInfo> regions = server.getValue(); List<HRegionInfo> regions = server.getValue();
int sz = regions.size(); int sz = regions.size();
if (sz == 0) emptyRegionServerPresent = true; if (sz == 0) emptyRegionServerPresent = true;

View File

@ -183,19 +183,26 @@ public class SimpleLoadBalancer extends BaseLoadBalancer {
public List<RegionPlan> balanceCluster( public List<RegionPlan> balanceCluster(
Map<ServerName, List<HRegionInfo>> clusterMap) { Map<ServerName, List<HRegionInfo>> clusterMap) {
List<RegionPlan> regionsToReturn = balanceMasterRegions(clusterMap); List<RegionPlan> regionsToReturn = balanceMasterRegions(clusterMap);
if (regionsToReturn != null) { if (regionsToReturn != null || clusterMap == null || clusterMap.size() <= 1) {
return regionsToReturn; return regionsToReturn;
} }
if (masterServerName != null && clusterMap.containsKey(masterServerName)) {
if (clusterMap.size() <= 2) {
return null;
}
clusterMap = new HashMap<ServerName, List<HRegionInfo>>(clusterMap);
clusterMap.remove(masterServerName);
}
boolean emptyRegionServerPresent = false; boolean emptyRegionServerPresent = false;
long startTime = System.currentTimeMillis(); long startTime = System.currentTimeMillis();
ClusterLoadState cs = new ClusterLoadState(masterServerName, clusterMap);
// construct a Cluster object with clusterMap and rest of the // construct a Cluster object with clusterMap and rest of the
// argument as defaults // argument as defaults
Cluster c = new Cluster(masterServerName, clusterMap, null, this.regionFinder, Cluster c = new Cluster(clusterMap, null, this.regionFinder, this.rackManager);
tablesOnMaster, this.rackManager);
if (!this.needsBalance(c)) return null; if (!this.needsBalance(c)) return null;
ClusterLoadState cs = new ClusterLoadState(clusterMap);
int numServers = cs.getNumServers(); int numServers = cs.getNumServers();
NavigableMap<ServerAndLoad, List<HRegionInfo>> serversByLoad = cs.getServersByLoad(); NavigableMap<ServerAndLoad, List<HRegionInfo>> serversByLoad = cs.getServersByLoad();
int numRegions = cs.getNumRegions(); int numRegions = cs.getNumRegions();

View File

@ -208,14 +208,21 @@ public class StochasticLoadBalancer extends BaseLoadBalancer {
@Override @Override
public List<RegionPlan> balanceCluster(Map<ServerName, List<HRegionInfo>> clusterState) { public List<RegionPlan> balanceCluster(Map<ServerName, List<HRegionInfo>> clusterState) {
List<RegionPlan> plans = balanceMasterRegions(clusterState); List<RegionPlan> plans = balanceMasterRegions(clusterState);
if (plans != null) { if (plans != null || clusterState == null || clusterState.size() <= 1) {
return plans; return plans;
} }
if (masterServerName != null && clusterState.containsKey(masterServerName)) {
if (clusterState.size() <= 2) {
return null;
}
clusterState = new HashMap<ServerName, List<HRegionInfo>>(clusterState);
clusterState.remove(masterServerName);
}
//The clusterState that is given to this method contains the state //The clusterState that is given to this method contains the state
//of all the regions in the table(s) (that's true today) //of all the regions in the table(s) (that's true today)
// Keep track of servers to iterate through them. // Keep track of servers to iterate through them.
Cluster cluster = new Cluster(masterServerName, Cluster cluster = new Cluster(clusterState, loads, regionFinder, rackManager);
clusterState, loads, regionFinder, tablesOnMaster, rackManager);
if (!needsBalance(cluster)) { if (!needsBalance(cluster)) {
return null; return null;
} }
@ -420,11 +427,7 @@ public class StochasticLoadBalancer extends BaseLoadBalancer {
return -1; return -1;
} }
int n = RANDOM.nextInt(cluster.numServers); return RANDOM.nextInt(cluster.numServers);
if (cluster.numServers > 1 && cluster.isActiveMaster(n)) {
n = (n + 1) % cluster.numServers;
}
return n;
} }
protected int pickRandomRack(Cluster cluster) { protected int pickRandomRack(Cluster cluster) {
@ -439,9 +442,6 @@ public class StochasticLoadBalancer extends BaseLoadBalancer {
if (cluster.numServers < 2) { if (cluster.numServers < 2) {
return -1; return -1;
} }
if (cluster.activeMasterIndex != -1 && cluster.numServers == 2) {
return -1;
}
while (true) { while (true) {
int otherServerIndex = pickRandomServer(cluster); int otherServerIndex = pickRandomServer(cluster);
if (otherServerIndex != serverIndex) { if (otherServerIndex != serverIndex) {
@ -530,8 +530,7 @@ public class StochasticLoadBalancer extends BaseLoadBalancer {
Integer[] servers = cluster.serverIndicesSortedByRegionCount; Integer[] servers = cluster.serverIndicesSortedByRegionCount;
int index = 0; int index = 0;
while (servers[index] == null || servers[index] == thisServer while (servers[index] == null || servers[index] == thisServer) {
|| cluster.isActiveMaster(index)) {
index++; index++;
if (index == servers.length) { if (index == servers.length) {
return -1; return -1;
@ -544,8 +543,7 @@ public class StochasticLoadBalancer extends BaseLoadBalancer {
Integer[] servers = cluster.serverIndicesSortedByRegionCount; Integer[] servers = cluster.serverIndicesSortedByRegionCount;
int index = servers.length - 1; int index = servers.length - 1;
while (servers[index] == null || servers[index] == thisServer while (servers[index] == null || servers[index] == thisServer) {
|| cluster.isActiveMaster(index)) {
index--; index--;
if (index < 0) { if (index < 0) {
return -1; return -1;
@ -801,9 +799,6 @@ public class StochasticLoadBalancer extends BaseLoadBalancer {
double total = getSum(stats); double total = getSum(stats);
double count = stats.length; double count = stats.length;
if (stats.length > 1 && cluster.activeMasterIndex != -1) {
count--; // Exclude the active master
}
double mean = total/count; double mean = total/count;
// Compute max as if all region servers had 0 and one had the sum of all costs. This must be // Compute max as if all region servers had 0 and one had the sum of all costs. This must be
@ -824,12 +819,6 @@ public class StochasticLoadBalancer extends BaseLoadBalancer {
} }
min = Math.max(0, min); min = Math.max(0, min);
for (int i=0; i<stats.length; i++) { for (int i=0; i<stats.length; i++) {
if (stats.length > 1 && cluster.isActiveMaster(i)) {
// Not count the active master load
continue;
}
double n = stats[i]; double n = stats[i];
double diff = Math.abs(mean - n); double diff = Math.abs(mean - n);
totalCost += diff; totalCost += diff;
@ -897,11 +886,9 @@ public class StochasticLoadBalancer extends BaseLoadBalancer {
double moveCost = cluster.numMovedRegions; double moveCost = cluster.numMovedRegions;
// Don't let this single balance move more than the max moves, // Don't let this single balance move more than the max moves.
// or move a region that should be on master away from the master.
// It is ok to move any master hosted region back to the master.
// This allows better scaling to accurately represent the actual cost of a move. // This allows better scaling to accurately represent the actual cost of a move.
if (moveCost > maxMoves || cluster.numMovedMasterHostedRegions > 0) { if (moveCost > maxMoves) {
return 1000000; // return a number much greater than any of the other cost return 1000000; // return a number much greater than any of the other cost
} }

View File

@ -42,6 +42,7 @@ import org.apache.hadoop.hbase.master.MasterServices;
import org.apache.hadoop.hbase.master.RegionState; import org.apache.hadoop.hbase.master.RegionState;
import org.apache.hadoop.hbase.master.RegionStates; import org.apache.hadoop.hbase.master.RegionStates;
import org.apache.hadoop.hbase.master.ServerManager; import org.apache.hadoop.hbase.master.ServerManager;
import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer;
import org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode; import org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode;
/** /**
@ -130,9 +131,10 @@ public class ServerShutdownHandler extends EventHandler {
// we are not ready to assign dead regions either. So we re-queue up // we are not ready to assign dead regions either. So we re-queue up
// the dead server for further processing too. // the dead server for further processing too.
AssignmentManager am = services.getAssignmentManager(); AssignmentManager am = services.getAssignmentManager();
ServerManager serverManager = services.getServerManager();
if (isCarryingMeta() // hbase:meta if (isCarryingMeta() // hbase:meta
|| !am.isFailoverCleanupDone()) { || !am.isFailoverCleanupDone()) {
this.services.getServerManager().processDeadServer(serverName, this.shouldSplitHlog); serverManager.processDeadServer(serverName, this.shouldSplitHlog);
return; return;
} }
@ -152,16 +154,22 @@ public class ServerShutdownHandler extends EventHandler {
// {@link SplitTransaction}. We'd also have to be figure another way for // {@link SplitTransaction}. We'd also have to be figure another way for
// doing the below hbase:meta daughters fixup. // doing the below hbase:meta daughters fixup.
Set<HRegionInfo> hris = null; Set<HRegionInfo> hris = null;
while (!this.server.isStopped()) {
try { try {
server.getMetaTableLocator().waitMetaRegionLocation(server.getZooKeeper()); server.getMetaTableLocator().waitMetaRegionLocation(server.getZooKeeper());
if (BaseLoadBalancer.tablesOnMaster(server.getConfiguration())) {
while (!this.server.isStopped() && serverManager.countOfRegionServers() < 2) {
// Wait till at least another regionserver is up besides the active master
// so that we don't assign all regions to the active master.
// This is best of efforts, because newly joined regionserver
// could crash right after that.
Thread.sleep(100);
}
}
hris = am.getRegionStates().getServerRegions(serverName); hris = am.getRegionStates().getServerRegions(serverName);
break;
} catch (InterruptedException e) { } catch (InterruptedException e) {
Thread.currentThread().interrupt(); Thread.currentThread().interrupt();
throw (InterruptedIOException)new InterruptedIOException().initCause(e); throw (InterruptedIOException)new InterruptedIOException().initCause(e);
} }
}
if (this.server.isStopped()) { if (this.server.isStopped()) {
throw new IOException("Server is stopped"); throw new IOException("Server is stopped");
} }

View File

@ -1120,11 +1120,8 @@ public class HRegionServer extends HasThread implements
} }
break; break;
} }
try { if (sleep(200)) {
Thread.sleep(200);
} catch (InterruptedException e) {
interrupted = true; interrupted = true;
LOG.warn("Interrupted while sleeping");
} }
} }
} finally { } finally {
@ -1134,6 +1131,17 @@ public class HRegionServer extends HasThread implements
} }
} }
private boolean sleep(long millis) {
boolean interrupted = false;
try {
Thread.sleep(millis);
} catch (InterruptedException e) {
LOG.warn("Interrupted while sleeping");
interrupted = true;
}
return interrupted;
}
private void closeWAL(final boolean delete) { private void closeWAL(final boolean delete) {
if (this.hlogForMeta != null) { if (this.hlogForMeta != null) {
// All hlogs (meta and non-meta) are in the same directory. Don't call // All hlogs (meta and non-meta) are in the same directory. Don't call
@ -2005,10 +2013,14 @@ public class HRegionServer extends HasThread implements
LOG.debug("No master found and cluster is stopped; bailing out"); LOG.debug("No master found and cluster is stopped; bailing out");
return null; return null;
} }
if (System.currentTimeMillis() > (previousLogTime + 1000)) {
LOG.debug("No master found; retry"); LOG.debug("No master found; retry");
previousLogTime = System.currentTimeMillis(); previousLogTime = System.currentTimeMillis();
}
refresh = true; // let's try pull it from ZK directly refresh = true; // let's try pull it from ZK directly
sleeper.sleep(); if (sleep(200)) {
interrupted = true;
}
continue; continue;
} }
@ -2023,24 +2035,18 @@ public class HRegionServer extends HasThread implements
intf = RegionServerStatusService.newBlockingStub(channel); intf = RegionServerStatusService.newBlockingStub(channel);
break; break;
} catch (IOException e) { } catch (IOException e) {
if (System.currentTimeMillis() > (previousLogTime + 1000)) {
e = e instanceof RemoteException ? e = e instanceof RemoteException ?
((RemoteException)e).unwrapRemoteException() : e; ((RemoteException)e).unwrapRemoteException() : e;
if (e instanceof ServerNotRunningYetException) { if (e instanceof ServerNotRunningYetException) {
if (System.currentTimeMillis() > (previousLogTime+1000)){
LOG.info("Master isn't available yet, retrying"); LOG.info("Master isn't available yet, retrying");
previousLogTime = System.currentTimeMillis();
}
} else { } else {
if (System.currentTimeMillis() > (previousLogTime + 1000)) {
LOG.warn("Unable to connect to master. Retrying. Error was:", e); LOG.warn("Unable to connect to master. Retrying. Error was:", e);
}
previousLogTime = System.currentTimeMillis(); previousLogTime = System.currentTimeMillis();
} }
} if (sleep(200)) {
try {
Thread.sleep(200);
} catch (InterruptedException ex) {
interrupted = true; interrupted = true;
LOG.warn("Interrupted while sleeping");
} }
} }
} }

View File

@ -245,7 +245,7 @@ public class TestBlockReorder {
byte[] sb = "sb".getBytes(); byte[] sb = "sb".getBytes();
htu.startMiniZKCluster(); htu.startMiniZKCluster();
MiniHBaseCluster hbm = htu.startMiniHBaseCluster(1, 0); MiniHBaseCluster hbm = htu.startMiniHBaseCluster(1, 1);
hbm.waitForActiveAndReadyMaster(); hbm.waitForActiveAndReadyMaster();
HRegionServer targetRs = hbm.getMaster(); HRegionServer targetRs = hbm.getMaster();

View File

@ -67,7 +67,7 @@ public class TestMasterMetrics {
public static void startCluster() throws Exception { public static void startCluster() throws Exception {
LOG.info("Starting cluster"); LOG.info("Starting cluster");
TEST_UTIL = new HBaseTestingUtility(); TEST_UTIL = new HBaseTestingUtility();
TEST_UTIL.startMiniCluster(1, 0, 1, null, MyMaster.class, null); TEST_UTIL.startMiniCluster(1, 1, 1, null, MyMaster.class, null);
cluster = TEST_UTIL.getHBaseCluster(); cluster = TEST_UTIL.getHBaseCluster();
LOG.info("Waiting for active/ready master"); LOG.info("Waiting for active/ready master");
cluster.waitForActiveAndReadyMaster(); cluster.waitForActiveAndReadyMaster();
@ -117,7 +117,7 @@ public class TestMasterMetrics {
@Test @Test
public void testDefaultMasterMetrics() throws Exception { public void testDefaultMasterMetrics() throws Exception {
MetricsMasterSource masterSource = master.getMasterMetrics().getMetricsSource(); MetricsMasterSource masterSource = master.getMasterMetrics().getMetricsSource();
metricsHelper.assertGauge( "numRegionServers", 1, masterSource); metricsHelper.assertGauge( "numRegionServers", 2, masterSource);
metricsHelper.assertGauge( "averageLoad", 2, masterSource); metricsHelper.assertGauge( "averageLoad", 2, masterSource);
metricsHelper.assertGauge( "numDeadRegionServers", 0, masterSource); metricsHelper.assertGauge( "numDeadRegionServers", 0, masterSource);

View File

@ -216,8 +216,8 @@ public class BalancerTestBase {
} }
protected BaseLoadBalancer.Cluster mockCluster(int[] mockCluster) { protected BaseLoadBalancer.Cluster mockCluster(int[] mockCluster) {
return new BaseLoadBalancer.Cluster(null, return new BaseLoadBalancer.Cluster(
mockClusterServers(mockCluster, -1), null, null, null, null); mockClusterServers(mockCluster, -1), null, null, null);
} }
protected TreeMap<ServerName, List<HRegionInfo>> mockClusterServers(int[] mockCluster, int numTables) { protected TreeMap<ServerName, List<HRegionInfo>> mockClusterServers(int[] mockCluster, int numTables) {

View File

@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.master.balancer;
import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertNotEquals; import static org.junit.Assert.assertNotEquals;
import static org.junit.Assert.assertNull;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import static org.mockito.Mockito.mock; import static org.mockito.Mockito.mock;
import static org.mockito.Mockito.when; import static org.mockito.Mockito.when;
@ -126,7 +127,7 @@ public class TestBaseLoadBalancer extends BalancerTestBase {
tmp = new ArrayList<ServerName>(); tmp = new ArrayList<ServerName>();
tmp.add(master); tmp.add(master);
sn = loadBalancer.randomAssignment(hri, tmp); sn = loadBalancer.randomAssignment(hri, tmp);
assertEquals(master, sn); assertNull("Should not assign user regions on master", sn);
for (int[] mock : regionsAndServersMocks) { for (int[] mock : regionsAndServersMocks) {
LOG.debug("testImmediateAssignment with " + mock[0] + " regions and " + mock[1] + " servers"); LOG.debug("testImmediateAssignment with " + mock[0] + " regions and " + mock[1] + " servers");
List<HRegionInfo> regions = randomRegions(mock[0]); List<HRegionInfo> regions = randomRegions(mock[0]);
@ -267,7 +268,7 @@ public class TestBaseLoadBalancer extends BalancerTestBase {
// cluster is created (constructor code) would make sure the indices of // cluster is created (constructor code) would make sure the indices of
// the servers are in the order in which it is inserted in the clusterState // the servers are in the order in which it is inserted in the clusterState
// map (linkedhashmap is important). A similar thing applies to the region lists // map (linkedhashmap is important). A similar thing applies to the region lists
Cluster cluster = new Cluster(master, clusterState, null, null, null, rackManager); Cluster cluster = new Cluster(clusterState, null, null, rackManager);
// check whether a move of region1 from servers[0] to servers[1] would lower // check whether a move of region1 from servers[0] to servers[1] would lower
// the availability of region1 // the availability of region1
assertTrue(cluster.wouldLowerAvailability(hri1, servers[1])); assertTrue(cluster.wouldLowerAvailability(hri1, servers[1]));
@ -284,7 +285,7 @@ public class TestBaseLoadBalancer extends BalancerTestBase {
// now lets have servers[1] host replica_of_region2 // now lets have servers[1] host replica_of_region2
list1.add(RegionReplicaUtil.getRegionInfoForReplica(hri3, 1)); list1.add(RegionReplicaUtil.getRegionInfoForReplica(hri3, 1));
// create a new clusterState with the above change // create a new clusterState with the above change
cluster = new Cluster(master, clusterState, null, null, null, rackManager); cluster = new Cluster(clusterState, null, null, rackManager);
// now check whether a move of a replica from servers[0] to servers[1] would lower // now check whether a move of a replica from servers[0] to servers[1] would lower
// the availability of region2 // the availability of region2
assertTrue(cluster.wouldLowerAvailability(hri3, servers[1])); assertTrue(cluster.wouldLowerAvailability(hri3, servers[1]));
@ -296,14 +297,14 @@ public class TestBaseLoadBalancer extends BalancerTestBase {
clusterState.put(servers[6], list2); //servers[6], rack2 hosts region2 clusterState.put(servers[6], list2); //servers[6], rack2 hosts region2
clusterState.put(servers[10], new ArrayList<HRegionInfo>()); //servers[10], rack3 hosts no region clusterState.put(servers[10], new ArrayList<HRegionInfo>()); //servers[10], rack3 hosts no region
// create a cluster with the above clusterState // create a cluster with the above clusterState
cluster = new Cluster(master, clusterState, null, null, null, rackManager); cluster = new Cluster(clusterState, null, null, rackManager);
// check whether a move of region1 from servers[0],rack1 to servers[6],rack2 would // check whether a move of region1 from servers[0],rack1 to servers[6],rack2 would
// lower the availability // lower the availability
assertTrue(cluster.wouldLowerAvailability(hri1, servers[0])); assertTrue(cluster.wouldLowerAvailability(hri1, servers[0]));
// now create a cluster without the rack manager // now create a cluster without the rack manager
cluster = new Cluster(master, clusterState, null, null, null, null); cluster = new Cluster(clusterState, null, null, null);
// now repeat check whether a move of region1 from servers[0] to servers[6] would // now repeat check whether a move of region1 from servers[0] to servers[6] would
// lower the availability // lower the availability
assertTrue(!cluster.wouldLowerAvailability(hri1, servers[6])); assertTrue(!cluster.wouldLowerAvailability(hri1, servers[6]));
@ -336,7 +337,7 @@ public class TestBaseLoadBalancer extends BalancerTestBase {
// cluster is created (constructor code) would make sure the indices of // cluster is created (constructor code) would make sure the indices of
// the servers are in the order in which it is inserted in the clusterState // the servers are in the order in which it is inserted in the clusterState
// map (linkedhashmap is important). // map (linkedhashmap is important).
Cluster cluster = new Cluster(master, clusterState, null, null, null, rackManager); Cluster cluster = new Cluster(clusterState, null, null, rackManager);
// check whether moving region1 from servers[1] to servers[2] would lower availability // check whether moving region1 from servers[1] to servers[2] would lower availability
assertTrue(!cluster.wouldLowerAvailability(hri1, servers[2])); assertTrue(!cluster.wouldLowerAvailability(hri1, servers[2]));
@ -356,7 +357,7 @@ public class TestBaseLoadBalancer extends BalancerTestBase {
clusterState.put(servers[6], list2); //servers[6], rack2 hosts region2 clusterState.put(servers[6], list2); //servers[6], rack2 hosts region2
clusterState.put(servers[12], list3); //servers[12], rack3 hosts replica_of_region2 clusterState.put(servers[12], list3); //servers[12], rack3 hosts replica_of_region2
// create a cluster with the above clusterState // create a cluster with the above clusterState
cluster = new Cluster(master, clusterState, null, null, null, rackManager); cluster = new Cluster(clusterState, null, null, rackManager);
// check whether a move of replica_of_region2 from servers[12],rack3 to servers[0],rack1 would // check whether a move of replica_of_region2 from servers[12],rack3 to servers[0],rack1 would
// lower the availability // lower the availability
assertTrue(!cluster.wouldLowerAvailability(hri4, servers[0])); assertTrue(!cluster.wouldLowerAvailability(hri4, servers[0]));
@ -442,7 +443,7 @@ public class TestBaseLoadBalancer extends BalancerTestBase {
assignRegions(regions, oldServers, clusterState); assignRegions(regions, oldServers, clusterState);
// should not throw exception: // should not throw exception:
BaseLoadBalancer.Cluster cluster = new Cluster(null, clusterState, null, null, null, null); BaseLoadBalancer.Cluster cluster = new Cluster(clusterState, null, null, null);
assertEquals(101 + 9, cluster.numRegions); assertEquals(101 + 9, cluster.numRegions);
assertEquals(10, cluster.numServers); // only 10 servers because they share the same host + port assertEquals(10, cluster.numServers); // only 10 servers because they share the same host + port
} }
@ -484,7 +485,7 @@ public class TestBaseLoadBalancer extends BalancerTestBase {
when(locationFinder.getTopBlockLocations(regions.get(43))).thenReturn( when(locationFinder.getTopBlockLocations(regions.get(43))).thenReturn(
Lists.newArrayList(ServerName.valueOf("foo", 0, 0))); // this server does not exists in clusterStatus Lists.newArrayList(ServerName.valueOf("foo", 0, 0))); // this server does not exists in clusterStatus
BaseLoadBalancer.Cluster cluster = new Cluster(null, clusterState, null, locationFinder, null, null); BaseLoadBalancer.Cluster cluster = new Cluster(clusterState, null, locationFinder, null);
int r0 = ArrayUtils.indexOf(cluster.regions, regions.get(0)); // this is ok, it is just a test int r0 = ArrayUtils.indexOf(cluster.regions, regions.get(0)); // this is ok, it is just a test
int r1 = ArrayUtils.indexOf(cluster.regions, regions.get(1)); int r1 = ArrayUtils.indexOf(cluster.regions, regions.get(1));

View File

@ -60,7 +60,6 @@ public class TestStochasticLoadBalancer extends BalancerTestBase {
private static StochasticLoadBalancer loadBalancer; private static StochasticLoadBalancer loadBalancer;
private static final Log LOG = LogFactory.getLog(TestStochasticLoadBalancer.class); private static final Log LOG = LogFactory.getLog(TestStochasticLoadBalancer.class);
private static Configuration conf; private static Configuration conf;
private static final ServerName master = ServerName.valueOf("fake-master", 0, 1L);
@BeforeClass @BeforeClass
public static void beforeAllTests() throws Exception { public static void beforeAllTests() throws Exception {
@ -332,7 +331,7 @@ public class TestStochasticLoadBalancer extends BalancerTestBase {
BaseLoadBalancer.Cluster cluster; BaseLoadBalancer.Cluster cluster;
cluster = new BaseLoadBalancer.Cluster(master, clusterState, null, null, null, null); cluster = new BaseLoadBalancer.Cluster(clusterState, null, null, null);
costFunction.init(cluster); costFunction.init(cluster);
double costWithoutReplicas = costFunction.cost(); double costWithoutReplicas = costFunction.cost();
assertEquals(0, costWithoutReplicas, 0); assertEquals(0, costWithoutReplicas, 0);
@ -342,7 +341,7 @@ public class TestStochasticLoadBalancer extends BalancerTestBase {
clusterState.firstEntry().getValue().get(0),1); clusterState.firstEntry().getValue().get(0),1);
clusterState.lastEntry().getValue().add(replica1); clusterState.lastEntry().getValue().add(replica1);
cluster = new BaseLoadBalancer.Cluster(master, clusterState, null, null, null, null); cluster = new BaseLoadBalancer.Cluster(clusterState, null, null, null);
costFunction.init(cluster); costFunction.init(cluster);
double costWith1ReplicaDifferentServer = costFunction.cost(); double costWith1ReplicaDifferentServer = costFunction.cost();
@ -352,7 +351,7 @@ public class TestStochasticLoadBalancer extends BalancerTestBase {
HRegionInfo replica2 = RegionReplicaUtil.getRegionInfoForReplica(replica1, 2); HRegionInfo replica2 = RegionReplicaUtil.getRegionInfoForReplica(replica1, 2);
clusterState.lastEntry().getValue().add(replica2); clusterState.lastEntry().getValue().add(replica2);
cluster = new BaseLoadBalancer.Cluster(master, clusterState, null, null, null, null); cluster = new BaseLoadBalancer.Cluster(clusterState, null, null, null);
costFunction.init(cluster); costFunction.init(cluster);
double costWith1ReplicaSameServer = costFunction.cost(); double costWith1ReplicaSameServer = costFunction.cost();
@ -375,7 +374,7 @@ public class TestStochasticLoadBalancer extends BalancerTestBase {
entry.getValue().add(replica2); entry.getValue().add(replica2);
it.next().getValue().add(replica3); //2nd server it.next().getValue().add(replica3); //2nd server
cluster = new BaseLoadBalancer.Cluster(master, clusterState, null, null, null, null); cluster = new BaseLoadBalancer.Cluster(clusterState, null, null, null);
costFunction.init(cluster); costFunction.init(cluster);
double costWith3ReplicasSameServer = costFunction.cost(); double costWith3ReplicasSameServer = costFunction.cost();
@ -389,7 +388,7 @@ public class TestStochasticLoadBalancer extends BalancerTestBase {
clusterState.lastEntry().getValue().add(replica2); clusterState.lastEntry().getValue().add(replica2);
clusterState.lastEntry().getValue().add(replica3); clusterState.lastEntry().getValue().add(replica3);
cluster = new BaseLoadBalancer.Cluster(master, clusterState, null, null, null, null); cluster = new BaseLoadBalancer.Cluster(clusterState, null, null, null);
costFunction.init(cluster); costFunction.init(cluster);
double costWith2ReplicasOnTwoServers = costFunction.cost(); double costWith2ReplicasOnTwoServers = costFunction.cost();
@ -409,7 +408,7 @@ public class TestStochasticLoadBalancer extends BalancerTestBase {
// until the step above s1 holds two replicas of a region // until the step above s1 holds two replicas of a region
regions = randomRegions(1); regions = randomRegions(1);
map.put(s2, regions); map.put(s2, regions);
assertTrue(loadBalancer.needsBalance(new Cluster(master, map, null, null, null, null))); assertTrue(loadBalancer.needsBalance(new Cluster(map, null, null, null)));
// check for the case where there are two hosts on the same rack and there are two racks // check for the case where there are two hosts on the same rack and there are two racks
// and both the replicas are on the same rack // and both the replicas are on the same rack
map.clear(); map.clear();
@ -420,7 +419,7 @@ public class TestStochasticLoadBalancer extends BalancerTestBase {
map.put(s2, regionsOnS2); map.put(s2, regionsOnS2);
// add another server so that the cluster has some host on another rack // add another server so that the cluster has some host on another rack
map.put(ServerName.valueOf("host2", 1000, 11111), randomRegions(1)); map.put(ServerName.valueOf("host2", 1000, 11111), randomRegions(1));
assertTrue(loadBalancer.needsBalance(new Cluster(master, map, null, null, null, assertTrue(loadBalancer.needsBalance(new Cluster(map, null, null,
new ForTestRackManagerOne()))); new ForTestRackManagerOne())));
} }