HBASE-2896 Retain assignment information between cluster shutdown/startup

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1032166 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Jonathan Gray 2010-11-06 22:11:19 +00:00
parent ce7567c23b
commit ea5395d62b
5 changed files with 170 additions and 19 deletions

View File

@ -1172,6 +1172,8 @@ Release 0.21.0 - Unreleased
HBASE-2201 JRuby shell for replication HBASE-2201 JRuby shell for replication
HBASE-2946 Increment multiple columns in a row at once HBASE-2946 Increment multiple columns in a row at once
HBASE-3013 Tool to verify data in two clusters HBASE-3013 Tool to verify data in two clusters
HBASE-2896 Retain assignment information between cluster
shutdown/startup
OPTIMIZATIONS OPTIMIZATIONS

View File

@ -24,7 +24,9 @@ import java.util.ArrayList;
import java.util.List; import java.util.List;
import java.util.Map; import java.util.Map;
import java.util.NavigableMap; import java.util.NavigableMap;
import java.util.Set;
import java.util.TreeMap; import java.util.TreeMap;
import java.util.TreeSet;
import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HRegionInfo;
@ -128,6 +130,25 @@ public class MetaReader {
*/ */
public static Map<HRegionInfo,HServerAddress> fullScan( public static Map<HRegionInfo,HServerAddress> fullScan(
CatalogTracker catalogTracker) CatalogTracker catalogTracker)
throws IOException {
return fullScan(catalogTracker, new TreeSet<String>());
}
/**
* Performs a full scan of <code>.META.</code>, skipping regions from any
* tables in the specified set of disabled tables.
* <p>
* Returns a map of every region to it's currently assigned server, according
* to META. If the region does not have an assignment it will have a null
* value in the map.
*
* @param catalogTracker
* @param disabledTables set of disabled tables that will not be returned
* @return map of regions to their currently assigned server
* @throws IOException
*/
public static Map<HRegionInfo,HServerAddress> fullScan(
CatalogTracker catalogTracker, final Set<String> disabledTables)
throws IOException { throws IOException {
final Map<HRegionInfo,HServerAddress> regions = final Map<HRegionInfo,HServerAddress> regions =
new TreeMap<HRegionInfo,HServerAddress>(); new TreeMap<HRegionInfo,HServerAddress>();
@ -137,6 +158,8 @@ public class MetaReader {
if (r == null || r.isEmpty()) return true; if (r == null || r.isEmpty()) return true;
Pair<HRegionInfo,HServerAddress> region = metaRowToRegionPair(r); Pair<HRegionInfo,HServerAddress> region = metaRowToRegionPair(r);
if (region == null) return true; if (region == null) return true;
if (disabledTables.contains(
region.getFirst().getTableDesc().getNameAsString())) return true;
regions.put(region.getFirst(), region.getSecond()); regions.put(region.getFirst(), region.getSecond());
return true; return true;
} }

View File

@ -46,6 +46,7 @@ import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.Chore; import org.apache.hadoop.hbase.Chore;
import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HRegionInfo;
import org.apache.hadoop.hbase.HServerAddress;
import org.apache.hadoop.hbase.HServerInfo; import org.apache.hadoop.hbase.HServerInfo;
import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.NotServingRegionException; import org.apache.hadoop.hbase.NotServingRegionException;
@ -54,7 +55,6 @@ import org.apache.hadoop.hbase.Stoppable;
import org.apache.hadoop.hbase.catalog.CatalogTracker; import org.apache.hadoop.hbase.catalog.CatalogTracker;
import org.apache.hadoop.hbase.catalog.MetaReader; import org.apache.hadoop.hbase.catalog.MetaReader;
import org.apache.hadoop.hbase.catalog.RootLocationEditor; import org.apache.hadoop.hbase.catalog.RootLocationEditor;
import org.apache.hadoop.hbase.client.MetaScanner;
import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Result;
import org.apache.hadoop.hbase.executor.ExecutorService; import org.apache.hadoop.hbase.executor.ExecutorService;
import org.apache.hadoop.hbase.executor.RegionTransitionData; import org.apache.hadoop.hbase.executor.RegionTransitionData;
@ -1106,23 +1106,33 @@ public class AssignmentManager extends ZooKeeperListener {
* should be shutdown. * should be shutdown.
*/ */
public void assignAllUserRegions() throws IOException { public void assignAllUserRegions() throws IOException {
// First experiment at synchronous assignment
// Simpler because just wait for no regions in transition
// Scan META for all user regions; do not include offlined regions in list. Map<HServerInfo, List<HRegionInfo>> bulkPlan = null;
List<HRegionInfo> allRegions =
MetaScanner.listAllRegions(master.getConfiguration(), false);
if (allRegions == null || allRegions.isEmpty()) return;
// Get all available servers // Get all available servers
List<HServerInfo> servers = serverManager.getOnlineServersList(); List<HServerInfo> servers = serverManager.getOnlineServersList();
// Scan META for all user regions, skipping any disabled tables
Map<HRegionInfo,HServerAddress> allRegions =
MetaReader.fullScan(catalogTracker, disabledTables);
if (allRegions == null || allRegions.isEmpty()) return;
// Determine what type of assignment to do on startup
boolean retainAssignment = master.getConfiguration().getBoolean(
"hbase.master.startup.retainassign", true);
if (retainAssignment) {
// Reuse existing assignment info
bulkPlan = LoadBalancer.retainAssignment(allRegions, servers);
} else {
// Generate a round-robin bulk assignment plan
bulkPlan = LoadBalancer.roundRobinAssignment(
new ArrayList<HRegionInfo>(allRegions.keySet()), servers);
}
LOG.info("Bulk assigning " + allRegions.size() + " region(s) across " + LOG.info("Bulk assigning " + allRegions.size() + " region(s) across " +
servers.size() + " server(s)"); servers.size() + " server(s)");
// Generate a cluster startup region placement plan
Map<HServerInfo, List<HRegionInfo>> bulkPlan =
LoadBalancer.bulkAssignment(allRegions, servers);
// Make a fixed thread count pool to run bulk assignments. Thought is that // Make a fixed thread count pool to run bulk assignments. Thought is that
// if a 1k cluster, running 1k bulk concurrent assignment threads will kill // if a 1k cluster, running 1k bulk concurrent assignment threads will kill
// master, HDFS or ZK? // master, HDFS or ZK?

View File

@ -36,6 +36,7 @@ import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HRegionInfo;
import org.apache.hadoop.hbase.HServerAddress;
import org.apache.hadoop.hbase.HServerInfo; import org.apache.hadoop.hbase.HServerInfo;
/** /**
@ -51,7 +52,7 @@ import org.apache.hadoop.hbase.HServerInfo;
* *
* <p>On cluster startup, {@link #bulkAssignment} can be used to determine * <p>On cluster startup, {@link #bulkAssignment} can be used to determine
* locations for all Regions in a cluster. * locations for all Regions in a cluster.
* *
* <p>This classes produces plans for the {@link AssignmentManager} to execute. * <p>This classes produces plans for the {@link AssignmentManager} to execute.
*/ */
public class LoadBalancer { public class LoadBalancer {
@ -337,11 +338,12 @@ public class LoadBalancer {
} }
/** /**
* Generates a bulk assignment plan to be used on cluster startup. * Generates a bulk assignment plan to be used on cluster startup using a
* * simple round-robin assignment.
* <p>
* Takes a list of all the regions and all the servers in the cluster and * Takes a list of all the regions and all the servers in the cluster and
* returns a map of each server to the regions that it should be assigned. * returns a map of each server to the regions that it should be assigned.
* * <p>
* Currently implemented as a round-robin assignment. Same invariant as * Currently implemented as a round-robin assignment. Same invariant as
* load balancing, all servers holding floor(avg) or ceiling(avg). * load balancing, all servers holding floor(avg) or ceiling(avg).
* *
@ -352,7 +354,7 @@ public class LoadBalancer {
* @return map of server to the regions it should take, or null if no * @return map of server to the regions it should take, or null if no
* assignment is possible (ie. no regions or no servers) * assignment is possible (ie. no regions or no servers)
*/ */
public static Map<HServerInfo,List<HRegionInfo>> bulkAssignment( public static Map<HServerInfo,List<HRegionInfo>> roundRobinAssignment(
List<HRegionInfo> regions, List<HServerInfo> servers) { List<HRegionInfo> regions, List<HServerInfo> servers) {
if(regions.size() == 0 || servers.size() == 0) { if(regions.size() == 0 || servers.size() == 0) {
return null; return null;
@ -374,6 +376,45 @@ public class LoadBalancer {
return assignments; return assignments;
} }
/**
* Generates a bulk assignment startup plan, attempting to reuse the existing
* assignment information from META, but adjusting for the specified list of
* available/online servers available for assignment.
* <p>
* Takes a map of all regions to their existing assignment from META. Also
* takes a list of online servers for regions to be assigned to. Attempts to
* retain all assignment, so in some instances initial assignment will not be
* completely balanced.
* <p>
* Any leftover regions without an existing server to be assigned to will be
* assigned randomly to available servers.
* @param regions regions and existing assignment from meta
* @param servers available servers
* @return map of servers and regions to be assigned to them
*/
public static Map<HServerInfo, List<HRegionInfo>> retainAssignment(
Map<HRegionInfo, HServerAddress> regions, List<HServerInfo> servers) {
Map<HServerInfo, List<HRegionInfo>> assignments =
new TreeMap<HServerInfo, List<HRegionInfo>>();
// Build a map of server addresses to server info so we can match things up
Map<HServerAddress, HServerInfo> serverMap =
new TreeMap<HServerAddress, HServerInfo>();
for (HServerInfo server : servers) {
serverMap.put(server.getServerAddress(), server);
assignments.put(server, new ArrayList<HRegionInfo>());
}
for (Map.Entry<HRegionInfo, HServerAddress> region : regions.entrySet()) {
HServerInfo server = serverMap.get(region.getValue());
if (server != null) {
assignments.get(server).add(region.getKey());
} else {
assignments.get(servers.get(rand.nextInt(assignments.size()))).add(
region.getKey());
}
}
return assignments;
}
/** /**
* Find the block locations for all of the files for the specified region. * Find the block locations for all of the files for the specified region.
* *
@ -573,7 +614,7 @@ public class LoadBalancer {
public String getRegionName() { public String getRegionName() {
return this.hri.getEncodedName(); return this.hri.getEncodedName();
} }
public HRegionInfo getRegionInfo() { public HRegionInfo getRegionInfo() {
return this.hri; return this.hri;
} }

View File

@ -19,6 +19,7 @@
*/ */
package org.apache.hadoop.hbase.master; package org.apache.hadoop.hbase.master;
import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import java.util.ArrayList; import java.util.ArrayList;
@ -28,11 +29,11 @@ import java.util.List;
import java.util.Map; import java.util.Map;
import java.util.Queue; import java.util.Queue;
import java.util.Random; import java.util.Random;
import java.util.Set;
import java.util.SortedSet; import java.util.SortedSet;
import java.util.TreeMap; import java.util.TreeMap;
import java.util.TreeSet; import java.util.TreeSet;
import org.apache.commons.lang.RandomStringUtils;
import org.apache.commons.logging.Log; import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HRegionInfo;
@ -239,7 +240,7 @@ public class TestLoadBalancer {
List<HRegionInfo> regions = randomRegions(mock[0]); List<HRegionInfo> regions = randomRegions(mock[0]);
List<HServerInfo> servers = randomServers(mock[1], 0); List<HServerInfo> servers = randomServers(mock[1], 0);
Map<HServerInfo,List<HRegionInfo>> assignments = Map<HServerInfo,List<HRegionInfo>> assignments =
LoadBalancer.bulkAssignment(regions, servers); LoadBalancer.roundRobinAssignment(regions, servers);
float average = (float)regions.size()/servers.size(); float average = (float)regions.size()/servers.size();
int min = (int)Math.floor(average); int min = (int)Math.floor(average);
int max = (int)Math.ceil(average); int max = (int)Math.ceil(average);
@ -253,6 +254,80 @@ public class TestLoadBalancer {
} }
} }
/**
* Test the cluster startup bulk assignment which attempts to retain
* assignment info.
* @throws Exception
*/
@Test
public void testRetainAssignment() throws Exception {
// Test simple case where all same servers are there
List<HServerInfo> servers = randomServers(10, 10);
List<HRegionInfo> regions = randomRegions(100);
Map<HRegionInfo, HServerAddress> existing =
new TreeMap<HRegionInfo, HServerAddress>();
for (int i=0;i<regions.size();i++) {
existing.put(regions.get(i),
servers.get(i % servers.size()).getServerAddress());
}
Map<HServerInfo, List<HRegionInfo>> assignment =
LoadBalancer.retainAssignment(existing, servers);
assertRetainedAssignment(existing, servers, assignment);
// Include two new servers that were not there before
List<HServerInfo> servers2 = new ArrayList<HServerInfo>(servers);
servers2.add(randomServer(10));
servers2.add(randomServer(10));
assignment = LoadBalancer.retainAssignment(existing, servers2);
assertRetainedAssignment(existing, servers2, assignment);
// Remove two of the servers that were previously there
List<HServerInfo> servers3 = new ArrayList<HServerInfo>(servers);
servers3.remove(servers3.size()-1);
servers3.remove(servers3.size()-2);
assignment = LoadBalancer.retainAssignment(existing, servers3);
assertRetainedAssignment(existing, servers3, assignment);
}
/**
* Asserts a valid retained assignment plan.
* <p>
* Must meet the following conditions:
* <ul>
* <li>Every input region has an assignment, and to an online server
* <li>If a region had an existing assignment to a server with the same
* address a a currently online server, it will be assigned to it
* </ul>
* @param existing
* @param servers
* @param assignment
*/
private void assertRetainedAssignment(
Map<HRegionInfo, HServerAddress> existing, List<HServerInfo> servers,
Map<HServerInfo, List<HRegionInfo>> assignment) {
// Verify condition 1, every region assigned, and to online server
Set<HServerInfo> onlineServerSet = new TreeSet<HServerInfo>(servers);
Set<HRegionInfo> assignedRegions = new TreeSet<HRegionInfo>();
for (Map.Entry<HServerInfo, List<HRegionInfo>> a : assignment.entrySet()) {
assertTrue("Region assigned to server that was not listed as online",
onlineServerSet.contains(a.getKey()));
for (HRegionInfo r : a.getValue()) assignedRegions.add(r);
}
assertEquals(existing.size(), assignedRegions.size());
// Verify condition 2, if server had existing assignment, must have same
Set<HServerAddress> onlineAddresses = new TreeSet<HServerAddress>();
for (HServerInfo s : servers) onlineAddresses.add(s.getServerAddress());
for (Map.Entry<HServerInfo, List<HRegionInfo>> a : assignment.entrySet()) {
for (HRegionInfo r : a.getValue()) {
HServerAddress address = existing.get(r);
if (address != null && onlineAddresses.contains(address)) {
assertTrue(a.getKey().getServerAddress().equals(address));
}
}
}
}
private String printStats(Map<HServerInfo, List<HRegionInfo>> servers) { private String printStats(Map<HServerInfo, List<HRegionInfo>> servers) {
int numServers = servers.size(); int numServers = servers.size();
int totalRegions = 0; int totalRegions = 0;