HBASE-25216 The client zk syncer should deal with meta replica count change (#2614)
Signed-off-by: Yu Li <liyu@apache.org>
This commit is contained in:
parent
f37cd05c32
commit
49774c7e18
|
@ -322,8 +322,9 @@ public class HMaster extends HRegionServer implements MasterServices {
|
||||||
// Tracker for load balancer state
|
// Tracker for load balancer state
|
||||||
LoadBalancerTracker loadBalancerTracker;
|
LoadBalancerTracker loadBalancerTracker;
|
||||||
// Tracker for meta location, if any client ZK quorum specified
|
// Tracker for meta location, if any client ZK quorum specified
|
||||||
MetaLocationSyncer metaLocationSyncer;
|
private MetaLocationSyncer metaLocationSyncer;
|
||||||
// Tracker for active master location, if any client ZK quorum specified
|
// Tracker for active master location, if any client ZK quorum specified
|
||||||
|
@VisibleForTesting
|
||||||
MasterAddressSyncer masterAddressSyncer;
|
MasterAddressSyncer masterAddressSyncer;
|
||||||
// Tracker for auto snapshot cleanup state
|
// Tracker for auto snapshot cleanup state
|
||||||
SnapshotCleanupTracker snapshotCleanupTracker;
|
SnapshotCleanupTracker snapshotCleanupTracker;
|
||||||
|
@ -3852,4 +3853,9 @@ public class HMaster extends HRegionServer implements MasterServices {
|
||||||
}
|
}
|
||||||
return compactionState;
|
return compactionState;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public MetaLocationSyncer getMetaLocationSyncer() {
|
||||||
|
return metaLocationSyncer;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -40,6 +40,7 @@ import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
||||||
import org.apache.hadoop.hbase.master.replication.ReplicationPeerManager;
|
import org.apache.hadoop.hbase.master.replication.ReplicationPeerManager;
|
||||||
import org.apache.hadoop.hbase.master.replication.SyncReplicationReplayWALManager;
|
import org.apache.hadoop.hbase.master.replication.SyncReplicationReplayWALManager;
|
||||||
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
||||||
|
import org.apache.hadoop.hbase.master.zksyncer.MetaLocationSyncer;
|
||||||
import org.apache.hadoop.hbase.procedure.MasterProcedureManagerHost;
|
import org.apache.hadoop.hbase.procedure.MasterProcedureManagerHost;
|
||||||
import org.apache.hadoop.hbase.procedure2.LockedResource;
|
import org.apache.hadoop.hbase.procedure2.LockedResource;
|
||||||
import org.apache.hadoop.hbase.procedure2.Procedure;
|
import org.apache.hadoop.hbase.procedure2.Procedure;
|
||||||
|
@ -570,4 +571,11 @@ public interface MasterServices extends Server {
|
||||||
*/
|
*/
|
||||||
boolean normalizeRegions(
|
boolean normalizeRegions(
|
||||||
final NormalizeTableFilterParams ntfp, final boolean isHighPriority) throws IOException;
|
final NormalizeTableFilterParams ntfp, final boolean isHighPriority) throws IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Get the meta location syncer.
|
||||||
|
* <p/>
|
||||||
|
* We need to get this in MTP to tell the syncer the new meta replica count.
|
||||||
|
*/
|
||||||
|
MetaLocationSyncer getMetaLocationSyncer();
|
||||||
}
|
}
|
||||||
|
|
|
@ -36,6 +36,7 @@ import org.apache.hadoop.hbase.client.RegionInfo;
|
||||||
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
||||||
import org.apache.hadoop.hbase.client.TableDescriptor;
|
import org.apache.hadoop.hbase.client.TableDescriptor;
|
||||||
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
||||||
|
import org.apache.hadoop.hbase.master.zksyncer.MetaLocationSyncer;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer;
|
import org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationException;
|
import org.apache.hadoop.hbase.replication.ReplicationException;
|
||||||
import org.apache.hadoop.hbase.rsgroup.RSGroupInfo;
|
import org.apache.hadoop.hbase.rsgroup.RSGroupInfo;
|
||||||
|
@ -157,6 +158,12 @@ public class ModifyTableProcedure
|
||||||
break;
|
break;
|
||||||
case MODIFY_TABLE_ASSIGN_NEW_REPLICAS:
|
case MODIFY_TABLE_ASSIGN_NEW_REPLICAS:
|
||||||
assignNewReplicasIfNeeded(env);
|
assignNewReplicasIfNeeded(env);
|
||||||
|
if (TableName.isMetaTableName(getTableName())) {
|
||||||
|
MetaLocationSyncer syncer = env.getMasterServices().getMetaLocationSyncer();
|
||||||
|
if (syncer != null) {
|
||||||
|
syncer.setMetaReplicaCount(modifiedTableDescriptor.getRegionReplication());
|
||||||
|
}
|
||||||
|
}
|
||||||
if (deleteColumnFamilyInModify) {
|
if (deleteColumnFamilyInModify) {
|
||||||
setNextState(ModifyTableState.MODIFY_TABLE_DELETE_FS_LAYOUT);
|
setNextState(ModifyTableState.MODIFY_TABLE_DELETE_FS_LAYOUT);
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -19,12 +19,11 @@
|
||||||
package org.apache.hadoop.hbase.master.zksyncer;
|
package org.apache.hadoop.hbase.master.zksyncer;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.Collection;
|
import java.util.Iterator;
|
||||||
import java.util.HashMap;
|
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
import java.util.concurrent.ArrayBlockingQueue;
|
import java.util.Set;
|
||||||
import java.util.concurrent.BlockingQueue;
|
import java.util.concurrent.ConcurrentHashMap;
|
||||||
|
import java.util.concurrent.ConcurrentMap;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.Server;
|
import org.apache.hadoop.hbase.Server;
|
||||||
import org.apache.hadoop.hbase.util.Threads;
|
import org.apache.hadoop.hbase.util.Threads;
|
||||||
|
@ -34,7 +33,6 @@ import org.apache.hadoop.hbase.zookeeper.ZKWatcher;
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
import org.apache.zookeeper.CreateMode;
|
import org.apache.zookeeper.CreateMode;
|
||||||
import org.apache.zookeeper.KeeperException;
|
import org.apache.zookeeper.KeeperException;
|
||||||
|
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
@ -42,22 +40,68 @@ import org.slf4j.LoggerFactory;
|
||||||
* Tracks the target znode(s) on server ZK cluster and synchronize them to client ZK cluster if
|
* Tracks the target znode(s) on server ZK cluster and synchronize them to client ZK cluster if
|
||||||
* changed
|
* changed
|
||||||
* <p/>
|
* <p/>
|
||||||
* The target znode(s) is given through {@link #getNodesToWatch()} method
|
* The target znode(s) is given through {@link #getPathsToWatch()} method
|
||||||
*/
|
*/
|
||||||
@InterfaceAudience.Private
|
@InterfaceAudience.Private
|
||||||
public abstract class ClientZKSyncer extends ZKListener {
|
public abstract class ClientZKSyncer extends ZKListener {
|
||||||
private static final Logger LOG = LoggerFactory.getLogger(ClientZKSyncer.class);
|
private static final Logger LOG = LoggerFactory.getLogger(ClientZKSyncer.class);
|
||||||
private final Server server;
|
private final Server server;
|
||||||
private final ZKWatcher clientZkWatcher;
|
private final ZKWatcher clientZkWatcher;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Used to store the newest data which we want to sync to client zk.
|
||||||
|
* <p/>
|
||||||
|
* For meta location, since we may reduce the replica number, so here we add a {@code delete} flag
|
||||||
|
* to tell the updater delete the znode on client zk and quit.
|
||||||
|
*/
|
||||||
|
private static final class ZKData {
|
||||||
|
|
||||||
|
byte[] data;
|
||||||
|
|
||||||
|
boolean delete = false;
|
||||||
|
|
||||||
|
synchronized void set(byte[] data) {
|
||||||
|
this.data = data;
|
||||||
|
notifyAll();
|
||||||
|
}
|
||||||
|
|
||||||
|
synchronized byte[] get() throws InterruptedException {
|
||||||
|
while (!delete && data == null) {
|
||||||
|
wait();
|
||||||
|
}
|
||||||
|
byte[] d = data;
|
||||||
|
data = null;
|
||||||
|
return d;
|
||||||
|
}
|
||||||
|
|
||||||
|
synchronized void delete() {
|
||||||
|
this.delete = true;
|
||||||
|
notifyAll();
|
||||||
|
}
|
||||||
|
|
||||||
|
synchronized boolean isDeleted() {
|
||||||
|
return delete;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// We use queues and daemon threads to synchronize the data to client ZK cluster
|
// We use queues and daemon threads to synchronize the data to client ZK cluster
|
||||||
// to avoid blocking the single event thread for watchers
|
// to avoid blocking the single event thread for watchers
|
||||||
private final Map<String, BlockingQueue<byte[]>> queues;
|
private final ConcurrentMap<String, ZKData> queues;
|
||||||
|
|
||||||
public ClientZKSyncer(ZKWatcher watcher, ZKWatcher clientZkWatcher, Server server) {
|
public ClientZKSyncer(ZKWatcher watcher, ZKWatcher clientZkWatcher, Server server) {
|
||||||
super(watcher);
|
super(watcher);
|
||||||
this.server = server;
|
this.server = server;
|
||||||
this.clientZkWatcher = clientZkWatcher;
|
this.clientZkWatcher = clientZkWatcher;
|
||||||
this.queues = new HashMap<>();
|
this.queues = new ConcurrentHashMap<>();
|
||||||
|
}
|
||||||
|
|
||||||
|
private void startNewSyncThread(String path) {
|
||||||
|
ZKData zkData = new ZKData();
|
||||||
|
queues.put(path, zkData);
|
||||||
|
Thread updater = new ClientZkUpdater(path, zkData);
|
||||||
|
updater.setDaemon(true);
|
||||||
|
updater.start();
|
||||||
|
watchAndCheckExists(path);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -69,17 +113,12 @@ public abstract class ClientZKSyncer extends ZKListener {
|
||||||
this.watcher.registerListener(this);
|
this.watcher.registerListener(this);
|
||||||
// create base znode on remote ZK
|
// create base znode on remote ZK
|
||||||
ZKUtil.createWithParents(clientZkWatcher, watcher.getZNodePaths().baseZNode);
|
ZKUtil.createWithParents(clientZkWatcher, watcher.getZNodePaths().baseZNode);
|
||||||
// set meta znodes for client ZK
|
// set znodes for client ZK
|
||||||
Collection<String> nodes = getNodesToWatch();
|
Set<String> paths = getPathsToWatch();
|
||||||
LOG.debug("Znodes to watch: " + nodes);
|
LOG.debug("ZNodes to watch: {}", paths);
|
||||||
// initialize queues and threads
|
// initialize queues and threads
|
||||||
for (String node : nodes) {
|
for (String path : paths) {
|
||||||
BlockingQueue<byte[]> queue = new ArrayBlockingQueue<>(1);
|
startNewSyncThread(path);
|
||||||
queues.put(node, queue);
|
|
||||||
Thread updater = new ClientZkUpdater(node, queue);
|
|
||||||
updater.setDaemon(true);
|
|
||||||
updater.start();
|
|
||||||
watchAndCheckExists(node);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -112,10 +151,9 @@ public abstract class ClientZKSyncer extends ZKListener {
|
||||||
* @param data the data to write to queue
|
* @param data the data to write to queue
|
||||||
*/
|
*/
|
||||||
private void upsertQueue(String node, byte[] data) {
|
private void upsertQueue(String node, byte[] data) {
|
||||||
BlockingQueue<byte[]> queue = queues.get(node);
|
ZKData zkData = queues.get(node);
|
||||||
synchronized (queue) {
|
if (zkData != null) {
|
||||||
queue.poll();
|
zkData.set(data);
|
||||||
queue.offer(data);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -126,35 +164,49 @@ public abstract class ClientZKSyncer extends ZKListener {
|
||||||
* @param data the data to set to client ZK
|
* @param data the data to set to client ZK
|
||||||
* @throws InterruptedException if the thread is interrupted during process
|
* @throws InterruptedException if the thread is interrupted during process
|
||||||
*/
|
*/
|
||||||
private final void setDataForClientZkUntilSuccess(String node, byte[] data)
|
private void setDataForClientZkUntilSuccess(String node, byte[] data)
|
||||||
throws InterruptedException {
|
throws InterruptedException {
|
||||||
|
boolean create = false;
|
||||||
while (!server.isStopped()) {
|
while (!server.isStopped()) {
|
||||||
try {
|
try {
|
||||||
LOG.debug("Set data for remote " + node + ", client zk wather: " + clientZkWatcher);
|
LOG.debug("Set data for remote " + node + ", client zk wather: " + clientZkWatcher);
|
||||||
ZKUtil.setData(clientZkWatcher, node, data);
|
if (create) {
|
||||||
break;
|
|
||||||
} catch (KeeperException.NoNodeException nne) {
|
|
||||||
// Node doesn't exist, create it and set value
|
|
||||||
try {
|
|
||||||
ZKUtil.createNodeIfNotExistsNoWatch(clientZkWatcher, node, data, CreateMode.PERSISTENT);
|
ZKUtil.createNodeIfNotExistsNoWatch(clientZkWatcher, node, data, CreateMode.PERSISTENT);
|
||||||
break;
|
} else {
|
||||||
} catch (KeeperException.ConnectionLossException
|
ZKUtil.setData(clientZkWatcher, node, data);
|
||||||
| KeeperException.SessionExpiredException ee) {
|
|
||||||
reconnectAfterExpiration();
|
|
||||||
} catch (KeeperException e) {
|
|
||||||
LOG.warn(
|
|
||||||
"Failed to create znode " + node + " due to: " + e.getMessage() + ", will retry later");
|
|
||||||
}
|
}
|
||||||
} catch (KeeperException.ConnectionLossException
|
break;
|
||||||
| KeeperException.SessionExpiredException ee) {
|
|
||||||
reconnectAfterExpiration();
|
|
||||||
} catch (KeeperException e) {
|
} catch (KeeperException e) {
|
||||||
LOG.debug("Failed to set data to client ZK, will retry later", e);
|
LOG.debug("Failed to set data for {} to client ZK, will retry later", node, e);
|
||||||
|
if (e.code() == KeeperException.Code.SESSIONEXPIRED) {
|
||||||
|
reconnectAfterExpiration();
|
||||||
|
}
|
||||||
|
if (e.code() == KeeperException.Code.NONODE) {
|
||||||
|
create = true;
|
||||||
|
}
|
||||||
|
if (e.code() == KeeperException.Code.NODEEXISTS) {
|
||||||
|
create = false;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
Threads.sleep(HConstants.SOCKET_RETRY_WAIT_MS);
|
Threads.sleep(HConstants.SOCKET_RETRY_WAIT_MS);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private void deleteDataForClientZkUntilSuccess(String node) throws InterruptedException {
|
||||||
|
while (!server.isStopped()) {
|
||||||
|
LOG.debug("Delete remote " + node + ", client zk wather: " + clientZkWatcher);
|
||||||
|
try {
|
||||||
|
ZKUtil.deleteNode(clientZkWatcher, node);
|
||||||
|
} catch (KeeperException e) {
|
||||||
|
LOG.debug("Failed to delete node from client ZK, will retry later", e);
|
||||||
|
if (e.code() == KeeperException.Code.SESSIONEXPIRED) {
|
||||||
|
reconnectAfterExpiration();
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private final void reconnectAfterExpiration() throws InterruptedException {
|
private final void reconnectAfterExpiration() throws InterruptedException {
|
||||||
LOG.warn("ZK session expired or lost. Retry a new connection...");
|
LOG.warn("ZK session expired or lost. Retry a new connection...");
|
||||||
try {
|
try {
|
||||||
|
@ -164,11 +216,7 @@ public abstract class ClientZKSyncer extends ZKListener {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
private void getDataAndWatch(String path) {
|
||||||
public void nodeCreated(String path) {
|
|
||||||
if (!validate(path)) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
try {
|
try {
|
||||||
byte[] data = ZKUtil.getDataAndWatch(watcher, path);
|
byte[] data = ZKUtil.getDataAndWatch(watcher, path);
|
||||||
upsertQueue(path, data);
|
upsertQueue(path, data);
|
||||||
|
@ -177,11 +225,25 @@ public abstract class ClientZKSyncer extends ZKListener {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private void removeQueue(String path) {
|
||||||
|
ZKData zkData = queues.remove(path);
|
||||||
|
if (zkData != null) {
|
||||||
|
zkData.delete();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void nodeCreated(String path) {
|
||||||
|
if (validate(path)) {
|
||||||
|
getDataAndWatch(path);
|
||||||
|
} else {
|
||||||
|
removeQueue(path);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void nodeDataChanged(String path) {
|
public void nodeDataChanged(String path) {
|
||||||
if (validate(path)) {
|
nodeCreated(path);
|
||||||
nodeCreated(path);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -189,11 +251,13 @@ public abstract class ClientZKSyncer extends ZKListener {
|
||||||
if (validate(path)) {
|
if (validate(path)) {
|
||||||
try {
|
try {
|
||||||
if (ZKUtil.watchAndCheckExists(watcher, path)) {
|
if (ZKUtil.watchAndCheckExists(watcher, path)) {
|
||||||
nodeCreated(path);
|
getDataAndWatch(path);
|
||||||
}
|
}
|
||||||
} catch (KeeperException e) {
|
} catch (KeeperException e) {
|
||||||
LOG.warn("Unexpected exception handling nodeDeleted event for path: " + path, e);
|
LOG.warn("Unexpected exception handling nodeDeleted event for path: " + path, e);
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
removeQueue(path);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -202,41 +266,67 @@ public abstract class ClientZKSyncer extends ZKListener {
|
||||||
* @param path the path to validate
|
* @param path the path to validate
|
||||||
* @return true if the znode is watched by us
|
* @return true if the znode is watched by us
|
||||||
*/
|
*/
|
||||||
abstract boolean validate(String path);
|
protected abstract boolean validate(String path);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @return the znode(s) to watch
|
* @return the zk path(s) to watch
|
||||||
*/
|
*/
|
||||||
abstract Collection<String> getNodesToWatch() throws KeeperException;
|
protected abstract Set<String> getPathsToWatch();
|
||||||
|
|
||||||
|
protected final void refreshWatchingList() {
|
||||||
|
Set<String> newPaths = getPathsToWatch();
|
||||||
|
LOG.debug("New ZNodes to watch: {}", newPaths);
|
||||||
|
Iterator<Map.Entry<String, ZKData>> iter = queues.entrySet().iterator();
|
||||||
|
// stop unused syncers
|
||||||
|
while (iter.hasNext()) {
|
||||||
|
Map.Entry<String, ZKData> entry = iter.next();
|
||||||
|
if (!newPaths.contains(entry.getKey())) {
|
||||||
|
iter.remove();
|
||||||
|
entry.getValue().delete();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// start new syncers
|
||||||
|
for (String newPath : newPaths) {
|
||||||
|
if (!queues.containsKey(newPath)) {
|
||||||
|
startNewSyncThread(newPath);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Thread to synchronize znode data to client ZK cluster
|
* Thread to synchronize znode data to client ZK cluster
|
||||||
*/
|
*/
|
||||||
class ClientZkUpdater extends Thread {
|
private final class ClientZkUpdater extends Thread {
|
||||||
final String znode;
|
private final String znode;
|
||||||
final BlockingQueue<byte[]> queue;
|
private final ZKData zkData;
|
||||||
|
|
||||||
public ClientZkUpdater(String znode, BlockingQueue<byte[]> queue) {
|
public ClientZkUpdater(String znode, ZKData zkData) {
|
||||||
this.znode = znode;
|
this.znode = znode;
|
||||||
this.queue = queue;
|
this.zkData = zkData;
|
||||||
setName("ClientZKUpdater-" + znode);
|
setName("ClientZKUpdater-" + znode);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void run() {
|
public void run() {
|
||||||
|
LOG.debug("Client zk updater for znode {} started", znode);
|
||||||
while (!server.isStopped()) {
|
while (!server.isStopped()) {
|
||||||
try {
|
try {
|
||||||
byte[] data = queue.take();
|
byte[] data = zkData.get();
|
||||||
setDataForClientZkUntilSuccess(znode, data);
|
if (data != null) {
|
||||||
} catch (InterruptedException e) {
|
setDataForClientZkUntilSuccess(znode, data);
|
||||||
if (LOG.isDebugEnabled()) {
|
} else {
|
||||||
LOG.debug(
|
if (zkData.isDeleted()) {
|
||||||
"Interrupted while checking whether need to update meta location to client zk");
|
deleteDataForClientZkUntilSuccess(znode);
|
||||||
|
break;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
} catch (InterruptedException e) {
|
||||||
|
LOG.debug("Interrupted while checking whether need to update meta location to client zk");
|
||||||
Thread.currentThread().interrupt();
|
Thread.currentThread().interrupt();
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
LOG.debug("Client zk updater for znode {} stopped", znode);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -18,9 +18,8 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.master.zksyncer;
|
package org.apache.hadoop.hbase.master.zksyncer;
|
||||||
|
|
||||||
import java.util.ArrayList;
|
import java.util.Collections;
|
||||||
import java.util.Collection;
|
import java.util.Set;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.Server;
|
import org.apache.hadoop.hbase.Server;
|
||||||
import org.apache.hadoop.hbase.zookeeper.ZKWatcher;
|
import org.apache.hadoop.hbase.zookeeper.ZKWatcher;
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
|
@ -39,14 +38,12 @@ public class MasterAddressSyncer extends ClientZKSyncer {
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
boolean validate(String path) {
|
protected boolean validate(String path) {
|
||||||
return path.equals(masterAddressZNode);
|
return path.equals(masterAddressZNode);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
Collection<String> getNodesToWatch() {
|
protected Set<String> getPathsToWatch() {
|
||||||
ArrayList<String> toReturn = new ArrayList<>();
|
return Collections.singleton(masterAddressZNode);
|
||||||
toReturn.add(masterAddressZNode);
|
|
||||||
return toReturn;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -18,13 +18,12 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.master.zksyncer;
|
package org.apache.hadoop.hbase.master.zksyncer;
|
||||||
|
|
||||||
import java.util.Collection;
|
import java.util.Set;
|
||||||
import java.util.stream.Collectors;
|
import java.util.stream.Collectors;
|
||||||
|
import java.util.stream.IntStream;
|
||||||
import org.apache.hadoop.hbase.Server;
|
import org.apache.hadoop.hbase.Server;
|
||||||
import org.apache.hadoop.hbase.zookeeper.ZKWatcher;
|
import org.apache.hadoop.hbase.zookeeper.ZKWatcher;
|
||||||
import org.apache.hadoop.hbase.zookeeper.ZNodePaths;
|
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
import org.apache.zookeeper.KeeperException;
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tracks the meta region locations on server ZK cluster and synchronize them to client ZK cluster
|
* Tracks the meta region locations on server ZK cluster and synchronize them to client ZK cluster
|
||||||
|
@ -32,19 +31,28 @@ import org.apache.zookeeper.KeeperException;
|
||||||
*/
|
*/
|
||||||
@InterfaceAudience.Private
|
@InterfaceAudience.Private
|
||||||
public class MetaLocationSyncer extends ClientZKSyncer {
|
public class MetaLocationSyncer extends ClientZKSyncer {
|
||||||
|
|
||||||
|
private volatile int metaReplicaCount = 1;
|
||||||
|
|
||||||
public MetaLocationSyncer(ZKWatcher watcher, ZKWatcher clientZkWatcher, Server server) {
|
public MetaLocationSyncer(ZKWatcher watcher, ZKWatcher clientZkWatcher, Server server) {
|
||||||
super(watcher, clientZkWatcher, server);
|
super(watcher, clientZkWatcher, server);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
boolean validate(String path) {
|
protected boolean validate(String path) {
|
||||||
return watcher.getZNodePaths().isMetaZNodePath(path);
|
return watcher.getZNodePaths().isMetaZNodePath(path);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
Collection<String> getNodesToWatch() throws KeeperException {
|
protected Set<String> getPathsToWatch() {
|
||||||
return watcher.getMetaReplicaNodes().stream()
|
return IntStream.range(0, metaReplicaCount)
|
||||||
.map(znode -> ZNodePaths.joinZNode(watcher.getZNodePaths().baseZNode, znode))
|
.mapToObj(watcher.getZNodePaths()::getZNodeForReplica).collect(Collectors.toSet());
|
||||||
.collect(Collectors.toList());
|
}
|
||||||
|
|
||||||
|
public void setMetaReplicaCount(int replicaCount) {
|
||||||
|
if (replicaCount != metaReplicaCount) {
|
||||||
|
metaReplicaCount = replicaCount;
|
||||||
|
refreshWatchingList();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
|
@ -17,6 +17,10 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.client;
|
package org.apache.hadoop.hbase.client;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertArrayEquals;
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
|
||||||
import java.io.File;
|
import java.io.File;
|
||||||
import org.apache.commons.io.FileUtils;
|
import org.apache.commons.io.FileUtils;
|
||||||
import org.apache.hadoop.hbase.HBaseClassTestRule;
|
import org.apache.hadoop.hbase.HBaseClassTestRule;
|
||||||
|
@ -26,6 +30,7 @@ import org.apache.hadoop.hbase.MiniHBaseCluster;
|
||||||
import org.apache.hadoop.hbase.ServerName;
|
import org.apache.hadoop.hbase.ServerName;
|
||||||
import org.apache.hadoop.hbase.StartMiniClusterOption;
|
import org.apache.hadoop.hbase.StartMiniClusterOption;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
|
import org.apache.hadoop.hbase.TableNameTestRule;
|
||||||
import org.apache.hadoop.hbase.master.HMaster;
|
import org.apache.hadoop.hbase.master.HMaster;
|
||||||
import org.apache.hadoop.hbase.master.assignment.AssignmentTestingUtil;
|
import org.apache.hadoop.hbase.master.assignment.AssignmentTestingUtil;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
||||||
|
@ -35,13 +40,11 @@ import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.hbase.util.JVMClusterUtil.RegionServerThread;
|
import org.apache.hadoop.hbase.util.JVMClusterUtil.RegionServerThread;
|
||||||
import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster;
|
import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster;
|
||||||
import org.junit.AfterClass;
|
import org.junit.AfterClass;
|
||||||
import org.junit.Assert;
|
|
||||||
import org.junit.BeforeClass;
|
import org.junit.BeforeClass;
|
||||||
import org.junit.ClassRule;
|
import org.junit.ClassRule;
|
||||||
import org.junit.Rule;
|
import org.junit.Rule;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.junit.experimental.categories.Category;
|
import org.junit.experimental.categories.Category;
|
||||||
import org.junit.rules.TestName;
|
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
@ -61,11 +64,11 @@ public class TestSeparateClientZKCluster {
|
||||||
private final byte[] newVal = Bytes.toBytes("v2");
|
private final byte[] newVal = Bytes.toBytes("v2");
|
||||||
|
|
||||||
@Rule
|
@Rule
|
||||||
public TestName name = new TestName();
|
public TableNameTestRule name = new TableNameTestRule();
|
||||||
|
|
||||||
@ClassRule
|
@ClassRule
|
||||||
public static final HBaseClassTestRule CLASS_RULE =
|
public static final HBaseClassTestRule CLASS_RULE =
|
||||||
HBaseClassTestRule.forClass(TestSeparateClientZKCluster.class);
|
HBaseClassTestRule.forClass(TestSeparateClientZKCluster.class);
|
||||||
|
|
||||||
@BeforeClass
|
@BeforeClass
|
||||||
public static void beforeAllTests() throws Exception {
|
public static void beforeAllTests() throws Exception {
|
||||||
|
@ -78,13 +81,15 @@ public class TestSeparateClientZKCluster {
|
||||||
TEST_UTIL.getConfiguration().setInt("hbase.client.start.log.errors.counter", -1);
|
TEST_UTIL.getConfiguration().setInt("hbase.client.start.log.errors.counter", -1);
|
||||||
TEST_UTIL.getConfiguration().setInt("zookeeper.recovery.retry", 1);
|
TEST_UTIL.getConfiguration().setInt("zookeeper.recovery.retry", 1);
|
||||||
// core settings for testing client ZK cluster
|
// core settings for testing client ZK cluster
|
||||||
|
TEST_UTIL.getConfiguration().setClass(HConstants.CLIENT_CONNECTION_REGISTRY_IMPL_CONF_KEY,
|
||||||
|
ZKConnectionRegistry.class, ConnectionRegistry.class);
|
||||||
TEST_UTIL.getConfiguration().set(HConstants.CLIENT_ZOOKEEPER_QUORUM, HConstants.LOCALHOST);
|
TEST_UTIL.getConfiguration().set(HConstants.CLIENT_ZOOKEEPER_QUORUM, HConstants.LOCALHOST);
|
||||||
TEST_UTIL.getConfiguration().setInt(HConstants.CLIENT_ZOOKEEPER_CLIENT_PORT, clientZkPort);
|
TEST_UTIL.getConfiguration().setInt(HConstants.CLIENT_ZOOKEEPER_CLIENT_PORT, clientZkPort);
|
||||||
// reduce zk session timeout to easier trigger session expiration
|
// reduce zk session timeout to easier trigger session expiration
|
||||||
TEST_UTIL.getConfiguration().setInt(HConstants.ZK_SESSION_TIMEOUT, ZK_SESSION_TIMEOUT);
|
TEST_UTIL.getConfiguration().setInt(HConstants.ZK_SESSION_TIMEOUT, ZK_SESSION_TIMEOUT);
|
||||||
// Start a cluster with 2 masters and 3 regionservers.
|
// Start a cluster with 2 masters and 3 regionservers.
|
||||||
StartMiniClusterOption option = StartMiniClusterOption.builder()
|
StartMiniClusterOption option =
|
||||||
.numMasters(2).numRegionServers(3).numDataNodes(3).build();
|
StartMiniClusterOption.builder().numMasters(2).numRegionServers(3).numDataNodes(3).build();
|
||||||
TEST_UTIL.startMiniCluster(option);
|
TEST_UTIL.startMiniCluster(option);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -97,7 +102,7 @@ public class TestSeparateClientZKCluster {
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testBasicOperation() throws Exception {
|
public void testBasicOperation() throws Exception {
|
||||||
TableName tn = TableName.valueOf(name.getMethodName());
|
TableName tn = name.getTableName();
|
||||||
// create table
|
// create table
|
||||||
Connection conn = TEST_UTIL.getConnection();
|
Connection conn = TEST_UTIL.getConnection();
|
||||||
try (Admin admin = conn.getAdmin(); Table table = conn.getTable(tn)) {
|
try (Admin admin = conn.getAdmin(); Table table = conn.getTable(tn)) {
|
||||||
|
@ -113,7 +118,7 @@ public class TestSeparateClientZKCluster {
|
||||||
Get get = new Get(row);
|
Get get = new Get(row);
|
||||||
Result result = table.get(get);
|
Result result = table.get(get);
|
||||||
LOG.debug("Result: " + Bytes.toString(result.getValue(family, qualifier)));
|
LOG.debug("Result: " + Bytes.toString(result.getValue(family, qualifier)));
|
||||||
Assert.assertArrayEquals(value, result.getValue(family, qualifier));
|
assertArrayEquals(value, result.getValue(family, qualifier));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -133,24 +138,24 @@ public class TestSeparateClientZKCluster {
|
||||||
}
|
}
|
||||||
LOG.info("Shutdown master {}", master.getServerName());
|
LOG.info("Shutdown master {}", master.getServerName());
|
||||||
while (cluster.getMaster() == null || !cluster.getMaster().isInitialized()) {
|
while (cluster.getMaster() == null || !cluster.getMaster().isInitialized()) {
|
||||||
LOG.info("Get master {}", cluster.getMaster() == null? "null":
|
LOG.info("Get master {}",
|
||||||
cluster.getMaster().getServerName());
|
cluster.getMaster() == null ? "null" : cluster.getMaster().getServerName());
|
||||||
Thread.sleep(200);
|
Thread.sleep(200);
|
||||||
}
|
}
|
||||||
LOG.info("Got master {}", cluster.getMaster().getServerName());
|
LOG.info("Got master {}", cluster.getMaster().getServerName());
|
||||||
// confirm client access still works
|
// confirm client access still works
|
||||||
Assert.assertTrue(admin.balance(false));
|
assertTrue(admin.balance(false));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testMetaRegionMove() throws Exception {
|
public void testMetaRegionMove() throws Exception {
|
||||||
TableName tn = TableName.valueOf(name.getMethodName());
|
TableName tn = name.getTableName();
|
||||||
// create table
|
// create table
|
||||||
Connection conn = TEST_UTIL.getConnection();
|
Connection conn = TEST_UTIL.getConnection();
|
||||||
try (Admin admin = conn.getAdmin();
|
try (Admin admin = conn.getAdmin();
|
||||||
Table table = conn.getTable(tn);
|
Table table = conn.getTable(tn);
|
||||||
RegionLocator locator = conn.getRegionLocator(tn)) {
|
RegionLocator locator = conn.getRegionLocator(tn)) {
|
||||||
MiniHBaseCluster cluster = TEST_UTIL.getHBaseCluster();
|
MiniHBaseCluster cluster = TEST_UTIL.getHBaseCluster();
|
||||||
ColumnFamilyDescriptorBuilder cfDescBuilder =
|
ColumnFamilyDescriptorBuilder cfDescBuilder =
|
||||||
ColumnFamilyDescriptorBuilder.newBuilder(family);
|
ColumnFamilyDescriptorBuilder.newBuilder(family);
|
||||||
|
@ -191,13 +196,13 @@ public class TestSeparateClientZKCluster {
|
||||||
table.put(put);
|
table.put(put);
|
||||||
result = table.get(get);
|
result = table.get(get);
|
||||||
LOG.debug("Result: " + Bytes.toString(result.getValue(family, qualifier)));
|
LOG.debug("Result: " + Bytes.toString(result.getValue(family, qualifier)));
|
||||||
Assert.assertArrayEquals(newVal, result.getValue(family, qualifier));
|
assertArrayEquals(newVal, result.getValue(family, qualifier));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testMetaMoveDuringClientZkClusterRestart() throws Exception {
|
public void testMetaMoveDuringClientZkClusterRestart() throws Exception {
|
||||||
TableName tn = TableName.valueOf(name.getMethodName());
|
TableName tn = name.getTableName();
|
||||||
// create table
|
// create table
|
||||||
Connection conn = TEST_UTIL.getConnection();
|
Connection conn = TEST_UTIL.getConnection();
|
||||||
try (Admin admin = conn.getAdmin(); Table table = conn.getTable(tn)) {
|
try (Admin admin = conn.getAdmin(); Table table = conn.getTable(tn)) {
|
||||||
|
@ -233,18 +238,18 @@ public class TestSeparateClientZKCluster {
|
||||||
Get get = new Get(row);
|
Get get = new Get(row);
|
||||||
Result result = table.get(get);
|
Result result = table.get(get);
|
||||||
LOG.debug("Result: " + Bytes.toString(result.getValue(family, qualifier)));
|
LOG.debug("Result: " + Bytes.toString(result.getValue(family, qualifier)));
|
||||||
Assert.assertArrayEquals(value, result.getValue(family, qualifier));
|
assertArrayEquals(value, result.getValue(family, qualifier));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testAsyncTable() throws Exception {
|
public void testAsyncTable() throws Exception {
|
||||||
TableName tn = TableName.valueOf(name.getMethodName());
|
TableName tn = name.getTableName();
|
||||||
ColumnFamilyDescriptorBuilder cfDescBuilder = ColumnFamilyDescriptorBuilder.newBuilder(family);
|
ColumnFamilyDescriptorBuilder cfDescBuilder = ColumnFamilyDescriptorBuilder.newBuilder(family);
|
||||||
TableDescriptorBuilder tableDescBuilder =
|
TableDescriptorBuilder tableDescBuilder =
|
||||||
TableDescriptorBuilder.newBuilder(tn).setColumnFamily(cfDescBuilder.build());
|
TableDescriptorBuilder.newBuilder(tn).setColumnFamily(cfDescBuilder.build());
|
||||||
try (AsyncConnection ASYNC_CONN =
|
try (AsyncConnection ASYNC_CONN =
|
||||||
ConnectionFactory.createAsyncConnection(TEST_UTIL.getConfiguration()).get()) {
|
ConnectionFactory.createAsyncConnection(TEST_UTIL.getConfiguration()).get()) {
|
||||||
ASYNC_CONN.getAdmin().createTable(tableDescBuilder.build()).get();
|
ASYNC_CONN.getAdmin().createTable(tableDescBuilder.build()).get();
|
||||||
AsyncTable<?> table = ASYNC_CONN.getTable(tn);
|
AsyncTable<?> table = ASYNC_CONN.getTable(tn);
|
||||||
// put some data
|
// put some data
|
||||||
|
@ -255,7 +260,22 @@ public class TestSeparateClientZKCluster {
|
||||||
Get get = new Get(row);
|
Get get = new Get(row);
|
||||||
Result result = table.get(get).get();
|
Result result = table.get(get).get();
|
||||||
LOG.debug("Result: " + Bytes.toString(result.getValue(family, qualifier)));
|
LOG.debug("Result: " + Bytes.toString(result.getValue(family, qualifier)));
|
||||||
Assert.assertArrayEquals(value, result.getValue(family, qualifier));
|
assertArrayEquals(value, result.getValue(family, qualifier));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testChangeMetaReplicaCount() throws Exception {
|
||||||
|
Admin admin = TEST_UTIL.getAdmin();
|
||||||
|
try (RegionLocator locator =
|
||||||
|
TEST_UTIL.getConnection().getRegionLocator(TableName.META_TABLE_NAME)) {
|
||||||
|
assertEquals(1, locator.getAllRegionLocations().size());
|
||||||
|
HBaseTestingUtility.setReplicas(admin, TableName.META_TABLE_NAME, 3);
|
||||||
|
TEST_UTIL.waitFor(30000, () -> locator.getAllRegionLocations().size() == 3);
|
||||||
|
HBaseTestingUtility.setReplicas(admin, TableName.META_TABLE_NAME, 2);
|
||||||
|
TEST_UTIL.waitFor(30000, () -> locator.getAllRegionLocations().size() == 2);
|
||||||
|
HBaseTestingUtility.setReplicas(admin, TableName.META_TABLE_NAME, 1);
|
||||||
|
TEST_UTIL.waitFor(30000, () -> locator.getAllRegionLocations().size() == 1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -44,6 +44,7 @@ import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
||||||
import org.apache.hadoop.hbase.master.replication.ReplicationPeerManager;
|
import org.apache.hadoop.hbase.master.replication.ReplicationPeerManager;
|
||||||
import org.apache.hadoop.hbase.master.replication.SyncReplicationReplayWALManager;
|
import org.apache.hadoop.hbase.master.replication.SyncReplicationReplayWALManager;
|
||||||
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
|
||||||
|
import org.apache.hadoop.hbase.master.zksyncer.MetaLocationSyncer;
|
||||||
import org.apache.hadoop.hbase.procedure.MasterProcedureManagerHost;
|
import org.apache.hadoop.hbase.procedure.MasterProcedureManagerHost;
|
||||||
import org.apache.hadoop.hbase.procedure2.LockedResource;
|
import org.apache.hadoop.hbase.procedure2.LockedResource;
|
||||||
import org.apache.hadoop.hbase.procedure2.Procedure;
|
import org.apache.hadoop.hbase.procedure2.Procedure;
|
||||||
|
@ -514,4 +515,9 @@ public class MockNoopMasterServices implements MasterServices {
|
||||||
public boolean normalizeRegions(NormalizeTableFilterParams ntfp, boolean isHighPriority) {
|
public boolean normalizeRegions(NormalizeTableFilterParams ntfp, boolean isHighPriority) {
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public MetaLocationSyncer getMetaLocationSyncer() {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -160,7 +160,7 @@ public class TestMasterNoCluster {
|
||||||
while (!master.isInitialized()) {
|
while (!master.isInitialized()) {
|
||||||
Threads.sleep(200);
|
Threads.sleep(200);
|
||||||
}
|
}
|
||||||
Assert.assertNull(master.metaLocationSyncer);
|
Assert.assertNull(master.getMetaLocationSyncer());
|
||||||
Assert.assertNull(master.masterAddressSyncer);
|
Assert.assertNull(master.masterAddressSyncer);
|
||||||
master.stopMaster();
|
master.stopMaster();
|
||||||
master.join();
|
master.join();
|
||||||
|
|
Loading…
Reference in New Issue