HBASE-17915 Implement async replication admin methods
This commit is contained in:
parent
b1ef8dd43a
commit
40cc666ac9
|
@ -18,6 +18,8 @@
|
||||||
package org.apache.hadoop.hbase.client;
|
package org.apache.hadoop.hbase.client;
|
||||||
|
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
import java.util.Collection;
|
||||||
|
import java.util.Map;
|
||||||
import java.util.concurrent.CompletableFuture;
|
import java.util.concurrent.CompletableFuture;
|
||||||
import java.util.regex.Pattern;
|
import java.util.regex.Pattern;
|
||||||
|
|
||||||
|
@ -30,6 +32,9 @@ import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
import org.apache.hadoop.hbase.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaFilter;
|
import org.apache.hadoop.hbase.quotas.QuotaFilter;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaSettings;
|
import org.apache.hadoop.hbase.quotas.QuotaSettings;
|
||||||
|
import org.apache.hadoop.hbase.client.replication.TableCFs;
|
||||||
|
import org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
|
||||||
|
import org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
|
||||||
import org.apache.hadoop.hbase.util.Pair;
|
import org.apache.hadoop.hbase.util.Pair;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -481,4 +486,91 @@ public interface AsyncAdmin {
|
||||||
* @return the QuotaSetting list, which wrapped by a CompletableFuture.
|
* @return the QuotaSetting list, which wrapped by a CompletableFuture.
|
||||||
*/
|
*/
|
||||||
CompletableFuture<List<QuotaSettings>> getQuota(QuotaFilter filter);
|
CompletableFuture<List<QuotaSettings>> getQuota(QuotaFilter filter);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Add a new replication peer for replicating data to slave cluster
|
||||||
|
* @param peerId a short name that identifies the peer
|
||||||
|
* @param peerConfig configuration for the replication slave cluster
|
||||||
|
*/
|
||||||
|
CompletableFuture<Void> addReplicationPeer(final String peerId,
|
||||||
|
final ReplicationPeerConfig peerConfig);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Remove a peer and stop the replication
|
||||||
|
* @param peerId a short name that identifies the peer
|
||||||
|
*/
|
||||||
|
CompletableFuture<Void> removeReplicationPeer(final String peerId);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Restart the replication stream to the specified peer
|
||||||
|
* @param peerId a short name that identifies the peer
|
||||||
|
*/
|
||||||
|
CompletableFuture<Void> enableReplicationPeer(final String peerId);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Stop the replication stream to the specified peer
|
||||||
|
* @param peerId a short name that identifies the peer
|
||||||
|
*/
|
||||||
|
CompletableFuture<Void> disableReplicationPeer(final String peerId);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Returns the configured ReplicationPeerConfig for the specified peer
|
||||||
|
* @param peerId a short name that identifies the peer
|
||||||
|
* @return ReplicationPeerConfig for the peer wrapped by a {@link CompletableFuture}.
|
||||||
|
*/
|
||||||
|
CompletableFuture<ReplicationPeerConfig> getReplicationPeerConfig(final String peerId);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Update the peerConfig for the specified peer
|
||||||
|
* @param peerId a short name that identifies the peer
|
||||||
|
* @param peerConfig new config for the peer
|
||||||
|
*/
|
||||||
|
CompletableFuture<Void> updateReplicationPeerConfig(final String peerId,
|
||||||
|
final ReplicationPeerConfig peerConfig);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Append the replicable table-cf config of the specified peer
|
||||||
|
* @param id a short that identifies the cluster
|
||||||
|
* @param tableCfs A map from tableName to column family names
|
||||||
|
*/
|
||||||
|
CompletableFuture<Void> appendReplicationPeerTableCFs(String id,
|
||||||
|
Map<TableName, ? extends Collection<String>> tableCfs);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Remove some table-cfs from config of the specified peer
|
||||||
|
* @param id a short name that identifies the cluster
|
||||||
|
* @param tableCfs A map from tableName to column family names
|
||||||
|
*/
|
||||||
|
CompletableFuture<Void> removeReplicationPeerTableCFs(String id,
|
||||||
|
Map<TableName, ? extends Collection<String>> tableCfs);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Return a list of replication peers.
|
||||||
|
* @return a list of replication peers description. The return value will be wrapped by a
|
||||||
|
* {@link CompletableFuture}.
|
||||||
|
*/
|
||||||
|
CompletableFuture<List<ReplicationPeerDescription>> listReplicationPeers();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Return a list of replication peers.
|
||||||
|
* @param regex The regular expression to match peer id
|
||||||
|
* @return a list of replication peers description. The return value will be wrapped by a
|
||||||
|
* {@link CompletableFuture}.
|
||||||
|
*/
|
||||||
|
CompletableFuture<List<ReplicationPeerDescription>> listReplicationPeers(String regex);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Return a list of replication peers.
|
||||||
|
* @param pattern The compiled regular expression to match peer id
|
||||||
|
* @return a list of replication peers description. The return value will be wrapped by a
|
||||||
|
* {@link CompletableFuture}.
|
||||||
|
*/
|
||||||
|
CompletableFuture<List<ReplicationPeerDescription>> listReplicationPeers(Pattern pattern);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Find all table and column families that are replicated from this cluster
|
||||||
|
* @return the replicated table-cfs list of this cluster. The return value will be wrapped by a
|
||||||
|
* {@link CompletableFuture}.
|
||||||
|
*/
|
||||||
|
CompletableFuture<List<TableCFs>> listReplicatedTableCFs();
|
||||||
}
|
}
|
||||||
|
|
|
@ -22,16 +22,21 @@ import static org.apache.hadoop.hbase.TableName.META_TABLE_NAME;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Arrays;
|
import java.util.Arrays;
|
||||||
|
import java.util.Collection;
|
||||||
|
import java.util.HashMap;
|
||||||
import java.util.LinkedList;
|
import java.util.LinkedList;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
import java.util.Optional;
|
import java.util.Optional;
|
||||||
import java.util.concurrent.CompletableFuture;
|
import java.util.concurrent.CompletableFuture;
|
||||||
import java.util.concurrent.TimeUnit;
|
import java.util.concurrent.TimeUnit;
|
||||||
import java.util.concurrent.atomic.AtomicReference;
|
import java.util.concurrent.atomic.AtomicReference;
|
||||||
import java.util.function.BiConsumer;
|
import java.util.function.BiConsumer;
|
||||||
import java.util.regex.Pattern;
|
import java.util.regex.Pattern;
|
||||||
|
import java.util.stream.Collectors;
|
||||||
|
|
||||||
import com.google.common.annotations.VisibleForTesting;
|
import com.google.common.annotations.VisibleForTesting;
|
||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.hbase.HColumnDescriptor;
|
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||||
|
@ -54,11 +59,16 @@ import org.apache.hadoop.hbase.classification.InterfaceStability;
|
||||||
import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
|
import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.AdminRequestCallerBuilder;
|
||||||
import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
|
import org.apache.hadoop.hbase.client.AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder;
|
||||||
import org.apache.hadoop.hbase.client.Scan.ReadType;
|
import org.apache.hadoop.hbase.client.Scan.ReadType;
|
||||||
|
import org.apache.hadoop.hbase.client.replication.ReplicationSerDeHelper;
|
||||||
|
import org.apache.hadoop.hbase.client.replication.TableCFs;
|
||||||
import org.apache.hadoop.hbase.exceptions.DeserializationException;
|
import org.apache.hadoop.hbase.exceptions.DeserializationException;
|
||||||
import org.apache.hadoop.hbase.ipc.HBaseRpcController;
|
import org.apache.hadoop.hbase.ipc.HBaseRpcController;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaFilter;
|
import org.apache.hadoop.hbase.quotas.QuotaFilter;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaSettings;
|
import org.apache.hadoop.hbase.quotas.QuotaSettings;
|
||||||
import org.apache.hadoop.hbase.quotas.QuotaTableUtil;
|
import org.apache.hadoop.hbase.quotas.QuotaTableUtil;
|
||||||
|
import org.apache.hadoop.hbase.replication.ReplicationException;
|
||||||
|
import org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
|
||||||
|
import org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
|
||||||
import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback;
|
import org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
|
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
|
import org.apache.hadoop.hbase.shaded.protobuf.RequestConverter;
|
||||||
|
@ -121,6 +131,20 @@ import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTa
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse;
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest;
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse;
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest;
|
||||||
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.hbase.util.ForeignExceptionUtil;
|
import org.apache.hadoop.hbase.util.ForeignExceptionUtil;
|
||||||
import org.apache.hadoop.hbase.util.Pair;
|
import org.apache.hadoop.hbase.util.Pair;
|
||||||
|
@ -1155,42 +1179,209 @@ public class AsyncHBaseAdmin implements AsyncAdmin {
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public CompletableFuture<Void> setQuota(QuotaSettings quota){
|
public CompletableFuture<Void> setQuota(QuotaSettings quota) {
|
||||||
return this.<Void> newMasterCaller()
|
return this
|
||||||
.action((controller, stub) -> this.<SetQuotaRequest, SetQuotaResponse, Void> call(
|
.<Void> newMasterCaller()
|
||||||
controller, stub, QuotaSettings.buildSetQuotaRequestProto(quota),
|
.action(
|
||||||
(s, c, req, done) -> s.setQuota(c, req, done), (resp) -> null))
|
(controller, stub) -> this.<SetQuotaRequest, SetQuotaResponse, Void> call(controller,
|
||||||
.call();
|
stub, QuotaSettings.buildSetQuotaRequestProto(quota),
|
||||||
|
(s, c, req, done) -> s.setQuota(c, req, done), (resp) -> null)).call();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public CompletableFuture<List<QuotaSettings>> getQuota(QuotaFilter filter) {
|
public CompletableFuture<List<QuotaSettings>> getQuota(QuotaFilter filter) {
|
||||||
CompletableFuture<List<QuotaSettings>> future = new CompletableFuture<>();
|
CompletableFuture<List<QuotaSettings>> future = new CompletableFuture<>();
|
||||||
Scan scan = QuotaTableUtil.makeScan(filter);
|
Scan scan = QuotaTableUtil.makeScan(filter);
|
||||||
this.connection.getRawTableBuilder(QuotaTableUtil.QUOTA_TABLE_NAME).build().scan(scan,
|
this.connection.getRawTableBuilder(QuotaTableUtil.QUOTA_TABLE_NAME).build()
|
||||||
new RawScanResultConsumer() {
|
.scan(scan, new RawScanResultConsumer() {
|
||||||
List<QuotaSettings> settings = new ArrayList<>();
|
List<QuotaSettings> settings = new ArrayList<>();
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void onNext(Result[] results, ScanController controller) {
|
public void onNext(Result[] results, ScanController controller) {
|
||||||
for (Result result : results) {
|
for (Result result : results) {
|
||||||
try {
|
try {
|
||||||
QuotaTableUtil.parseResultToCollection(result, settings);
|
QuotaTableUtil.parseResultToCollection(result, settings);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
controller.terminate();
|
controller.terminate();
|
||||||
future.completeExceptionally(e);
|
future.completeExceptionally(e);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void onError(Throwable error) {
|
public void onError(Throwable error) {
|
||||||
future.completeExceptionally(error);
|
future.completeExceptionally(error);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void onComplete() {
|
public void onComplete() {
|
||||||
future.complete(settings);
|
future.complete(settings);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
return future;
|
||||||
|
}
|
||||||
|
|
||||||
|
public CompletableFuture<Void> addReplicationPeer(String peerId, ReplicationPeerConfig peerConfig) {
|
||||||
|
return this
|
||||||
|
.<Void> newMasterCaller()
|
||||||
|
.action(
|
||||||
|
(controller, stub) -> this
|
||||||
|
.<AddReplicationPeerRequest, AddReplicationPeerResponse, Void> call(controller, stub,
|
||||||
|
RequestConverter.buildAddReplicationPeerRequest(peerId, peerConfig), (s, c, req,
|
||||||
|
done) -> s.addReplicationPeer(c, req, done), (resp) -> null)).call();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<Void> removeReplicationPeer(String peerId) {
|
||||||
|
return this
|
||||||
|
.<Void> newMasterCaller()
|
||||||
|
.action(
|
||||||
|
(controller, stub) -> this
|
||||||
|
.<RemoveReplicationPeerRequest, RemoveReplicationPeerResponse, Void> call(controller,
|
||||||
|
stub, RequestConverter.buildRemoveReplicationPeerRequest(peerId),
|
||||||
|
(s, c, req, done) -> s.removeReplicationPeer(c, req, done), (resp) -> null)).call();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<Void> enableReplicationPeer(String peerId) {
|
||||||
|
return this
|
||||||
|
.<Void> newMasterCaller()
|
||||||
|
.action(
|
||||||
|
(controller, stub) -> this
|
||||||
|
.<EnableReplicationPeerRequest, EnableReplicationPeerResponse, Void> call(controller,
|
||||||
|
stub, RequestConverter.buildEnableReplicationPeerRequest(peerId),
|
||||||
|
(s, c, req, done) -> s.enableReplicationPeer(c, req, done), (resp) -> null)).call();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<Void> disableReplicationPeer(String peerId) {
|
||||||
|
return this
|
||||||
|
.<Void> newMasterCaller()
|
||||||
|
.action(
|
||||||
|
(controller, stub) -> this
|
||||||
|
.<DisableReplicationPeerRequest, DisableReplicationPeerResponse, Void> call(
|
||||||
|
controller, stub, RequestConverter.buildDisableReplicationPeerRequest(peerId), (s,
|
||||||
|
c, req, done) -> s.disableReplicationPeer(c, req, done), (resp) -> null))
|
||||||
|
.call();
|
||||||
|
}
|
||||||
|
|
||||||
|
public CompletableFuture<ReplicationPeerConfig> getReplicationPeerConfig(String peerId) {
|
||||||
|
return this
|
||||||
|
.<ReplicationPeerConfig> newMasterCaller()
|
||||||
|
.action(
|
||||||
|
(controller, stub) -> this
|
||||||
|
.<GetReplicationPeerConfigRequest, GetReplicationPeerConfigResponse, ReplicationPeerConfig> call(
|
||||||
|
controller, stub, RequestConverter.buildGetReplicationPeerConfigRequest(peerId), (
|
||||||
|
s, c, req, done) -> s.getReplicationPeerConfig(c, req, done),
|
||||||
|
(resp) -> ReplicationSerDeHelper.convert(resp.getPeerConfig()))).call();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<Void> updateReplicationPeerConfig(String peerId,
|
||||||
|
ReplicationPeerConfig peerConfig) {
|
||||||
|
return this
|
||||||
|
.<Void> newMasterCaller()
|
||||||
|
.action(
|
||||||
|
(controller, stub) -> this
|
||||||
|
.<UpdateReplicationPeerConfigRequest, UpdateReplicationPeerConfigResponse, Void> call(
|
||||||
|
controller, stub, RequestConverter.buildUpdateReplicationPeerConfigRequest(peerId,
|
||||||
|
peerConfig), (s, c, req, done) -> s.updateReplicationPeerConfig(c, req, done), (
|
||||||
|
resp) -> null)).call();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<Void> appendReplicationPeerTableCFs(String id,
|
||||||
|
Map<TableName, ? extends Collection<String>> tableCfs) {
|
||||||
|
if (tableCfs == null) {
|
||||||
|
return failedFuture(new ReplicationException("tableCfs is null"));
|
||||||
|
}
|
||||||
|
|
||||||
|
CompletableFuture<Void> future = new CompletableFuture<Void>();
|
||||||
|
getReplicationPeerConfig(id).whenComplete((peerConfig, error) -> {
|
||||||
|
if (!completeExceptionally(future, error)) {
|
||||||
|
ReplicationSerDeHelper.appendTableCFsToReplicationPeerConfig(tableCfs, peerConfig);
|
||||||
|
updateReplicationPeerConfig(id, peerConfig).whenComplete((result, err) -> {
|
||||||
|
if (!completeExceptionally(future, error)) {
|
||||||
|
future.complete(result);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
});
|
||||||
|
return future;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<Void> removeReplicationPeerTableCFs(String id,
|
||||||
|
Map<TableName, ? extends Collection<String>> tableCfs) {
|
||||||
|
if (tableCfs == null) {
|
||||||
|
return failedFuture(new ReplicationException("tableCfs is null"));
|
||||||
|
}
|
||||||
|
|
||||||
|
CompletableFuture<Void> future = new CompletableFuture<Void>();
|
||||||
|
getReplicationPeerConfig(id).whenComplete((peerConfig, error) -> {
|
||||||
|
if (!completeExceptionally(future, error)) {
|
||||||
|
try {
|
||||||
|
ReplicationSerDeHelper.removeTableCFsFromReplicationPeerConfig(tableCfs, peerConfig, id);
|
||||||
|
} catch (ReplicationException e) {
|
||||||
|
future.completeExceptionally(e);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
updateReplicationPeerConfig(id, peerConfig).whenComplete((result, err) -> {
|
||||||
|
if (!completeExceptionally(future, error)) {
|
||||||
|
future.complete(result);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
});
|
||||||
|
return future;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<List<ReplicationPeerDescription>> listReplicationPeers() {
|
||||||
|
return listReplicationPeers((Pattern) null);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<List<ReplicationPeerDescription>> listReplicationPeers(String regex) {
|
||||||
|
return listReplicationPeers(Pattern.compile(regex));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<List<ReplicationPeerDescription>> listReplicationPeers(Pattern pattern) {
|
||||||
|
return this
|
||||||
|
.<List<ReplicationPeerDescription>> newMasterCaller()
|
||||||
|
.action(
|
||||||
|
(controller, stub) -> this
|
||||||
|
.<ListReplicationPeersRequest, ListReplicationPeersResponse, List<ReplicationPeerDescription>> call(
|
||||||
|
controller,
|
||||||
|
stub,
|
||||||
|
RequestConverter.buildListReplicationPeersRequest(pattern),
|
||||||
|
(s, c, req, done) -> s.listReplicationPeers(c, req, done),
|
||||||
|
(resp) -> resp.getPeerDescList().stream()
|
||||||
|
.map(ReplicationSerDeHelper::toReplicationPeerDescription)
|
||||||
|
.collect(Collectors.toList()))).call();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public CompletableFuture<List<TableCFs>> listReplicatedTableCFs() {
|
||||||
|
CompletableFuture<List<TableCFs>> future = new CompletableFuture<List<TableCFs>>();
|
||||||
|
listTables().whenComplete(
|
||||||
|
(tables, error) -> {
|
||||||
|
if (!completeExceptionally(future, error)) {
|
||||||
|
List<TableCFs> replicatedTableCFs = new ArrayList<>();
|
||||||
|
Arrays.asList(tables).forEach(
|
||||||
|
table -> {
|
||||||
|
Map<String, Integer> cfs = new HashMap<>();
|
||||||
|
Arrays.asList(table.getColumnFamilies()).stream()
|
||||||
|
.filter(column -> column.getScope() != HConstants.REPLICATION_SCOPE_LOCAL)
|
||||||
|
.forEach(column -> {
|
||||||
|
cfs.put(column.getNameAsString(), column.getScope());
|
||||||
|
});
|
||||||
|
if (!cfs.isEmpty()) {
|
||||||
|
replicatedTableCFs.add(new TableCFs(table.getTableName(), cfs));
|
||||||
|
}
|
||||||
|
});
|
||||||
|
future.complete(replicatedTableCFs);
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
return future;
|
return future;
|
||||||
|
@ -1470,4 +1661,12 @@ public class AsyncHBaseAdmin implements AsyncAdmin {
|
||||||
future.completeExceptionally(error);
|
future.completeExceptionally(error);
|
||||||
return future;
|
return future;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private <T> boolean completeExceptionally(CompletableFuture<T> future, Throwable error) {
|
||||||
|
if (error != null) {
|
||||||
|
future.completeExceptionally(error);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
return false;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -3869,31 +3869,7 @@ public class HBaseAdmin implements Admin {
|
||||||
throw new ReplicationException("tableCfs is null");
|
throw new ReplicationException("tableCfs is null");
|
||||||
}
|
}
|
||||||
ReplicationPeerConfig peerConfig = getReplicationPeerConfig(id);
|
ReplicationPeerConfig peerConfig = getReplicationPeerConfig(id);
|
||||||
Map<TableName, List<String>> preTableCfs = peerConfig.getTableCFsMap();
|
ReplicationSerDeHelper.appendTableCFsToReplicationPeerConfig(tableCfs, peerConfig);
|
||||||
if (preTableCfs == null) {
|
|
||||||
peerConfig.setTableCFsMap(tableCfs);
|
|
||||||
} else {
|
|
||||||
for (Map.Entry<TableName, ? extends Collection<String>> entry : tableCfs.entrySet()) {
|
|
||||||
TableName table = entry.getKey();
|
|
||||||
Collection<String> appendCfs = entry.getValue();
|
|
||||||
if (preTableCfs.containsKey(table)) {
|
|
||||||
List<String> cfs = preTableCfs.get(table);
|
|
||||||
if (cfs == null || appendCfs == null || appendCfs.isEmpty()) {
|
|
||||||
preTableCfs.put(table, null);
|
|
||||||
} else {
|
|
||||||
Set<String> cfSet = new HashSet<String>(cfs);
|
|
||||||
cfSet.addAll(appendCfs);
|
|
||||||
preTableCfs.put(table, Lists.newArrayList(cfSet));
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
if (appendCfs == null || appendCfs.isEmpty()) {
|
|
||||||
preTableCfs.put(table, null);
|
|
||||||
} else {
|
|
||||||
preTableCfs.put(table, Lists.newArrayList(appendCfs));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
updateReplicationPeerConfig(id, peerConfig);
|
updateReplicationPeerConfig(id, peerConfig);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -3905,37 +3881,7 @@ public class HBaseAdmin implements Admin {
|
||||||
throw new ReplicationException("tableCfs is null");
|
throw new ReplicationException("tableCfs is null");
|
||||||
}
|
}
|
||||||
ReplicationPeerConfig peerConfig = getReplicationPeerConfig(id);
|
ReplicationPeerConfig peerConfig = getReplicationPeerConfig(id);
|
||||||
Map<TableName, List<String>> preTableCfs = peerConfig.getTableCFsMap();
|
ReplicationSerDeHelper.removeTableCFsFromReplicationPeerConfig(tableCfs, peerConfig, id);
|
||||||
if (preTableCfs == null) {
|
|
||||||
throw new ReplicationException("Table-Cfs for peer" + id + " is null");
|
|
||||||
}
|
|
||||||
for (Map.Entry<TableName, ? extends Collection<String>> entry : tableCfs.entrySet()) {
|
|
||||||
|
|
||||||
TableName table = entry.getKey();
|
|
||||||
Collection<String> removeCfs = entry.getValue();
|
|
||||||
if (preTableCfs.containsKey(table)) {
|
|
||||||
List<String> cfs = preTableCfs.get(table);
|
|
||||||
if (cfs == null && (removeCfs == null || removeCfs.isEmpty())) {
|
|
||||||
preTableCfs.remove(table);
|
|
||||||
} else if (cfs != null && (removeCfs != null && !removeCfs.isEmpty())) {
|
|
||||||
Set<String> cfSet = new HashSet<String>(cfs);
|
|
||||||
cfSet.removeAll(removeCfs);
|
|
||||||
if (cfSet.isEmpty()) {
|
|
||||||
preTableCfs.remove(table);
|
|
||||||
} else {
|
|
||||||
preTableCfs.put(table, Lists.newArrayList(cfSet));
|
|
||||||
}
|
|
||||||
} else if (cfs == null && (removeCfs != null && !removeCfs.isEmpty())) {
|
|
||||||
throw new ReplicationException("Cannot remove cf of table: " + table
|
|
||||||
+ " which doesn't specify cfs from table-cfs config in peer: " + id);
|
|
||||||
} else if (cfs != null && (removeCfs == null || removeCfs.isEmpty())) {
|
|
||||||
throw new ReplicationException("Cannot remove table: " + table
|
|
||||||
+ " which has specified cfs from table-cfs config in peer: " + id);
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
throw new ReplicationException("No table: " + table + " in table-cfs config of peer: " + id);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
updateReplicationPeerConfig(id, peerConfig);
|
updateReplicationPeerConfig(id, peerConfig);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -30,11 +30,14 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
|
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
|
||||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos;
|
import org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos;
|
||||||
|
import org.apache.hadoop.hbase.replication.ReplicationException;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
|
import org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
|
import org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.hbase.util.Strings;
|
import org.apache.hadoop.hbase.util.Strings;
|
||||||
|
|
||||||
|
import com.google.common.collect.Lists;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.Collection;
|
import java.util.Collection;
|
||||||
import java.util.HashSet;
|
import java.util.HashSet;
|
||||||
|
@ -367,4 +370,68 @@ public final class ReplicationSerDeHelper {
|
||||||
builder.setConfig(convert(desc.getPeerConfig()));
|
builder.setConfig(convert(desc.getPeerConfig()));
|
||||||
return builder.build();
|
return builder.build();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static void appendTableCFsToReplicationPeerConfig(
|
||||||
|
Map<TableName, ? extends Collection<String>> tableCfs, ReplicationPeerConfig peerConfig) {
|
||||||
|
Map<TableName, List<String>> preTableCfs = peerConfig.getTableCFsMap();
|
||||||
|
if (preTableCfs == null) {
|
||||||
|
peerConfig.setTableCFsMap(tableCfs);
|
||||||
|
} else {
|
||||||
|
for (Map.Entry<TableName, ? extends Collection<String>> entry : tableCfs.entrySet()) {
|
||||||
|
TableName table = entry.getKey();
|
||||||
|
Collection<String> appendCfs = entry.getValue();
|
||||||
|
if (preTableCfs.containsKey(table)) {
|
||||||
|
List<String> cfs = preTableCfs.get(table);
|
||||||
|
if (cfs == null || appendCfs == null || appendCfs.isEmpty()) {
|
||||||
|
preTableCfs.put(table, null);
|
||||||
|
} else {
|
||||||
|
Set<String> cfSet = new HashSet<String>(cfs);
|
||||||
|
cfSet.addAll(appendCfs);
|
||||||
|
preTableCfs.put(table, Lists.newArrayList(cfSet));
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if (appendCfs == null || appendCfs.isEmpty()) {
|
||||||
|
preTableCfs.put(table, null);
|
||||||
|
} else {
|
||||||
|
preTableCfs.put(table, Lists.newArrayList(appendCfs));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public static void removeTableCFsFromReplicationPeerConfig(
|
||||||
|
Map<TableName, ? extends Collection<String>> tableCfs, ReplicationPeerConfig peerConfig,
|
||||||
|
String id) throws ReplicationException {
|
||||||
|
Map<TableName, List<String>> preTableCfs = peerConfig.getTableCFsMap();
|
||||||
|
if (preTableCfs == null) {
|
||||||
|
throw new ReplicationException("Table-Cfs for peer: " + id + " is null");
|
||||||
|
}
|
||||||
|
for (Map.Entry<TableName, ? extends Collection<String>> entry : tableCfs.entrySet()) {
|
||||||
|
TableName table = entry.getKey();
|
||||||
|
Collection<String> removeCfs = entry.getValue();
|
||||||
|
if (preTableCfs.containsKey(table)) {
|
||||||
|
List<String> cfs = preTableCfs.get(table);
|
||||||
|
if (cfs == null && (removeCfs == null || removeCfs.isEmpty())) {
|
||||||
|
preTableCfs.remove(table);
|
||||||
|
} else if (cfs != null && (removeCfs != null && !removeCfs.isEmpty())) {
|
||||||
|
Set<String> cfSet = new HashSet<String>(cfs);
|
||||||
|
cfSet.removeAll(removeCfs);
|
||||||
|
if (cfSet.isEmpty()) {
|
||||||
|
preTableCfs.remove(table);
|
||||||
|
} else {
|
||||||
|
preTableCfs.put(table, Lists.newArrayList(cfSet));
|
||||||
|
}
|
||||||
|
} else if (cfs == null && (removeCfs != null && !removeCfs.isEmpty())) {
|
||||||
|
throw new ReplicationException("Cannot remove cf of table: " + table
|
||||||
|
+ " which doesn't specify cfs from table-cfs config in peer: " + id);
|
||||||
|
} else if (cfs != null && (removeCfs == null || removeCfs.isEmpty())) {
|
||||||
|
throw new ReplicationException("Cannot remove table: " + table
|
||||||
|
+ " which has specified cfs from table-cfs config in peer: " + id);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
throw new ReplicationException("No table: " + table + " in table-cfs config of peer: " + id);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -34,7 +34,7 @@ import org.junit.BeforeClass;
|
||||||
*/
|
*/
|
||||||
public abstract class TestAsyncAdminBase {
|
public abstract class TestAsyncAdminBase {
|
||||||
|
|
||||||
protected static final Log LOG = LogFactory.getLog(TestAdmin1.class);
|
protected static final Log LOG = LogFactory.getLog(TestAsyncAdminBase.class);
|
||||||
protected final static HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
|
protected final static HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
|
||||||
protected static byte[] FAMILY = Bytes.toBytes("testFamily");
|
protected static byte[] FAMILY = Bytes.toBytes("testFamily");
|
||||||
protected static final byte[] FAMILY_0 = Bytes.toBytes("cf0");
|
protected static final byte[] FAMILY_0 = Bytes.toBytes("cf0");
|
||||||
|
|
|
@ -0,0 +1,416 @@
|
||||||
|
/**
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hbase.client;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertFalse;
|
||||||
|
import static org.junit.Assert.assertNotNull;
|
||||||
|
import static org.junit.Assert.assertNull;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
import static org.junit.Assert.fail;
|
||||||
|
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.HashMap;
|
||||||
|
import java.util.HashSet;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
import java.util.Set;
|
||||||
|
import java.util.concurrent.CompletionException;
|
||||||
|
|
||||||
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
|
import org.apache.hadoop.hbase.TableName;
|
||||||
|
import org.apache.hadoop.hbase.replication.ReplicationException;
|
||||||
|
import org.apache.hadoop.hbase.replication.ReplicationPeerConfig;
|
||||||
|
import org.apache.hadoop.hbase.replication.ReplicationPeerDescription;
|
||||||
|
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
||||||
|
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||||
|
import org.junit.BeforeClass;
|
||||||
|
import org.junit.Rule;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.experimental.categories.Category;
|
||||||
|
import org.junit.rules.TestName;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Class to test asynchronous replication admin operations.
|
||||||
|
*/
|
||||||
|
@Category({MediumTests.class, ClientTests.class})
|
||||||
|
public class TestAsyncReplicationAdminApi extends TestAsyncAdminBase {
|
||||||
|
|
||||||
|
private final String ID_ONE = "1";
|
||||||
|
private final String KEY_ONE = "127.0.0.1:2181:/hbase";
|
||||||
|
private final String ID_SECOND = "2";
|
||||||
|
private final String KEY_SECOND = "127.0.0.1:2181:/hbase2";
|
||||||
|
|
||||||
|
@Rule
|
||||||
|
public TestName name = new TestName();
|
||||||
|
|
||||||
|
@BeforeClass
|
||||||
|
public static void setUpBeforeClass() throws Exception {
|
||||||
|
TEST_UTIL.getConfiguration().setInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, 1);
|
||||||
|
TEST_UTIL.startMiniCluster();
|
||||||
|
ASYNC_CONN = ConnectionFactory.createAsyncConnection(TEST_UTIL.getConfiguration()).get();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testAddRemovePeer() throws Exception {
|
||||||
|
ReplicationPeerConfig rpc1 = new ReplicationPeerConfig();
|
||||||
|
rpc1.setClusterKey(KEY_ONE);
|
||||||
|
ReplicationPeerConfig rpc2 = new ReplicationPeerConfig();
|
||||||
|
rpc2.setClusterKey(KEY_SECOND);
|
||||||
|
// Add a valid peer
|
||||||
|
admin.addReplicationPeer(ID_ONE, rpc1).join();
|
||||||
|
// try adding the same (fails)
|
||||||
|
try {
|
||||||
|
admin.addReplicationPeer(ID_ONE, rpc1).join();
|
||||||
|
fail("Test case should fail as adding a same peer.");
|
||||||
|
} catch (CompletionException e) {
|
||||||
|
// OK!
|
||||||
|
}
|
||||||
|
assertEquals(1, admin.listReplicationPeers().get().size());
|
||||||
|
// Try to remove an inexisting peer
|
||||||
|
try {
|
||||||
|
admin.removeReplicationPeer(ID_SECOND).join();
|
||||||
|
fail("Test case should fail as removing a inexisting peer.");
|
||||||
|
} catch (CompletionException e) {
|
||||||
|
// OK!
|
||||||
|
}
|
||||||
|
assertEquals(1, admin.listReplicationPeers().get().size());
|
||||||
|
// Add a second since multi-slave is supported
|
||||||
|
admin.addReplicationPeer(ID_SECOND, rpc2).join();
|
||||||
|
assertEquals(2, admin.listReplicationPeers().get().size());
|
||||||
|
// Remove the first peer we added
|
||||||
|
admin.removeReplicationPeer(ID_ONE).join();
|
||||||
|
assertEquals(1, admin.listReplicationPeers().get().size());
|
||||||
|
admin.removeReplicationPeer(ID_SECOND).join();
|
||||||
|
assertEquals(0, admin.listReplicationPeers().get().size());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testPeerConfig() throws Exception {
|
||||||
|
ReplicationPeerConfig config = new ReplicationPeerConfig();
|
||||||
|
config.setClusterKey(KEY_ONE);
|
||||||
|
config.getConfiguration().put("key1", "value1");
|
||||||
|
config.getConfiguration().put("key2", "value2");
|
||||||
|
admin.addReplicationPeer(ID_ONE, config).join();
|
||||||
|
|
||||||
|
List<ReplicationPeerDescription> peers = admin.listReplicationPeers().get();
|
||||||
|
assertEquals(1, peers.size());
|
||||||
|
ReplicationPeerDescription peerOne = peers.get(0);
|
||||||
|
assertNotNull(peerOne);
|
||||||
|
assertEquals("value1", peerOne.getPeerConfig().getConfiguration().get("key1"));
|
||||||
|
assertEquals("value2", peerOne.getPeerConfig().getConfiguration().get("key2"));
|
||||||
|
|
||||||
|
admin.removeReplicationPeer(ID_ONE).join();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testEnableDisablePeer() throws Exception {
|
||||||
|
ReplicationPeerConfig rpc1 = new ReplicationPeerConfig();
|
||||||
|
rpc1.setClusterKey(KEY_ONE);
|
||||||
|
admin.addReplicationPeer(ID_ONE, rpc1).join();
|
||||||
|
List<ReplicationPeerDescription> peers = admin.listReplicationPeers().get();
|
||||||
|
assertEquals(1, peers.size());
|
||||||
|
assertTrue(peers.get(0).isEnabled());
|
||||||
|
|
||||||
|
admin.disableReplicationPeer(ID_ONE).join();
|
||||||
|
peers = admin.listReplicationPeers().get();
|
||||||
|
assertEquals(1, peers.size());
|
||||||
|
assertFalse(peers.get(0).isEnabled());
|
||||||
|
admin.removeReplicationPeer(ID_ONE).join();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testAppendPeerTableCFs() throws Exception {
|
||||||
|
ReplicationPeerConfig rpc1 = new ReplicationPeerConfig();
|
||||||
|
rpc1.setClusterKey(KEY_ONE);
|
||||||
|
final TableName tableName1 = TableName.valueOf(name.getMethodName() + "t1");
|
||||||
|
final TableName tableName2 = TableName.valueOf(name.getMethodName() + "t2");
|
||||||
|
final TableName tableName3 = TableName.valueOf(name.getMethodName() + "t3");
|
||||||
|
final TableName tableName4 = TableName.valueOf(name.getMethodName() + "t4");
|
||||||
|
final TableName tableName5 = TableName.valueOf(name.getMethodName() + "t5");
|
||||||
|
final TableName tableName6 = TableName.valueOf(name.getMethodName() + "t6");
|
||||||
|
|
||||||
|
// Add a valid peer
|
||||||
|
admin.addReplicationPeer(ID_ONE, rpc1).join();
|
||||||
|
|
||||||
|
Map<TableName, List<String>> tableCFs = new HashMap<>();
|
||||||
|
|
||||||
|
// append table t1 to replication
|
||||||
|
tableCFs.put(tableName1, null);
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
Map<TableName, List<String>> result = admin.getReplicationPeerConfig(ID_ONE).get()
|
||||||
|
.getTableCFsMap();
|
||||||
|
assertEquals(1, result.size());
|
||||||
|
assertEquals(true, result.containsKey(tableName1));
|
||||||
|
assertNull(result.get(tableName1));
|
||||||
|
|
||||||
|
// append table t2 to replication
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName2, null);
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
result = admin.getReplicationPeerConfig(ID_ONE).get().getTableCFsMap();
|
||||||
|
assertEquals(2, result.size());
|
||||||
|
assertTrue("Should contain t1", result.containsKey(tableName1));
|
||||||
|
assertTrue("Should contain t2", result.containsKey(tableName2));
|
||||||
|
assertNull(result.get(tableName1));
|
||||||
|
assertNull(result.get(tableName2));
|
||||||
|
|
||||||
|
// append table column family: f1 of t3 to replication
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName3, new ArrayList<>());
|
||||||
|
tableCFs.get(tableName3).add("f1");
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
result = admin.getReplicationPeerConfig(ID_ONE).get().getTableCFsMap();
|
||||||
|
assertEquals(3, result.size());
|
||||||
|
assertTrue("Should contain t1", result.containsKey(tableName1));
|
||||||
|
assertTrue("Should contain t2", result.containsKey(tableName2));
|
||||||
|
assertTrue("Should contain t3", result.containsKey(tableName3));
|
||||||
|
assertNull(result.get(tableName1));
|
||||||
|
assertNull(result.get(tableName2));
|
||||||
|
assertEquals(1, result.get(tableName3).size());
|
||||||
|
assertEquals("f1", result.get(tableName3).get(0));
|
||||||
|
|
||||||
|
// append table column family: f1,f2 of t4 to replication
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName4, new ArrayList<>());
|
||||||
|
tableCFs.get(tableName4).add("f1");
|
||||||
|
tableCFs.get(tableName4).add("f2");
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
result = admin.getReplicationPeerConfig(ID_ONE).get().getTableCFsMap();
|
||||||
|
assertEquals(4, result.size());
|
||||||
|
assertTrue("Should contain t1", result.containsKey(tableName1));
|
||||||
|
assertTrue("Should contain t2", result.containsKey(tableName2));
|
||||||
|
assertTrue("Should contain t3", result.containsKey(tableName3));
|
||||||
|
assertTrue("Should contain t4", result.containsKey(tableName4));
|
||||||
|
assertNull(result.get(tableName1));
|
||||||
|
assertNull(result.get(tableName2));
|
||||||
|
assertEquals(1, result.get(tableName3).size());
|
||||||
|
assertEquals("f1", result.get(tableName3).get(0));
|
||||||
|
assertEquals(2, result.get(tableName4).size());
|
||||||
|
assertEquals("f1", result.get(tableName4).get(0));
|
||||||
|
assertEquals("f2", result.get(tableName4).get(1));
|
||||||
|
|
||||||
|
// append "table5" => [], then append "table5" => ["f1"]
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName5, new ArrayList<>());
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName5, new ArrayList<>());
|
||||||
|
tableCFs.get(tableName5).add("f1");
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
result = admin.getReplicationPeerConfig(ID_ONE).get().getTableCFsMap();
|
||||||
|
assertEquals(5, result.size());
|
||||||
|
assertTrue("Should contain t5", result.containsKey(tableName5));
|
||||||
|
// null means replication all cfs of tab5
|
||||||
|
assertNull(result.get(tableName5));
|
||||||
|
|
||||||
|
// append "table6" => ["f1"], then append "table6" => []
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName6, new ArrayList<>());
|
||||||
|
tableCFs.get(tableName6).add("f1");
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName6, new ArrayList<>());
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
result = admin.getReplicationPeerConfig(ID_ONE).get().getTableCFsMap();
|
||||||
|
assertEquals(6, result.size());
|
||||||
|
assertTrue("Should contain t6", result.containsKey(tableName6));
|
||||||
|
// null means replication all cfs of tab6
|
||||||
|
assertNull(result.get(tableName6));
|
||||||
|
|
||||||
|
admin.removeReplicationPeer(ID_ONE).join();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testRemovePeerTableCFs() throws Exception {
|
||||||
|
ReplicationPeerConfig rpc1 = new ReplicationPeerConfig();
|
||||||
|
rpc1.setClusterKey(KEY_ONE);
|
||||||
|
final TableName tableName1 = TableName.valueOf(name.getMethodName() + "t1");
|
||||||
|
final TableName tableName2 = TableName.valueOf(name.getMethodName() + "t2");
|
||||||
|
final TableName tableName3 = TableName.valueOf(name.getMethodName() + "t3");
|
||||||
|
final TableName tableName4 = TableName.valueOf(name.getMethodName() + "t4");
|
||||||
|
// Add a valid peer
|
||||||
|
admin.addReplicationPeer(ID_ONE, rpc1).join();
|
||||||
|
Map<TableName, List<String>> tableCFs = new HashMap<>();
|
||||||
|
try {
|
||||||
|
tableCFs.put(tableName3, null);
|
||||||
|
admin.removeReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
fail("Test case should fail as removing table-cfs from a peer whose table-cfs is null");
|
||||||
|
} catch (CompletionException e) {
|
||||||
|
assertTrue(e.getCause() instanceof ReplicationException);
|
||||||
|
}
|
||||||
|
assertNull(admin.getReplicationPeerConfig(ID_ONE).get().getTableCFsMap());
|
||||||
|
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName1, null);
|
||||||
|
tableCFs.put(tableName2, new ArrayList<>());
|
||||||
|
tableCFs.get(tableName2).add("cf1");
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
try {
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName3, null);
|
||||||
|
admin.removeReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
fail("Test case should fail as removing table-cfs from a peer whose table-cfs didn't contain t3");
|
||||||
|
} catch (CompletionException e) {
|
||||||
|
assertTrue(e.getCause() instanceof ReplicationException);
|
||||||
|
}
|
||||||
|
Map<TableName, List<String>> result = admin.getReplicationPeerConfig(ID_ONE).get()
|
||||||
|
.getTableCFsMap();
|
||||||
|
assertEquals(2, result.size());
|
||||||
|
assertTrue("Should contain t1", result.containsKey(tableName1));
|
||||||
|
assertTrue("Should contain t2", result.containsKey(tableName2));
|
||||||
|
assertNull(result.get(tableName1));
|
||||||
|
assertEquals(1, result.get(tableName2).size());
|
||||||
|
assertEquals("cf1", result.get(tableName2).get(0));
|
||||||
|
|
||||||
|
try {
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName1, new ArrayList<>());
|
||||||
|
tableCFs.get(tableName1).add("cf1");
|
||||||
|
admin.removeReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
fail("Test case should fail, because table t1 didn't specify cfs in peer config");
|
||||||
|
} catch (CompletionException e) {
|
||||||
|
assertTrue(e.getCause() instanceof ReplicationException);
|
||||||
|
}
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName1, null);
|
||||||
|
admin.removeReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
result = admin.getReplicationPeerConfig(ID_ONE).get().getTableCFsMap();
|
||||||
|
assertEquals(1, result.size());
|
||||||
|
assertEquals(1, result.get(tableName2).size());
|
||||||
|
assertEquals("cf1", result.get(tableName2).get(0));
|
||||||
|
|
||||||
|
try {
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName2, null);
|
||||||
|
admin.removeReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
fail("Test case should fail, because table t2 hase specified cfs in peer config");
|
||||||
|
} catch (CompletionException e) {
|
||||||
|
assertTrue(e.getCause() instanceof ReplicationException);
|
||||||
|
}
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName2, new ArrayList<>());
|
||||||
|
tableCFs.get(tableName2).add("cf1");
|
||||||
|
admin.removeReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
assertNull(admin.getReplicationPeerConfig(ID_ONE).get().getTableCFsMap());
|
||||||
|
|
||||||
|
tableCFs.clear();
|
||||||
|
tableCFs.put(tableName4, new ArrayList<>());
|
||||||
|
admin.appendReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
admin.removeReplicationPeerTableCFs(ID_ONE, tableCFs).join();
|
||||||
|
assertNull(admin.getReplicationPeerConfig(ID_ONE).get().getTableCFsMap());
|
||||||
|
|
||||||
|
admin.removeReplicationPeer(ID_ONE);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testSetPeerNamespaces() throws Exception {
|
||||||
|
String ns1 = "ns1";
|
||||||
|
String ns2 = "ns2";
|
||||||
|
|
||||||
|
ReplicationPeerConfig rpc = new ReplicationPeerConfig();
|
||||||
|
rpc.setClusterKey(KEY_ONE);
|
||||||
|
admin.addReplicationPeer(ID_ONE, rpc).join();
|
||||||
|
|
||||||
|
// add ns1 and ns2 to peer config
|
||||||
|
rpc = admin.getReplicationPeerConfig(ID_ONE).get();
|
||||||
|
Set<String> namespaces = new HashSet<>();
|
||||||
|
namespaces.add(ns1);
|
||||||
|
namespaces.add(ns2);
|
||||||
|
rpc.setNamespaces(namespaces);
|
||||||
|
admin.updateReplicationPeerConfig(ID_ONE, rpc).join();
|
||||||
|
namespaces = admin.getReplicationPeerConfig(ID_ONE).get().getNamespaces();
|
||||||
|
assertEquals(2, namespaces.size());
|
||||||
|
assertTrue(namespaces.contains(ns1));
|
||||||
|
assertTrue(namespaces.contains(ns2));
|
||||||
|
|
||||||
|
// update peer config only contains ns1
|
||||||
|
rpc = admin.getReplicationPeerConfig(ID_ONE).get();
|
||||||
|
namespaces.clear();
|
||||||
|
namespaces.add(ns1);
|
||||||
|
rpc.setNamespaces(namespaces);
|
||||||
|
admin.updateReplicationPeerConfig(ID_ONE, rpc).join();
|
||||||
|
namespaces = admin.getReplicationPeerConfig(ID_ONE).get().getNamespaces();
|
||||||
|
assertEquals(1, namespaces.size());
|
||||||
|
assertTrue(namespaces.contains(ns1));
|
||||||
|
|
||||||
|
admin.removeReplicationPeer(ID_ONE).join();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testNamespacesAndTableCfsConfigConflict() throws Exception {
|
||||||
|
String ns1 = "ns1";
|
||||||
|
String ns2 = "ns2";
|
||||||
|
final TableName tableName1 = TableName.valueOf(ns1 + ":" + name.getMethodName());
|
||||||
|
final TableName tableName2 = TableName.valueOf(ns2 + ":" + name.getMethodName() + "2");
|
||||||
|
|
||||||
|
ReplicationPeerConfig rpc = new ReplicationPeerConfig();
|
||||||
|
rpc.setClusterKey(KEY_ONE);
|
||||||
|
admin.addReplicationPeer(ID_ONE, rpc).join();
|
||||||
|
|
||||||
|
rpc = admin.getReplicationPeerConfig(ID_ONE).get();
|
||||||
|
Set<String> namespaces = new HashSet<String>();
|
||||||
|
namespaces.add(ns1);
|
||||||
|
rpc.setNamespaces(namespaces);
|
||||||
|
admin.updateReplicationPeerConfig(ID_ONE, rpc).get();
|
||||||
|
rpc = admin.getReplicationPeerConfig(ID_ONE).get();
|
||||||
|
Map<TableName, List<String>> tableCfs = new HashMap<>();
|
||||||
|
tableCfs.put(tableName1, new ArrayList<>());
|
||||||
|
rpc.setTableCFsMap(tableCfs);
|
||||||
|
try {
|
||||||
|
admin.updateReplicationPeerConfig(ID_ONE, rpc).join();
|
||||||
|
fail("Test case should fail, because table " + tableName1 + " conflict with namespace " + ns1);
|
||||||
|
} catch (CompletionException e) {
|
||||||
|
// OK
|
||||||
|
}
|
||||||
|
|
||||||
|
rpc = admin.getReplicationPeerConfig(ID_ONE).get();
|
||||||
|
tableCfs.clear();
|
||||||
|
tableCfs.put(tableName2, new ArrayList<>());
|
||||||
|
rpc.setTableCFsMap(tableCfs);
|
||||||
|
admin.updateReplicationPeerConfig(ID_ONE, rpc).get();
|
||||||
|
rpc = admin.getReplicationPeerConfig(ID_ONE).get();
|
||||||
|
namespaces.clear();
|
||||||
|
namespaces.add(ns2);
|
||||||
|
rpc.setNamespaces(namespaces);
|
||||||
|
try {
|
||||||
|
admin.updateReplicationPeerConfig(ID_ONE, rpc).join();
|
||||||
|
fail("Test case should fail, because namespace " + ns2 + " conflict with table " + tableName2);
|
||||||
|
} catch (CompletionException e) {
|
||||||
|
// OK
|
||||||
|
}
|
||||||
|
|
||||||
|
admin.removeReplicationPeer(ID_ONE).join();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testPeerBandwidth() throws Exception {
|
||||||
|
ReplicationPeerConfig rpc = new ReplicationPeerConfig();
|
||||||
|
rpc.setClusterKey(KEY_ONE);
|
||||||
|
|
||||||
|
admin.addReplicationPeer(ID_ONE, rpc).join();
|
||||||
|
rpc = admin.getReplicationPeerConfig(ID_ONE).get();
|
||||||
|
assertEquals(0, rpc.getBandwidth());
|
||||||
|
|
||||||
|
rpc.setBandwidth(2097152);
|
||||||
|
admin.updateReplicationPeerConfig(ID_ONE, rpc).join();
|
||||||
|
assertEquals(2097152, admin.getReplicationPeerConfig(ID_ONE).join().getBandwidth());
|
||||||
|
|
||||||
|
admin.removeReplicationPeer(ID_ONE).join();
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue