HBASE-24964 Remove MetaTableAccessor.tableExists (#2330)
Signed-off-by: Viraj Jasani <vjasani@apache.org>
This commit is contained in:
parent
54454f8de6
commit
e022094400
|
@ -446,21 +446,6 @@ public class MetaTableAccessor {
|
|||
PrivateCellUtil.qualifierStartsWith(cell, HConstants.MERGE_QUALIFIER_PREFIX);
|
||||
}
|
||||
|
||||
/**
|
||||
* Checks if the specified table exists. Looks at the hbase:meta table hosted on
|
||||
* the specified server.
|
||||
* @param connection connection we're using
|
||||
* @param tableName table to check
|
||||
* @return true if the table exists in meta, false if not
|
||||
*/
|
||||
public static boolean tableExists(Connection connection,
|
||||
final TableName tableName)
|
||||
throws IOException {
|
||||
// Catalog tables always exist.
|
||||
return tableName.equals(TableName.META_TABLE_NAME) ||
|
||||
getTableState(connection, tableName) != null;
|
||||
}
|
||||
|
||||
/**
|
||||
* Lists all of the regions currently in META.
|
||||
*
|
||||
|
|
|
@ -464,7 +464,7 @@ public class HBaseAdmin implements Admin {
|
|||
return executeCallable(new RpcRetryingCallable<Boolean>() {
|
||||
@Override
|
||||
protected Boolean rpcCall(int callTimeout) throws Exception {
|
||||
return MetaTableAccessor.tableExists(connection, tableName);
|
||||
return MetaTableAccessor.getTableState(getConnection(), tableName) != null;
|
||||
}
|
||||
});
|
||||
}
|
||||
|
@ -2026,7 +2026,7 @@ public class HBaseAdmin implements Admin {
|
|||
return executeCallable(new RpcRetryingCallable<TableName>() {
|
||||
@Override
|
||||
protected TableName rpcCall(int callTimeout) throws Exception {
|
||||
if (!MetaTableAccessor.tableExists(connection, tableName)) {
|
||||
if (MetaTableAccessor.getTableState(getConnection(), tableName) == null) {
|
||||
throw new TableNotFoundException(tableName);
|
||||
}
|
||||
return tableName;
|
||||
|
|
|
@ -28,6 +28,14 @@ import org.apache.hadoop.hbase.client.TableDescriptor;
|
|||
*/
|
||||
@InterfaceAudience.Private
|
||||
public interface TableDescriptors {
|
||||
|
||||
/**
|
||||
* Test whether a given table exists, i.e, has a table descriptor.
|
||||
*/
|
||||
default boolean exists(TableName tableName) throws IOException {
|
||||
return get(tableName) != null;
|
||||
}
|
||||
|
||||
/**
|
||||
* @return TableDescriptor for tablename
|
||||
*/
|
||||
|
|
|
@ -2577,7 +2577,7 @@ public class HMaster extends HRegionServer implements MasterServices {
|
|||
|
||||
private void checkTableExists(final TableName tableName)
|
||||
throws IOException, TableNotFoundException {
|
||||
if (!MetaTableAccessor.tableExists(getConnection(), tableName)) {
|
||||
if (!tableDescriptors.exists(tableName)) {
|
||||
throw new TableNotFoundException(tableName);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -21,7 +21,6 @@ package org.apache.hadoop.hbase.master;
|
|||
import java.io.IOException;
|
||||
import java.io.InterruptedIOException;
|
||||
import java.util.NavigableSet;
|
||||
|
||||
import org.apache.commons.lang3.StringUtils;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hbase.Cell;
|
||||
|
@ -30,7 +29,6 @@ import org.apache.hadoop.hbase.CellBuilderType;
|
|||
import org.apache.hadoop.hbase.CellUtil;
|
||||
import org.apache.hadoop.hbase.DoNotRetryIOException;
|
||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.NamespaceDescriptor;
|
||||
import org.apache.hadoop.hbase.Stoppable;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
|
@ -48,9 +46,6 @@ import org.apache.hadoop.hbase.exceptions.TimeoutIOException;
|
|||
import org.apache.hadoop.hbase.master.procedure.MasterProcedureEnv;
|
||||
import org.apache.hadoop.hbase.master.procedure.ProcedurePrepareLatch;
|
||||
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
||||
import org.apache.hbase.thirdparty.com.google.common.collect.Sets;
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
||||
import org.apache.hadoop.hbase.util.Threads;
|
||||
|
@ -58,6 +53,11 @@ import org.apache.yetus.audience.InterfaceAudience;
|
|||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import org.apache.hbase.thirdparty.com.google.common.collect.Sets;
|
||||
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
|
||||
|
||||
/**
|
||||
* This is a helper class used internally to manage the namespace metadata that is stored in
|
||||
* TableName.NAMESPACE_TABLE_NAME. It also mirrors updates to the ZK store by forwarding updates to
|
||||
|
@ -91,8 +91,7 @@ public class TableNamespaceManager implements Stoppable {
|
|||
}
|
||||
|
||||
public void start() throws IOException {
|
||||
if (!MetaTableAccessor.tableExists(masterServices.getConnection(),
|
||||
TableName.NAMESPACE_TABLE_NAME)) {
|
||||
if (!masterServices.getTableDescriptors().exists(TableName.NAMESPACE_TABLE_NAME)) {
|
||||
LOG.info("Namespace table not found. Creating...");
|
||||
createNamespaceTable(masterServices);
|
||||
}
|
||||
|
|
|
@ -19,8 +19,6 @@
|
|||
package org.apache.hadoop.hbase.master.procedure;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.TableNotFoundException;
|
||||
import org.apache.hadoop.hbase.client.RegionInfo;
|
||||
|
@ -78,19 +76,6 @@ public abstract class AbstractStateMachineRegionProcedure<TState>
|
|||
sb.append(", region=").append(getRegion().getShortNameToLog());
|
||||
}
|
||||
|
||||
/**
|
||||
* Check whether a table is modifiable - exists and either offline or online with config set
|
||||
* @param env MasterProcedureEnv
|
||||
* @throws IOException
|
||||
*/
|
||||
@Override
|
||||
protected void checkTableModifiable(final MasterProcedureEnv env) throws IOException {
|
||||
// Checks whether the table exists
|
||||
if (!MetaTableAccessor.tableExists(env.getMasterServices().getConnection(), getTableName())) {
|
||||
throw new TableNotFoundException(getTableName());
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
protected boolean holdLock(MasterProcedureEnv env) {
|
||||
return true;
|
||||
|
|
|
@ -20,7 +20,6 @@ package org.apache.hadoop.hbase.master.procedure;
|
|||
import java.io.IOException;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hbase.HBaseIOException;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.TableNotDisabledException;
|
||||
import org.apache.hadoop.hbase.TableNotEnabledException;
|
||||
|
@ -120,11 +119,10 @@ public abstract class AbstractStateMachineTableProcedure<TState>
|
|||
/**
|
||||
* Check whether a table is modifiable - exists and either offline or online with config set
|
||||
* @param env MasterProcedureEnv
|
||||
* @throws IOException
|
||||
*/
|
||||
protected void checkTableModifiable(final MasterProcedureEnv env) throws IOException {
|
||||
// Checks whether the table exists
|
||||
if (!MetaTableAccessor.tableExists(env.getMasterServices().getConnection(), getTableName())) {
|
||||
if (!env.getMasterServices().getTableDescriptors().exists(getTableName())) {
|
||||
throw new TableNotFoundException(getTableName());
|
||||
}
|
||||
}
|
||||
|
|
|
@ -27,7 +27,6 @@ import java.util.Map;
|
|||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.TableExistsException;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.RegionInfo;
|
||||
|
@ -330,12 +329,11 @@ public class CloneSnapshotProcedure
|
|||
/**
|
||||
* Action before any real action of cloning from snapshot.
|
||||
* @param env MasterProcedureEnv
|
||||
* @throws IOException
|
||||
*/
|
||||
private void prepareClone(final MasterProcedureEnv env) throws IOException {
|
||||
final TableName tableName = getTableName();
|
||||
if (MetaTableAccessor.tableExists(env.getMasterServices().getConnection(), tableName)) {
|
||||
throw new TableExistsException(getTableName());
|
||||
if (env.getMasterServices().getTableDescriptors().exists(tableName)) {
|
||||
throw new TableExistsException(tableName);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -240,7 +240,7 @@ public class CreateTableProcedure
|
|||
|
||||
private boolean prepareCreate(final MasterProcedureEnv env) throws IOException {
|
||||
final TableName tableName = getTableName();
|
||||
if (MetaTableAccessor.tableExists(env.getMasterServices().getConnection(), tableName)) {
|
||||
if (env.getMasterServices().getTableDescriptors().exists(tableName)) {
|
||||
setFailure("master-create-table", new TableExistsException(getTableName()));
|
||||
return false;
|
||||
}
|
||||
|
|
|
@ -246,7 +246,7 @@ public class DisableTableProcedure
|
|||
setFailure("master-disable-table",
|
||||
new ConstraintException("Cannot disable " + this.tableName));
|
||||
canTableBeDisabled = false;
|
||||
} else if (!MetaTableAccessor.tableExists(env.getMasterServices().getConnection(), tableName)) {
|
||||
} else if (!env.getMasterServices().getTableDescriptors().exists(tableName)) {
|
||||
setFailure("master-disable-table", new TableNotFoundException(tableName));
|
||||
canTableBeDisabled = false;
|
||||
} else if (!skipTableStateCheck) {
|
||||
|
|
|
@ -281,9 +281,8 @@ public class EnableTableProcedure
|
|||
*/
|
||||
private boolean prepareEnable(final MasterProcedureEnv env) throws IOException {
|
||||
boolean canTableBeEnabled = true;
|
||||
|
||||
// Check whether table exists
|
||||
if (!MetaTableAccessor.tableExists(env.getMasterServices().getConnection(), tableName)) {
|
||||
if (!env.getMasterServices().getTableDescriptors().exists(tableName)) {
|
||||
setFailure("master-enable-table", new TableNotFoundException(tableName));
|
||||
canTableBeEnabled = false;
|
||||
} else {
|
||||
|
|
|
@ -275,7 +275,7 @@ public class ModifyTableProcedure
|
|||
*/
|
||||
private void prepareModify(final MasterProcedureEnv env) throws IOException {
|
||||
// Checks whether the table exists
|
||||
if (!MetaTableAccessor.tableExists(env.getMasterServices().getConnection(), getTableName())) {
|
||||
if (!env.getMasterServices().getTableDescriptors().exists(getTableName())) {
|
||||
throw new TableNotFoundException(getTableName());
|
||||
}
|
||||
|
||||
|
|
|
@ -24,7 +24,6 @@ import java.util.HashMap;
|
|||
import java.util.Iterator;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hbase.DoNotRetryIOException;
|
||||
|
@ -54,6 +53,7 @@ import org.apache.hadoop.hbase.util.Pair;
|
|||
import org.apache.yetus.audience.InterfaceAudience;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos;
|
||||
|
@ -330,7 +330,7 @@ public class RestoreSnapshotProcedure
|
|||
private void prepareRestore(final MasterProcedureEnv env) throws IOException {
|
||||
final TableName tableName = getTableName();
|
||||
// Checks whether the table exists
|
||||
if (!MetaTableAccessor.tableExists(env.getMasterServices().getConnection(), tableName)) {
|
||||
if (!env.getMasterServices().getTableDescriptors().exists(tableName)) {
|
||||
throw new TableNotFoundException(tableName);
|
||||
}
|
||||
|
||||
|
|
|
@ -63,7 +63,7 @@ public class SlowLogMasterService {
|
|||
LOG.info("Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting.");
|
||||
return;
|
||||
}
|
||||
if (!MetaTableAccessor.tableExists(masterServices.getConnection(),
|
||||
if (!masterServices.getTableDescriptors().exists(
|
||||
SlowLogTableAccessor.SLOW_LOG_TABLE_NAME)) {
|
||||
LOG.info("slowlog table not found. Creating.");
|
||||
this.masterServices.createSystemTable(TABLE_DESCRIPTOR_BUILDER.build());
|
||||
|
|
|
@ -42,7 +42,6 @@ import org.apache.hadoop.fs.FileSystem;
|
|||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
|
||||
import org.apache.hadoop.hbase.HConstants;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.Stoppable;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.TableDescriptor;
|
||||
|
@ -851,7 +850,7 @@ public class SnapshotManager extends MasterProcedureManager implements Stoppable
|
|||
|
||||
// Execute the restore/clone operation
|
||||
long procId;
|
||||
if (MetaTableAccessor.tableExists(master.getConnection(), tableName)) {
|
||||
if (master.getTableDescriptors().exists(tableName)) {
|
||||
procId = restoreSnapshot(reqSnapshot, tableName, snapshot, snapshotTableDesc, nonceKey,
|
||||
restoreAcl);
|
||||
} else {
|
||||
|
|
|
@ -66,7 +66,7 @@ public class NamespaceAuditor {
|
|||
public void checkQuotaToCreateTable(TableName tName, int regions) throws IOException {
|
||||
if (stateManager.isInitialized()) {
|
||||
// We do this check to fail fast.
|
||||
if (MetaTableAccessor.tableExists(this.masterServices.getConnection(), tName)) {
|
||||
if (masterServices.getTableDescriptors().exists(tName)) {
|
||||
throw new TableExistsException(tName);
|
||||
}
|
||||
stateManager.checkAndUpdateNamespaceTableCount(tName, regions);
|
||||
|
|
|
@ -28,12 +28,10 @@ import java.util.Map.Entry;
|
|||
import java.util.Optional;
|
||||
import java.util.Set;
|
||||
import java.util.concurrent.ConcurrentHashMap;
|
||||
|
||||
import org.apache.commons.lang3.builder.HashCodeBuilder;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.hbase.DoNotRetryIOException;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.NamespaceDescriptor;
|
||||
import org.apache.hadoop.hbase.RegionStateListener;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
|
@ -105,8 +103,7 @@ public class MasterQuotaManager implements RegionStateListener {
|
|||
}
|
||||
|
||||
// Create the quota table if missing
|
||||
if (!MetaTableAccessor.tableExists(masterServices.getConnection(),
|
||||
QuotaUtil.QUOTA_TABLE_NAME)) {
|
||||
if (!masterServices.getTableDescriptors().exists(QuotaUtil.QUOTA_TABLE_NAME)) {
|
||||
LOG.info("Quota table not found. Creating...");
|
||||
createQuotaTable();
|
||||
}
|
||||
|
|
|
@ -21,19 +21,18 @@ import java.util.HashMap;
|
|||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.ScheduledChore;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.yetus.audience.InterfaceAudience;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
import org.apache.hadoop.hbase.client.Connection;
|
||||
import org.apache.hadoop.hbase.client.Result;
|
||||
import org.apache.hadoop.hbase.client.ResultScanner;
|
||||
import org.apache.hadoop.hbase.client.Table;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.yetus.audience.InterfaceAudience;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* A {@link ScheduledChore} which periodically updates the {@link RegionServerSpaceQuotaManager}
|
||||
|
@ -160,7 +159,9 @@ public class SpaceQuotaRefresherChore extends ScheduledChore {
|
|||
* @throws IOException throws IOException
|
||||
*/
|
||||
boolean checkQuotaTableExists() throws IOException {
|
||||
return MetaTableAccessor.tableExists(getConnection(), QuotaUtil.QUOTA_TABLE_NAME);
|
||||
try (Admin admin = getConnection().getAdmin()) {
|
||||
return admin.tableExists(QuotaUtil.QUOTA_TABLE_NAME);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -122,6 +122,6 @@ public class SnapshotScannerHDFSAclCleaner extends BaseHFileCleanerDelegate {
|
|||
}
|
||||
|
||||
private boolean tableExists(TableName tableName) throws IOException {
|
||||
return master != null && MetaTableAccessor.tableExists(master.getConnection(), tableName);
|
||||
return master != null && master.getTableDescriptors().exists(tableName);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -18,16 +18,15 @@
|
|||
|
||||
package org.apache.hadoop.hbase.security.visibility;
|
||||
|
||||
import com.google.protobuf.ByteString;
|
||||
import com.google.protobuf.RpcCallback;
|
||||
import com.google.protobuf.RpcController;
|
||||
import com.google.protobuf.Service;
|
||||
|
||||
import static org.apache.hadoop.hbase.HConstants.OperationStatusCode.SANITY_CHECK_FAILURE;
|
||||
import static org.apache.hadoop.hbase.HConstants.OperationStatusCode.SUCCESS;
|
||||
import static org.apache.hadoop.hbase.security.visibility.VisibilityConstants.LABELS_TABLE_FAMILY;
|
||||
import static org.apache.hadoop.hbase.security.visibility.VisibilityConstants.LABELS_TABLE_NAME;
|
||||
|
||||
import com.google.protobuf.ByteString;
|
||||
import com.google.protobuf.RpcCallback;
|
||||
import com.google.protobuf.RpcController;
|
||||
import com.google.protobuf.Service;
|
||||
import java.io.IOException;
|
||||
import java.net.InetAddress;
|
||||
import java.util.ArrayList;
|
||||
|
@ -38,7 +37,6 @@ import java.util.List;
|
|||
import java.util.Map;
|
||||
import java.util.Objects;
|
||||
import java.util.Optional;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hbase.AuthUtil;
|
||||
import org.apache.hadoop.hbase.Cell;
|
||||
|
@ -46,15 +44,13 @@ import org.apache.hadoop.hbase.CellScanner;
|
|||
import org.apache.hadoop.hbase.CoprocessorEnvironment;
|
||||
import org.apache.hadoop.hbase.DoNotRetryIOException;
|
||||
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
|
||||
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.PrivateCellUtil;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.Tag;
|
||||
import org.apache.hadoop.hbase.TagType;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
import org.apache.hadoop.hbase.client.Append;
|
||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
|
||||
import org.apache.hadoop.hbase.client.Delete;
|
||||
import org.apache.hadoop.hbase.client.Get;
|
||||
import org.apache.hadoop.hbase.client.Increment;
|
||||
|
@ -64,6 +60,7 @@ import org.apache.hadoop.hbase.client.Put;
|
|||
import org.apache.hadoop.hbase.client.Result;
|
||||
import org.apache.hadoop.hbase.client.Scan;
|
||||
import org.apache.hadoop.hbase.client.TableDescriptor;
|
||||
import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
|
||||
import org.apache.hadoop.hbase.constraint.ConstraintException;
|
||||
import org.apache.hadoop.hbase.coprocessor.CoprocessorException;
|
||||
import org.apache.hadoop.hbase.coprocessor.CoprocessorHost;
|
||||
|
@ -108,8 +105,6 @@ import org.apache.hadoop.hbase.security.Superusers;
|
|||
import org.apache.hadoop.hbase.security.User;
|
||||
import org.apache.hadoop.hbase.security.access.AccessChecker;
|
||||
import org.apache.hadoop.hbase.security.access.AccessController;
|
||||
import org.apache.hbase.thirdparty.com.google.common.collect.Lists;
|
||||
import org.apache.hbase.thirdparty.com.google.common.collect.MapMaker;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.hadoop.hbase.util.Pair;
|
||||
import org.apache.hadoop.util.StringUtils;
|
||||
|
@ -117,6 +112,9 @@ import org.apache.yetus.audience.InterfaceAudience;
|
|||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import org.apache.hbase.thirdparty.com.google.common.collect.Lists;
|
||||
import org.apache.hbase.thirdparty.com.google.common.collect.MapMaker;
|
||||
|
||||
/**
|
||||
* Coprocessor that has both the MasterObserver and RegionObserver implemented that supports in
|
||||
* visibility labels
|
||||
|
@ -207,21 +205,21 @@ public class VisibilityController implements MasterCoprocessor, RegionCoprocesso
|
|||
/********************************* Master related hooks **********************************/
|
||||
|
||||
@Override
|
||||
public void postStartMaster(ObserverContext<MasterCoprocessorEnvironment> ctx) throws IOException {
|
||||
public void postStartMaster(ObserverContext<MasterCoprocessorEnvironment> ctx)
|
||||
throws IOException {
|
||||
// Need to create the new system table for labels here
|
||||
if (!MetaTableAccessor.tableExists(ctx.getEnvironment().getConnection(), LABELS_TABLE_NAME)) {
|
||||
HTableDescriptor labelsTable = new HTableDescriptor(LABELS_TABLE_NAME);
|
||||
HColumnDescriptor labelsColumn = new HColumnDescriptor(LABELS_TABLE_FAMILY);
|
||||
labelsColumn.setBloomFilterType(BloomType.NONE);
|
||||
labelsColumn.setBlockCacheEnabled(false); // We will cache all the labels. No need of normal
|
||||
// table block cache.
|
||||
labelsTable.addFamily(labelsColumn);
|
||||
// Let the "labels" table having only one region always. We are not expecting too many labels in
|
||||
// the system.
|
||||
labelsTable.setValue(HTableDescriptor.SPLIT_POLICY,
|
||||
DisabledRegionSplitPolicy.class.getName());
|
||||
try (Admin admin = ctx.getEnvironment().getConnection().getAdmin()) {
|
||||
admin.createTable(labelsTable);
|
||||
if (!admin.tableExists(LABELS_TABLE_NAME)) {
|
||||
// We will cache all the labels. No need of normal table block cache.
|
||||
// Let the "labels" table having only one region always. We are not expecting too many
|
||||
// labels in the system.
|
||||
TableDescriptor tableDescriptor = TableDescriptorBuilder.newBuilder(LABELS_TABLE_NAME)
|
||||
.setColumnFamily(ColumnFamilyDescriptorBuilder.newBuilder(LABELS_TABLE_FAMILY)
|
||||
.setBloomFilterType(BloomType.NONE).setBlockCacheEnabled(false).build())
|
||||
.setValue(TableDescriptorBuilder.SPLIT_POLICY, DisabledRegionSplitPolicy.class.getName())
|
||||
.build();
|
||||
|
||||
admin.createTable(tableDescriptor);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,7 +19,6 @@ package org.apache.hadoop.hbase;
|
|||
|
||||
import static org.junit.Assert.assertArrayEquals;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertNotEquals;
|
||||
import static org.junit.Assert.assertNotNull;
|
||||
import static org.junit.Assert.assertNull;
|
||||
|
@ -328,25 +327,8 @@ public class TestMetaTableAccessor {
|
|||
MetaTableLocator.getMetaRegionsAndLocations(UTIL.getZooKeeperWatcher()).size() >= 1);
|
||||
}
|
||||
|
||||
@Test public void testTableExists() throws IOException {
|
||||
final TableName tableName = TableName.valueOf(name.getMethodName());
|
||||
assertFalse(MetaTableAccessor.tableExists(connection, tableName));
|
||||
UTIL.createTable(tableName, HConstants.CATALOG_FAMILY);
|
||||
assertTrue(MetaTableAccessor.tableExists(connection, tableName));
|
||||
Admin admin = UTIL.getAdmin();
|
||||
admin.disableTable(tableName);
|
||||
admin.deleteTable(tableName);
|
||||
assertFalse(MetaTableAccessor.tableExists(connection, tableName));
|
||||
assertTrue(MetaTableAccessor.tableExists(connection,
|
||||
TableName.META_TABLE_NAME));
|
||||
UTIL.createTable(tableName, HConstants.CATALOG_FAMILY);
|
||||
assertTrue(MetaTableAccessor.tableExists(connection, tableName));
|
||||
admin.disableTable(tableName);
|
||||
admin.deleteTable(tableName);
|
||||
assertFalse(MetaTableAccessor.tableExists(connection, tableName));
|
||||
}
|
||||
|
||||
@Test public void testGetRegion() throws IOException, InterruptedException {
|
||||
@Test
|
||||
public void testGetRegion() throws IOException, InterruptedException {
|
||||
final String name = this.name.getMethodName();
|
||||
LOG.info("Started " + name);
|
||||
// Test get on non-existent region.
|
||||
|
|
|
@ -37,6 +37,7 @@ import org.apache.hadoop.hbase.MiniHBaseCluster;
|
|||
import org.apache.hadoop.hbase.RegionLocations;
|
||||
import org.apache.hadoop.hbase.ServerName;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
import org.apache.hadoop.hbase.client.BufferedMutator;
|
||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
|
||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
|
||||
|
@ -211,7 +212,7 @@ public class MasterProcedureTestingUtility {
|
|||
LOG.debug("Table directory layout is as expected.");
|
||||
|
||||
// check meta
|
||||
assertTrue(MetaTableAccessor.tableExists(master.getConnection(), tableName));
|
||||
assertTrue(tableExists(master.getConnection(), tableName));
|
||||
assertEquals(regions.length, countMetaRegions(master, tableName));
|
||||
|
||||
// check htd
|
||||
|
@ -232,7 +233,7 @@ public class MasterProcedureTestingUtility {
|
|||
assertFalse(fs.exists(tableDir));
|
||||
|
||||
// check meta
|
||||
assertFalse(MetaTableAccessor.tableExists(master.getConnection(), tableName));
|
||||
assertFalse(tableExists(master.getConnection(), tableName));
|
||||
assertEquals(0, countMetaRegions(master, tableName));
|
||||
|
||||
// check htd
|
||||
|
@ -577,6 +578,12 @@ public class MasterProcedureTestingUtility {
|
|||
}
|
||||
}
|
||||
|
||||
public static boolean tableExists(Connection conn, TableName tableName) throws IOException {
|
||||
try (Admin admin = conn.getAdmin()) {
|
||||
return admin.tableExists(tableName);
|
||||
}
|
||||
}
|
||||
|
||||
public static class InjectAbortOnLoadListener
|
||||
implements ProcedureExecutor.ProcedureExecutorListener {
|
||||
private final ProcedureExecutor<MasterProcedureEnv> procExec;
|
||||
|
|
|
@ -28,7 +28,6 @@ import org.apache.hadoop.conf.Configuration;
|
|||
import org.apache.hadoop.hbase.HBaseClassTestRule;
|
||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||
import org.apache.hadoop.hbase.HConstants;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.client.Connection;
|
||||
import org.apache.hadoop.hbase.client.Result;
|
||||
import org.apache.hadoop.hbase.client.ResultScanner;
|
||||
|
@ -37,12 +36,9 @@ import org.apache.hadoop.hbase.client.Table;
|
|||
import org.apache.hadoop.hbase.namequeues.request.NamedQueueGetRequest;
|
||||
import org.apache.hadoop.hbase.namequeues.response.NamedQueueGetResponse;
|
||||
import org.apache.hadoop.hbase.regionserver.HRegionServer;
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos;
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.TooSlowLog;
|
||||
import org.apache.hadoop.hbase.slowlog.SlowLogTableAccessor;
|
||||
import org.apache.hadoop.hbase.testclassification.MasterTests;
|
||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||
import org.apache.hbase.thirdparty.com.google.common.util.concurrent.Uninterruptibles;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Before;
|
||||
|
@ -53,6 +49,11 @@ import org.junit.experimental.categories.Category;
|
|||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import org.apache.hbase.thirdparty.com.google.common.util.concurrent.Uninterruptibles;
|
||||
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos;
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.TooSlowLog;
|
||||
|
||||
/**
|
||||
* Tests for SlowLog System Table
|
||||
*/
|
||||
|
@ -167,17 +168,16 @@ public class TestSlowLogAccessor {
|
|||
}
|
||||
}
|
||||
|
||||
private Connection waitForSlowLogTableCreation() {
|
||||
Connection connection =
|
||||
HBASE_TESTING_UTILITY.getMiniHBaseCluster().getRegionServer(0).getConnection();
|
||||
private Connection waitForSlowLogTableCreation() throws IOException {
|
||||
Assert.assertNotEquals(-1, HBASE_TESTING_UTILITY.waitFor(2000, () -> {
|
||||
try {
|
||||
return MetaTableAccessor.tableExists(connection, SlowLogTableAccessor.SLOW_LOG_TABLE_NAME);
|
||||
return HBASE_TESTING_UTILITY.getAdmin()
|
||||
.tableExists(SlowLogTableAccessor.SLOW_LOG_TABLE_NAME);
|
||||
} catch (IOException e) {
|
||||
return false;
|
||||
}
|
||||
}));
|
||||
return connection;
|
||||
return HBASE_TESTING_UTILITY.getConnection();
|
||||
}
|
||||
|
||||
@Test
|
||||
|
|
|
@ -19,17 +19,16 @@ package org.apache.hadoop.hbase.quotas;
|
|||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Arrays;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hbase.HBaseClassTestRule;
|
||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.Waiter;
|
||||
import org.apache.hadoop.hbase.client.Admin;
|
||||
|
@ -43,7 +42,6 @@ import org.apache.hadoop.hbase.client.Table;
|
|||
import org.apache.hadoop.hbase.client.TableDescriptor;
|
||||
import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
|
||||
import org.apache.hadoop.hbase.master.HMaster;
|
||||
import org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos;
|
||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.junit.After;
|
||||
|
@ -96,9 +94,9 @@ public class TestQuotaObserverChoreRegionReports {
|
|||
TEST_UTIL.startMiniCluster(1);
|
||||
// Wait till quota table onlined.
|
||||
TEST_UTIL.waitFor(10000, new Waiter.Predicate<Exception>() {
|
||||
@Override public boolean evaluate() throws Exception {
|
||||
return MetaTableAccessor.tableExists(TEST_UTIL.getConnection(),
|
||||
QuotaTableUtil.QUOTA_TABLE_NAME);
|
||||
@Override
|
||||
public boolean evaluate() throws Exception {
|
||||
return TEST_UTIL.getAdmin().tableExists(QuotaTableUtil.QUOTA_TABLE_NAME);
|
||||
}
|
||||
});
|
||||
|
||||
|
@ -155,9 +153,9 @@ public class TestQuotaObserverChoreRegionReports {
|
|||
TEST_UTIL.startMiniCluster(1);
|
||||
// Wait till quota table onlined.
|
||||
TEST_UTIL.waitFor(10000, new Waiter.Predicate<Exception>() {
|
||||
@Override public boolean evaluate() throws Exception {
|
||||
return MetaTableAccessor.tableExists(TEST_UTIL.getConnection(),
|
||||
QuotaTableUtil.QUOTA_TABLE_NAME);
|
||||
@Override
|
||||
public boolean evaluate() throws Exception {
|
||||
return TEST_UTIL.getAdmin().tableExists(QuotaTableUtil.QUOTA_TABLE_NAME);
|
||||
}
|
||||
});
|
||||
|
||||
|
|
|
@ -30,7 +30,6 @@ import org.apache.hadoop.hbase.Cell;
|
|||
import org.apache.hadoop.hbase.CellScanner;
|
||||
import org.apache.hadoop.hbase.HBaseClassTestRule;
|
||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.Waiter;
|
||||
import org.apache.hadoop.hbase.Waiter.Predicate;
|
||||
|
@ -88,9 +87,9 @@ public class TestSpaceQuotasWithSnapshots {
|
|||
TEST_UTIL.startMiniCluster(1);
|
||||
// Wait till quota table onlined.
|
||||
TEST_UTIL.waitFor(10000, new Waiter.Predicate<Exception>() {
|
||||
@Override public boolean evaluate() throws Exception {
|
||||
return MetaTableAccessor.tableExists(TEST_UTIL.getConnection(),
|
||||
QuotaTableUtil.QUOTA_TABLE_NAME);
|
||||
@Override
|
||||
public boolean evaluate() throws Exception {
|
||||
return TEST_UTIL.getAdmin().tableExists(QuotaTableUtil.QUOTA_TABLE_NAME);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
|
|
@ -50,7 +50,6 @@ import org.apache.hadoop.hbase.HBaseTestingUtility;
|
|||
import org.apache.hadoop.hbase.HConstants;
|
||||
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||
import org.apache.hadoop.hbase.MasterNotRunningException;
|
||||
import org.apache.hadoop.hbase.MetaTableAccessor;
|
||||
import org.apache.hadoop.hbase.MiniHBaseCluster;
|
||||
import org.apache.hadoop.hbase.ServerName;
|
||||
import org.apache.hadoop.hbase.StartMiniClusterOption;
|
||||
|
@ -670,8 +669,7 @@ public class TestSplitTransactionOnCluster {
|
|||
admin.balancerSwitch(false, true);
|
||||
// Turn off the meta scanner so it don't remove parent on us.
|
||||
cluster.getMaster().setCatalogJanitorEnabled(false);
|
||||
boolean tableExists = MetaTableAccessor.tableExists(regionServer.getConnection(),
|
||||
tableName);
|
||||
boolean tableExists = TESTING_UTIL.getAdmin().tableExists(tableName);
|
||||
assertEquals("The specified table should be present.", true, tableExists);
|
||||
final HRegion region = findSplittableRegion(oldRegions);
|
||||
regionServerIndex = cluster.getServerWith(region.getRegionInfo().getRegionName());
|
||||
|
@ -691,8 +689,7 @@ public class TestSplitTransactionOnCluster {
|
|||
Thread.sleep(1000);
|
||||
} while ((newRegions.contains(oldRegions.get(0)) || newRegions.contains(oldRegions.get(1)))
|
||||
|| newRegions.size() != 4);
|
||||
tableExists = MetaTableAccessor.tableExists(regionServer.getConnection(),
|
||||
tableName);
|
||||
tableExists = TESTING_UTIL.getAdmin().tableExists(tableName);
|
||||
assertEquals("The specified table should be present.", true, tableExists);
|
||||
// exists works on stale and we see the put after the flush
|
||||
byte[] b1 = "row1".getBytes();
|
||||
|
|
Loading…
Reference in New Issue