HBASE-19541 Removed unnecessary semicolons in hbase-server
This commit is contained in:
parent
afc2cdbaff
commit
2af61718b1
|
@ -103,7 +103,7 @@ public interface Server extends Abortable, Stoppable {
|
|||
// If an exception, just return null
|
||||
}
|
||||
return fs;
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* @return True is the server is Stopping
|
||||
|
|
|
@ -223,7 +223,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
|
|||
@Override
|
||||
public void checkTasks() {
|
||||
rescan(Long.MAX_VALUE);
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* signal the workers that a task was resubmitted by creating the RESCAN node.
|
||||
|
@ -554,7 +554,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
|
|||
/**
|
||||
* task completed with error
|
||||
*/
|
||||
ERR();
|
||||
ERR()
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -58,7 +58,7 @@ public class ObserverContextImpl<E extends CoprocessorEnvironment> implements Ob
|
|||
|
||||
public boolean isBypassable() {
|
||||
return this.bypassable;
|
||||
};
|
||||
}
|
||||
|
||||
public void bypass() {
|
||||
if (!this.bypassable) {
|
||||
|
|
|
@ -212,7 +212,7 @@ public final class FanOutOneBlockAsyncDFSOutputHelper {
|
|||
throw new RuntimeException(e.getCause());
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
Object createObject(ClientProtocol instance, String src, FsPermission masked, String clientName,
|
||||
EnumSetWritable<CreateFlag> flag, boolean createParent, short replication, long blockSize,
|
||||
|
|
|
@ -143,5 +143,5 @@ public interface BlockCache extends Iterable<CachedBlock> {
|
|||
* @param cacheKey the cache key of the block
|
||||
* @param block the hfileblock to be returned
|
||||
*/
|
||||
default void returnBlock(BlockCacheKey cacheKey, Cacheable block){};
|
||||
default void returnBlock(BlockCacheKey cacheKey, Cacheable block){}
|
||||
}
|
||||
|
|
|
@ -73,6 +73,6 @@ public interface Cacheable extends HeapSize {
|
|||
* exclusive memory area of this Cacheable.
|
||||
*/
|
||||
public static enum MemoryType {
|
||||
SHARED, EXCLUSIVE;
|
||||
SHARED, EXCLUSIVE
|
||||
}
|
||||
}
|
||||
|
|
|
@ -834,7 +834,7 @@ public class HFileBlock implements Cacheable {
|
|||
INIT,
|
||||
WRITING,
|
||||
BLOCK_READY
|
||||
};
|
||||
}
|
||||
|
||||
/** Writer state. Used to ensure the correct usage protocol. */
|
||||
private State state = State.INIT;
|
||||
|
|
|
@ -3429,7 +3429,7 @@ public class HMaster extends HRegionServer implements MasterServices {
|
|||
}
|
||||
if (this.serverManager.addServerToDrainList(server)) {
|
||||
serversAdded.add(server);
|
||||
};
|
||||
}
|
||||
}
|
||||
// Move the regions off the decommissioned servers.
|
||||
if (offload) {
|
||||
|
|
|
@ -30,17 +30,17 @@ public interface ServerListener {
|
|||
/**
|
||||
* Started waiting on RegionServers to check-in.
|
||||
*/
|
||||
default void waiting() {};
|
||||
default void waiting() {}
|
||||
|
||||
/**
|
||||
* The server has joined the cluster.
|
||||
* @param serverName The remote servers name.
|
||||
*/
|
||||
default void serverAdded(final ServerName serverName) {};
|
||||
default void serverAdded(final ServerName serverName) {}
|
||||
|
||||
/**
|
||||
* The server was removed from the cluster.
|
||||
* @param serverName The remote servers name.
|
||||
*/
|
||||
default void serverRemoved(final ServerName serverName) {};
|
||||
default void serverRemoved(final ServerName serverName) {}
|
||||
}
|
||||
|
|
|
@ -30,7 +30,7 @@ import org.apache.yetus.audience.InterfaceStability;
|
|||
public interface ServerProcedureInterface {
|
||||
public enum ServerOperationType {
|
||||
CRASH_HANDLER
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* @return Name of this server instance.
|
||||
|
|
|
@ -33,7 +33,7 @@ public interface TableProcedureInterface {
|
|||
CREATE, DELETE, DISABLE, EDIT, ENABLE, READ,
|
||||
REGION_EDIT, REGION_SPLIT, REGION_MERGE, REGION_ASSIGN, REGION_UNASSIGN,
|
||||
REGION_GC, MERGED_REGIONS_GC/* region operations */
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* @return the name of the table the procedure is operating on
|
||||
|
|
|
@ -59,6 +59,6 @@ public abstract class MobCompactionRequest {
|
|||
/**
|
||||
* All of mob files are selected.
|
||||
*/
|
||||
ALL_FILES;
|
||||
ALL_FILES
|
||||
}
|
||||
}
|
||||
|
|
|
@ -29,7 +29,7 @@ public interface MonitoredTask extends Cloneable {
|
|||
RUNNING,
|
||||
WAITING,
|
||||
COMPLETE,
|
||||
ABORTED;
|
||||
ABORTED
|
||||
}
|
||||
|
||||
long getStartTime();
|
||||
|
|
|
@ -337,5 +337,5 @@ abstract public class Subprocedure implements Callable<Void> {
|
|||
|
||||
@Override
|
||||
public void cleanup(Exception e) {}
|
||||
};
|
||||
}
|
||||
}
|
||||
|
|
|
@ -71,7 +71,7 @@ public class MasterFlushTableProcedureManager extends MasterProcedureManager {
|
|||
private Map<TableName, Procedure> procMap = new HashMap<>();
|
||||
private boolean stopped;
|
||||
|
||||
public MasterFlushTableProcedureManager() {};
|
||||
public MasterFlushTableProcedureManager() {}
|
||||
|
||||
@Override
|
||||
public void stop(String why) {
|
||||
|
|
|
@ -33,5 +33,5 @@ enum FlushType {
|
|||
ABOVE_ONHEAP_HIGHER_MARK,/* happens due to higher mark breach of onheap memstore settings
|
||||
An offheap memstore can even breach the onheap_higher_mark*/
|
||||
ABOVE_OFFHEAP_LOWER_MARK,/* happens due to lower mark breach of offheap memstore settings*/
|
||||
ABOVE_OFFHEAP_HIGHER_MARK;/*/* happens due to higer mark breach of offheap memstore settings*/
|
||||
ABOVE_OFFHEAP_HIGHER_MARK /* happens due to higer mark breach of offheap memstore settings*/
|
||||
}
|
||||
|
|
|
@ -552,8 +552,7 @@ public class StoreFileInfo {
|
|||
if (link != o.link && link != null && !link.equals(o.link)) return false;
|
||||
|
||||
return true;
|
||||
};
|
||||
|
||||
}
|
||||
|
||||
@Override
|
||||
public int hashCode() {
|
||||
|
|
|
@ -290,7 +290,7 @@ public class FSHLog extends AbstractFSWAL<Writer> {
|
|||
*/
|
||||
@VisibleForTesting
|
||||
protected void beforeWaitOnSafePoint() {
|
||||
};
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void doAppend(Writer writer, FSWALEntry entry) throws IOException {
|
||||
|
|
|
@ -87,7 +87,7 @@ class FSWALEntry extends Entry {
|
|||
|
||||
public String toString() {
|
||||
return "sequence=" + this.txid + ", " + super.toString();
|
||||
};
|
||||
}
|
||||
|
||||
boolean isInMemStore() {
|
||||
return this.inMemstore;
|
||||
|
|
|
@ -55,7 +55,7 @@ class AccessControlFilter extends FilterBase {
|
|||
CHECK_TABLE_AND_CF_ONLY,
|
||||
/** Cell permissions can override table or CF permissions */
|
||||
CHECK_CELL_DEFAULT,
|
||||
};
|
||||
}
|
||||
|
||||
private TableAuthManager authManager;
|
||||
private TableName table;
|
||||
|
|
|
@ -31,5 +31,5 @@ public enum Operator {
|
|||
|
||||
public String toString() {
|
||||
return String.valueOf(this.rep);
|
||||
};
|
||||
}
|
||||
}
|
||||
|
|
|
@ -229,4 +229,4 @@ public final class BloomFilterFactory {
|
|||
writer.addInlineBlockWriter(bloomWriter);
|
||||
return bloomWriter;
|
||||
}
|
||||
};
|
||||
}
|
||||
|
|
|
@ -2765,8 +2765,7 @@ public class HBaseFsck extends Configured implements Closeable {
|
|||
handler.handleOverlapGroup(overlapgroup);
|
||||
return null;
|
||||
}
|
||||
};
|
||||
|
||||
}
|
||||
|
||||
/**
|
||||
* Maintain information about a particular table.
|
||||
|
@ -2936,7 +2935,7 @@ public class HBaseFsck extends Configured implements Closeable {
|
|||
+ ". You need to create a new .regioninfo and region "
|
||||
+ "dir in hdfs to plug the hole.");
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* This handler fixes integrity errors from hdfs information. There are
|
||||
|
@ -4429,7 +4428,7 @@ public class HBaseFsck extends Configured implements Closeable {
|
|||
} catch (ExecutionException e) {
|
||||
LOG.error("Unexpected exec exception! Should've been caught already. (Bug?)", e);
|
||||
// Shouldn't happen, we already logged/caught any exceptions in the Runnable
|
||||
};
|
||||
}
|
||||
}
|
||||
} catch (IOException e) {
|
||||
LOG.error("Cannot execute WorkItemHdfsDir for " + tableDir, e);
|
||||
|
@ -4489,7 +4488,7 @@ public class HBaseFsck extends Configured implements Closeable {
|
|||
}
|
||||
return null;
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Display the full report from fsck. This displays all live and dead region
|
||||
|
@ -4858,8 +4857,7 @@ public class HBaseFsck extends Configured implements Closeable {
|
|||
hbck.close();
|
||||
return hbck.getRetCode();
|
||||
}
|
||||
};
|
||||
|
||||
}
|
||||
|
||||
public HBaseFsck exec(ExecutorService exec, String[] args) throws KeeperException, IOException,
|
||||
InterruptedException {
|
||||
|
|
|
@ -64,7 +64,7 @@ abstract public class ShutdownHookManager {
|
|||
|
||||
return Runtime.getRuntime().removeShutdownHook(shutdownHookThread);
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
private static class ShutdownHookManagerV2 extends ShutdownHookManager {
|
||||
public void addShutdownHook(Thread shutdownHookThread, int priority) {
|
||||
|
@ -91,6 +91,5 @@ abstract public class ShutdownHookManager {
|
|||
throw new RuntimeException("we could not use ShutdownHookManager", ex);
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
}
|
||||
}
|
||||
|
|
|
@ -182,8 +182,8 @@ public class TestMetaWithReplicas {
|
|||
util.getAdmin().deleteTable(TABLE);
|
||||
}
|
||||
ServerName master = null;
|
||||
try (Connection c = ConnectionFactory.createConnection(util.getConfiguration());) {
|
||||
try (Table htable = util.createTable(TABLE, FAMILIES);) {
|
||||
try (Connection c = ConnectionFactory.createConnection(util.getConfiguration())) {
|
||||
try (Table htable = util.createTable(TABLE, FAMILIES)) {
|
||||
util.getAdmin().flush(TableName.META_TABLE_NAME);
|
||||
Thread.sleep(conf.getInt(StorefileRefresherChore.REGIONSERVER_STOREFILE_REFRESH_PERIOD,
|
||||
30000) * 6);
|
||||
|
@ -219,7 +219,7 @@ public class TestMetaWithReplicas {
|
|||
Get get = null;
|
||||
Result r = null;
|
||||
byte[] row = "test".getBytes();
|
||||
try (Table htable = c.getTable(TABLE);) {
|
||||
try (Table htable = c.getTable(TABLE)) {
|
||||
Put put = new Put(row);
|
||||
put.addColumn("foo".getBytes(), row, row);
|
||||
BufferedMutator m = c.getBufferedMutator(TABLE);
|
||||
|
@ -237,7 +237,7 @@ public class TestMetaWithReplicas {
|
|||
((ClusterConnection)c).clearRegionCache();
|
||||
}
|
||||
conf.setBoolean(HConstants.USE_META_REPLICAS, false);
|
||||
try (Table htable = c.getTable(TABLE);) {
|
||||
try (Table htable = c.getTable(TABLE)) {
|
||||
r = htable.get(get);
|
||||
assertTrue(Arrays.equals(r.getRow(), row));
|
||||
}
|
||||
|
@ -252,7 +252,7 @@ public class TestMetaWithReplicas {
|
|||
TEST_UTIL.getAdmin().disableTable(tableName);
|
||||
TEST_UTIL.getAdmin().deleteTable(tableName);
|
||||
}
|
||||
try (Table htable = TEST_UTIL.createTable(tableName, FAMILIES);) {
|
||||
try (Table htable = TEST_UTIL.createTable(tableName, FAMILIES)) {
|
||||
byte[] row = "test".getBytes();
|
||||
ConnectionImplementation c = ((ConnectionImplementation) TEST_UTIL.getConnection());
|
||||
// check that metalookup pool would get created
|
||||
|
@ -440,7 +440,7 @@ public class TestMetaWithReplicas {
|
|||
// checks that the when the server holding meta replica is shut down, the meta replica
|
||||
// can be recovered
|
||||
try (ClusterConnection conn = (ClusterConnection)
|
||||
ConnectionFactory.createConnection(TEST_UTIL.getConfiguration());) {
|
||||
ConnectionFactory.createConnection(TEST_UTIL.getConfiguration())) {
|
||||
RegionLocations rl = conn.
|
||||
locateRegion(TableName.META_TABLE_NAME, Bytes.toBytes(""), false, true);
|
||||
HRegionLocation hrl = rl.getRegionLocation(1);
|
||||
|
|
|
@ -87,7 +87,7 @@ public class TestSnapshotWithAcl extends SecureTestUtil {
|
|||
}
|
||||
return null;
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
static class AccessWriteAction implements AccessTestAction {
|
||||
private TableName tableName;
|
||||
|
|
|
@ -90,7 +90,7 @@ public class TestSplitOrMergeStatus {
|
|||
assertFalse(results[0]);
|
||||
admin.split(t.getName());
|
||||
while ((count = admin.getTableRegions(tableName).size()) == originalCount) {
|
||||
Threads.sleep(1);;
|
||||
Threads.sleep(1);
|
||||
}
|
||||
count = admin.getTableRegions(tableName).size();
|
||||
assertTrue(originalCount < count);
|
||||
|
@ -110,7 +110,7 @@ public class TestSplitOrMergeStatus {
|
|||
admin.split(t.getName());
|
||||
int postSplitCount = -1;
|
||||
while ((postSplitCount = admin.getTableRegions(tableName).size()) == originalCount) {
|
||||
Threads.sleep(1);;
|
||||
Threads.sleep(1);
|
||||
}
|
||||
assertTrue("originalCount=" + originalCount + ", newCount=" + postSplitCount,
|
||||
originalCount != postSplitCount);
|
||||
|
|
|
@ -349,7 +349,7 @@ public class TestWALObserver {
|
|||
// Make a new wal for new region open.
|
||||
final WALFactory wals2 = new WALFactory(conf, null,
|
||||
ServerName.valueOf(currentTest.getMethodName()+"2", 16010, System.currentTimeMillis()).toString());
|
||||
WAL wal2 = wals2.getWAL(UNSPECIFIED_REGION, null);;
|
||||
WAL wal2 = wals2.getWAL(UNSPECIFIED_REGION, null);
|
||||
HRegion region = HRegion.openHRegion(newConf, FileSystem.get(newConf), hbaseRootDir,
|
||||
hri, htd, wal2, TEST_UTIL.getHBaseCluster().getRegionServer(0), null);
|
||||
long seqid2 = region.getOpenSeqNum();
|
||||
|
|
|
@ -30,6 +30,5 @@ public enum TagUsage {
|
|||
// KVs with tags
|
||||
ONLY_TAG,
|
||||
// kvs with and without tags
|
||||
PARTIAL_TAG;
|
||||
|
||||
PARTIAL_TAG
|
||||
}
|
||||
|
|
|
@ -87,7 +87,7 @@ public class TestCacheConfig {
|
|||
LOG.info("Deserialized " + b);
|
||||
return cacheable;
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
static class IndexCacheEntry extends DataCacheEntry {
|
||||
private static IndexCacheEntry SINGLETON = new IndexCacheEntry();
|
||||
|
@ -118,7 +118,7 @@ public class TestCacheConfig {
|
|||
@Override
|
||||
public String toString() {
|
||||
return "size=" + SIZE + ", type=" + getBlockType();
|
||||
};
|
||||
}
|
||||
|
||||
@Override
|
||||
public long heapSize() {
|
||||
|
@ -149,7 +149,7 @@ public class TestCacheConfig {
|
|||
public MemoryType getMemoryType() {
|
||||
return MemoryType.EXCLUSIVE;
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
static class MetaCacheEntry extends DataCacheEntry {
|
||||
@Override
|
||||
|
|
|
@ -88,7 +88,7 @@ public class TestHFileDataBlockEncoder {
|
|||
HeapSize heapSize = blockCache.getBlock(cacheKey, false, false, true);
|
||||
assertTrue(heapSize instanceof HFileBlock);
|
||||
|
||||
HFileBlock returnedBlock = (HFileBlock) heapSize;;
|
||||
HFileBlock returnedBlock = (HFileBlock) heapSize;
|
||||
|
||||
if (blockEncoder.getDataBlockEncoding() ==
|
||||
DataBlockEncoding.NONE) {
|
||||
|
|
|
@ -126,7 +126,7 @@ abstract public class MapreduceTestingShim {
|
|||
public String obtainMROutputDirProp() {
|
||||
return "mapred.output.dir";
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
private static class MapreduceV2Shim extends MapreduceTestingShim {
|
||||
public JobContext newJobContext(Configuration jobConf) {
|
||||
|
@ -166,6 +166,5 @@ abstract public class MapreduceTestingShim {
|
|||
// from Hadoop 0.23.x. If we use the source directly we break the hadoop 1.x compile.
|
||||
return "mapreduce.output.fileoutputformat.outputdir";
|
||||
}
|
||||
};
|
||||
|
||||
}
|
||||
}
|
||||
|
|
|
@ -262,7 +262,7 @@ public abstract class AbstractTestDLS {
|
|||
master.balanceSwitch(false);
|
||||
|
||||
try (ZKWatcher zkw = new ZKWatcher(conf, "table-creation", null);
|
||||
Table ht = installTable(zkw, numRegionsToCreate);) {
|
||||
Table ht = installTable(zkw, numRegionsToCreate)) {
|
||||
HRegionServer hrs = findRSToKill(false);
|
||||
List<RegionInfo> regions = ProtobufUtil.getOnlineRegions(hrs.getRSRpcServices());
|
||||
makeWAL(hrs, regions, numLogLines, 100);
|
||||
|
|
|
@ -228,7 +228,7 @@ public class TestMasterNoCluster {
|
|||
while (!master.serviceStarted) Threads.sleep(10);
|
||||
// Fake master that there are regionservers out there. Report in.
|
||||
for (int i = 0; i < sns.length; i++) {
|
||||
RegionServerReportRequest.Builder request = RegionServerReportRequest.newBuilder();;
|
||||
RegionServerReportRequest.Builder request = RegionServerReportRequest.newBuilder();
|
||||
ServerName sn = ServerName.parseVersionedServerName(sns[i].getVersionedBytes());
|
||||
request.setServer(ProtobufUtil.toServerName(sn));
|
||||
request.setLoad(ServerMetricsBuilder.toServerLoad(ServerMetricsBuilder.of(sn)));
|
||||
|
|
|
@ -134,7 +134,7 @@ public class TestMasterShutdown {
|
|||
cluster.waitOnMaster(MASTER_INDEX);
|
||||
} catch (Exception e) {
|
||||
}
|
||||
};
|
||||
}
|
||||
};
|
||||
shutdownThread.start();
|
||||
LOG.info("Called master join on " + master.getName());
|
||||
|
|
|
@ -542,7 +542,7 @@ public class TestSplitLogManager {
|
|||
}
|
||||
}
|
||||
}
|
||||
};
|
||||
}
|
||||
}.start();
|
||||
|
||||
slm.splitLogDistributed(logDirPath);
|
||||
|
|
|
@ -433,7 +433,7 @@ public class TestCleanerChore {
|
|||
protected boolean validate(Path file) {
|
||||
return true;
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
public static class AlwaysDelete extends BaseHFileCleanerDelegate {
|
||||
@Override
|
||||
|
|
|
@ -116,7 +116,7 @@ public class TestAssignProcedure {
|
|||
// Happens after the code we are testing has been called.
|
||||
this.remoteCallFailedWasCalled.set(true);
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Test that we deal with ServerCrashProcedure zero'ing out the targetServer in the
|
||||
|
|
|
@ -208,7 +208,7 @@ public class TestSnapshotFileCache {
|
|||
files.addAll(SnapshotReferenceUtil.getHFileNames(UTIL.getConfiguration(), fs, snapshotDir));
|
||||
return files;
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
private SnapshotMock.SnapshotBuilder createAndTestSnapshotV1(final SnapshotFileCache cache,
|
||||
final String name, final boolean tmp, final boolean removeOnExit) throws IOException {
|
||||
|
|
|
@ -77,7 +77,7 @@ public class TestProcedure {
|
|||
public void sendGlobalBarrierComplete() {
|
||||
completedProcedure.countDown();
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* With a single member, verify ordered execution. The Coordinator side is run in a separate
|
||||
|
|
|
@ -186,7 +186,7 @@ public class RegionAsTable implements Table {
|
|||
public ScanMetrics getScanMetrics() {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
@Override
|
||||
public ResultScanner getScanner(Scan scan) throws IOException {
|
||||
|
|
|
@ -120,7 +120,7 @@ public class TestBulkLoad {
|
|||
walKey.setWriteEntry(we);
|
||||
}
|
||||
return 01L;
|
||||
};
|
||||
}
|
||||
});
|
||||
testRegionWithFamiliesAndSpecifiedTableName(tableName, family1)
|
||||
.bulkLoadHFiles(familyPaths, false, null);
|
||||
|
@ -145,7 +145,7 @@ public class TestBulkLoad {
|
|||
walKey.setWriteEntry(we);
|
||||
}
|
||||
return 01L;
|
||||
};
|
||||
}
|
||||
});
|
||||
testRegionWithFamilies(family1).bulkLoadHFiles(withFamilyPathsFor(family1), false, null);
|
||||
verify(log).sync(anyLong());
|
||||
|
@ -164,7 +164,7 @@ public class TestBulkLoad {
|
|||
walKey.setWriteEntry(we);
|
||||
}
|
||||
return 01L;
|
||||
};
|
||||
}
|
||||
});
|
||||
testRegionWithFamilies(family1, family2).bulkLoadHFiles(withFamilyPathsFor(family1, family2),
|
||||
false, null);
|
||||
|
@ -184,7 +184,7 @@ public class TestBulkLoad {
|
|||
walKey.setWriteEntry(we);
|
||||
}
|
||||
return 01L;
|
||||
};
|
||||
}
|
||||
});
|
||||
TableName tableName = TableName.valueOf("test", "test");
|
||||
testRegionWithFamiliesAndSpecifiedTableName(tableName, family1, family2)
|
||||
|
|
|
@ -352,7 +352,7 @@ public class TestHRegionFileSystem {
|
|||
|
||||
// Create a Region
|
||||
String familyName = "cf";
|
||||
;
|
||||
|
||||
RegionInfo hri = RegionInfoBuilder.newBuilder(TableName.valueOf(name.getMethodName())).build();
|
||||
HRegionFileSystem regionFs = HRegionFileSystem.createRegionOnFileSystem(conf, fs, rootDir, hri);
|
||||
|
||||
|
|
|
@ -414,7 +414,7 @@ public class TestHRegionReplayEvents {
|
|||
}
|
||||
}
|
||||
} else {
|
||||
lastReplayed = replayEdit(secondaryRegion, entry);;
|
||||
lastReplayed = replayEdit(secondaryRegion, entry);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -304,7 +304,7 @@ public class TestMobStoreScanner {
|
|||
long ts1 = System.currentTimeMillis();
|
||||
long ts2 = ts1 + 1;
|
||||
long ts3 = ts1 + 2;
|
||||
byte [] value = generateMobValue((int)defaultThreshold+1);;
|
||||
byte [] value = generateMobValue((int)defaultThreshold+1);
|
||||
|
||||
Put put1 = new Put(row1);
|
||||
put1.addColumn(family, qf1, ts3, value);
|
||||
|
@ -387,7 +387,7 @@ public class TestMobStoreScanner {
|
|||
long ts1 = System.currentTimeMillis();
|
||||
long ts2 = ts1 + 1;
|
||||
long ts3 = ts1 + 2;
|
||||
byte [] value = generateMobValue((int)defaultThreshold+1);;
|
||||
byte [] value = generateMobValue((int)defaultThreshold+1);
|
||||
// Put some data
|
||||
Put put1 = new Put(row1);
|
||||
put1.addColumn(family, qf1, ts3, value);
|
||||
|
|
|
@ -75,7 +75,7 @@ public class TestMultiLogThreshold {
|
|||
}
|
||||
|
||||
private enum ActionType {
|
||||
REGION_ACTIONS, ACTIONS;
|
||||
REGION_ACTIONS, ACTIONS
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -293,7 +293,7 @@ public class TestRegionReplicaFailover {
|
|||
} catch (Throwable e) {
|
||||
ex.compareAndSet(null, e);
|
||||
}
|
||||
};
|
||||
}
|
||||
};
|
||||
|
||||
aborter.start();
|
||||
|
|
|
@ -203,7 +203,7 @@ public class TestStoreScanner {
|
|||
PrivateCellUtil.createFirstOnRow(CELL_GRID[CELL_GRID_BLOCK3_BOUNDARY]):
|
||||
PrivateCellUtil.createFirstOnRow(CELL_GRID[CELL_GRID_BLOCK2_BOUNDARY]);
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
private static final int CELL_WITH_VERSIONS_BLOCK2_BOUNDARY = 4;
|
||||
|
||||
|
@ -243,7 +243,7 @@ public class TestStoreScanner {
|
|||
return PrivateCellUtil
|
||||
.createFirstOnRow(CELL_WITH_VERSIONS[CELL_WITH_VERSIONS_BLOCK2_BOUNDARY]);
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
private static class CellWithVersionsNoOptimizeStoreScanner extends StoreScanner {
|
||||
// Count of how often optimize is called and of how often it does an optimize.
|
||||
|
@ -269,7 +269,7 @@ public class TestStoreScanner {
|
|||
public Cell getNextIndexedKey() {
|
||||
return null;
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testWithColumnCountGetFilter() throws Exception {
|
||||
|
|
|
@ -270,7 +270,7 @@ public class TestWALLockup {
|
|||
LOG.info("In flush", e);
|
||||
}
|
||||
LOG.info("Exiting");
|
||||
};
|
||||
}
|
||||
};
|
||||
t.setDaemon(true);
|
||||
t.start();
|
||||
|
@ -452,7 +452,7 @@ public class TestWALLockup {
|
|||
}
|
||||
latch.countDown();
|
||||
LOG.info("Sync exiting");
|
||||
};
|
||||
}
|
||||
};
|
||||
t.setDaemon(true);
|
||||
t.start();
|
||||
|
|
|
@ -657,7 +657,7 @@ public abstract class AbstractTestWALReplay {
|
|||
}
|
||||
return super.flushSnapshot(snapshot, cacheFlushId, status, throughputController, tracker);
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Test that we could recover the data correctly after aborting flush. In the
|
||||
|
|
|
@ -909,7 +909,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
Increment inc = new Increment(TEST_ROW);
|
||||
inc.addColumn(TEST_FAMILY, TEST_QUALIFIER, 1);
|
||||
try(Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table t = conn.getTable(TEST_TABLE);) {
|
||||
Table t = conn.getTable(TEST_TABLE)) {
|
||||
t.increment(inc);
|
||||
}
|
||||
return null;
|
||||
|
@ -927,7 +927,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
Delete d = new Delete(TEST_ROW);
|
||||
d.addFamily(TEST_FAMILY);
|
||||
try(Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table t = conn.getTable(TEST_TABLE);) {
|
||||
Table t = conn.getTable(TEST_TABLE)) {
|
||||
t.checkAndMutate(TEST_ROW, TEST_FAMILY).qualifier(TEST_QUALIFIER)
|
||||
.ifEquals(Bytes.toBytes("test_value")).thenDelete(d);
|
||||
}
|
||||
|
@ -1147,7 +1147,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
@Override
|
||||
public Object run() throws Exception {
|
||||
try(Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table acl = conn.getTable(AccessControlLists.ACL_TABLE_NAME);) {
|
||||
Table acl = conn.getTable(AccessControlLists.ACL_TABLE_NAME)) {
|
||||
BlockingRpcChannel service = acl.coprocessorService(HConstants.EMPTY_START_ROW);
|
||||
AccessControlService.BlockingInterface protocol =
|
||||
AccessControlService.newBlockingStub(service);
|
||||
|
@ -1212,7 +1212,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
p.addColumn(family2, qualifier, Bytes.toBytes("v2"));
|
||||
|
||||
try (Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table t = conn.getTable(tableName);) {
|
||||
Table t = conn.getTable(tableName)) {
|
||||
t.put(p);
|
||||
}
|
||||
return null;
|
||||
|
@ -1239,7 +1239,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
Put p = new Put(Bytes.toBytes("a"));
|
||||
p.addColumn(family2, qualifier, Bytes.toBytes("v2"));
|
||||
try (Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table t = conn.getTable(tableName);) {
|
||||
Table t = conn.getTable(tableName)) {
|
||||
t.put(p);
|
||||
}
|
||||
return null;
|
||||
|
@ -1253,7 +1253,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
g.addFamily(family1);
|
||||
g.addFamily(family2);
|
||||
try (Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table t = conn.getTable(tableName);) {
|
||||
Table t = conn.getTable(tableName)) {
|
||||
t.get(g);
|
||||
}
|
||||
return null;
|
||||
|
@ -2127,7 +2127,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
@Override
|
||||
public Object run() throws Exception {
|
||||
try (Connection conn = ConnectionFactory.createConnection(TEST_UTIL.getConfiguration());
|
||||
Admin admin = conn.getAdmin();) {
|
||||
Admin admin = conn.getAdmin()) {
|
||||
return admin.getTableDescriptor(TEST_TABLE);
|
||||
}
|
||||
}
|
||||
|
@ -2216,7 +2216,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
@Override
|
||||
public Object run() throws Exception {
|
||||
try(Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table t = conn.getTable(TEST_TABLE);) {
|
||||
Table t = conn.getTable(TEST_TABLE)) {
|
||||
return t.get(new Get(TEST_ROW));
|
||||
}
|
||||
}
|
||||
|
@ -2242,7 +2242,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
@Override
|
||||
public Object run() throws Exception {
|
||||
try(Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table t = conn.getTable(TEST_TABLE);) {
|
||||
Table t = conn.getTable(TEST_TABLE)) {
|
||||
return t.get(new Get(TEST_ROW));
|
||||
}
|
||||
}
|
||||
|
@ -2450,7 +2450,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
@Override
|
||||
public Object run() throws Exception {
|
||||
try(Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table t = conn.getTable(TEST_TABLE);) {
|
||||
Table t = conn.getTable(TEST_TABLE)) {
|
||||
return t.get(new Get(TEST_ROW));
|
||||
}
|
||||
}
|
||||
|
@ -2558,7 +2558,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
@Override
|
||||
public Object run() throws Exception {
|
||||
try (Connection conn = ConnectionFactory.createConnection(conf);
|
||||
Table t = conn.getTable(TEST_TABLE);) {
|
||||
Table t = conn.getTable(TEST_TABLE)) {
|
||||
BlockingRpcChannel service = t.coprocessorService(HConstants.EMPTY_BYTE_ARRAY);
|
||||
PingCoprocessor.newBlockingStub(service).noop(null, NoopRequest.newBuilder().build());
|
||||
}
|
||||
|
@ -2716,7 +2716,7 @@ public class TestAccessController extends SecureTestUtil {
|
|||
return new PrivilegedAction<List<UserPermission>>() {
|
||||
@Override
|
||||
public List<UserPermission> run() {
|
||||
try(Connection conn = ConnectionFactory.createConnection(conf);) {
|
||||
try(Connection conn = ConnectionFactory.createConnection(conf)) {
|
||||
return AccessControlClient.getUserPermissions(conn, regex);
|
||||
} catch (Throwable e) {
|
||||
LOG.error("error during call of AccessControlClient.getUserPermissions.", e);
|
||||
|
|
|
@ -348,9 +348,9 @@ public class TestAccessController2 extends SecureTestUtil {
|
|||
@Override
|
||||
public Void run() throws Exception {
|
||||
try (Connection connection = ConnectionFactory.createConnection(conf);
|
||||
Table table = connection.getTable(tableName);) {
|
||||
Table table = connection.getTable(tableName)) {
|
||||
Scan s1 = new Scan();
|
||||
try (ResultScanner scanner1 = table.getScanner(s1);) {
|
||||
try (ResultScanner scanner1 = table.getScanner(s1)) {
|
||||
Result[] next1 = scanner1.next(5);
|
||||
assertTrue("User having table level access should be able to scan all "
|
||||
+ "the data in the table.", next1.length == 3);
|
||||
|
@ -364,9 +364,9 @@ public class TestAccessController2 extends SecureTestUtil {
|
|||
@Override
|
||||
public Void run() throws Exception {
|
||||
try (Connection connection = ConnectionFactory.createConnection(conf);
|
||||
Table table = connection.getTable(tableName);) {
|
||||
Table table = connection.getTable(tableName)) {
|
||||
Scan s1 = new Scan();
|
||||
try (ResultScanner scanner1 = table.getScanner(s1);) {
|
||||
try (ResultScanner scanner1 = table.getScanner(s1)) {
|
||||
Result[] next1 = scanner1.next(5);
|
||||
assertTrue("User having column family level access should be able to scan all "
|
||||
+ "the data belonging to that family.", next1.length == 2);
|
||||
|
@ -380,10 +380,10 @@ public class TestAccessController2 extends SecureTestUtil {
|
|||
@Override
|
||||
public Void run() throws Exception {
|
||||
try (Connection connection = ConnectionFactory.createConnection(conf);
|
||||
Table table = connection.getTable(tableName);) {
|
||||
Table table = connection.getTable(tableName)) {
|
||||
Scan s1 = new Scan();
|
||||
s1.addFamily(TEST_FAMILY_2);
|
||||
try (ResultScanner scanner1 = table.getScanner(s1);) {
|
||||
try (ResultScanner scanner1 = table.getScanner(s1)) {
|
||||
scanner1.next();
|
||||
}
|
||||
}
|
||||
|
@ -395,9 +395,9 @@ public class TestAccessController2 extends SecureTestUtil {
|
|||
@Override
|
||||
public Void run() throws Exception {
|
||||
try (Connection connection = ConnectionFactory.createConnection(conf);
|
||||
Table table = connection.getTable(tableName);) {
|
||||
Table table = connection.getTable(tableName)) {
|
||||
Scan s1 = new Scan();
|
||||
try (ResultScanner scanner1 = table.getScanner(s1);) {
|
||||
try (ResultScanner scanner1 = table.getScanner(s1)) {
|
||||
Result[] next1 = scanner1.next(5);
|
||||
assertTrue("User having column qualifier level access should be able to scan "
|
||||
+ "that column family qualifier data.", next1.length == 1);
|
||||
|
@ -411,10 +411,10 @@ public class TestAccessController2 extends SecureTestUtil {
|
|||
@Override
|
||||
public Void run() throws Exception {
|
||||
try (Connection connection = ConnectionFactory.createConnection(conf);
|
||||
Table table = connection.getTable(tableName);) {
|
||||
Table table = connection.getTable(tableName)) {
|
||||
Scan s1 = new Scan();
|
||||
s1.addFamily(TEST_FAMILY_2);
|
||||
try (ResultScanner scanner1 = table.getScanner(s1);) {
|
||||
try (ResultScanner scanner1 = table.getScanner(s1)) {
|
||||
scanner1.next();
|
||||
}
|
||||
}
|
||||
|
@ -426,10 +426,10 @@ public class TestAccessController2 extends SecureTestUtil {
|
|||
@Override
|
||||
public Void run() throws Exception {
|
||||
try (Connection connection = ConnectionFactory.createConnection(conf);
|
||||
Table table = connection.getTable(tableName);) {
|
||||
Table table = connection.getTable(tableName)) {
|
||||
Scan s1 = new Scan();
|
||||
s1.addColumn(TEST_FAMILY, Q2);
|
||||
try (ResultScanner scanner1 = table.getScanner(s1);) {
|
||||
try (ResultScanner scanner1 = table.getScanner(s1)) {
|
||||
scanner1.next();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -375,7 +375,7 @@ public abstract class TestVisibilityLabels {
|
|||
RegionServerThread rs = TEST_UTIL.getHBaseCluster().startRegionServer();
|
||||
waitForLabelsRegionAvailability(rs.getRegionServer());
|
||||
try (Table table = createTableAndWriteDataWithLabels(tableName, "(" + SECRET + "|" + CONFIDENTIAL
|
||||
+ ")", PRIVATE);) {
|
||||
+ ")", PRIVATE)) {
|
||||
Scan s = new Scan();
|
||||
s.setAuthorizations(new Authorizations(SECRET));
|
||||
ResultScanner scanner = table.getScanner(s);
|
||||
|
@ -425,7 +425,7 @@ public abstract class TestVisibilityLabels {
|
|||
}
|
||||
};
|
||||
SUPERUSER.runAs(action);
|
||||
try (Table ht = TEST_UTIL.getConnection().getTable(LABELS_TABLE_NAME);) {
|
||||
try (Table ht = TEST_UTIL.getConnection().getTable(LABELS_TABLE_NAME)) {
|
||||
Scan scan = new Scan();
|
||||
scan.setAuthorizations(new Authorizations(VisibilityUtils.SYSTEM_LABEL));
|
||||
ResultScanner scanner = ht.getScanner(scan);
|
||||
|
@ -616,7 +616,7 @@ public abstract class TestVisibilityLabels {
|
|||
@Test
|
||||
public void testLabelsWithAppend() throws Throwable {
|
||||
TableName tableName = TableName.valueOf(TEST_NAME.getMethodName());
|
||||
try (Table table = TEST_UTIL.createTable(tableName, fam);) {
|
||||
try (Table table = TEST_UTIL.createTable(tableName, fam)) {
|
||||
byte[] row1 = Bytes.toBytes("row1");
|
||||
byte[] val = Bytes.toBytes("a");
|
||||
Put put = new Put(row1);
|
||||
|
|
|
@ -222,7 +222,7 @@ public class TestVisibilityLabelsReplication {
|
|||
try (Table table = writeData(TABLE_NAME, "(" + SECRET + "&" + PUBLIC + ")" + "|(" + CONFIDENTIAL
|
||||
+ ")&(" + TOPSECRET + ")", "(" + PRIVATE + "|" + CONFIDENTIAL + ")&(" + PUBLIC + "|"
|
||||
+ TOPSECRET + ")", "(" + SECRET + "|" + CONFIDENTIAL + ")" + "&" + "!" + TOPSECRET,
|
||||
CellVisibility.quote(UNICODE_VIS_TAG) + "&" + SECRET);) {
|
||||
CellVisibility.quote(UNICODE_VIS_TAG) + "&" + SECRET)) {
|
||||
Scan s = new Scan();
|
||||
s.setAuthorizations(new Authorizations(SECRET, CONFIDENTIAL, PRIVATE, TOPSECRET,
|
||||
UNICODE_VIS_TAG));
|
||||
|
@ -250,7 +250,7 @@ public class TestVisibilityLabelsReplication {
|
|||
current = cellScanner.current();
|
||||
assertTrue(Bytes.equals(current.getRowArray(), current.getRowOffset(),
|
||||
current.getRowLength(), row4, 0, row4.length));
|
||||
try (Table table2 = TEST_UTIL1.getConnection().getTable(TABLE_NAME);) {
|
||||
try (Table table2 = TEST_UTIL1.getConnection().getTable(TABLE_NAME)) {
|
||||
s = new Scan();
|
||||
// Ensure both rows are replicated
|
||||
scanner = table2.getScanner(s);
|
||||
|
|
|
@ -222,7 +222,7 @@ public class TestVisibilityLabelsWithDefaultVisLabelService extends TestVisibili
|
|||
public void testVisibilityLabelsOnWALReplay() throws Exception {
|
||||
final TableName tableName = TableName.valueOf(TEST_NAME.getMethodName());
|
||||
try (Table table = createTableAndWriteDataWithLabels(tableName,
|
||||
"(" + SECRET + "|" + CONFIDENTIAL + ")", PRIVATE);) {
|
||||
"(" + SECRET + "|" + CONFIDENTIAL + ")", PRIVATE)) {
|
||||
List<RegionServerThread> regionServerThreads = TEST_UTIL.getHBaseCluster()
|
||||
.getRegionServerThreads();
|
||||
for (RegionServerThread rsThread : regionServerThreads) {
|
||||
|
|
|
@ -169,7 +169,7 @@ public class TestVisibilityLabelsWithDeletes {
|
|||
public void testVisibilityLabelsWithDeleteFamily() throws Exception {
|
||||
setAuths();
|
||||
final TableName tableName = TableName.valueOf(TEST_NAME.getMethodName());
|
||||
try (Table table = createTableAndWriteDataWithLabels(SECRET, CONFIDENTIAL + "|" + TOPSECRET);) {
|
||||
try (Table table = createTableAndWriteDataWithLabels(SECRET, CONFIDENTIAL + "|" + TOPSECRET)) {
|
||||
PrivilegedExceptionAction<Void> actiona = new PrivilegedExceptionAction<Void>() {
|
||||
@Override
|
||||
public Void run() throws Exception {
|
||||
|
@ -245,7 +245,7 @@ public class TestVisibilityLabelsWithDeletes {
|
|||
final TableName tableName = TableName.valueOf(TEST_NAME.getMethodName());
|
||||
long[] ts = new long[] { 123l, 125l };
|
||||
try (Table table = createTableAndWriteDataWithLabels(ts,
|
||||
CONFIDENTIAL + "|" + TOPSECRET, SECRET);) {
|
||||
CONFIDENTIAL + "|" + TOPSECRET, SECRET)) {
|
||||
PrivilegedExceptionAction<Void> actiona = new PrivilegedExceptionAction<Void>() {
|
||||
@Override
|
||||
public Void run() throws Exception {
|
||||
|
@ -2092,7 +2092,7 @@ public class TestVisibilityLabelsWithDeletes {
|
|||
};
|
||||
VisibilityLabelsResponse response = SUPERUSER.runAs(action);
|
||||
final TableName tableName = TableName.valueOf(TEST_NAME.getMethodName());
|
||||
try (Table table = doPuts(tableName);) {
|
||||
try (Table table = doPuts(tableName)) {
|
||||
PrivilegedExceptionAction<Void> actiona = new PrivilegedExceptionAction<Void>() {
|
||||
@Override
|
||||
public Void run() throws Exception {
|
||||
|
|
|
@ -444,7 +444,7 @@ public class TestFlushSnapshotFromClient {
|
|||
.toString(ProtobufUtil.createHBaseProtosSnapshotDesc(ss)));
|
||||
toBeSubmitted.countDown();
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// build descriptions
|
||||
SnapshotDescription[] descs = new SnapshotDescription[ssNum];
|
||||
|
|
|
@ -126,7 +126,7 @@ abstract public class MapreduceTestingShim {
|
|||
public String obtainMROutputDirProp() {
|
||||
return "mapred.output.dir";
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
private static class MapreduceV2Shim extends MapreduceTestingShim {
|
||||
public JobContext newJobContext(Configuration jobConf) {
|
||||
|
@ -166,6 +166,5 @@ abstract public class MapreduceTestingShim {
|
|||
// from Hadoop 0.23.x. If we use the source directly we break the hadoop 1.x compile.
|
||||
return "mapreduce.output.fileoutputformat.outputdir";
|
||||
}
|
||||
};
|
||||
|
||||
}
|
||||
}
|
||||
|
|
|
@ -73,7 +73,7 @@ public class ProcessBasedLocalHBaseCluster {
|
|||
ProcessBasedLocalHBaseCluster.class);
|
||||
|
||||
private List<String> daemonPidFiles =
|
||||
Collections.synchronizedList(new ArrayList<String>());;
|
||||
Collections.synchronizedList(new ArrayList<String>());
|
||||
|
||||
private boolean shutdownHookInstalled;
|
||||
|
||||
|
|
|
@ -47,7 +47,7 @@ public class TestFSHDFSUtils {
|
|||
Configuration conf = HTU.getConfiguration();
|
||||
conf.setInt("hbase.lease.recovery.first.pause", 10);
|
||||
conf.setInt("hbase.lease.recovery.pause", 10);
|
||||
};
|
||||
}
|
||||
private FSHDFSUtils fsHDFSUtils = new FSHDFSUtils();
|
||||
private static Path FILE = new Path(HTU.getDataTestDir(), "file.txt");
|
||||
long startTime = -1;
|
||||
|
|
|
@ -78,7 +78,7 @@ public class IOTestProvider implements WALProvider {
|
|||
append,
|
||||
sync,
|
||||
fileroll,
|
||||
none;
|
||||
none
|
||||
}
|
||||
|
||||
private FSHLog log = null;
|
||||
|
|
|
@ -52,7 +52,7 @@ import org.junit.experimental.categories.Category;
|
|||
*/
|
||||
@Category({RegionServerTests.class, SmallTests.class})
|
||||
public class TestWALMethods {
|
||||
private static final byte[] TEST_REGION = Bytes.toBytes("test_region");;
|
||||
private static final byte[] TEST_REGION = Bytes.toBytes("test_region");
|
||||
private static final TableName TEST_TABLE =
|
||||
TableName.valueOf("test_table");
|
||||
|
||||
|
|
Loading…
Reference in New Issue