HBASE-26246 Persist the StoreFileTracker configurations to TableDescriptor when creating table (#3666)
Signed-off-by: Duo Zhang <zhangduo@apache.org>
This commit is contained in:
parent
f298f9cacd
commit
26f263b792
|
@ -581,6 +581,10 @@ public class TableDescriptorBuilder {
|
||||||
return this;
|
return this;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public String getValue(String key) {
|
||||||
|
return desc.getValue(key);
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Sets replication scope all & only the columns already in the builder. Columns added later won't
|
* Sets replication scope all & only the columns already in the builder. Columns added later won't
|
||||||
* be backfilled with replication scope.
|
* be backfilled with replication scope.
|
||||||
|
|
|
@ -18,6 +18,7 @@
|
||||||
|
|
||||||
package org.apache.hadoop.hbase.master.procedure;
|
package org.apache.hadoop.hbase.master.procedure;
|
||||||
|
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
@ -31,10 +32,12 @@ import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.client.RegionInfo;
|
import org.apache.hadoop.hbase.client.RegionInfo;
|
||||||
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
import org.apache.hadoop.hbase.client.RegionReplicaUtil;
|
||||||
import org.apache.hadoop.hbase.client.TableDescriptor;
|
import org.apache.hadoop.hbase.client.TableDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
|
||||||
import org.apache.hadoop.hbase.client.TableState;
|
import org.apache.hadoop.hbase.client.TableState;
|
||||||
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
import org.apache.hadoop.hbase.master.MasterCoprocessorHost;
|
||||||
import org.apache.hadoop.hbase.master.MasterFileSystem;
|
import org.apache.hadoop.hbase.master.MasterFileSystem;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer;
|
import org.apache.hadoop.hbase.procedure2.ProcedureStateSerializer;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationException;
|
import org.apache.hadoop.hbase.replication.ReplicationException;
|
||||||
import org.apache.hadoop.hbase.util.CommonFSUtils;
|
import org.apache.hadoop.hbase.util.CommonFSUtils;
|
||||||
import org.apache.hadoop.hbase.util.FSTableDescriptors;
|
import org.apache.hadoop.hbase.util.FSTableDescriptors;
|
||||||
|
@ -265,6 +268,10 @@ public class CreateTableProcedure
|
||||||
getTableName(), (newRegions != null ? newRegions.size() : 0));
|
getTableName(), (newRegions != null ? newRegions.size() : 0));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
TableDescriptorBuilder builder = TableDescriptorBuilder.newBuilder(tableDescriptor);
|
||||||
|
StoreFileTrackerFactory.persistTrackerConfig(env.getMasterConfiguration(), builder);
|
||||||
|
tableDescriptor = builder.build();
|
||||||
|
|
||||||
final MasterCoprocessorHost cpHost = env.getMasterCoprocessorHost();
|
final MasterCoprocessorHost cpHost = env.getMasterCoprocessorHost();
|
||||||
if (cpHost != null) {
|
if (cpHost != null) {
|
||||||
final RegionInfo[] regions = newRegions == null ? null :
|
final RegionInfo[] regions = newRegions == null ? null :
|
||||||
|
|
|
@ -610,7 +610,7 @@ public class HRegionFileSystem {
|
||||||
writeRegionInfoFileContent(conf, fs, regionInfoFile, regionInfoContent);
|
writeRegionInfoFileContent(conf, fs, regionInfoFile, regionInfoContent);
|
||||||
HRegionFileSystem regionFs = HRegionFileSystem.openRegionFromFileSystem(
|
HRegionFileSystem regionFs = HRegionFileSystem.openRegionFromFileSystem(
|
||||||
env.getMasterConfiguration(), fs, getTableDir(), regionInfo, false);
|
env.getMasterConfiguration(), fs, getTableDir(), regionInfo, false);
|
||||||
insertRegionFilesIntoStoreTracker(allRegionFiles, env, regionFs);
|
insertRegionFilesIntoStoreTracker(allRegionFiles, env, regionFs);
|
||||||
}
|
}
|
||||||
return regionDir;
|
return regionDir;
|
||||||
}
|
}
|
||||||
|
|
|
@ -21,6 +21,7 @@ import java.io.IOException;
|
||||||
import java.util.Collection;
|
import java.util.Collection;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreContext;
|
import org.apache.hadoop.hbase.regionserver.StoreContext;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
|
@ -85,4 +86,11 @@ class MigrationStoreFileTracker extends StoreFileTrackerBase {
|
||||||
throw new UnsupportedOperationException(
|
throw new UnsupportedOperationException(
|
||||||
"Should not call this method on " + getClass().getSimpleName());
|
"Should not call this method on " + getClass().getSimpleName());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void persistConfiguration(TableDescriptorBuilder builder) {
|
||||||
|
super.persistConfiguration(builder);
|
||||||
|
builder.setValue(SRC_IMPL, src.getClass().getName());
|
||||||
|
builder.setValue(DST_IMPL, dst.getClass().getName());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.regionserver.storefiletracker;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.Collection;
|
import java.util.Collection;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
|
||||||
|
import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
|
||||||
import org.apache.hadoop.hbase.regionserver.CreateStoreFileWriterParams;
|
import org.apache.hadoop.hbase.regionserver.CreateStoreFileWriterParams;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
|
import org.apache.hadoop.hbase.regionserver.StoreFileWriter;
|
||||||
|
@ -71,4 +73,10 @@ public interface StoreFileTracker {
|
||||||
* @return Writer for a new StoreFile
|
* @return Writer for a new StoreFile
|
||||||
*/
|
*/
|
||||||
StoreFileWriter createWriter(CreateStoreFileWriterParams params) throws IOException;
|
StoreFileWriter createWriter(CreateStoreFileWriterParams params) throws IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Saves StoreFileTracker implementations specific configs into the table descriptors.
|
||||||
|
* @param builder The table descriptor builder for the given table.
|
||||||
|
*/
|
||||||
|
void persistConfiguration(TableDescriptorBuilder builder);
|
||||||
}
|
}
|
||||||
|
|
|
@ -17,18 +17,22 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.regionserver.storefiletracker;
|
package org.apache.hadoop.hbase.regionserver.storefiletracker;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory.TRACK_IMPL;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.Collection;
|
import java.util.Collection;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
|
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
|
||||||
import org.apache.hadoop.hbase.io.compress.Compression;
|
import org.apache.hadoop.hbase.io.compress.Compression;
|
||||||
import org.apache.hadoop.hbase.io.crypto.Encryption;
|
import org.apache.hadoop.hbase.io.crypto.Encryption;
|
||||||
import org.apache.hadoop.hbase.io.hfile.CacheConfig;
|
import org.apache.hadoop.hbase.io.hfile.CacheConfig;
|
||||||
import org.apache.hadoop.hbase.io.hfile.HFile;
|
import org.apache.hadoop.hbase.io.hfile.HFile;
|
||||||
import org.apache.hadoop.hbase.io.hfile.HFileContext;
|
import org.apache.hadoop.hbase.io.hfile.HFileContext;
|
||||||
import org.apache.hadoop.hbase.io.hfile.HFileContextBuilder;
|
import org.apache.hadoop.hbase.io.hfile.HFileContextBuilder;
|
||||||
|
import org.apache.hadoop.hbase.procedure2.util.StringUtils;
|
||||||
import org.apache.hadoop.hbase.regionserver.CreateStoreFileWriterParams;
|
import org.apache.hadoop.hbase.regionserver.CreateStoreFileWriterParams;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreContext;
|
import org.apache.hadoop.hbase.regionserver.StoreContext;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
||||||
|
@ -78,6 +82,15 @@ abstract class StoreFileTrackerBase implements StoreFileTracker {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void persistConfiguration(TableDescriptorBuilder builder) {
|
||||||
|
if (StringUtils.isEmpty(builder.getValue(TRACK_IMPL))) {
|
||||||
|
String trackerImpl = StoreFileTrackerFactory.
|
||||||
|
getStoreFileTrackerImpl(conf).getName();
|
||||||
|
builder.setValue(TRACK_IMPL, trackerImpl).build();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
private HFileContext createFileContext(Compression.Algorithm compression,
|
private HFileContext createFileContext(Compression.Algorithm compression,
|
||||||
boolean includeMVCCReadpoint, boolean includesTag, Encryption.Context encryptionContext) {
|
boolean includeMVCCReadpoint, boolean includesTag, Encryption.Context encryptionContext) {
|
||||||
if (compression == null) {
|
if (compression == null) {
|
||||||
|
|
|
@ -6,9 +6,7 @@
|
||||||
* to you under the Apache License, Version 2.0 (the
|
* to you under the Apache License, Version 2.0 (the
|
||||||
* "License"); you may not use this file except in compliance
|
* "License"); you may not use this file except in compliance
|
||||||
* with the License. You may obtain a copy of the License at
|
* with the License. You may obtain a copy of the License at
|
||||||
*
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
@ -21,6 +19,7 @@ import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
|
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor;
|
||||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
|
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
|
||||||
import org.apache.hadoop.hbase.client.TableDescriptor;
|
import org.apache.hadoop.hbase.client.TableDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.client.TableDescriptorBuilder;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
|
import org.apache.hadoop.hbase.regionserver.HRegionFileSystem;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreContext;
|
import org.apache.hadoop.hbase.regionserver.StoreContext;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreUtils;
|
import org.apache.hadoop.hbase.regionserver.StoreUtils;
|
||||||
|
@ -35,15 +34,17 @@ import org.apache.hbase.thirdparty.com.google.common.base.Preconditions;
|
||||||
/**
|
/**
|
||||||
* Factory method for creating store file tracker.
|
* Factory method for creating store file tracker.
|
||||||
*/
|
*/
|
||||||
@InterfaceAudience.Private
|
@InterfaceAudience.Private public final class StoreFileTrackerFactory {
|
||||||
public final class StoreFileTrackerFactory {
|
|
||||||
public static final String TRACK_IMPL = "hbase.store.file-tracker.impl";
|
public static final String TRACK_IMPL = "hbase.store.file-tracker.impl";
|
||||||
private static final Logger LOG = LoggerFactory.getLogger(StoreFileTrackerFactory.class);
|
private static final Logger LOG = LoggerFactory.getLogger(StoreFileTrackerFactory.class);
|
||||||
|
|
||||||
|
public static Class<? extends StoreFileTracker> getStoreFileTrackerImpl(Configuration conf) {
|
||||||
|
return conf.getClass(TRACK_IMPL, DefaultStoreFileTracker.class, StoreFileTracker.class);
|
||||||
|
}
|
||||||
|
|
||||||
public static StoreFileTracker create(Configuration conf, boolean isPrimaryReplica,
|
public static StoreFileTracker create(Configuration conf, boolean isPrimaryReplica,
|
||||||
StoreContext ctx) {
|
StoreContext ctx) {
|
||||||
Class<? extends StoreFileTracker> tracker =
|
Class<? extends StoreFileTracker> tracker = getStoreFileTrackerImpl(conf);
|
||||||
conf.getClass(TRACK_IMPL, DefaultStoreFileTracker.class, StoreFileTracker.class);
|
|
||||||
LOG.info("instantiating StoreFileTracker impl {}", tracker.getName());
|
LOG.info("instantiating StoreFileTracker impl {}", tracker.getName());
|
||||||
return ReflectionUtils.newInstance(tracker, conf, isPrimaryReplica, ctx);
|
return ReflectionUtils.newInstance(tracker, conf, isPrimaryReplica, ctx);
|
||||||
}
|
}
|
||||||
|
@ -54,7 +55,7 @@ public final class StoreFileTrackerFactory {
|
||||||
ColumnFamilyDescriptorBuilder.newBuilder(Bytes.toBytes(family));
|
ColumnFamilyDescriptorBuilder.newBuilder(Bytes.toBytes(family));
|
||||||
StoreContext ctx = StoreContext.getBuilder().withColumnFamilyDescriptor(fDescBuilder.build())
|
StoreContext ctx = StoreContext.getBuilder().withColumnFamilyDescriptor(fDescBuilder.build())
|
||||||
.withRegionFileSystem(regionFs).build();
|
.withRegionFileSystem(regionFs).build();
|
||||||
return StoreFileTrackerFactory.create(conf, TRACK_IMPL, isPrimaryReplica, ctx);
|
return StoreFileTrackerFactory.create(conf, isPrimaryReplica, ctx);
|
||||||
}
|
}
|
||||||
|
|
||||||
public static Configuration mergeConfigurations(Configuration global, TableDescriptor table,
|
public static Configuration mergeConfigurations(Configuration global, TableDescriptor table,
|
||||||
|
@ -75,4 +76,12 @@ public final class StoreFileTrackerFactory {
|
||||||
LOG.info("instantiating StoreFileTracker impl {} as {}", tracker.getName(), configName);
|
LOG.info("instantiating StoreFileTracker impl {} as {}", tracker.getName(), configName);
|
||||||
return ReflectionUtils.newInstance(tracker, conf, isPrimaryReplica, ctx);
|
return ReflectionUtils.newInstance(tracker, conf, isPrimaryReplica, ctx);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static void persistTrackerConfig(Configuration conf, TableDescriptorBuilder builder) {
|
||||||
|
TableDescriptor tableDescriptor = builder.build();
|
||||||
|
ColumnFamilyDescriptor cfDesc = tableDescriptor.getColumnFamilies()[0];
|
||||||
|
StoreContext context = StoreContext.getBuilder().withColumnFamilyDescriptor(cfDesc).build();
|
||||||
|
StoreFileTracker tracker = StoreFileTrackerFactory.create(conf, true, context);
|
||||||
|
tracker.persistConfiguration(builder);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -17,6 +17,7 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.client;
|
package org.apache.hadoop.hbase.client;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory.TRACK_IMPL;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertFalse;
|
import static org.junit.Assert.assertFalse;
|
||||||
import static org.junit.Assert.assertNotEquals;
|
import static org.junit.Assert.assertNotEquals;
|
||||||
|
@ -39,6 +40,7 @@ import org.apache.hadoop.hbase.TableExistsException;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.TableNotFoundException;
|
import org.apache.hadoop.hbase.TableNotFoundException;
|
||||||
import org.apache.hadoop.hbase.master.LoadBalancer;
|
import org.apache.hadoop.hbase.master.LoadBalancer;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory;
|
||||||
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
||||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
@ -420,6 +422,10 @@ public class TestAdmin extends TestAdminBase {
|
||||||
assertEquals(BLOCK_SIZE, newTableDesc.getColumnFamily(FAMILY_1).getBlocksize());
|
assertEquals(BLOCK_SIZE, newTableDesc.getColumnFamily(FAMILY_1).getBlocksize());
|
||||||
assertEquals(BLOCK_CACHE, newTableDesc.getColumnFamily(FAMILY_1).isBlockCacheEnabled());
|
assertEquals(BLOCK_CACHE, newTableDesc.getColumnFamily(FAMILY_1).isBlockCacheEnabled());
|
||||||
assertEquals(TTL, newTableDesc.getColumnFamily(FAMILY_1).getTimeToLive());
|
assertEquals(TTL, newTableDesc.getColumnFamily(FAMILY_1).getTimeToLive());
|
||||||
|
// HBASE-26246 introduced persist of store file tracker into table descriptor
|
||||||
|
tableDesc = TableDescriptorBuilder.newBuilder(tableDesc).setValue(TRACK_IMPL,
|
||||||
|
StoreFileTrackerFactory.getStoreFileTrackerImpl(TEST_UTIL.getConfiguration()).getName()).
|
||||||
|
build();
|
||||||
TEST_UTIL.verifyTableDescriptorIgnoreTableName(tableDesc, newTableDesc);
|
TEST_UTIL.verifyTableDescriptorIgnoreTableName(tableDesc, newTableDesc);
|
||||||
|
|
||||||
if (preserveSplits) {
|
if (preserveSplits) {
|
||||||
|
|
|
@ -17,6 +17,7 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.client;
|
package org.apache.hadoop.hbase.client;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory.TRACK_IMPL;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertFalse;
|
import static org.junit.Assert.assertFalse;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
|
@ -34,6 +35,7 @@ import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.TableNotDisabledException;
|
import org.apache.hadoop.hbase.TableNotDisabledException;
|
||||||
import org.apache.hadoop.hbase.TableNotEnabledException;
|
import org.apache.hadoop.hbase.TableNotEnabledException;
|
||||||
import org.apache.hadoop.hbase.TableNotFoundException;
|
import org.apache.hadoop.hbase.TableNotFoundException;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory;
|
||||||
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
||||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
@ -236,6 +238,10 @@ public class TestAdmin3 extends TestAdminBase {
|
||||||
ADMIN.createTable(htd);
|
ADMIN.createTable(htd);
|
||||||
Table table = TEST_UTIL.getConnection().getTable(htd.getTableName());
|
Table table = TEST_UTIL.getConnection().getTable(htd.getTableName());
|
||||||
TableDescriptor confirmedHtd = table.getDescriptor();
|
TableDescriptor confirmedHtd = table.getDescriptor();
|
||||||
|
//HBASE-26246 introduced persist of store file tracker into table descriptor
|
||||||
|
htd = TableDescriptorBuilder.newBuilder(htd).setValue(TRACK_IMPL,
|
||||||
|
StoreFileTrackerFactory.getStoreFileTrackerImpl(TEST_UTIL.getConfiguration()).getName()).
|
||||||
|
build();
|
||||||
assertEquals(0, TableDescriptor.COMPARATOR.compare(htd, confirmedHtd));
|
assertEquals(0, TableDescriptor.COMPARATOR.compare(htd, confirmedHtd));
|
||||||
MetaTableAccessor.fullScanMetaAndPrint(TEST_UTIL.getConnection());
|
MetaTableAccessor.fullScanMetaAndPrint(TEST_UTIL.getConnection());
|
||||||
table.close();
|
table.close();
|
||||||
|
|
|
@ -18,10 +18,12 @@
|
||||||
package org.apache.hadoop.hbase.client;
|
package org.apache.hadoop.hbase.client;
|
||||||
|
|
||||||
import static org.apache.hadoop.hbase.TableName.META_TABLE_NAME;
|
import static org.apache.hadoop.hbase.TableName.META_TABLE_NAME;
|
||||||
|
import static org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory.TRACK_IMPL;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertFalse;
|
import static org.junit.Assert.assertFalse;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
import static org.junit.Assert.fail;
|
import static org.junit.Assert.fail;
|
||||||
|
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
import java.util.Iterator;
|
import java.util.Iterator;
|
||||||
|
@ -38,6 +40,7 @@ import org.apache.hadoop.hbase.TableExistsException;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.TableNotFoundException;
|
import org.apache.hadoop.hbase.TableNotFoundException;
|
||||||
import org.apache.hadoop.hbase.master.LoadBalancer;
|
import org.apache.hadoop.hbase.master.LoadBalancer;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory;
|
||||||
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
||||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
@ -409,6 +412,10 @@ public class TestAsyncTableAdminApi extends TestAsyncAdminBase {
|
||||||
assertEquals(BLOCK_SIZE, newTableDesc.getColumnFamily(FAMILY_1).getBlocksize());
|
assertEquals(BLOCK_SIZE, newTableDesc.getColumnFamily(FAMILY_1).getBlocksize());
|
||||||
assertEquals(BLOCK_CACHE, newTableDesc.getColumnFamily(FAMILY_1).isBlockCacheEnabled());
|
assertEquals(BLOCK_CACHE, newTableDesc.getColumnFamily(FAMILY_1).isBlockCacheEnabled());
|
||||||
assertEquals(TTL, newTableDesc.getColumnFamily(FAMILY_1).getTimeToLive());
|
assertEquals(TTL, newTableDesc.getColumnFamily(FAMILY_1).getTimeToLive());
|
||||||
|
//HBASE-26246 introduced persist of store file tracker into table descriptor
|
||||||
|
tableDesc = TableDescriptorBuilder.newBuilder(tableDesc).setValue(TRACK_IMPL,
|
||||||
|
StoreFileTrackerFactory.getStoreFileTrackerImpl(TEST_UTIL.getConfiguration()).getName()).
|
||||||
|
build();
|
||||||
TEST_UTIL.verifyTableDescriptorIgnoreTableName(tableDesc, newTableDesc);
|
TEST_UTIL.verifyTableDescriptorIgnoreTableName(tableDesc, newTableDesc);
|
||||||
|
|
||||||
if (preserveSplits) {
|
if (preserveSplits) {
|
||||||
|
|
|
@ -18,6 +18,7 @@
|
||||||
package org.apache.hadoop.hbase.client;
|
package org.apache.hadoop.hbase.client;
|
||||||
|
|
||||||
import static org.apache.hadoop.hbase.TableName.META_TABLE_NAME;
|
import static org.apache.hadoop.hbase.TableName.META_TABLE_NAME;
|
||||||
|
import static org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory.TRACK_IMPL;
|
||||||
import static org.hamcrest.CoreMatchers.instanceOf;
|
import static org.hamcrest.CoreMatchers.instanceOf;
|
||||||
import static org.hamcrest.MatcherAssert.assertThat;
|
import static org.hamcrest.MatcherAssert.assertThat;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
|
@ -36,6 +37,7 @@ import org.apache.hadoop.hbase.HBaseClassTestRule;
|
||||||
import org.apache.hadoop.hbase.HRegionLocation;
|
import org.apache.hadoop.hbase.HRegionLocation;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.client.TableDescriptorBuilder.ModifyableTableDescriptor;
|
import org.apache.hadoop.hbase.client.TableDescriptorBuilder.ModifyableTableDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory;
|
||||||
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
||||||
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
@ -149,7 +151,11 @@ public class TestAsyncTableAdminApi3 extends TestAsyncAdminBase {
|
||||||
admin.createTable(desc).join();
|
admin.createTable(desc).join();
|
||||||
ModifyableTableDescriptor modifyableDesc = ((ModifyableTableDescriptor) desc);
|
ModifyableTableDescriptor modifyableDesc = ((ModifyableTableDescriptor) desc);
|
||||||
TableDescriptor confirmedHtd = admin.getDescriptor(tableName).get();
|
TableDescriptor confirmedHtd = admin.getDescriptor(tableName).get();
|
||||||
assertEquals(0, modifyableDesc.compareTo((ModifyableTableDescriptor) confirmedHtd));
|
//HBASE-26246 introduced persist of store file tracker into table descriptor
|
||||||
|
desc = TableDescriptorBuilder.newBuilder(desc).setValue(TRACK_IMPL,
|
||||||
|
StoreFileTrackerFactory.getStoreFileTrackerImpl(TEST_UTIL.getConfiguration()).getName()).
|
||||||
|
build();
|
||||||
|
assertEquals(0, TableDescriptor.COMPARATOR.compare(desc, confirmedHtd));
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
|
|
@ -18,6 +18,7 @@
|
||||||
|
|
||||||
package org.apache.hadoop.hbase.master.procedure;
|
package org.apache.hadoop.hbase.master.procedure;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory.TRACK_IMPL;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertFalse;
|
import static org.junit.Assert.assertFalse;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
|
@ -58,6 +59,7 @@ import org.apache.hadoop.hbase.procedure2.Procedure;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility;
|
import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility;
|
||||||
import org.apache.hadoop.hbase.procedure2.StateMachineProcedure;
|
import org.apache.hadoop.hbase.procedure2.StateMachineProcedure;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.hbase.util.CommonFSUtils;
|
import org.apache.hadoop.hbase.util.CommonFSUtils;
|
||||||
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
||||||
|
@ -224,6 +226,11 @@ public class MasterProcedureTestingUtility {
|
||||||
assertTrue("family not found " + family[i], htd.getColumnFamily(Bytes.toBytes(family[i])) != null);
|
assertTrue("family not found " + family[i], htd.getColumnFamily(Bytes.toBytes(family[i])) != null);
|
||||||
}
|
}
|
||||||
assertEquals(family.length, htd.getColumnFamilyCount());
|
assertEquals(family.length, htd.getColumnFamilyCount());
|
||||||
|
|
||||||
|
// checks store file tracker impl has been properly set in htd
|
||||||
|
String storeFileTrackerImpl =
|
||||||
|
StoreFileTrackerFactory.getStoreFileTrackerImpl(master.getConfiguration()).getName();
|
||||||
|
assertEquals(storeFileTrackerImpl, htd.getValue(TRACK_IMPL));
|
||||||
}
|
}
|
||||||
|
|
||||||
public static void validateTableDeletion(
|
public static void validateTableDeletion(
|
||||||
|
|
|
@ -17,6 +17,7 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.master.procedure;
|
package org.apache.hadoop.hbase.master.procedure;
|
||||||
|
|
||||||
|
import static org.apache.hadoop.hbase.regionserver.storefiletracker.StoreFileTrackerFactory.TRACK_IMPL;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
import static org.junit.Assert.fail;
|
import static org.junit.Assert.fail;
|
||||||
|
@ -38,6 +39,7 @@ import org.apache.hadoop.hbase.master.MasterFileSystem;
|
||||||
import org.apache.hadoop.hbase.procedure2.Procedure;
|
import org.apache.hadoop.hbase.procedure2.Procedure;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
|
||||||
import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility;
|
import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.storefiletracker.TestStoreFileTracker;
|
||||||
import org.apache.hadoop.hbase.testclassification.MasterTests;
|
import org.apache.hadoop.hbase.testclassification.MasterTests;
|
||||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
@ -88,6 +90,21 @@ public class TestCreateTableProcedure extends TestTableDDLProcedureBase {
|
||||||
MasterProcedureTestingUtility.validateTableCreation(getMaster(), tableName, regions, F1, F2);
|
MasterProcedureTestingUtility.validateTableCreation(getMaster(), tableName, regions, F1, F2);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testCreateWithTrackImpl() throws Exception {
|
||||||
|
final TableName tableName = TableName.valueOf(name.getMethodName());
|
||||||
|
ProcedureExecutor<MasterProcedureEnv> procExec = getMasterProcedureExecutor();
|
||||||
|
TableDescriptor htd = MasterProcedureTestingUtility.createHTD(tableName, F1);
|
||||||
|
String trackerName = TestStoreFileTracker.class.getName();
|
||||||
|
htd = TableDescriptorBuilder.newBuilder(htd).setValue(TRACK_IMPL, trackerName).build();
|
||||||
|
RegionInfo[] regions = ModifyRegionUtils.createRegionInfos(htd, null);
|
||||||
|
long procId = ProcedureTestingUtility.submitAndWait(procExec,
|
||||||
|
new CreateTableProcedure(procExec.getEnvironment(), htd, regions));
|
||||||
|
ProcedureTestingUtility.assertProcNotFailed(procExec.getResult(procId));
|
||||||
|
htd = getMaster().getTableDescriptors().get(tableName);
|
||||||
|
assertEquals(trackerName, htd.getValue(TRACK_IMPL));
|
||||||
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testCreateWithoutColumnFamily() throws Exception {
|
public void testCreateWithoutColumnFamily() throws Exception {
|
||||||
final ProcedureExecutor<MasterProcedureEnv> procExec = getMasterProcedureExecutor();
|
final ProcedureExecutor<MasterProcedureEnv> procExec = getMasterProcedureExecutor();
|
||||||
|
|
|
@ -29,7 +29,7 @@ import java.util.stream.Collectors;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.HBaseClassTestRule;
|
import org.apache.hadoop.hbase.HBaseClassTestRule;
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtil;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
|
import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder;
|
||||||
import org.apache.hadoop.hbase.client.Get;
|
import org.apache.hadoop.hbase.client.Get;
|
||||||
|
@ -70,7 +70,7 @@ public class TestMigrationStoreFileTracker {
|
||||||
public static final HBaseClassTestRule CLASS_RULE =
|
public static final HBaseClassTestRule CLASS_RULE =
|
||||||
HBaseClassTestRule.forClass(TestMigrationStoreFileTracker.class);
|
HBaseClassTestRule.forClass(TestMigrationStoreFileTracker.class);
|
||||||
|
|
||||||
private static final HBaseTestingUtil UTIL = new HBaseTestingUtil();
|
private static final HBaseTestingUtility UTIL = new HBaseTestingUtility();
|
||||||
|
|
||||||
private static final byte[] CF = Bytes.toBytes("cf");
|
private static final byte[] CF = Bytes.toBytes("cf");
|
||||||
|
|
||||||
|
@ -125,7 +125,7 @@ public class TestMigrationStoreFileTracker {
|
||||||
conf.setClass(MigrationStoreFileTracker.SRC_IMPL, srcImplClass, StoreFileTrackerBase.class);
|
conf.setClass(MigrationStoreFileTracker.SRC_IMPL, srcImplClass, StoreFileTrackerBase.class);
|
||||||
conf.setClass(MigrationStoreFileTracker.DST_IMPL, dstImplClass, StoreFileTrackerBase.class);
|
conf.setClass(MigrationStoreFileTracker.DST_IMPL, dstImplClass, StoreFileTrackerBase.class);
|
||||||
rootDir = UTIL.getDataTestDir(name.getMethodName().replaceAll("[=:\\[ ]", "_"));
|
rootDir = UTIL.getDataTestDir(name.getMethodName().replaceAll("[=:\\[ ]", "_"));
|
||||||
wal = HBaseTestingUtil.createWal(conf, rootDir, RI);
|
wal = HBaseTestingUtility.createWal(conf, rootDir, RI);
|
||||||
}
|
}
|
||||||
|
|
||||||
@After
|
@After
|
||||||
|
|
|
@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.regionserver.storefiletracker;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Collection;
|
import java.util.Collection;
|
||||||
|
import java.util.Collections;
|
||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
|
@ -27,6 +28,7 @@ import java.util.Map;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreContext;
|
import org.apache.hadoop.hbase.regionserver.StoreContext;
|
||||||
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
import org.apache.hadoop.hbase.regionserver.StoreFileInfo;
|
||||||
|
import org.apache.hbase.thirdparty.org.apache.commons.collections4.CollectionUtils;
|
||||||
import org.slf4j.Logger;
|
import org.slf4j.Logger;
|
||||||
import org.slf4j.LoggerFactory;
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
|
@ -38,15 +40,21 @@ public class TestStoreFileTracker extends DefaultStoreFileTracker {
|
||||||
|
|
||||||
public TestStoreFileTracker(Configuration conf, boolean isPrimaryReplica, StoreContext ctx) {
|
public TestStoreFileTracker(Configuration conf, boolean isPrimaryReplica, StoreContext ctx) {
|
||||||
super(conf, isPrimaryReplica, ctx);
|
super(conf, isPrimaryReplica, ctx);
|
||||||
this.storeId = ctx.getRegionInfo().getEncodedName() + "-" + ctx.getFamily().getNameAsString();
|
if (ctx.getRegionFileSystem() != null) {
|
||||||
LOG.info("created storeId: {}", storeId);
|
this.storeId = ctx.getRegionInfo().getEncodedName() + "-" + ctx.getFamily().getNameAsString();
|
||||||
trackedFiles.computeIfAbsent(storeId, v -> new ArrayList<>());
|
LOG.info("created storeId: {}", storeId);
|
||||||
|
trackedFiles.computeIfAbsent(storeId, v -> new ArrayList<>());
|
||||||
|
} else {
|
||||||
|
LOG.info("ctx.getRegionFileSystem() returned null. Leaving storeId null.");
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
protected void doAddNewStoreFiles(Collection<StoreFileInfo> newFiles) throws IOException {
|
protected void doAddNewStoreFiles(Collection<StoreFileInfo> newFiles) throws IOException {
|
||||||
LOG.info("adding to storeId: {}", storeId);
|
LOG.info("adding to storeId: {}", storeId);
|
||||||
trackedFiles.get(storeId).addAll(newFiles);
|
trackedFiles.get(storeId).addAll(newFiles);
|
||||||
|
trackedFiles.putIfAbsent(storeId, (List<StoreFileInfo>)newFiles);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
Loading…
Reference in New Issue