HBASE-7484 Fix Restore with schema changes (Matteo Bertozzi)

git-svn-id: https://svn.apache.org/repos/asf/hbase/branches/hbase-7290@1445840 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Jonathan Hsieh 2013-02-13 18:49:13 +00:00
parent 0dd232a66b
commit db5100a75f
3 changed files with 136 additions and 20 deletions

View File

@ -209,17 +209,17 @@ public class RestoreSnapshotHelper {
Path snapshotRegionDir = new Path(snapshotDir, regionInfo.getEncodedName()); Path snapshotRegionDir = new Path(snapshotDir, regionInfo.getEncodedName());
Map<String, List<String>> snapshotFiles = Map<String, List<String>> snapshotFiles =
SnapshotReferenceUtil.getRegionHFileReferences(fs, snapshotRegionDir); SnapshotReferenceUtil.getRegionHFileReferences(fs, snapshotRegionDir);
Path regionDir = new Path(tableDir, regionInfo.getEncodedName()); Path regionDir = new Path(tableDir, regionInfo.getEncodedName());
String tableName = tableDesc.getNameAsString(); String tableName = tableDesc.getNameAsString();
for (Map.Entry<String, List<String>> familyEntry: snapshotFiles.entrySet()) { // Restore families present in the table
byte[] family = Bytes.toBytes(familyEntry.getKey()); for (Path familyDir: FSUtils.getFamilyDirs(fs, regionDir)) {
Path familyDir = new Path(regionDir, familyEntry.getKey()); byte[] family = Bytes.toBytes(familyDir.getName());
Set<String> familyFiles = getTableRegionFamilyFiles(familyDir); Set<String> familyFiles = getTableRegionFamilyFiles(familyDir);
List<String> snapshotFamilyFiles = snapshotFiles.remove(familyDir.getName());
if (snapshotFamilyFiles != null) {
List<String> hfilesToAdd = new LinkedList<String>(); List<String> hfilesToAdd = new LinkedList<String>();
for (String hfileName: familyEntry.getValue()) { for (String hfileName: snapshotFamilyFiles) {
if (familyFiles.contains(hfileName)) { if (familyFiles.contains(hfileName)) {
// HFile already present // HFile already present
familyFiles.remove(hfileName); familyFiles.remove(hfileName);
@ -229,15 +229,39 @@ public class RestoreSnapshotHelper {
} }
} }
// Restore Missing files
for (String hfileName: hfilesToAdd) {
LOG.trace("Adding HFileLink " + hfileName +
" to region=" + regionInfo.getEncodedName() + " table=" + tableName);
restoreStoreFile(familyDir, regionInfo, hfileName);
}
// Remove hfiles not present in the snapshot // Remove hfiles not present in the snapshot
for (String hfileName: familyFiles) { for (String hfileName: familyFiles) {
Path hfile = new Path(familyDir, hfileName); Path hfile = new Path(familyDir, hfileName);
LOG.trace("Removing hfile=" + hfile + " from table=" + tableName); LOG.trace("Removing hfile=" + hfile +
" from region=" + regionInfo.getEncodedName() + " table=" + tableName);
HFileArchiver.archiveStoreFile(fs, regionInfo, conf, tableDir, family, hfile); HFileArchiver.archiveStoreFile(fs, regionInfo, conf, tableDir, family, hfile);
} }
} else {
// Family doesn't exists in the snapshot
LOG.trace("Removing family=" + Bytes.toString(family) +
" from region=" + regionInfo.getEncodedName() + " table=" + tableName);
HFileArchiver.archiveFamily(fs, conf, regionInfo, tableDir, family);
fs.delete(familyDir, true);
}
}
// Restore Missing files // Add families not present in the table
for (String hfileName: hfilesToAdd) { for (Map.Entry<String, List<String>> familyEntry: snapshotFiles.entrySet()) {
byte[] family = Bytes.toBytes(familyEntry.getKey());
Path familyDir = new Path(regionDir, familyEntry.getKey());
if (!fs.mkdirs(familyDir)) {
throw new IOException("Unable to create familyDir=" + familyDir);
}
List<String> hfilesToAdd = new LinkedList<String>();
for (String hfileName: familyEntry.getValue()) {
LOG.trace("Adding HFileLink " + hfileName + " to table=" + tableName); LOG.trace("Adding HFileLink " + hfileName + " to table=" + tableName);
restoreStoreFile(familyDir, regionInfo, hfileName); restoreStoreFile(familyDir, regionInfo, hfileName);
} }

View File

@ -1218,6 +1218,20 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
return count; return count;
} }
public int countRows(final HTable table, final byte[]... families) throws IOException {
Scan scan = new Scan();
for (byte[] family: families) {
scan.addFamily(family);
}
ResultScanner results = table.getScanner(scan);
int count = 0;
for (@SuppressWarnings("unused") Result res : results) {
count++;
}
results.close();
return count;
}
/** /**
* Return an md5 digest of the entire contents of a table. * Return an md5 digest of the entire contents of a table.
*/ */

View File

@ -20,9 +20,12 @@ package org.apache.hadoop.hbase.client;
import static org.junit.Assert.assertArrayEquals; import static org.junit.Assert.assertArrayEquals;
import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue; import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
import java.io.IOException; import java.io.IOException;
import java.util.HashSet;
import java.util.List; import java.util.List;
import java.util.Set;
import org.apache.commons.logging.Log; import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
@ -37,10 +40,12 @@ import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.LargeTests; import org.apache.hadoop.hbase.LargeTests;
import org.apache.hadoop.hbase.master.HMaster; import org.apache.hadoop.hbase.master.HMaster;
import org.apache.hadoop.hbase.master.MasterFileSystem;
import org.apache.hadoop.hbase.master.snapshot.SnapshotManager; import org.apache.hadoop.hbase.master.snapshot.SnapshotManager;
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription; import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription;
import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.HRegion;
import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.regionserver.HRegionServer;
import org.apache.hadoop.hbase.regionserver.NoSuchColumnFamilyException;
import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils; import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils;
import org.apache.hadoop.hbase.snapshot.SnapshotDoesNotExistException; import org.apache.hadoop.hbase.snapshot.SnapshotDoesNotExistException;
import org.apache.hadoop.hbase.snapshot.SnapshotTestingUtils; import org.apache.hadoop.hbase.snapshot.SnapshotTestingUtils;
@ -62,6 +67,7 @@ public class TestRestoreSnapshotFromClient {
private final byte[] FAMILY = Bytes.toBytes("cf"); private final byte[] FAMILY = Bytes.toBytes("cf");
private byte[] emptySnapshot;
private byte[] snapshotName0; private byte[] snapshotName0;
private byte[] snapshotName1; private byte[] snapshotName1;
private byte[] snapshotName2; private byte[] snapshotName2;
@ -99,13 +105,22 @@ public class TestRestoreSnapshotFromClient {
long tid = System.currentTimeMillis(); long tid = System.currentTimeMillis();
tableName = Bytes.toBytes("testtb-" + tid); tableName = Bytes.toBytes("testtb-" + tid);
emptySnapshot = Bytes.toBytes("emptySnaptb-" + tid);
snapshotName0 = Bytes.toBytes("snaptb0-" + tid); snapshotName0 = Bytes.toBytes("snaptb0-" + tid);
snapshotName1 = Bytes.toBytes("snaptb1-" + tid); snapshotName1 = Bytes.toBytes("snaptb1-" + tid);
snapshotName2 = Bytes.toBytes("snaptb2-" + tid); snapshotName2 = Bytes.toBytes("snaptb2-" + tid);
// create Table and disable it // create Table and disable it
createTable(tableName, FAMILY); createTable(tableName, FAMILY);
admin.disableTable(tableName);
// take an empty snapshot
admin.snapshot(emptySnapshot, tableName);
HTable table = new HTable(TEST_UTIL.getConfiguration(), tableName); HTable table = new HTable(TEST_UTIL.getConfiguration(), tableName);
// enable table and insert data
admin.enableTable(tableName);
loadData(table, 500, FAMILY); loadData(table, 500, FAMILY);
snapshot0Rows = TEST_UTIL.countRows(table); snapshot0Rows = TEST_UTIL.countRows(table);
admin.disableTable(tableName); admin.disableTable(tableName);
@ -149,6 +164,13 @@ public class TestRestoreSnapshotFromClient {
table = new HTable(TEST_UTIL.getConfiguration(), tableName); table = new HTable(TEST_UTIL.getConfiguration(), tableName);
assertEquals(snapshot0Rows, TEST_UTIL.countRows(table)); assertEquals(snapshot0Rows, TEST_UTIL.countRows(table));
// Restore from emptySnapshot
admin.disableTable(tableName);
admin.restoreSnapshot(emptySnapshot);
admin.enableTable(tableName);
table = new HTable(TEST_UTIL.getConfiguration(), tableName);
assertEquals(0, TEST_UTIL.countRows(table));
// Restore from snapshot-1 // Restore from snapshot-1
admin.disableTable(tableName); admin.disableTable(tableName);
admin.restoreSnapshot(snapshotName1); admin.restoreSnapshot(snapshotName1);
@ -157,6 +179,49 @@ public class TestRestoreSnapshotFromClient {
assertEquals(snapshot1Rows, TEST_UTIL.countRows(table)); assertEquals(snapshot1Rows, TEST_UTIL.countRows(table));
} }
@Test
public void testRestoreSchemaChange() throws IOException {
byte[] TEST_FAMILY2 = Bytes.toBytes("cf2");
// Add one column family and put some data in it
admin.disableTable(tableName);
admin.addColumn(tableName, new HColumnDescriptor(TEST_FAMILY2));
admin.enableTable(tableName);
HTable table = new HTable(TEST_UTIL.getConfiguration(), tableName);
loadData(table, 500, TEST_FAMILY2);
long snapshot2Rows = snapshot1Rows + 500;
assertEquals(snapshot2Rows, TEST_UTIL.countRows(table));
assertEquals(500, TEST_UTIL.countRows(table, TEST_FAMILY2));
// Take a snapshot
admin.disableTable(tableName);
admin.snapshot(snapshotName2, tableName);
// Restore the snapshot (without the cf)
admin.restoreSnapshot(snapshotName0);
admin.enableTable(tableName);
table = new HTable(TEST_UTIL.getConfiguration(), tableName);
try {
TEST_UTIL.countRows(table, TEST_FAMILY2);
fail("family '" + Bytes.toString(TEST_FAMILY2) + "' should not exists");
} catch (NoSuchColumnFamilyException e) {
// expected
}
assertEquals(snapshot0Rows, TEST_UTIL.countRows(table));
Set<String> fsFamilies = getFamiliesFromFS(tableName);
assertEquals(1, fsFamilies.size());
// Restore back the snapshot (with the cf)
admin.disableTable(tableName);
admin.restoreSnapshot(snapshotName2);
admin.enableTable(tableName);
table = new HTable(TEST_UTIL.getConfiguration(), tableName);
assertEquals(500, TEST_UTIL.countRows(table, TEST_FAMILY2));
assertEquals(snapshot2Rows, TEST_UTIL.countRows(table));
fsFamilies = getFamiliesFromFS(tableName);
assertEquals(2, fsFamilies.size());
}
@Test(expected=SnapshotDoesNotExistException.class) @Test(expected=SnapshotDoesNotExistException.class)
public void testCloneNonExistentSnapshot() throws IOException, InterruptedException { public void testCloneNonExistentSnapshot() throws IOException, InterruptedException {
String snapshotName = "random-snapshot-" + System.currentTimeMillis(); String snapshotName = "random-snapshot-" + System.currentTimeMillis();
@ -169,6 +234,7 @@ public class TestRestoreSnapshotFromClient {
byte[] clonedTableName = Bytes.toBytes("clonedtb-" + System.currentTimeMillis()); byte[] clonedTableName = Bytes.toBytes("clonedtb-" + System.currentTimeMillis());
testCloneSnapshot(clonedTableName, snapshotName0, snapshot0Rows); testCloneSnapshot(clonedTableName, snapshotName0, snapshot0Rows);
testCloneSnapshot(clonedTableName, snapshotName1, snapshot1Rows); testCloneSnapshot(clonedTableName, snapshotName1, snapshot1Rows);
testCloneSnapshot(clonedTableName, emptySnapshot, 0);
} }
private void testCloneSnapshot(final byte[] tableName, final byte[] snapshotName, private void testCloneSnapshot(final byte[] tableName, final byte[] snapshotName,
@ -298,4 +364,16 @@ public class TestRestoreSnapshotFromClient {
private void waitCleanerRun() throws InterruptedException { private void waitCleanerRun() throws InterruptedException {
TEST_UTIL.getMiniHBaseCluster().getMaster().getHFileCleaner().choreForTesting(); TEST_UTIL.getMiniHBaseCluster().getMaster().getHFileCleaner().choreForTesting();
} }
private Set<String> getFamiliesFromFS(final byte[] tableName) throws IOException {
MasterFileSystem mfs = TEST_UTIL.getMiniHBaseCluster().getMaster().getMasterFileSystem();
Set<String> families = new HashSet<String>();
Path tableDir = HTableDescriptor.getTableDir(mfs.getRootDir(), tableName);
for (Path regionDir: FSUtils.getRegionDirs(mfs.getFileSystem(), tableDir)) {
for (Path familyDir: FSUtils.getFamilyDirs(mfs.getFileSystem(), regionDir)) {
families.add(familyDir.getName());
}
}
return families;
}
} }