HBASE-4797 [availability] Skip recovered.edits files with edits we know older than what region currently has (Jimmy Jiang)
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1205290 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
ef52263373
commit
351e75629f
|
@ -2506,12 +2506,32 @@ public class HRegion implements HeapSize { // , Writable{
|
||||||
long seqid = minSeqId;
|
long seqid = minSeqId;
|
||||||
NavigableSet<Path> files = HLog.getSplitEditFilesSorted(this.fs, regiondir);
|
NavigableSet<Path> files = HLog.getSplitEditFilesSorted(this.fs, regiondir);
|
||||||
if (files == null || files.isEmpty()) return seqid;
|
if (files == null || files.isEmpty()) return seqid;
|
||||||
|
boolean checkSafeToSkip = true;
|
||||||
for (Path edits: files) {
|
for (Path edits: files) {
|
||||||
if (edits == null || !this.fs.exists(edits)) {
|
if (edits == null || !this.fs.exists(edits)) {
|
||||||
LOG.warn("Null or non-existent edits file: " + edits);
|
LOG.warn("Null or non-existent edits file: " + edits);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if (isZeroLengthThenDelete(this.fs, edits)) continue;
|
if (isZeroLengthThenDelete(this.fs, edits)) continue;
|
||||||
|
|
||||||
|
if (checkSafeToSkip) {
|
||||||
|
Path higher = files.higher(edits);
|
||||||
|
long maxSeqId = Long.MAX_VALUE;
|
||||||
|
if (higher != null) {
|
||||||
|
// Edit file name pattern, HLog.EDITFILES_NAME_PATTERN: "-?[0-9]+"
|
||||||
|
String fileName = higher.getName();
|
||||||
|
maxSeqId = Math.abs(Long.parseLong(fileName));
|
||||||
|
}
|
||||||
|
if (maxSeqId <= minSeqId) {
|
||||||
|
String msg = "Maximum possible sequenceid for this log is " + maxSeqId
|
||||||
|
+ ", skipped the whole file, path=" + edits;
|
||||||
|
LOG.debug(msg);
|
||||||
|
continue;
|
||||||
|
} else {
|
||||||
|
checkSafeToSkip = false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
seqid = replayRecoveredEdits(edits, seqid, reporter);
|
seqid = replayRecoveredEdits(edits, seqid, reporter);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
|
|
|
@ -34,16 +34,30 @@ import java.util.concurrent.atomic.AtomicReference;
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FSDataOutputStream;
|
||||||
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.*;
|
import org.apache.hadoop.hbase.DoNotRetryIOException;
|
||||||
|
import org.apache.hadoop.hbase.HBaseConfiguration;
|
||||||
|
import org.apache.hadoop.hbase.HBaseTestCase;
|
||||||
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
|
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.HConstants.OperationStatusCode;
|
import org.apache.hadoop.hbase.HConstants.OperationStatusCode;
|
||||||
|
import org.apache.hadoop.hbase.HDFSBlocksDistribution;
|
||||||
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.KeyValue;
|
||||||
|
import org.apache.hadoop.hbase.MediumTests;
|
||||||
|
import org.apache.hadoop.hbase.MiniHBaseCluster;
|
||||||
|
import org.apache.hadoop.hbase.MultithreadedTestUtil;
|
||||||
import org.apache.hadoop.hbase.MultithreadedTestUtil.TestThread;
|
import org.apache.hadoop.hbase.MultithreadedTestUtil.TestThread;
|
||||||
import org.apache.hadoop.hbase.client.Delete;
|
import org.apache.hadoop.hbase.client.Delete;
|
||||||
import org.apache.hadoop.hbase.client.Get;
|
import org.apache.hadoop.hbase.client.Get;
|
||||||
|
import org.apache.hadoop.hbase.client.HTable;
|
||||||
import org.apache.hadoop.hbase.client.Put;
|
import org.apache.hadoop.hbase.client.Put;
|
||||||
import org.apache.hadoop.hbase.client.Result;
|
import org.apache.hadoop.hbase.client.Result;
|
||||||
import org.apache.hadoop.hbase.client.Scan;
|
import org.apache.hadoop.hbase.client.Scan;
|
||||||
import org.apache.hadoop.hbase.client.HTable;
|
|
||||||
import org.apache.hadoop.hbase.filter.BinaryComparator;
|
import org.apache.hadoop.hbase.filter.BinaryComparator;
|
||||||
import org.apache.hadoop.hbase.filter.ColumnCountGetFilter;
|
import org.apache.hadoop.hbase.filter.ColumnCountGetFilter;
|
||||||
import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
|
import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
|
||||||
|
@ -52,9 +66,13 @@ import org.apache.hadoop.hbase.filter.FilterList;
|
||||||
import org.apache.hadoop.hbase.filter.NullComparator;
|
import org.apache.hadoop.hbase.filter.NullComparator;
|
||||||
import org.apache.hadoop.hbase.filter.PrefixFilter;
|
import org.apache.hadoop.hbase.filter.PrefixFilter;
|
||||||
import org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
|
import org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
|
||||||
|
import org.apache.hadoop.hbase.monitoring.MonitoredTask;
|
||||||
|
import org.apache.hadoop.hbase.monitoring.TaskMonitor;
|
||||||
import org.apache.hadoop.hbase.regionserver.HRegion.RegionScannerImpl;
|
import org.apache.hadoop.hbase.regionserver.HRegion.RegionScannerImpl;
|
||||||
import org.apache.hadoop.hbase.regionserver.metrics.SchemaMetrics;
|
import org.apache.hadoop.hbase.regionserver.metrics.SchemaMetrics;
|
||||||
import org.apache.hadoop.hbase.regionserver.wal.HLog;
|
import org.apache.hadoop.hbase.regionserver.wal.HLog;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
||||||
import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper;
|
import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper;
|
||||||
|
@ -64,9 +82,9 @@ import org.apache.hadoop.hbase.util.Pair;
|
||||||
import org.apache.hadoop.hbase.util.PairOfSameType;
|
import org.apache.hadoop.hbase.util.PairOfSameType;
|
||||||
import org.apache.hadoop.hbase.util.Threads;
|
import org.apache.hadoop.hbase.util.Threads;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
|
import org.junit.experimental.categories.Category;
|
||||||
|
|
||||||
import com.google.common.collect.Lists;
|
import com.google.common.collect.Lists;
|
||||||
import org.junit.experimental.categories.Category;
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -2809,6 +2827,115 @@ public class TestHRegion extends HBaseTestCase {
|
||||||
region.get(g, null);
|
region.get(g, null);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void testSkipRecoveredEditsReplay() throws Exception {
|
||||||
|
String method = "testSkipRecoveredEditsReplay";
|
||||||
|
byte[] tableName = Bytes.toBytes(method);
|
||||||
|
byte[] family = Bytes.toBytes("family");
|
||||||
|
Configuration conf = HBaseConfiguration.create();
|
||||||
|
initHRegion(tableName, method, conf, family);
|
||||||
|
Path regiondir = region.getRegionDir();
|
||||||
|
FileSystem fs = region.getFilesystem();
|
||||||
|
byte[] regionName = region.getRegionInfo().getEncodedNameAsBytes();
|
||||||
|
|
||||||
|
Path recoveredEditsDir = HLog.getRegionDirRecoveredEditsDir(regiondir);
|
||||||
|
|
||||||
|
long maxSeqId = 1050;
|
||||||
|
long minSeqId = 1000;
|
||||||
|
|
||||||
|
for (long i = minSeqId; i <= maxSeqId; i += 10) {
|
||||||
|
Path recoveredEdits = new Path(recoveredEditsDir, String.format("%019d", i));
|
||||||
|
HLog.Writer writer = HLog.createWriter(fs, recoveredEdits, conf);
|
||||||
|
|
||||||
|
long time = System.nanoTime();
|
||||||
|
WALEdit edit = new WALEdit();
|
||||||
|
edit.add(new KeyValue(row, family, Bytes.toBytes(i),
|
||||||
|
time, KeyValue.Type.Put, Bytes.toBytes(i)));
|
||||||
|
writer.append(new HLog.Entry(new HLogKey(regionName, tableName,
|
||||||
|
i, time, HConstants.DEFAULT_CLUSTER_ID), edit));
|
||||||
|
|
||||||
|
writer.close();
|
||||||
|
}
|
||||||
|
MonitoredTask status = TaskMonitor.get().createStatus(method);
|
||||||
|
long seqId = region.replayRecoveredEditsIfAny(regiondir, minSeqId-1, null, status);
|
||||||
|
assertEquals(maxSeqId, seqId);
|
||||||
|
Get get = new Get(row);
|
||||||
|
Result result = region.get(get, null);
|
||||||
|
for (long i = minSeqId; i <= maxSeqId; i += 10) {
|
||||||
|
List<KeyValue> kvs = result.getColumn(family, Bytes.toBytes(i));
|
||||||
|
assertEquals(1, kvs.size());
|
||||||
|
assertEquals(Bytes.toBytes(i), kvs.get(0).getValue());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public void testSkipRecoveredEditsReplaySomeIgnored() throws Exception {
|
||||||
|
String method = "testSkipRecoveredEditsReplaySomeIgnored";
|
||||||
|
byte[] tableName = Bytes.toBytes(method);
|
||||||
|
byte[] family = Bytes.toBytes("family");
|
||||||
|
initHRegion(tableName, method, HBaseConfiguration.create(), family);
|
||||||
|
Path regiondir = region.getRegionDir();
|
||||||
|
FileSystem fs = region.getFilesystem();
|
||||||
|
byte[] regionName = region.getRegionInfo().getEncodedNameAsBytes();
|
||||||
|
|
||||||
|
Path recoveredEditsDir = HLog.getRegionDirRecoveredEditsDir(regiondir);
|
||||||
|
|
||||||
|
long maxSeqId = 1050;
|
||||||
|
long minSeqId = 1000;
|
||||||
|
|
||||||
|
for (long i = minSeqId; i <= maxSeqId; i += 10) {
|
||||||
|
Path recoveredEdits = new Path(recoveredEditsDir, String.format("%019d", i));
|
||||||
|
HLog.Writer writer = HLog.createWriter(fs, recoveredEdits, conf);
|
||||||
|
|
||||||
|
long time = System.nanoTime();
|
||||||
|
WALEdit edit = new WALEdit();
|
||||||
|
edit.add(new KeyValue(row, family, Bytes.toBytes(i),
|
||||||
|
time, KeyValue.Type.Put, Bytes.toBytes(i)));
|
||||||
|
writer.append(new HLog.Entry(new HLogKey(regionName, tableName,
|
||||||
|
i, time, HConstants.DEFAULT_CLUSTER_ID), edit));
|
||||||
|
|
||||||
|
writer.close();
|
||||||
|
}
|
||||||
|
long recoverSeqId = 1030;
|
||||||
|
MonitoredTask status = TaskMonitor.get().createStatus(method);
|
||||||
|
long seqId = region.replayRecoveredEditsIfAny(regiondir, recoverSeqId-1, null, status);
|
||||||
|
assertEquals(maxSeqId, seqId);
|
||||||
|
Get get = new Get(row);
|
||||||
|
Result result = region.get(get, null);
|
||||||
|
for (long i = minSeqId; i <= maxSeqId; i += 10) {
|
||||||
|
List<KeyValue> kvs = result.getColumn(family, Bytes.toBytes(i));
|
||||||
|
if (i < recoverSeqId) {
|
||||||
|
assertEquals(0, kvs.size());
|
||||||
|
} else {
|
||||||
|
assertEquals(1, kvs.size());
|
||||||
|
assertEquals(Bytes.toBytes(i), kvs.get(0).getValue());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
public void testSkipRecoveredEditsReplayAllIgnored() throws Exception {
|
||||||
|
String method = "testSkipRecoveredEditsReplayAllIgnored";
|
||||||
|
byte[] tableName = Bytes.toBytes(method);
|
||||||
|
byte[] family = Bytes.toBytes("family");
|
||||||
|
initHRegion(tableName, method, HBaseConfiguration.create(), family);
|
||||||
|
Path regiondir = region.getRegionDir();
|
||||||
|
FileSystem fs = region.getFilesystem();
|
||||||
|
|
||||||
|
Path recoveredEditsDir = HLog.getRegionDirRecoveredEditsDir(regiondir);
|
||||||
|
for (int i = 1000; i < 1050; i += 10) {
|
||||||
|
Path recoveredEdits = new Path(
|
||||||
|
recoveredEditsDir, String.format("%019d", i));
|
||||||
|
FSDataOutputStream dos= fs.create(recoveredEdits);
|
||||||
|
dos.writeInt(i);
|
||||||
|
dos.close();
|
||||||
|
}
|
||||||
|
long minSeqId = 2000;
|
||||||
|
Path recoveredEdits = new Path(
|
||||||
|
recoveredEditsDir, String.format("%019d", minSeqId-1));
|
||||||
|
FSDataOutputStream dos= fs.create(recoveredEdits);
|
||||||
|
dos.close();
|
||||||
|
long seqId = region.replayRecoveredEditsIfAny(regiondir, minSeqId, null, null);
|
||||||
|
assertEquals(minSeqId, seqId);
|
||||||
|
}
|
||||||
|
|
||||||
public void testIndexesScanWithOneDeletedRow() throws IOException {
|
public void testIndexesScanWithOneDeletedRow() throws IOException {
|
||||||
byte[] tableName = Bytes.toBytes("testIndexesScanWithOneDeletedRow");
|
byte[] tableName = Bytes.toBytes("testIndexesScanWithOneDeletedRow");
|
||||||
byte[] family = Bytes.toBytes("family");
|
byte[] family = Bytes.toBytes("family");
|
||||||
|
|
Loading…
Reference in New Issue