HBASE-12600: Remove REPLAY tag dependency in Distributed Replay Mode
This commit is contained in:
parent
86dc3cab96
commit
a184620fb3
|
@ -68,19 +68,6 @@ public class CellComparator implements Comparator<Cell>, Serializable {
|
||||||
|
|
||||||
if (!ignoreSequenceid) {
|
if (!ignoreSequenceid) {
|
||||||
// Negate following comparisons so later edits show up first
|
// Negate following comparisons so later edits show up first
|
||||||
|
|
||||||
// compare log replay tag value if there is any
|
|
||||||
// when either keyvalue tagged with log replay sequence number, we need to compare them:
|
|
||||||
// 1) when both keyvalues have the tag, then use the tag values for comparison
|
|
||||||
// 2) when one has and the other doesn't have, the one without the log
|
|
||||||
// replay tag wins because
|
|
||||||
// it means the edit isn't from recovery but new one coming from clients during recovery
|
|
||||||
// 3) when both doesn't have, then skip to the next mvcc comparison
|
|
||||||
long leftChangeSeqNum = getReplaySeqNum(a);
|
|
||||||
long RightChangeSeqNum = getReplaySeqNum(b);
|
|
||||||
if (leftChangeSeqNum != Long.MAX_VALUE || RightChangeSeqNum != Long.MAX_VALUE) {
|
|
||||||
return Longs.compare(RightChangeSeqNum, leftChangeSeqNum);
|
|
||||||
}
|
|
||||||
// mvccVersion: later sorts first
|
// mvccVersion: later sorts first
|
||||||
return Longs.compare(b.getMvccVersion(), a.getMvccVersion());
|
return Longs.compare(b.getMvccVersion(), a.getMvccVersion());
|
||||||
} else {
|
} else {
|
||||||
|
@ -88,22 +75,6 @@ public class CellComparator implements Comparator<Cell>, Serializable {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Return replay log sequence number for the cell
|
|
||||||
*
|
|
||||||
* @param c
|
|
||||||
* @return Long.MAX_VALUE if there is no LOG_REPLAY_TAG
|
|
||||||
*/
|
|
||||||
private static long getReplaySeqNum(final Cell c) {
|
|
||||||
Tag tag = Tag.getTag(c.getTagsArray(), c.getTagsOffset(), c.getTagsLength(),
|
|
||||||
TagType.LOG_REPLAY_TAG_TYPE);
|
|
||||||
|
|
||||||
if (tag != null) {
|
|
||||||
return Bytes.toLong(tag.getBuffer(), tag.getTagOffset(), tag.getTagLength());
|
|
||||||
}
|
|
||||||
return Long.MAX_VALUE;
|
|
||||||
}
|
|
||||||
|
|
||||||
public static int findCommonPrefixInRowPart(Cell left, Cell right, int rowCommonPrefix) {
|
public static int findCommonPrefixInRowPart(Cell left, Cell right, int rowCommonPrefix) {
|
||||||
return findCommonPrefix(left.getRowArray(), right.getRowArray(), left.getRowLength()
|
return findCommonPrefix(left.getRowArray(), right.getRowArray(), left.getRowLength()
|
||||||
- rowCommonPrefix, right.getRowLength() - rowCommonPrefix, left.getRowOffset()
|
- rowCommonPrefix, right.getRowLength() - rowCommonPrefix, left.getRowOffset()
|
||||||
|
|
|
@ -746,6 +746,7 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
|
||||||
c.getQualifierArray(), c.getQualifierOffset(), (int) c.getQualifierLength(),
|
c.getQualifierArray(), c.getQualifierOffset(), (int) c.getQualifierLength(),
|
||||||
c.getTimestamp(), Type.codeToType(c.getTypeByte()), c.getValueArray(), c.getValueOffset(),
|
c.getTimestamp(), Type.codeToType(c.getTypeByte()), c.getValueArray(), c.getValueOffset(),
|
||||||
c.getValueLength(), c.getTagsArray(), c.getTagsOffset(), c.getTagsLength());
|
c.getValueLength(), c.getTagsArray(), c.getTagsOffset(), c.getTagsLength());
|
||||||
|
this.seqId = c.getSequenceId();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -1020,8 +1021,8 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
|
||||||
checkForTagsLength(tagsLength);
|
checkForTagsLength(tagsLength);
|
||||||
// Allocate right-sized byte array.
|
// Allocate right-sized byte array.
|
||||||
int keyLength = (int) getKeyDataStructureSize(rlength, flength, qlength);
|
int keyLength = (int) getKeyDataStructureSize(rlength, flength, qlength);
|
||||||
byte [] bytes =
|
byte[] bytes = new byte[(int) getKeyValueDataStructureSize(rlength, flength, qlength, vlength,
|
||||||
new byte[(int) getKeyValueDataStructureSize(rlength, flength, qlength, vlength, tagsLength)];
|
tagsLength)];
|
||||||
// Write key, value and key row length.
|
// Write key, value and key row length.
|
||||||
int pos = 0;
|
int pos = 0;
|
||||||
pos = Bytes.putInt(bytes, pos, keyLength);
|
pos = Bytes.putInt(bytes, pos, keyLength);
|
||||||
|
@ -2551,8 +2552,8 @@ public class KeyValue implements Cell, HeapSize, Cloneable, SettableSequenceId,
|
||||||
* Compare two keys assuming that the first n bytes are the same.
|
* Compare two keys assuming that the first n bytes are the same.
|
||||||
* @param commonPrefix How many bytes are the same.
|
* @param commonPrefix How many bytes are the same.
|
||||||
*/
|
*/
|
||||||
int compareIgnoringPrefix(
|
int compareIgnoringPrefix(int commonPrefix, byte[] left, int loffset, int llength,
|
||||||
int commonPrefix, byte[] left, int loffset, int llength, byte[] right, int roffset, int rlength
|
byte[] right, int roffset, int rlength
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -26,7 +26,7 @@ public final class TagType {
|
||||||
// Please declare new Tag Types here to avoid step on pre-existing tag types.
|
// Please declare new Tag Types here to avoid step on pre-existing tag types.
|
||||||
public static final byte ACL_TAG_TYPE = (byte) 1;
|
public static final byte ACL_TAG_TYPE = (byte) 1;
|
||||||
public static final byte VISIBILITY_TAG_TYPE = (byte) 2;
|
public static final byte VISIBILITY_TAG_TYPE = (byte) 2;
|
||||||
public static final byte LOG_REPLAY_TAG_TYPE = (byte) 3;
|
// public static final byte LOG_REPLAY_TAG_TYPE = (byte) 3; // deprecated
|
||||||
public static final byte VISIBILITY_EXP_SERIALIZATION_FORMAT_TAG_TYPE = (byte)4;
|
public static final byte VISIBILITY_EXP_SERIALIZATION_FORMAT_TAG_TYPE = (byte)4;
|
||||||
// String based tag type used in replication
|
// String based tag type used in replication
|
||||||
public static final byte STRING_VIS_TAG_TYPE = (byte) 7;
|
public static final byte STRING_VIS_TAG_TYPE = (byte) 7;
|
||||||
|
|
|
@ -774,8 +774,8 @@ public class ZKSplitLogManagerCoordination extends ZooKeeperListener implements
|
||||||
public void setRecoveryMode(boolean isForInitialization) throws IOException {
|
public void setRecoveryMode(boolean isForInitialization) throws IOException {
|
||||||
synchronized(this) {
|
synchronized(this) {
|
||||||
if (this.isDrainingDone) {
|
if (this.isDrainingDone) {
|
||||||
// when there is no outstanding splitlogtask after master start up, we already have up to date
|
// when there is no outstanding splitlogtask after master start up, we already have up to
|
||||||
// recovery mode
|
// date recovery mode
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -867,12 +867,10 @@ public class ZKSplitLogManagerCoordination extends ZooKeeperListener implements
|
||||||
boolean dlr =
|
boolean dlr =
|
||||||
conf.getBoolean(HConstants.DISTRIBUTED_LOG_REPLAY_KEY,
|
conf.getBoolean(HConstants.DISTRIBUTED_LOG_REPLAY_KEY,
|
||||||
HConstants.DEFAULT_DISTRIBUTED_LOG_REPLAY_CONFIG);
|
HConstants.DEFAULT_DISTRIBUTED_LOG_REPLAY_CONFIG);
|
||||||
int version = conf.getInt(HFile.FORMAT_VERSION_KEY, HFile.MAX_FORMAT_VERSION);
|
|
||||||
if (LOG.isDebugEnabled()) {
|
if (LOG.isDebugEnabled()) {
|
||||||
LOG.debug("Distributed log replay=" + dlr + ", " + HFile.FORMAT_VERSION_KEY + "=" + version);
|
LOG.debug("Distributed log replay=" + dlr);
|
||||||
}
|
}
|
||||||
// For distributed log replay, hfile version must be 3 at least; we need tag support.
|
return dlr;
|
||||||
return dlr && (version >= 3);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private boolean resubmit(ServerName serverName, String path, int version) {
|
private boolean resubmit(ServerName serverName, String path, int version) {
|
||||||
|
|
|
@ -2688,7 +2688,7 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver { //
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
doRollBackMemstore = true; // If we have a failure, we need to clean what we wrote
|
doRollBackMemstore = true; // If we have a failure, we need to clean what we wrote
|
||||||
addedSize += applyFamilyMapToMemstore(familyMaps[i], mvccNum, memstoreCells);
|
addedSize += applyFamilyMapToMemstore(familyMaps[i], mvccNum, memstoreCells, isInReplay);
|
||||||
}
|
}
|
||||||
|
|
||||||
// ------------------------------------
|
// ------------------------------------
|
||||||
|
@ -3191,12 +3191,13 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver { //
|
||||||
* @param localizedWriteEntry The WriteEntry of the MVCC for this transaction.
|
* @param localizedWriteEntry The WriteEntry of the MVCC for this transaction.
|
||||||
* If null, then this method internally creates a mvcc transaction.
|
* If null, then this method internally creates a mvcc transaction.
|
||||||
* @param output newly added KVs into memstore
|
* @param output newly added KVs into memstore
|
||||||
|
* @param isInReplay true when adding replayed KVs into memstore
|
||||||
* @return the additional memory usage of the memstore caused by the
|
* @return the additional memory usage of the memstore caused by the
|
||||||
* new entries.
|
* new entries.
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
*/
|
*/
|
||||||
private long applyFamilyMapToMemstore(Map<byte[], List<Cell>> familyMap,
|
private long applyFamilyMapToMemstore(Map<byte[], List<Cell>> familyMap,
|
||||||
long mvccNum, List<Cell> memstoreCells) throws IOException {
|
long mvccNum, List<Cell> memstoreCells, boolean isInReplay) throws IOException {
|
||||||
long size = 0;
|
long size = 0;
|
||||||
|
|
||||||
for (Map.Entry<byte[], List<Cell>> e : familyMap.entrySet()) {
|
for (Map.Entry<byte[], List<Cell>> e : familyMap.entrySet()) {
|
||||||
|
@ -3211,6 +3212,10 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver { //
|
||||||
Pair<Long, Cell> ret = store.add(cell);
|
Pair<Long, Cell> ret = store.add(cell);
|
||||||
size += ret.getFirst();
|
size += ret.getFirst();
|
||||||
memstoreCells.add(ret.getSecond());
|
memstoreCells.add(ret.getSecond());
|
||||||
|
if(isInReplay) {
|
||||||
|
// set memstore newly added cells with replay mvcc number
|
||||||
|
CellUtil.setSequenceId(ret.getSecond(), mvccNum);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -3409,7 +3414,8 @@ public class HRegion implements HeapSize, PropagatingConfigurationObserver { //
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
// replay the edits. Replay can return -1 if everything is skipped, only update if seqId is greater
|
// replay the edits. Replay can return -1 if everything is skipped, only update
|
||||||
|
// if seqId is greater
|
||||||
seqid = Math.max(seqid, replayRecoveredEdits(edits, maxSeqIdInStores, reporter));
|
seqid = Math.max(seqid, replayRecoveredEdits(edits, maxSeqIdInStores, reporter));
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
boolean skipErrors = conf.getBoolean(
|
boolean skipErrors = conf.getBoolean(
|
||||||
|
|
|
@ -1445,8 +1445,6 @@ public class RSRpcServices implements HBaseRPCErrorHandler,
|
||||||
entries.get(0).getKey().getEncodedRegionName().toStringUtf8());
|
entries.get(0).getKey().getEncodedRegionName().toStringUtf8());
|
||||||
RegionCoprocessorHost coprocessorHost = region.getCoprocessorHost();
|
RegionCoprocessorHost coprocessorHost = region.getCoprocessorHost();
|
||||||
List<Pair<HLogKey, WALEdit>> walEntries = new ArrayList<Pair<HLogKey, WALEdit>>();
|
List<Pair<HLogKey, WALEdit>> walEntries = new ArrayList<Pair<HLogKey, WALEdit>>();
|
||||||
// when tag is enabled, we need tag replay edits with log sequence number
|
|
||||||
boolean needAddReplayTag = (HFile.getFormatVersion(regionServer.conf) >= 3);
|
|
||||||
for (WALEntry entry : entries) {
|
for (WALEntry entry : entries) {
|
||||||
if (regionServer.nonceManager != null) {
|
if (regionServer.nonceManager != null) {
|
||||||
long nonceGroup = entry.getKey().hasNonceGroup()
|
long nonceGroup = entry.getKey().hasNonceGroup()
|
||||||
|
@ -1457,7 +1455,7 @@ public class RSRpcServices implements HBaseRPCErrorHandler,
|
||||||
Pair<HLogKey, WALEdit> walEntry = (coprocessorHost == null) ? null :
|
Pair<HLogKey, WALEdit> walEntry = (coprocessorHost == null) ? null :
|
||||||
new Pair<HLogKey, WALEdit>();
|
new Pair<HLogKey, WALEdit>();
|
||||||
List<HLogSplitter.MutationReplay> edits = HLogSplitter.getMutationsFromWALEntry(entry,
|
List<HLogSplitter.MutationReplay> edits = HLogSplitter.getMutationsFromWALEntry(entry,
|
||||||
cells, walEntry, needAddReplayTag);
|
cells, walEntry);
|
||||||
if (coprocessorHost != null) {
|
if (coprocessorHost != null) {
|
||||||
// Start coprocessor replay here. The coprocessor is for each WALEdit instead of a
|
// Start coprocessor replay here. The coprocessor is for each WALEdit instead of a
|
||||||
// KeyValue.
|
// KeyValue.
|
||||||
|
|
|
@ -92,8 +92,9 @@ public class DefaultCompactor extends Compactor {
|
||||||
smallestReadPoint = Math.min(fd.minSeqIdToKeep, smallestReadPoint);
|
smallestReadPoint = Math.min(fd.minSeqIdToKeep, smallestReadPoint);
|
||||||
cleanSeqId = true;
|
cleanSeqId = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
writer = store.createWriterInTmp(fd.maxKeyCount, this.compactionCompression, true,
|
writer = store.createWriterInTmp(fd.maxKeyCount, this.compactionCompression, true,
|
||||||
fd.maxMVCCReadpoint >= smallestReadPoint, fd.maxTagsLength > 0);
|
true, fd.maxTagsLength > 0);
|
||||||
boolean finished = performCompaction(scanner, writer, smallestReadPoint, cleanSeqId);
|
boolean finished = performCompaction(scanner, writer, smallestReadPoint, cleanSeqId);
|
||||||
if (!finished) {
|
if (!finished) {
|
||||||
writer.close();
|
writer.close();
|
||||||
|
|
|
@ -115,13 +115,13 @@ public class StripeCompactor extends Compactor {
|
||||||
smallestReadPoint = Math.min(fd.minSeqIdToKeep, smallestReadPoint);
|
smallestReadPoint = Math.min(fd.minSeqIdToKeep, smallestReadPoint);
|
||||||
cleanSeqId = true;
|
cleanSeqId = true;
|
||||||
}
|
}
|
||||||
final boolean needMvcc = fd.maxMVCCReadpoint >= smallestReadPoint;
|
|
||||||
final Compression.Algorithm compression = store.getFamily().getCompactionCompression();
|
final Compression.Algorithm compression = store.getFamily().getCompactionCompression();
|
||||||
StripeMultiFileWriter.WriterFactory factory = new StripeMultiFileWriter.WriterFactory() {
|
StripeMultiFileWriter.WriterFactory factory = new StripeMultiFileWriter.WriterFactory() {
|
||||||
@Override
|
@Override
|
||||||
public Writer createWriter() throws IOException {
|
public Writer createWriter() throws IOException {
|
||||||
return store.createWriterInTmp(
|
return store.createWriterInTmp(
|
||||||
fd.maxKeyCount, compression, true, needMvcc, fd.maxTagsLength > 0);
|
fd.maxKeyCount, compression, true, true, fd.maxTagsLength > 0);
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
@ -1906,34 +1906,6 @@ public class HLogSplitter {
|
||||||
public final long nonce;
|
public final long nonce;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Tag original sequence number for each edit to be replayed
|
|
||||||
* @param seqId
|
|
||||||
* @param cell
|
|
||||||
*/
|
|
||||||
private static Cell tagReplayLogSequenceNumber(long seqId, Cell cell) {
|
|
||||||
// Tag puts with original sequence number if there is no LOG_REPLAY_TAG yet
|
|
||||||
boolean needAddRecoveryTag = true;
|
|
||||||
if (cell.getTagsLength() > 0) {
|
|
||||||
Tag tmpTag = Tag.getTag(cell.getTagsArray(), cell.getTagsOffset(), cell.getTagsLength(),
|
|
||||||
TagType.LOG_REPLAY_TAG_TYPE);
|
|
||||||
if (tmpTag != null) {
|
|
||||||
// found an existing log replay tag so reuse it
|
|
||||||
needAddRecoveryTag = false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (needAddRecoveryTag) {
|
|
||||||
List<Tag> newTags = new ArrayList<Tag>();
|
|
||||||
Tag replayTag = new Tag(TagType.LOG_REPLAY_TAG_TYPE, Bytes.toBytes(seqId));
|
|
||||||
newTags.add(replayTag);
|
|
||||||
if (cell.getTagsLength() > 0) {
|
|
||||||
newTags.addAll(Tag.asList(cell.getTagsArray(), cell.getTagsOffset(), cell.getTagsLength()));
|
|
||||||
}
|
|
||||||
return new TagRewriteCell(cell, Tag.fromList(newTags));
|
|
||||||
}
|
|
||||||
return cell;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This function is used to construct mutations from a WALEntry. It also reconstructs HLogKey &
|
* This function is used to construct mutations from a WALEntry. It also reconstructs HLogKey &
|
||||||
* WALEdit from the passed in WALEntry
|
* WALEdit from the passed in WALEntry
|
||||||
|
@ -1941,12 +1913,11 @@ public class HLogSplitter {
|
||||||
* @param cells
|
* @param cells
|
||||||
* @param logEntry pair of HLogKey and WALEdit instance stores HLogKey and WALEdit instances
|
* @param logEntry pair of HLogKey and WALEdit instance stores HLogKey and WALEdit instances
|
||||||
* extracted from the passed in WALEntry.
|
* extracted from the passed in WALEntry.
|
||||||
* @param addLogReplayTag
|
|
||||||
* @return list of Pair<MutationType, Mutation> to be replayed
|
* @return list of Pair<MutationType, Mutation> to be replayed
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
*/
|
*/
|
||||||
public static List<MutationReplay> getMutationsFromWALEntry(WALEntry entry, CellScanner cells,
|
public static List<MutationReplay> getMutationsFromWALEntry(WALEntry entry, CellScanner cells,
|
||||||
Pair<HLogKey, WALEdit> logEntry, boolean addLogReplayTag) throws IOException {
|
Pair<HLogKey, WALEdit> logEntry) throws IOException {
|
||||||
|
|
||||||
if (entry == null) {
|
if (entry == null) {
|
||||||
// return an empty array
|
// return an empty array
|
||||||
|
@ -1993,11 +1964,7 @@ public class HLogSplitter {
|
||||||
if (CellUtil.isDelete(cell)) {
|
if (CellUtil.isDelete(cell)) {
|
||||||
((Delete) m).addDeleteMarker(cell);
|
((Delete) m).addDeleteMarker(cell);
|
||||||
} else {
|
} else {
|
||||||
Cell tmpNewCell = cell;
|
((Put) m).add(cell);
|
||||||
if (addLogReplayTag) {
|
|
||||||
tmpNewCell = tagReplayLogSequenceNumber(replaySeqId, cell);
|
|
||||||
}
|
|
||||||
((Put) m).add(tmpNewCell);
|
|
||||||
}
|
}
|
||||||
previousCell = cell;
|
previousCell = cell;
|
||||||
}
|
}
|
||||||
|
|
|
@ -1186,7 +1186,6 @@ public class TestDistributedLogSplitting {
|
||||||
LOG.info("testSameVersionUpdatesRecovery");
|
LOG.info("testSameVersionUpdatesRecovery");
|
||||||
conf.setLong("hbase.regionserver.hlog.blocksize", 15 * 1024);
|
conf.setLong("hbase.regionserver.hlog.blocksize", 15 * 1024);
|
||||||
conf.setBoolean(HConstants.DISTRIBUTED_LOG_REPLAY_KEY, true);
|
conf.setBoolean(HConstants.DISTRIBUTED_LOG_REPLAY_KEY, true);
|
||||||
conf.setInt("hfile.format.version", 3);
|
|
||||||
startCluster(NUM_RS);
|
startCluster(NUM_RS);
|
||||||
final AtomicLong sequenceId = new AtomicLong(100);
|
final AtomicLong sequenceId = new AtomicLong(100);
|
||||||
final int NUM_REGIONS_TO_CREATE = 40;
|
final int NUM_REGIONS_TO_CREATE = 40;
|
||||||
|
@ -1278,11 +1277,10 @@ public class TestDistributedLogSplitting {
|
||||||
conf.setBoolean(HConstants.DISTRIBUTED_LOG_REPLAY_KEY, true);
|
conf.setBoolean(HConstants.DISTRIBUTED_LOG_REPLAY_KEY, true);
|
||||||
conf.setInt(HConstants.HREGION_MEMSTORE_FLUSH_SIZE, 30 * 1024);
|
conf.setInt(HConstants.HREGION_MEMSTORE_FLUSH_SIZE, 30 * 1024);
|
||||||
conf.setInt("hbase.hstore.compactionThreshold", 3);
|
conf.setInt("hbase.hstore.compactionThreshold", 3);
|
||||||
conf.setInt("hfile.format.version", 3);
|
|
||||||
startCluster(NUM_RS);
|
startCluster(NUM_RS);
|
||||||
final AtomicLong sequenceId = new AtomicLong(100);
|
final AtomicLong sequenceId = new AtomicLong(100);
|
||||||
final int NUM_REGIONS_TO_CREATE = 40;
|
final int NUM_REGIONS_TO_CREATE = 40;
|
||||||
final int NUM_LOG_LINES = 1000;
|
final int NUM_LOG_LINES = 2000;
|
||||||
// turn off load balancing to prevent regions from moving around otherwise
|
// turn off load balancing to prevent regions from moving around otherwise
|
||||||
// they will consume recovered.edits
|
// they will consume recovered.edits
|
||||||
master.balanceSwitch(false);
|
master.balanceSwitch(false);
|
||||||
|
|
Loading…
Reference in New Issue