HBASE-2461 Split doesn't handle IOExceptions when creating new region reference files; forgot to svn add new files
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@982122 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
f7b8d9cf7a
commit
094197a180
|
@ -0,0 +1,41 @@
|
||||||
|
/**
|
||||||
|
* Copyright 2010 The Apache Software Foundation
|
||||||
|
*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hbase.regionserver;
|
||||||
|
|
||||||
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Add and remove online regions.
|
||||||
|
*/
|
||||||
|
interface OnlineRegions {
|
||||||
|
/**
|
||||||
|
* Add to online regions.
|
||||||
|
* @param r
|
||||||
|
*/
|
||||||
|
void addToOnlineRegions(final HRegion r);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* This method removes HRegion corresponding to hri from the Map of onlineRegions.
|
||||||
|
*
|
||||||
|
* @param hri the HRegionInfo corresponding to the HRegion to-be-removed.
|
||||||
|
* @return the removed HRegion, or null if the HRegion was not in onlineRegions.
|
||||||
|
*/
|
||||||
|
HRegion removeFromOnlineRegions(HRegionInfo hri);
|
||||||
|
}
|
|
@ -0,0 +1,548 @@
|
||||||
|
/**
|
||||||
|
* Copyright 2010 The Apache Software Foundation
|
||||||
|
*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hbase.regionserver;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.ListIterator;
|
||||||
|
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
|
import org.apache.hadoop.conf.Configuration;
|
||||||
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
|
import org.apache.hadoop.hbase.KeyValue.KVComparator;
|
||||||
|
import org.apache.hadoop.hbase.client.HTable;
|
||||||
|
import org.apache.hadoop.hbase.client.Put;
|
||||||
|
import org.apache.hadoop.hbase.io.Reference.Range;
|
||||||
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
||||||
|
import org.apache.hadoop.hbase.util.FSUtils;
|
||||||
|
import org.apache.hadoop.hbase.util.PairOfSameType;
|
||||||
|
import org.apache.hadoop.hbase.util.Writables;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Executes region split as a "transaction". Call {@link #prepare()} to setup
|
||||||
|
* the transaction, {@link #execute(OnlineRegions)} to run the transaction and
|
||||||
|
* {@link #rollback(OnlineRegions)} to cleanup if execute fails.
|
||||||
|
*
|
||||||
|
* <p>Here is an example of how you would use this class:
|
||||||
|
* <pre>
|
||||||
|
* SplitTransaction st = new SplitTransaction(this.conf, parent, midKey)
|
||||||
|
* if (!st.prepare()) return;
|
||||||
|
* try {
|
||||||
|
* st.execute(myOnlineRegions);
|
||||||
|
* } catch (IOException ioe) {
|
||||||
|
* try {
|
||||||
|
* st.rollback(myOnlineRegions);
|
||||||
|
* return;
|
||||||
|
* } catch (RuntimeException e) {
|
||||||
|
* myAbortable.abort("Failed split, abort");
|
||||||
|
* }
|
||||||
|
* }
|
||||||
|
* </Pre>
|
||||||
|
*/
|
||||||
|
class SplitTransaction {
|
||||||
|
private static final Log LOG = LogFactory.getLog(SplitTransaction.class);
|
||||||
|
private static final String SPLITDIR = "splits";
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Region to split
|
||||||
|
*/
|
||||||
|
private final HRegion parent;
|
||||||
|
private HRegionInfo hri_a;
|
||||||
|
private HRegionInfo hri_b;
|
||||||
|
private Path splitdir;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Row to split around
|
||||||
|
*/
|
||||||
|
private final byte [] splitrow;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Types to add to the transaction journal
|
||||||
|
*/
|
||||||
|
enum JournalEntry {
|
||||||
|
/**
|
||||||
|
* We created the temporary split data directory.
|
||||||
|
*/
|
||||||
|
CREATE_SPLIT_DIR,
|
||||||
|
/**
|
||||||
|
* Closed the parent region.
|
||||||
|
*/
|
||||||
|
CLOSED_PARENT_REGION,
|
||||||
|
/**
|
||||||
|
* The parent has been taken out of the server's online regions list.
|
||||||
|
*/
|
||||||
|
OFFLINED_PARENT,
|
||||||
|
/**
|
||||||
|
* Started in on creation of the first daughter region.
|
||||||
|
*/
|
||||||
|
STARTED_REGION_A_CREATION,
|
||||||
|
/**
|
||||||
|
* Started in on the creation of the second daughter region.
|
||||||
|
*/
|
||||||
|
STARTED_REGION_B_CREATION
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Journal of how far the split transaction has progressed.
|
||||||
|
*/
|
||||||
|
private final List<JournalEntry> journal = new ArrayList<JournalEntry>();
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructor
|
||||||
|
* @param c Configuration to use running split
|
||||||
|
* @param r Region to split
|
||||||
|
* @param splitrow Row to split around
|
||||||
|
*/
|
||||||
|
SplitTransaction(final HRegion r, final byte [] splitrow) {
|
||||||
|
this.parent = r;
|
||||||
|
this.splitrow = splitrow;
|
||||||
|
this.splitdir = getSplitDir(this.parent);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Does checks on split inputs.
|
||||||
|
* @return <code>true</code> if the region is splittable else
|
||||||
|
* <code>false</code> if it is not (e.g. its already closed, etc.). If we
|
||||||
|
* return <code>true</code>, we'll have taken out the parent's
|
||||||
|
* <code>splitsAndClosesLock</code> and only way to unlock is successful
|
||||||
|
* {@link #execute(OnlineRegions)} or {@link #rollback(OnlineRegions)}
|
||||||
|
*/
|
||||||
|
public boolean prepare() {
|
||||||
|
boolean prepared = false;
|
||||||
|
this.parent.splitsAndClosesLock.writeLock().lock();
|
||||||
|
try {
|
||||||
|
if (this.parent.isClosed() || this.parent.isClosing()) return prepared;
|
||||||
|
HRegionInfo hri = this.parent.getRegionInfo();
|
||||||
|
// Check splitrow.
|
||||||
|
byte [] startKey = hri.getStartKey();
|
||||||
|
byte [] endKey = hri.getEndKey();
|
||||||
|
KVComparator comparator = this.parent.comparator;
|
||||||
|
if (comparator.compareRows(startKey, 0, startKey.length,
|
||||||
|
this.splitrow, 0, this.splitrow.length) >= 0) {
|
||||||
|
LOG.info("Split row is <= start key row, not splitting: " +
|
||||||
|
Bytes.toString(this.splitrow));
|
||||||
|
return prepared;
|
||||||
|
}
|
||||||
|
if (comparator.compareRows(this.splitrow, 0, this.splitrow.length,
|
||||||
|
endKey, 0, endKey.length) >= 0) {
|
||||||
|
LOG.info("Split row is >= end key row, not splitting: " +
|
||||||
|
Bytes.toString(this.splitrow));
|
||||||
|
return prepared;
|
||||||
|
}
|
||||||
|
long rid = getDaughterRegionIdTimestamp(hri);
|
||||||
|
this.hri_a = new HRegionInfo(hri.getTableDesc(), startKey, this.splitrow,
|
||||||
|
false, rid);
|
||||||
|
this.hri_b = new HRegionInfo(hri.getTableDesc(), this.splitrow, endKey,
|
||||||
|
false, rid);
|
||||||
|
prepared = true;
|
||||||
|
} finally {
|
||||||
|
if (!prepared) this.parent.splitsAndClosesLock.writeLock().unlock();
|
||||||
|
}
|
||||||
|
return prepared;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Calculate daughter regionid to use.
|
||||||
|
* @param hri Parent {@link HRegionInfo}
|
||||||
|
* @return Daughter region id (timestamp) to use.
|
||||||
|
*/
|
||||||
|
private static long getDaughterRegionIdTimestamp(final HRegionInfo hri) {
|
||||||
|
long rid = EnvironmentEdgeManager.currentTimeMillis();
|
||||||
|
// Regionid is timestamp. Can't be less than that of parent else will insert
|
||||||
|
// at wrong location in .META. (See HBASE-710).
|
||||||
|
if (rid < hri.getRegionId()) {
|
||||||
|
LOG.warn("Clock skew; parent regions id is " + hri.getRegionId() +
|
||||||
|
" but current time here is " + rid);
|
||||||
|
rid = hri.getRegionId() + 1;
|
||||||
|
}
|
||||||
|
return rid;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Run the transaction.
|
||||||
|
* @param or Object that can online/offline parent region.
|
||||||
|
* @throws IOException If thrown, transaction failed. Call {@link #rollback(OnlineRegions)}
|
||||||
|
* @return Regions created
|
||||||
|
* @see #rollback(OnlineRegions)
|
||||||
|
*/
|
||||||
|
public PairOfSameType<HRegion> execute(final OnlineRegions or) throws IOException {
|
||||||
|
return execute(or, or != null);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Run the transaction.
|
||||||
|
* @param or Object that can online/offline parent region. Can be null (Tests
|
||||||
|
* will pass null).
|
||||||
|
* @param If <code>true</code>, update meta (set to false when testing).
|
||||||
|
* @throws IOException If thrown, transaction failed. Call {@link #rollback(OnlineRegions)}
|
||||||
|
* @return Regions created
|
||||||
|
* @see #rollback(OnlineRegions)
|
||||||
|
*/
|
||||||
|
PairOfSameType<HRegion> execute(final OnlineRegions or, final boolean updateMeta)
|
||||||
|
throws IOException {
|
||||||
|
LOG.info("Starting split of region " + this.parent);
|
||||||
|
if (!this.parent.splitsAndClosesLock.writeLock().isHeldByCurrentThread()) {
|
||||||
|
throw new SplitAndCloseWriteLockNotHeld();
|
||||||
|
}
|
||||||
|
|
||||||
|
// We'll need one of these later but get it now because if we fail there
|
||||||
|
// is nothing to undo.
|
||||||
|
HTable t = null;
|
||||||
|
if (updateMeta) t = getTable(this.parent.getConf());
|
||||||
|
|
||||||
|
createSplitDir(this.parent.getFilesystem(), this.splitdir);
|
||||||
|
this.journal.add(JournalEntry.CREATE_SPLIT_DIR);
|
||||||
|
|
||||||
|
List<StoreFile> hstoreFilesToSplit = this.parent.close(false);
|
||||||
|
this.journal.add(JournalEntry.CLOSED_PARENT_REGION);
|
||||||
|
|
||||||
|
if (or != null) or.removeFromOnlineRegions(this.parent.getRegionInfo());
|
||||||
|
this.journal.add(JournalEntry.OFFLINED_PARENT);
|
||||||
|
|
||||||
|
splitStoreFiles(this.splitdir, hstoreFilesToSplit);
|
||||||
|
// splitStoreFiles creates daughter region dirs under the parent splits dir
|
||||||
|
// Nothing to unroll here if failure -- clean up of CREATE_SPLIT_DIR will
|
||||||
|
// clean this up.
|
||||||
|
|
||||||
|
// Log to the journal that we are creating region A, the first daughter
|
||||||
|
// region. We could fail halfway through. If we do, we could have left
|
||||||
|
// stuff in fs that needs cleanup -- a storefile or two. Thats why we
|
||||||
|
// add entry to journal BEFORE rather than AFTER the change.
|
||||||
|
this.journal.add(JournalEntry.STARTED_REGION_A_CREATION);
|
||||||
|
HRegion a = createDaughterRegion(this.hri_a);
|
||||||
|
|
||||||
|
// Ditto
|
||||||
|
this.journal.add(JournalEntry.STARTED_REGION_B_CREATION);
|
||||||
|
HRegion b = createDaughterRegion(this.hri_b);
|
||||||
|
|
||||||
|
Put editParentPut = createOfflineParentPut();
|
||||||
|
if (t != null) t.put(editParentPut);
|
||||||
|
|
||||||
|
// The is the point of no return. We are committed to the split now. Up to
|
||||||
|
// a failure editing parent in meta or a crash of the hosting regionserver,
|
||||||
|
// we could rollback (or, if crash, we could cleanup on redeploy) but now
|
||||||
|
// meta has been changed, we can only go forward. If the below last steps
|
||||||
|
// do not complete, repair has to be done by another agent. For example,
|
||||||
|
// basescanner, at least up till master rewrite, would add daughter rows if
|
||||||
|
// missing from meta. It could do this because the parent edit includes the
|
||||||
|
// daughter specs. In Bigtable paper, they have another mechanism where
|
||||||
|
// some feedback to the master somehow flags it that split is incomplete and
|
||||||
|
// needs fixup. Whatever the mechanism, its a TODO that we have some fixup.
|
||||||
|
|
||||||
|
// I looked at writing the put of the parent edit above out to the WAL log
|
||||||
|
// before changing meta with the notion that should we fail, then on replay
|
||||||
|
// the offlining of the parent and addition of daughters up into meta could
|
||||||
|
// be reinserted. The edits would have to be 'special' and given how our
|
||||||
|
// splits work, splitting by region, I think the replay would have to happen
|
||||||
|
// inside in the split code -- as soon as it saw one of these special edits,
|
||||||
|
// rather than write the edit out a file for the .META. region to replay or
|
||||||
|
// somehow, write it out to this regions edits file for it to handle on
|
||||||
|
// redeploy -- this'd be whacky, we'd be telling meta about a split during
|
||||||
|
// the deploy of the parent -- instead we'd have to play the edit inside
|
||||||
|
// in the split code somehow; this would involve a stop-the-splitting till
|
||||||
|
// meta had been edited which might hold up splitting a good while.
|
||||||
|
|
||||||
|
// Finish up the meta edits. If these fail, another agent needs to do fixup
|
||||||
|
HRegionInfo hri = this.hri_a;
|
||||||
|
try {
|
||||||
|
if (t != null) t.put(createDaughterPut(hri));
|
||||||
|
hri = this.hri_b;
|
||||||
|
if (t != null) t.put(createDaughterPut(hri));
|
||||||
|
} catch (IOException e) {
|
||||||
|
// Don't let this out or we'll run rollback.
|
||||||
|
LOG.warn("Failed adding daughter " + hri.toString());
|
||||||
|
}
|
||||||
|
// This should not fail because the HTable instance we are using is not
|
||||||
|
// running a buffer -- its immediately flushing its puts.
|
||||||
|
if (t != null) t.close();
|
||||||
|
|
||||||
|
// Unlock if successful split.
|
||||||
|
this.parent.splitsAndClosesLock.writeLock().unlock();
|
||||||
|
|
||||||
|
// Leaving here, the splitdir with its dross will be in place but since the
|
||||||
|
// split was successful, just leave it; it'll be cleaned when parent is
|
||||||
|
// deleted and cleaned up.
|
||||||
|
return new PairOfSameType<HRegion>(a, b);
|
||||||
|
}
|
||||||
|
|
||||||
|
private static Path getSplitDir(final HRegion r) {
|
||||||
|
return new Path(r.getRegionDir(), SPLITDIR);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param fs Filesystem to use
|
||||||
|
* @param splitdir Directory to store temporary split data in
|
||||||
|
* @throws IOException If <code>splitdir</code> already exists or we fail
|
||||||
|
* to create it.
|
||||||
|
* @see #cleanupSplitDir(FileSystem, Path)
|
||||||
|
*/
|
||||||
|
private static void createSplitDir(final FileSystem fs, final Path splitdir)
|
||||||
|
throws IOException {
|
||||||
|
if (fs.exists(splitdir)) throw new IOException("Splitdir already exits? " + splitdir);
|
||||||
|
if (!fs.mkdirs(splitdir)) throw new IOException("Failed create of " + splitdir);
|
||||||
|
}
|
||||||
|
|
||||||
|
private static void cleanupSplitDir(final FileSystem fs, final Path splitdir)
|
||||||
|
throws IOException {
|
||||||
|
// Splitdir may have been cleaned up by reopen of the parent dir.
|
||||||
|
deleteDir(fs, splitdir, false);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param fs Filesystem to use
|
||||||
|
* @param dir Directory to delete
|
||||||
|
* @param mustPreExist If true, we'll throw exception if <code>dir</code>
|
||||||
|
* does not preexist, else we'll just pass.
|
||||||
|
* @throws IOException Thrown if we fail to delete passed <code>dir</code>
|
||||||
|
*/
|
||||||
|
private static void deleteDir(final FileSystem fs, final Path dir,
|
||||||
|
final boolean mustPreExist)
|
||||||
|
throws IOException {
|
||||||
|
if (!fs.exists(dir)) {
|
||||||
|
if (mustPreExist) throw new IOException(dir.toString() + " does not exist!");
|
||||||
|
} else if (!fs.delete(dir, true)) {
|
||||||
|
throw new IOException("Failed delete of " + dir);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void splitStoreFiles(final Path splitdir,
|
||||||
|
final List<StoreFile> hstoreFilesToSplit)
|
||||||
|
throws IOException {
|
||||||
|
if (hstoreFilesToSplit == null) {
|
||||||
|
// Could be null because close didn't succeed -- for now consider it fatal
|
||||||
|
throw new IOException("Close returned empty list of StoreFiles");
|
||||||
|
}
|
||||||
|
|
||||||
|
// Split each store file.
|
||||||
|
for (StoreFile sf: hstoreFilesToSplit) {
|
||||||
|
splitStoreFile(sf, splitdir);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
private void splitStoreFile(final StoreFile sf, final Path splitdir)
|
||||||
|
throws IOException {
|
||||||
|
FileSystem fs = this.parent.getFilesystem();
|
||||||
|
byte [] family = sf.getFamily();
|
||||||
|
String encoded = this.hri_a.getEncodedName();
|
||||||
|
Path storedir = Store.getStoreHomedir(splitdir, encoded, family);
|
||||||
|
StoreFile.split(fs, storedir, sf, this.splitrow, Range.bottom);
|
||||||
|
encoded = this.hri_b.getEncodedName();
|
||||||
|
storedir = Store.getStoreHomedir(splitdir, encoded, family);
|
||||||
|
StoreFile.split(fs, storedir, sf, this.splitrow, Range.top);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param hri
|
||||||
|
* @return Created daughter HRegion.
|
||||||
|
* @throws IOException
|
||||||
|
* @see #cleanupDaughterRegion(FileSystem, Path, HRegionInfo)
|
||||||
|
*/
|
||||||
|
HRegion createDaughterRegion(final HRegionInfo hri)
|
||||||
|
throws IOException {
|
||||||
|
// Package private so unit tests have access.
|
||||||
|
FileSystem fs = this.parent.getFilesystem();
|
||||||
|
Path regionDir = getSplitDirForDaughter(this.parent.getFilesystem(),
|
||||||
|
this.splitdir, hri);
|
||||||
|
HRegion r = HRegion.newHRegion(this.parent.getTableDir(),
|
||||||
|
this.parent.getLog(), fs, this.parent.getConf(),
|
||||||
|
hri, null);
|
||||||
|
HRegion.moveInitialFilesIntoPlace(fs, regionDir, r.getRegionDir());
|
||||||
|
return r;
|
||||||
|
}
|
||||||
|
|
||||||
|
private static void cleanupDaughterRegion(final FileSystem fs,
|
||||||
|
final Path tabledir, final String encodedName)
|
||||||
|
throws IOException {
|
||||||
|
Path regiondir = HRegion.getRegionDir(tabledir, encodedName);
|
||||||
|
// Dir may not preexist.
|
||||||
|
deleteDir(fs, regiondir, false);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Get the daughter directories in the splits dir. The splits dir is under
|
||||||
|
* the parent regions' directory.
|
||||||
|
* @param fs
|
||||||
|
* @param splitdir
|
||||||
|
* @param hri
|
||||||
|
* @return Path to daughter split dir.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
private static Path getSplitDirForDaughter(final FileSystem fs,
|
||||||
|
final Path splitdir, final HRegionInfo hri)
|
||||||
|
throws IOException {
|
||||||
|
return new Path(splitdir, hri.getEncodedName());
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* @param r Parent region we want to edit.
|
||||||
|
* @return An HTable instance against the meta table that holds passed
|
||||||
|
* <code>r</code>; it has autoFlush enabled so we immediately send puts (No
|
||||||
|
* buffering enabled).
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
private HTable getTable(final Configuration conf) throws IOException {
|
||||||
|
// When a region is split, the META table needs to updated if we're
|
||||||
|
// splitting a 'normal' region, and the ROOT table needs to be
|
||||||
|
// updated if we are splitting a META region.
|
||||||
|
HTable t = null;
|
||||||
|
if (this.parent.getRegionInfo().isMetaTable()) {
|
||||||
|
t = new HTable(conf, HConstants.ROOT_TABLE_NAME);
|
||||||
|
} else {
|
||||||
|
t = new HTable(conf, HConstants.META_TABLE_NAME);
|
||||||
|
}
|
||||||
|
// Flush puts as we send them -- no buffering.
|
||||||
|
t.setAutoFlush(true);
|
||||||
|
return t;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
private Put createOfflineParentPut() throws IOException {
|
||||||
|
HRegionInfo editedParentRegionInfo =
|
||||||
|
new HRegionInfo(this.parent.getRegionInfo());
|
||||||
|
editedParentRegionInfo.setOffline(true);
|
||||||
|
editedParentRegionInfo.setSplit(true);
|
||||||
|
Put put = new Put(editedParentRegionInfo.getRegionName());
|
||||||
|
put.add(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER,
|
||||||
|
Writables.getBytes(editedParentRegionInfo));
|
||||||
|
put.add(HConstants.CATALOG_FAMILY, HConstants.SERVER_QUALIFIER,
|
||||||
|
HConstants.EMPTY_BYTE_ARRAY);
|
||||||
|
put.add(HConstants.CATALOG_FAMILY, HConstants.STARTCODE_QUALIFIER,
|
||||||
|
HConstants.EMPTY_BYTE_ARRAY);
|
||||||
|
put.add(HConstants.CATALOG_FAMILY, HConstants.SPLITA_QUALIFIER,
|
||||||
|
Writables.getBytes(this.hri_a));
|
||||||
|
put.add(HConstants.CATALOG_FAMILY, HConstants.SPLITB_QUALIFIER,
|
||||||
|
Writables.getBytes(this.hri_b));
|
||||||
|
return put;
|
||||||
|
}
|
||||||
|
|
||||||
|
private Put createDaughterPut(final HRegionInfo daughter)
|
||||||
|
throws IOException {
|
||||||
|
Put p = new Put(daughter.getRegionName());
|
||||||
|
p.add(HConstants.CATALOG_FAMILY, HConstants.REGIONINFO_QUALIFIER,
|
||||||
|
Writables.getBytes(daughter));
|
||||||
|
return p;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param or Object that can online/offline parent region. Can be passed null
|
||||||
|
* by unit tests.
|
||||||
|
* @return The region we were splitting
|
||||||
|
* @throws IOException If thrown, rollback failed. Take drastic action.
|
||||||
|
*/
|
||||||
|
public void rollback(final OnlineRegions or) throws IOException {
|
||||||
|
if (!this.parent.splitsAndClosesLock.writeLock().isHeldByCurrentThread()) {
|
||||||
|
throw new SplitAndCloseWriteLockNotHeld();
|
||||||
|
}
|
||||||
|
FileSystem fs = this.parent.getFilesystem();
|
||||||
|
ListIterator<JournalEntry> iterator =
|
||||||
|
this.journal.listIterator(this.journal.size());
|
||||||
|
while (iterator.hasPrevious()) {
|
||||||
|
JournalEntry je = iterator.previous();
|
||||||
|
switch(je) {
|
||||||
|
case CREATE_SPLIT_DIR:
|
||||||
|
cleanupSplitDir(fs, this.splitdir);
|
||||||
|
break;
|
||||||
|
|
||||||
|
case CLOSED_PARENT_REGION:
|
||||||
|
// So, this returns a seqid but if we just closed and then reopened, we
|
||||||
|
// should be ok. On close, we flushed using sequenceid obtained from
|
||||||
|
// hosting regionserver so no need to propagate the sequenceid returned
|
||||||
|
// out of initialize below up into regionserver as we normally do.
|
||||||
|
// TODO: Verify.
|
||||||
|
this.parent.initialize();
|
||||||
|
break;
|
||||||
|
|
||||||
|
case STARTED_REGION_A_CREATION:
|
||||||
|
cleanupDaughterRegion(fs, this.parent.getTableDir(),
|
||||||
|
this.hri_a.getEncodedName());
|
||||||
|
break;
|
||||||
|
|
||||||
|
case STARTED_REGION_B_CREATION:
|
||||||
|
cleanupDaughterRegion(fs, this.parent.getTableDir(),
|
||||||
|
this.hri_b.getEncodedName());
|
||||||
|
break;
|
||||||
|
|
||||||
|
case OFFLINED_PARENT:
|
||||||
|
if (or != null) or.addToOnlineRegions(this.parent);
|
||||||
|
break;
|
||||||
|
|
||||||
|
default:
|
||||||
|
throw new RuntimeException("Unhandled journal entry: " + je);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (this.parent.splitsAndClosesLock.writeLock().isHeldByCurrentThread()) {
|
||||||
|
this.parent.splitsAndClosesLock.writeLock().unlock();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Thrown if lock not held.
|
||||||
|
*/
|
||||||
|
@SuppressWarnings("serial")
|
||||||
|
public class SplitAndCloseWriteLockNotHeld extends IOException {}
|
||||||
|
|
||||||
|
HRegionInfo getFirstDaughter() {
|
||||||
|
return hri_a;
|
||||||
|
}
|
||||||
|
|
||||||
|
HRegionInfo getSecondDaughter() {
|
||||||
|
return hri_b;
|
||||||
|
}
|
||||||
|
|
||||||
|
// For unit testing.
|
||||||
|
Path getSplitDir() {
|
||||||
|
return this.splitdir;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Clean up any split detritus that may have been left around from previous
|
||||||
|
* split attempts.
|
||||||
|
* Call this method on initial region deploy. Cleans up any mess
|
||||||
|
* left by previous deploys of passed <code>r</code> region.
|
||||||
|
* @param r
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
static void cleanupAnySplitDetritus(final HRegion r) throws IOException {
|
||||||
|
Path splitdir = getSplitDir(r);
|
||||||
|
FileSystem fs = r.getFilesystem();
|
||||||
|
if (!fs.exists(splitdir)) return;
|
||||||
|
// Look at the splitdir. It could have the encoded names of the daughter
|
||||||
|
// regions we tried to make. See if the daughter regions actually got made
|
||||||
|
// out under the tabledir. If here under splitdir still, then the split did
|
||||||
|
// not complete. Try and do cleanup. This code WILL NOT catch the case
|
||||||
|
// where we successfully created daughter a but regionserver crashed during
|
||||||
|
// the creation of region b. In this case, there'll be an orphan daughter
|
||||||
|
// dir in the filesystem. TOOD: Fix.
|
||||||
|
FileStatus [] daughters = fs.listStatus(splitdir, new FSUtils.DirFilter(fs));
|
||||||
|
for (int i = 0; i < daughters.length; i++) {
|
||||||
|
cleanupDaughterRegion(fs, r.getTableDir(),
|
||||||
|
daughters[i].getPath().getName());
|
||||||
|
}
|
||||||
|
cleanupSplitDir(r.getFilesystem(), splitdir);
|
||||||
|
LOG.info("Cleaned up old failed split transaction detritus: " + splitdir);
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,112 @@
|
||||||
|
/**
|
||||||
|
* Copyright 2010 The Apache Software Foundation
|
||||||
|
*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.hbase.util;
|
||||||
|
|
||||||
|
import java.util.Iterator;
|
||||||
|
|
||||||
|
import org.apache.commons.lang.NotImplementedException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* A generic, immutable class for pairs of objects both of type <code>T</code>.
|
||||||
|
* @param <T>
|
||||||
|
* @See {@link Pair} if Types differ.
|
||||||
|
*/
|
||||||
|
public class PairOfSameType<T> implements Iterable<T> {
|
||||||
|
private final T first;
|
||||||
|
private final T second;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Constructor
|
||||||
|
* @param a operand
|
||||||
|
* @param b operand
|
||||||
|
*/
|
||||||
|
public PairOfSameType(T a, T b) {
|
||||||
|
this.first = a;
|
||||||
|
this.second = b;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Return the first element stored in the pair.
|
||||||
|
* @return T
|
||||||
|
*/
|
||||||
|
public T getFirst() {
|
||||||
|
return first;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Return the second element stored in the pair.
|
||||||
|
* @return T
|
||||||
|
*/
|
||||||
|
public T getSecond() {
|
||||||
|
return second;
|
||||||
|
}
|
||||||
|
|
||||||
|
private static boolean equals(Object x, Object y) {
|
||||||
|
return (x == null && y == null) || (x != null && x.equals(y));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
@SuppressWarnings("unchecked")
|
||||||
|
public boolean equals(Object other) {
|
||||||
|
return other instanceof PairOfSameType &&
|
||||||
|
equals(first, ((PairOfSameType)other).first) &&
|
||||||
|
equals(second, ((PairOfSameType)other).second);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int hashCode() {
|
||||||
|
if (first == null)
|
||||||
|
return (second == null) ? 0 : second.hashCode() + 1;
|
||||||
|
else if (second == null)
|
||||||
|
return first.hashCode() + 2;
|
||||||
|
else
|
||||||
|
return first.hashCode() * 17 + second.hashCode();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public String toString() {
|
||||||
|
return "{" + getFirst() + "," + getSecond() + "}";
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Iterator<T> iterator() {
|
||||||
|
return new Iterator<T>() {
|
||||||
|
private int returned = 0;
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public boolean hasNext() {
|
||||||
|
return this.returned < 2;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public T next() {
|
||||||
|
if (++this.returned == 1) return getFirst();
|
||||||
|
else if (this.returned == 2) return getSecond();
|
||||||
|
else throw new IllegalAccessError("this.returned=" + this.returned);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void remove() {
|
||||||
|
throw new NotImplementedException();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,253 @@
|
||||||
|
/**
|
||||||
|
* Copyright 2010 The Apache Software Foundation
|
||||||
|
*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hbase.regionserver;
|
||||||
|
|
||||||
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertFalse;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
|
import static org.mockito.Mockito.spy;
|
||||||
|
import static org.mockito.Mockito.when;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.List;
|
||||||
|
|
||||||
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
|
import org.apache.hadoop.hbase.HColumnDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
|
import org.apache.hadoop.hbase.HRegionInfo;
|
||||||
|
import org.apache.hadoop.hbase.HTableDescriptor;
|
||||||
|
import org.apache.hadoop.hbase.KeyValue;
|
||||||
|
import org.apache.hadoop.hbase.client.Scan;
|
||||||
|
import org.apache.hadoop.hbase.regionserver.wal.HLog;
|
||||||
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
import org.apache.hadoop.hbase.util.PairOfSameType;
|
||||||
|
import org.junit.After;
|
||||||
|
import org.junit.Before;
|
||||||
|
import org.junit.Test;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test the {@link SplitTransaction} class against an HRegion (as opposed to
|
||||||
|
* running cluster).
|
||||||
|
*/
|
||||||
|
public class TestSplitTransaction {
|
||||||
|
private static final HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility();
|
||||||
|
private final Path testdir =
|
||||||
|
HBaseTestingUtility.getTestDir(this.getClass().getName());
|
||||||
|
private HRegion parent;
|
||||||
|
private HLog wal;
|
||||||
|
private FileSystem fs;
|
||||||
|
private static final byte [] STARTROW = new byte [] {'a', 'a', 'a'};
|
||||||
|
// '{' is next ascii after 'z'.
|
||||||
|
private static final byte [] ENDROW = new byte [] {'{', '{', '{'};
|
||||||
|
private static final byte [] GOOD_SPLIT_ROW = new byte [] {'d', 'd', 'd'};
|
||||||
|
private static final byte [] CF = HConstants.CATALOG_FAMILY;
|
||||||
|
|
||||||
|
@Before public void setup() throws IOException {
|
||||||
|
this.fs = FileSystem.get(TEST_UTIL.getConfiguration());
|
||||||
|
this.fs.delete(this.testdir, true);
|
||||||
|
this.wal = new HLog(fs, new Path(this.testdir, "logs"),
|
||||||
|
new Path(this.testdir, "archive"),
|
||||||
|
TEST_UTIL.getConfiguration(), null);
|
||||||
|
this.parent = createRegion(this.testdir, this.wal);
|
||||||
|
}
|
||||||
|
|
||||||
|
@After public void teardown() throws IOException {
|
||||||
|
if (this.parent != null && !this.parent.isClosed()) this.parent.close();
|
||||||
|
if (this.fs.exists(this.parent.getRegionDir()) &&
|
||||||
|
!this.fs.delete(this.parent.getRegionDir(), true)) {
|
||||||
|
throw new IOException("Failed delete of " + this.parent.getRegionDir());
|
||||||
|
}
|
||||||
|
if (this.wal != null) this.wal.closeAndDelete();
|
||||||
|
this.fs.delete(this.testdir, true);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test straight prepare works. Tries to split on {@link #GOOD_SPLIT_ROW}
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
@Test public void testPrepare() throws IOException {
|
||||||
|
prepareGOOD_SPLIT_ROW();
|
||||||
|
}
|
||||||
|
|
||||||
|
private SplitTransaction prepareGOOD_SPLIT_ROW() {
|
||||||
|
SplitTransaction st = new SplitTransaction(this.parent, GOOD_SPLIT_ROW);
|
||||||
|
assertTrue(st.prepare());
|
||||||
|
// Assert the write lock is held on successful prepare as the javadoc asserts.
|
||||||
|
assertTrue(this.parent.splitsAndClosesLock.writeLock().isHeldByCurrentThread());
|
||||||
|
return st;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Pass an unreasonable split row.
|
||||||
|
*/
|
||||||
|
@Test public void testPrepareWithBadSplitRow() throws IOException {
|
||||||
|
// Pass start row as split key.
|
||||||
|
SplitTransaction st = new SplitTransaction(this.parent, STARTROW);
|
||||||
|
assertFalse(st.prepare());
|
||||||
|
st = new SplitTransaction(this.parent, HConstants.EMPTY_BYTE_ARRAY);
|
||||||
|
assertFalse(st.prepare());
|
||||||
|
st = new SplitTransaction(this.parent, new byte [] {'A', 'A', 'A'});
|
||||||
|
assertFalse(st.prepare());
|
||||||
|
st = new SplitTransaction(this.parent, ENDROW);
|
||||||
|
assertFalse(st.prepare());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test public void testPrepareWithClosedRegion() throws IOException {
|
||||||
|
this.parent.close();
|
||||||
|
SplitTransaction st = new SplitTransaction(this.parent, GOOD_SPLIT_ROW);
|
||||||
|
assertFalse(st.prepare());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test public void testWholesomeSplit() throws IOException {
|
||||||
|
final int rowcount = TEST_UTIL.loadRegion(this.parent, CF);
|
||||||
|
assertTrue(rowcount > 0);
|
||||||
|
int parentRowCount = countRows(this.parent);
|
||||||
|
assertEquals(rowcount, parentRowCount);
|
||||||
|
|
||||||
|
// Start transaction.
|
||||||
|
SplitTransaction st = prepareGOOD_SPLIT_ROW();
|
||||||
|
|
||||||
|
// Run the execute. Look at what it returns.
|
||||||
|
PairOfSameType<HRegion> daughters = st.execute(null);
|
||||||
|
// Do some assertions about execution.
|
||||||
|
assertTrue(this.fs.exists(st.getSplitDir()));
|
||||||
|
// Assert the parent region is closed.
|
||||||
|
assertTrue(this.parent.isClosed());
|
||||||
|
|
||||||
|
// Assert splitdir is empty -- because its content will have been moved out
|
||||||
|
// to be under the daughter region dirs.
|
||||||
|
assertEquals(0, this.fs.listStatus(st.getSplitDir()).length);
|
||||||
|
// Check daughters have correct key span.
|
||||||
|
assertTrue(Bytes.equals(this.parent.getStartKey(),
|
||||||
|
daughters.getFirst().getStartKey()));
|
||||||
|
assertTrue(Bytes.equals(GOOD_SPLIT_ROW,
|
||||||
|
daughters.getFirst().getEndKey()));
|
||||||
|
assertTrue(Bytes.equals(daughters.getSecond().getStartKey(),
|
||||||
|
GOOD_SPLIT_ROW));
|
||||||
|
assertTrue(Bytes.equals(this.parent.getEndKey(),
|
||||||
|
daughters.getSecond().getEndKey()));
|
||||||
|
// Count rows.
|
||||||
|
int daughtersRowCount = 0;
|
||||||
|
for (HRegion r: daughters) {
|
||||||
|
// Open so can count its content.
|
||||||
|
HRegion openRegion = HRegion.openHRegion(r.getRegionInfo(), this.testdir,
|
||||||
|
r.getLog(), r.getConf());
|
||||||
|
try {
|
||||||
|
int count = countRows(openRegion);
|
||||||
|
assertTrue(count > 0 && count != rowcount);
|
||||||
|
daughtersRowCount += count;
|
||||||
|
} finally {
|
||||||
|
openRegion.close();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
assertEquals(rowcount, daughtersRowCount);
|
||||||
|
// Assert the write lock is no longer held on parent
|
||||||
|
assertTrue(!this.parent.splitsAndClosesLock.writeLock().isHeldByCurrentThread());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test public void testRollback() throws IOException {
|
||||||
|
final int rowcount = TEST_UTIL.loadRegion(this.parent, CF);
|
||||||
|
assertTrue(rowcount > 0);
|
||||||
|
int parentRowCount = countRows(this.parent);
|
||||||
|
assertEquals(rowcount, parentRowCount);
|
||||||
|
|
||||||
|
// Start transaction.
|
||||||
|
SplitTransaction st = prepareGOOD_SPLIT_ROW();
|
||||||
|
SplitTransaction spiedUponSt = spy(st);
|
||||||
|
when(spiedUponSt.createDaughterRegion(spiedUponSt.getSecondDaughter())).
|
||||||
|
thenThrow(new MockedFailedDaughterCreation());
|
||||||
|
// Run the execute. Look at what it returns.
|
||||||
|
boolean expectedException = false;
|
||||||
|
try {
|
||||||
|
spiedUponSt.execute(null);
|
||||||
|
} catch (MockedFailedDaughterCreation e) {
|
||||||
|
expectedException = true;
|
||||||
|
}
|
||||||
|
assertTrue(expectedException);
|
||||||
|
// Run rollback
|
||||||
|
spiedUponSt.rollback(null);
|
||||||
|
|
||||||
|
// Assert I can scan parent.
|
||||||
|
int parentRowCount2 = countRows(this.parent);
|
||||||
|
assertEquals(parentRowCount, parentRowCount2);
|
||||||
|
|
||||||
|
// Assert rollback cleaned up stuff in fs
|
||||||
|
assertTrue(!this.fs.exists(HRegion.getRegionDir(this.testdir, st.getFirstDaughter())));
|
||||||
|
assertTrue(!this.fs.exists(HRegion.getRegionDir(this.testdir, st.getSecondDaughter())));
|
||||||
|
assertTrue(!this.parent.splitsAndClosesLock.writeLock().isHeldByCurrentThread());
|
||||||
|
|
||||||
|
// Now retry the split but do not throw an exception this time.
|
||||||
|
assertTrue(st.prepare());
|
||||||
|
PairOfSameType<HRegion> daughters = st.execute(null);
|
||||||
|
// Count rows.
|
||||||
|
int daughtersRowCount = 0;
|
||||||
|
for (HRegion r: daughters) {
|
||||||
|
// Open so can count its content.
|
||||||
|
HRegion openRegion = HRegion.openHRegion(r.getRegionInfo(), this.testdir,
|
||||||
|
r.getLog(), r.getConf());
|
||||||
|
try {
|
||||||
|
int count = countRows(openRegion);
|
||||||
|
assertTrue(count > 0 && count != rowcount);
|
||||||
|
daughtersRowCount += count;
|
||||||
|
} finally {
|
||||||
|
openRegion.close();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
assertEquals(rowcount, daughtersRowCount);
|
||||||
|
// Assert the write lock is no longer held on parent
|
||||||
|
assertTrue(!this.parent.splitsAndClosesLock.writeLock().isHeldByCurrentThread());
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Exception used in this class only.
|
||||||
|
*/
|
||||||
|
@SuppressWarnings("serial")
|
||||||
|
private class MockedFailedDaughterCreation extends IOException {}
|
||||||
|
|
||||||
|
private int countRows(final HRegion r) throws IOException {
|
||||||
|
int rowcount = 0;
|
||||||
|
InternalScanner scanner = r.getScanner(new Scan());
|
||||||
|
try {
|
||||||
|
List<KeyValue> kvs = new ArrayList<KeyValue>();
|
||||||
|
boolean hasNext = true;
|
||||||
|
while (hasNext) {
|
||||||
|
hasNext = scanner.next(kvs);
|
||||||
|
if (!kvs.isEmpty()) rowcount++;
|
||||||
|
}
|
||||||
|
} finally {
|
||||||
|
scanner.close();
|
||||||
|
}
|
||||||
|
return rowcount;
|
||||||
|
}
|
||||||
|
|
||||||
|
static HRegion createRegion(final Path testdir, final HLog wal)
|
||||||
|
throws IOException {
|
||||||
|
// Make a region with start and end keys. Use 'aaa', to 'AAA'. The load
|
||||||
|
// region utility will add rows between 'aaa' and 'zzz'.
|
||||||
|
HTableDescriptor htd = new HTableDescriptor("table");
|
||||||
|
HColumnDescriptor hcd = new HColumnDescriptor(CF);
|
||||||
|
htd.addFamily(hcd);
|
||||||
|
HRegionInfo hri = new HRegionInfo(htd, STARTROW, ENDROW);
|
||||||
|
return HRegion.openHRegion(hri, testdir, wal, TEST_UTIL.getConfiguration());
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue