HBASE-10925 Do not OOME, throw RowTooBigException instead (Mikhail Antonov)

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1591154 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Michael Stack 2014-04-30 00:08:39 +00:00
parent 36ad5f8e98
commit c1d32df547
5 changed files with 210 additions and 0 deletions

View File

@ -311,6 +311,16 @@ public final class HConstants {
/** Default maximum file size */
public static final long DEFAULT_MAX_FILE_SIZE = 10 * 1024 * 1024 * 1024L;
/**
* Max size of single row for Get's or Scan's without in-row scanning flag set.
*/
public static final String TABLE_MAX_ROWSIZE_KEY = "hbase.table.max.rowsize";
/**
* Default max row size (1 Gb).
*/
public static final long TABLE_MAX_ROWSIZE_DEFAULT = 1024 * 1024 * 1024L;
/**
* The max number of threads used for opening and closing stores or store
* files in parallel

View File

@ -896,6 +896,15 @@ possible configurations would overwhelm and obscure the important.
Table locking from master prevents concurrent schema modifications to corrupt table
state.</description>
</property>
<property>
<name>hbase.table.max.rowsize</name>
<value>1073741824</value>
<description>
Maximum size of single row in bytes (default is 1 Gb) for Get'ting
or Scan'ning without in-row scan flag set. If row size exceeds this limit
RowTooBigException is thrown to client.
</description>
</property>
<property>
<name>hbase.thrift.minWorkerThreads</name>
<value>16</value>

View File

@ -0,0 +1,33 @@
/**
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hbase.regionserver;
import org.apache.hadoop.hbase.RegionException;
/**
* Gets or Scans throw this exception if running without in-row scan flag
* set and row size appears to exceed max configured size (configurable via
* hbase.table.max.rowsize).
*/
public class RowTooBigException extends RegionException {
public RowTooBigException(String message) {
super(message);
}
}

View File

@ -31,6 +31,7 @@ import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.hbase.Cell;
import org.apache.hadoop.hbase.CellUtil;
import org.apache.hadoop.hbase.DoNotRetryIOException;
import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.KeyValue;
@ -76,6 +77,7 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
protected final NavigableSet<byte[]> columns;
protected final long oldestUnexpiredTS;
protected final int minVersions;
protected final long maxRowSize;
/**
* The number of KVs seen by the scanner. Includes explicitly skipped KVs, but not
@ -123,6 +125,14 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
oldestUnexpiredTS = EnvironmentEdgeManager.currentTimeMillis() - ttl;
this.minVersions = minVersions;
if (store != null && ((HStore)store).getHRegion() != null
&& ((HStore)store).getHRegion().getBaseConf() != null) {
this.maxRowSize = ((HStore) store).getHRegion().getBaseConf().getLong(
HConstants.TABLE_MAX_ROWSIZE_KEY, HConstants.TABLE_MAX_ROWSIZE_DEFAULT);
} else {
this.maxRowSize = HConstants.TABLE_MAX_ROWSIZE_DEFAULT;
}
// We look up row-column Bloom filters for multi-column queries as part of
// the seek operation. However, we also look the row-column Bloom filter
// for multi-row (non-"get") scans because this is not done in
@ -313,8 +323,17 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
}
} else {
if (!isParallelSeek) {
long totalScannersSoughtBytes = 0;
for (KeyValueScanner scanner : scanners) {
if (totalScannersSoughtBytes >= maxRowSize) {
throw new RowTooBigException("Max row size allowed: " + maxRowSize
+ ", but row is bigger than that");
}
scanner.seek(seekKey);
Cell c = scanner.peek();
if (c != null ) {
totalScannersSoughtBytes += CellUtil.estimatedSizeOf(c);
}
}
} else {
parallelSeek(scanners, seekKey);
@ -461,6 +480,8 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
store != null ? store.getComparator() : null;
int count = 0;
long totalBytesRead = 0;
LOOP: while((cell = this.heap.peek()) != null) {
if (prevCell != cell) ++kvsScanned; // Do object compare - we set prevKV from the same heap.
checkScanOrder(prevCell, cell, comparator);
@ -494,6 +515,11 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
if (this.countPerRow > storeOffset) {
outResult.add(cell);
count++;
totalBytesRead += CellUtil.estimatedSizeOf(cell);
if (totalBytesRead > maxRowSize) {
throw new RowTooBigException("Max row size allowed: " + maxRowSize
+ ", but the row is bigger than that.");
}
}
if (qcode == ScanQueryMatcher.MatchCode.INCLUDE_AND_SEEK_NEXT_ROW) {

View File

@ -0,0 +1,132 @@
/**
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hbase.regionserver;
import org.apache.hadoop.hbase.*;
import org.apache.hadoop.hbase.client.Get;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.util.Bytes;
import org.junit.AfterClass;
import org.junit.BeforeClass;
import org.junit.Test;
import org.junit.experimental.categories.Category;
import java.io.IOException;
/**
* Test case to check HRS throws {@link RowTooBigException}
* when row size exceeds configured limits.
*/
@Category(MediumTests.class)
public class TestRowTooBig {
private final static HBaseTestingUtility HTU = HBaseTestingUtility.createLocalHTU();
private static final HTableDescriptor TEST_HTD =
new HTableDescriptor(TableName.valueOf(TestRowTooBig.class.getSimpleName()));
@BeforeClass
public static void before() throws Exception {
HTU.startMiniCluster();
HTU.getConfiguration().setLong(HConstants.TABLE_MAX_ROWSIZE_KEY,
10 * 1024 * 1024L);
}
@AfterClass
public static void after() throws Exception {
HTU.shutdownMiniCluster();
}
/**
* Usecase:
* - create a row with 5 large cells (5 Mb each)
* - flush memstore but don't compact storefiles.
* - try to Get whole row.
*
* OOME happened before we actually get to reading results, but
* during seeking, as each StoreFile gets it's own scanner,
* and each scanner seeks after the first KV.
* @throws IOException
*/
@Test(expected = RowTooBigException.class)
public void testScannersSeekOnFewLargeCells() throws IOException {
byte[] row1 = Bytes.toBytes("row1");
byte[] fam1 = Bytes.toBytes("fam1");
HTableDescriptor htd = TEST_HTD;
HColumnDescriptor hcd = new HColumnDescriptor(fam1);
htd.addFamily(hcd);
final HRegionInfo hri =
new HRegionInfo(htd.getTableName(), HConstants.EMPTY_END_ROW,
HConstants.EMPTY_END_ROW);
HRegion region = HTU.createLocalHRegion(hri, htd);
// Add 5 cells to memstore
for (int i = 0; i < 5 ; i++) {
Put put = new Put(row1);
put.add(fam1, Bytes.toBytes("col_" + i ), new byte[5 * 1024 * 1024]);
region.put(put);
region.flushcache();
}
Get get = new Get(row1);
region.get(get);
}
/**
* Usecase:
*
* - create a row with 1M cells, 10 bytes in each
* - flush & run major compaction
* - try to Get whole row.
*
* OOME happened in StoreScanner.next(..).
*
* @throws IOException
*/
@Test(expected = RowTooBigException.class)
public void testScanAcrossManySmallColumns() throws IOException {
byte[] row1 = Bytes.toBytes("row1");
byte[] fam1 = Bytes.toBytes("fam1");
HTableDescriptor htd = TEST_HTD;
HColumnDescriptor hcd = new HColumnDescriptor(fam1);
htd.addFamily(hcd);
final HRegionInfo hri =
new HRegionInfo(htd.getTableName(), HConstants.EMPTY_END_ROW,
HConstants.EMPTY_END_ROW);
HRegion region = HTU.createLocalHRegion(hri, htd);
// Add to memstore
for (int i = 0; i < 10; i++) {
Put put = new Put(row1);
for (int j = 0; j < 10 * 10000; j++) {
put.add(fam1, Bytes.toBytes("col_" + i + "_" + j), new byte[10]);
}
region.put(put);
region.flushcache();
}
region.compactStores(true);
Get get = new Get(row1);
region.get(get);
}
}