HBASE-10925 Do not OOME, throw RowTooBigException instead (Mikhail Antonov)
git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1591154 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
36ad5f8e98
commit
c1d32df547
|
@ -311,6 +311,16 @@ public final class HConstants {
|
||||||
/** Default maximum file size */
|
/** Default maximum file size */
|
||||||
public static final long DEFAULT_MAX_FILE_SIZE = 10 * 1024 * 1024 * 1024L;
|
public static final long DEFAULT_MAX_FILE_SIZE = 10 * 1024 * 1024 * 1024L;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Max size of single row for Get's or Scan's without in-row scanning flag set.
|
||||||
|
*/
|
||||||
|
public static final String TABLE_MAX_ROWSIZE_KEY = "hbase.table.max.rowsize";
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Default max row size (1 Gb).
|
||||||
|
*/
|
||||||
|
public static final long TABLE_MAX_ROWSIZE_DEFAULT = 1024 * 1024 * 1024L;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The max number of threads used for opening and closing stores or store
|
* The max number of threads used for opening and closing stores or store
|
||||||
* files in parallel
|
* files in parallel
|
||||||
|
|
|
@ -896,6 +896,15 @@ possible configurations would overwhelm and obscure the important.
|
||||||
Table locking from master prevents concurrent schema modifications to corrupt table
|
Table locking from master prevents concurrent schema modifications to corrupt table
|
||||||
state.</description>
|
state.</description>
|
||||||
</property>
|
</property>
|
||||||
|
<property>
|
||||||
|
<name>hbase.table.max.rowsize</name>
|
||||||
|
<value>1073741824</value>
|
||||||
|
<description>
|
||||||
|
Maximum size of single row in bytes (default is 1 Gb) for Get'ting
|
||||||
|
or Scan'ning without in-row scan flag set. If row size exceeds this limit
|
||||||
|
RowTooBigException is thrown to client.
|
||||||
|
</description>
|
||||||
|
</property>
|
||||||
<property>
|
<property>
|
||||||
<name>hbase.thrift.minWorkerThreads</name>
|
<name>hbase.thrift.minWorkerThreads</name>
|
||||||
<value>16</value>
|
<value>16</value>
|
||||||
|
|
|
@ -0,0 +1,33 @@
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
package org.apache.hadoop.hbase.regionserver;
|
||||||
|
|
||||||
|
import org.apache.hadoop.hbase.RegionException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Gets or Scans throw this exception if running without in-row scan flag
|
||||||
|
* set and row size appears to exceed max configured size (configurable via
|
||||||
|
* hbase.table.max.rowsize).
|
||||||
|
*/
|
||||||
|
public class RowTooBigException extends RegionException {
|
||||||
|
|
||||||
|
public RowTooBigException(String message) {
|
||||||
|
super(message);
|
||||||
|
}
|
||||||
|
}
|
|
@ -31,6 +31,7 @@ import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.classification.InterfaceAudience;
|
import org.apache.hadoop.classification.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.Cell;
|
import org.apache.hadoop.hbase.Cell;
|
||||||
|
import org.apache.hadoop.hbase.CellUtil;
|
||||||
import org.apache.hadoop.hbase.DoNotRetryIOException;
|
import org.apache.hadoop.hbase.DoNotRetryIOException;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.KeyValue;
|
import org.apache.hadoop.hbase.KeyValue;
|
||||||
|
@ -76,6 +77,7 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
|
||||||
protected final NavigableSet<byte[]> columns;
|
protected final NavigableSet<byte[]> columns;
|
||||||
protected final long oldestUnexpiredTS;
|
protected final long oldestUnexpiredTS;
|
||||||
protected final int minVersions;
|
protected final int minVersions;
|
||||||
|
protected final long maxRowSize;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The number of KVs seen by the scanner. Includes explicitly skipped KVs, but not
|
* The number of KVs seen by the scanner. Includes explicitly skipped KVs, but not
|
||||||
|
@ -123,6 +125,14 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
|
||||||
oldestUnexpiredTS = EnvironmentEdgeManager.currentTimeMillis() - ttl;
|
oldestUnexpiredTS = EnvironmentEdgeManager.currentTimeMillis() - ttl;
|
||||||
this.minVersions = minVersions;
|
this.minVersions = minVersions;
|
||||||
|
|
||||||
|
if (store != null && ((HStore)store).getHRegion() != null
|
||||||
|
&& ((HStore)store).getHRegion().getBaseConf() != null) {
|
||||||
|
this.maxRowSize = ((HStore) store).getHRegion().getBaseConf().getLong(
|
||||||
|
HConstants.TABLE_MAX_ROWSIZE_KEY, HConstants.TABLE_MAX_ROWSIZE_DEFAULT);
|
||||||
|
} else {
|
||||||
|
this.maxRowSize = HConstants.TABLE_MAX_ROWSIZE_DEFAULT;
|
||||||
|
}
|
||||||
|
|
||||||
// We look up row-column Bloom filters for multi-column queries as part of
|
// We look up row-column Bloom filters for multi-column queries as part of
|
||||||
// the seek operation. However, we also look the row-column Bloom filter
|
// the seek operation. However, we also look the row-column Bloom filter
|
||||||
// for multi-row (non-"get") scans because this is not done in
|
// for multi-row (non-"get") scans because this is not done in
|
||||||
|
@ -313,8 +323,17 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
if (!isParallelSeek) {
|
if (!isParallelSeek) {
|
||||||
|
long totalScannersSoughtBytes = 0;
|
||||||
for (KeyValueScanner scanner : scanners) {
|
for (KeyValueScanner scanner : scanners) {
|
||||||
|
if (totalScannersSoughtBytes >= maxRowSize) {
|
||||||
|
throw new RowTooBigException("Max row size allowed: " + maxRowSize
|
||||||
|
+ ", but row is bigger than that");
|
||||||
|
}
|
||||||
scanner.seek(seekKey);
|
scanner.seek(seekKey);
|
||||||
|
Cell c = scanner.peek();
|
||||||
|
if (c != null ) {
|
||||||
|
totalScannersSoughtBytes += CellUtil.estimatedSizeOf(c);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
parallelSeek(scanners, seekKey);
|
parallelSeek(scanners, seekKey);
|
||||||
|
@ -461,6 +480,8 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
|
||||||
store != null ? store.getComparator() : null;
|
store != null ? store.getComparator() : null;
|
||||||
|
|
||||||
int count = 0;
|
int count = 0;
|
||||||
|
long totalBytesRead = 0;
|
||||||
|
|
||||||
LOOP: while((cell = this.heap.peek()) != null) {
|
LOOP: while((cell = this.heap.peek()) != null) {
|
||||||
if (prevCell != cell) ++kvsScanned; // Do object compare - we set prevKV from the same heap.
|
if (prevCell != cell) ++kvsScanned; // Do object compare - we set prevKV from the same heap.
|
||||||
checkScanOrder(prevCell, cell, comparator);
|
checkScanOrder(prevCell, cell, comparator);
|
||||||
|
@ -494,6 +515,11 @@ public class StoreScanner extends NonReversedNonLazyKeyValueScanner
|
||||||
if (this.countPerRow > storeOffset) {
|
if (this.countPerRow > storeOffset) {
|
||||||
outResult.add(cell);
|
outResult.add(cell);
|
||||||
count++;
|
count++;
|
||||||
|
totalBytesRead += CellUtil.estimatedSizeOf(cell);
|
||||||
|
if (totalBytesRead > maxRowSize) {
|
||||||
|
throw new RowTooBigException("Max row size allowed: " + maxRowSize
|
||||||
|
+ ", but the row is bigger than that.");
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (qcode == ScanQueryMatcher.MatchCode.INCLUDE_AND_SEEK_NEXT_ROW) {
|
if (qcode == ScanQueryMatcher.MatchCode.INCLUDE_AND_SEEK_NEXT_ROW) {
|
||||||
|
|
|
@ -0,0 +1,132 @@
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hadoop.hbase.regionserver;
|
||||||
|
|
||||||
|
import org.apache.hadoop.hbase.*;
|
||||||
|
import org.apache.hadoop.hbase.client.Get;
|
||||||
|
import org.apache.hadoop.hbase.client.Put;
|
||||||
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
import org.junit.AfterClass;
|
||||||
|
import org.junit.BeforeClass;
|
||||||
|
import org.junit.Test;
|
||||||
|
import org.junit.experimental.categories.Category;
|
||||||
|
|
||||||
|
import java.io.IOException;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Test case to check HRS throws {@link RowTooBigException}
|
||||||
|
* when row size exceeds configured limits.
|
||||||
|
*/
|
||||||
|
@Category(MediumTests.class)
|
||||||
|
public class TestRowTooBig {
|
||||||
|
private final static HBaseTestingUtility HTU = HBaseTestingUtility.createLocalHTU();
|
||||||
|
private static final HTableDescriptor TEST_HTD =
|
||||||
|
new HTableDescriptor(TableName.valueOf(TestRowTooBig.class.getSimpleName()));
|
||||||
|
|
||||||
|
@BeforeClass
|
||||||
|
public static void before() throws Exception {
|
||||||
|
HTU.startMiniCluster();
|
||||||
|
HTU.getConfiguration().setLong(HConstants.TABLE_MAX_ROWSIZE_KEY,
|
||||||
|
10 * 1024 * 1024L);
|
||||||
|
}
|
||||||
|
|
||||||
|
@AfterClass
|
||||||
|
public static void after() throws Exception {
|
||||||
|
HTU.shutdownMiniCluster();
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Usecase:
|
||||||
|
* - create a row with 5 large cells (5 Mb each)
|
||||||
|
* - flush memstore but don't compact storefiles.
|
||||||
|
* - try to Get whole row.
|
||||||
|
*
|
||||||
|
* OOME happened before we actually get to reading results, but
|
||||||
|
* during seeking, as each StoreFile gets it's own scanner,
|
||||||
|
* and each scanner seeks after the first KV.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
@Test(expected = RowTooBigException.class)
|
||||||
|
public void testScannersSeekOnFewLargeCells() throws IOException {
|
||||||
|
byte[] row1 = Bytes.toBytes("row1");
|
||||||
|
byte[] fam1 = Bytes.toBytes("fam1");
|
||||||
|
|
||||||
|
HTableDescriptor htd = TEST_HTD;
|
||||||
|
HColumnDescriptor hcd = new HColumnDescriptor(fam1);
|
||||||
|
htd.addFamily(hcd);
|
||||||
|
|
||||||
|
final HRegionInfo hri =
|
||||||
|
new HRegionInfo(htd.getTableName(), HConstants.EMPTY_END_ROW,
|
||||||
|
HConstants.EMPTY_END_ROW);
|
||||||
|
HRegion region = HTU.createLocalHRegion(hri, htd);
|
||||||
|
|
||||||
|
// Add 5 cells to memstore
|
||||||
|
for (int i = 0; i < 5 ; i++) {
|
||||||
|
Put put = new Put(row1);
|
||||||
|
|
||||||
|
put.add(fam1, Bytes.toBytes("col_" + i ), new byte[5 * 1024 * 1024]);
|
||||||
|
region.put(put);
|
||||||
|
region.flushcache();
|
||||||
|
}
|
||||||
|
|
||||||
|
Get get = new Get(row1);
|
||||||
|
region.get(get);
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Usecase:
|
||||||
|
*
|
||||||
|
* - create a row with 1M cells, 10 bytes in each
|
||||||
|
* - flush & run major compaction
|
||||||
|
* - try to Get whole row.
|
||||||
|
*
|
||||||
|
* OOME happened in StoreScanner.next(..).
|
||||||
|
*
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
@Test(expected = RowTooBigException.class)
|
||||||
|
public void testScanAcrossManySmallColumns() throws IOException {
|
||||||
|
byte[] row1 = Bytes.toBytes("row1");
|
||||||
|
byte[] fam1 = Bytes.toBytes("fam1");
|
||||||
|
|
||||||
|
HTableDescriptor htd = TEST_HTD;
|
||||||
|
HColumnDescriptor hcd = new HColumnDescriptor(fam1);
|
||||||
|
htd.addFamily(hcd);
|
||||||
|
|
||||||
|
final HRegionInfo hri =
|
||||||
|
new HRegionInfo(htd.getTableName(), HConstants.EMPTY_END_ROW,
|
||||||
|
HConstants.EMPTY_END_ROW);
|
||||||
|
HRegion region = HTU.createLocalHRegion(hri, htd);
|
||||||
|
|
||||||
|
// Add to memstore
|
||||||
|
for (int i = 0; i < 10; i++) {
|
||||||
|
Put put = new Put(row1);
|
||||||
|
for (int j = 0; j < 10 * 10000; j++) {
|
||||||
|
put.add(fam1, Bytes.toBytes("col_" + i + "_" + j), new byte[10]);
|
||||||
|
}
|
||||||
|
region.put(put);
|
||||||
|
region.flushcache();
|
||||||
|
}
|
||||||
|
region.compactStores(true);
|
||||||
|
|
||||||
|
Get get = new Get(row1);
|
||||||
|
region.get(get);
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue