HBASE-15214 Valid mutate Ops fail with RPC Codec in use and region moves across.
This commit is contained in:
parent
f03c8801c9
commit
64c87f2476
|
@ -556,7 +556,7 @@ public final class ProtobufUtil {
|
|||
MutationType type = proto.getMutateType();
|
||||
assert type == MutationType.PUT: type.name();
|
||||
long timestamp = proto.hasTimestamp()? proto.getTimestamp(): HConstants.LATEST_TIMESTAMP;
|
||||
Put put = null;
|
||||
Put put = proto.hasRow() ? new Put(proto.getRow().toByteArray(), timestamp) : null;
|
||||
int cellCount = proto.hasAssociatedCellCount()? proto.getAssociatedCellCount(): 0;
|
||||
if (cellCount > 0) {
|
||||
// The proto has metadata only and the data is separate to be found in the cellScanner.
|
||||
|
@ -576,9 +576,7 @@ public final class ProtobufUtil {
|
|||
put.add(cell);
|
||||
}
|
||||
} else {
|
||||
if (proto.hasRow()) {
|
||||
put = new Put(proto.getRow().asReadOnlyByteBuffer(), timestamp);
|
||||
} else {
|
||||
if (put == null) {
|
||||
throw new IllegalArgumentException("row cannot be null");
|
||||
}
|
||||
// The proto has the metadata and the data itself
|
||||
|
@ -655,12 +653,8 @@ public final class ProtobufUtil {
|
|||
throws IOException {
|
||||
MutationType type = proto.getMutateType();
|
||||
assert type == MutationType.DELETE : type.name();
|
||||
byte [] row = proto.hasRow()? proto.getRow().toByteArray(): null;
|
||||
long timestamp = HConstants.LATEST_TIMESTAMP;
|
||||
if (proto.hasTimestamp()) {
|
||||
timestamp = proto.getTimestamp();
|
||||
}
|
||||
Delete delete = null;
|
||||
long timestamp = proto.hasTimestamp() ? proto.getTimestamp() : HConstants.LATEST_TIMESTAMP;
|
||||
Delete delete = proto.hasRow() ? new Delete(proto.getRow().toByteArray(), timestamp) : null;
|
||||
int cellCount = proto.hasAssociatedCellCount()? proto.getAssociatedCellCount(): 0;
|
||||
if (cellCount > 0) {
|
||||
// The proto has metadata only and the data is separate to be found in the cellScanner.
|
||||
|
@ -683,7 +677,9 @@ public final class ProtobufUtil {
|
|||
delete.addDeleteMarker(cell);
|
||||
}
|
||||
} else {
|
||||
delete = new Delete(row, timestamp);
|
||||
if (delete == null) {
|
||||
throw new IllegalArgumentException("row cannot be null");
|
||||
}
|
||||
for (ColumnValue column: proto.getColumnValueList()) {
|
||||
byte[] family = column.getFamily().toByteArray();
|
||||
for (QualifierValue qv: column.getQualifierValueList()) {
|
||||
|
|
|
@ -129,6 +129,7 @@ import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.WALEntry;
|
|||
import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.WarmupRegionRequest;
|
||||
import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.WarmupRegionResponse;
|
||||
import org.apache.hadoop.hbase.protobuf.generated.ClientProtos;
|
||||
import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Action;
|
||||
import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest;
|
||||
import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath;
|
||||
import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileResponse;
|
||||
|
@ -620,6 +621,9 @@ public class RSRpcServices implements HBaseRPCErrorHandler,
|
|||
setException(ResponseConverter.buildException(sizeIOE));
|
||||
resultOrExceptionBuilder.setIndex(action.getIndex());
|
||||
builder.addResultOrException(resultOrExceptionBuilder.build());
|
||||
if (cellScanner != null) {
|
||||
skipCellsForMutation(action, cellScanner);
|
||||
}
|
||||
continue;
|
||||
}
|
||||
if (action.hasGet()) {
|
||||
|
@ -2127,6 +2131,12 @@ public class RSRpcServices implements HBaseRPCErrorHandler,
|
|||
rpcServer.getMetrics().exception(e);
|
||||
regionActionResultBuilder.setException(ResponseConverter.buildException(e));
|
||||
responseBuilder.addRegionActionResult(regionActionResultBuilder.build());
|
||||
// All Mutations in this RegionAction not executed as we can not see the Region online here
|
||||
// in this RS. Will be retried from Client. Skipping all the Cells in CellScanner
|
||||
// corresponding to these Mutations.
|
||||
if (cellScanner != null) {
|
||||
skipCellsForMutations(regionAction.getActionList(), cellScanner);
|
||||
}
|
||||
continue; // For this region it's a failure.
|
||||
}
|
||||
|
||||
|
@ -2177,6 +2187,30 @@ public class RSRpcServices implements HBaseRPCErrorHandler,
|
|||
return responseBuilder.build();
|
||||
}
|
||||
|
||||
private void skipCellsForMutations(List<Action> actions, CellScanner cellScanner) {
|
||||
for (Action action : actions) {
|
||||
skipCellsForMutation(action, cellScanner);
|
||||
}
|
||||
}
|
||||
|
||||
private void skipCellsForMutation(Action action, CellScanner cellScanner) {
|
||||
try {
|
||||
if (action.hasMutation()) {
|
||||
MutationProto m = action.getMutation();
|
||||
if (m.hasAssociatedCellCount()) {
|
||||
for (int i = 0; i < m.getAssociatedCellCount(); i++) {
|
||||
cellScanner.advance();
|
||||
}
|
||||
}
|
||||
}
|
||||
} catch (IOException e) {
|
||||
// No need to handle these Individual Muatation level issue. Any way this entire RegionAction
|
||||
// marked as failed as we could not see the Region here. At client side the top level
|
||||
// RegionAction exception will be considered first.
|
||||
LOG.error("Error while skipping Cells in CellScanner for invalid Region Mutations", e);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Mutate data in a table.
|
||||
*
|
||||
|
|
|
@ -35,11 +35,13 @@ import org.apache.commons.logging.LogFactory;
|
|||
import org.apache.hadoop.hbase.Cell;
|
||||
import org.apache.hadoop.hbase.CellUtil;
|
||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||
import org.apache.hadoop.hbase.HConstants;
|
||||
import org.apache.hadoop.hbase.HRegionLocation;
|
||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||
import org.apache.hadoop.hbase.ServerName;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.Waiter;
|
||||
import org.apache.hadoop.hbase.codec.KeyValueCodec;
|
||||
import org.apache.hadoop.hbase.exceptions.OperationConflictException;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.hadoop.hbase.util.JVMClusterUtil;
|
||||
|
@ -74,6 +76,8 @@ public class TestMultiParallel {
|
|||
//((Log4JLogger)RpcServer.LOG).getLogger().setLevel(Level.ALL);
|
||||
//((Log4JLogger)RpcClient.LOG).getLogger().setLevel(Level.ALL);
|
||||
//((Log4JLogger)ScannerCallable.LOG).getLogger().setLevel(Level.ALL);
|
||||
UTIL.getConfiguration().set(HConstants.RPC_CODEC_CONF_KEY,
|
||||
KeyValueCodec.class.getCanonicalName());
|
||||
UTIL.startMiniCluster(slaves);
|
||||
HTable t = UTIL.createMultiRegionTable(TEST_TABLE, Bytes.toBytes(FAMILY));
|
||||
UTIL.waitTableEnabled(TEST_TABLE);
|
||||
|
|
Loading…
Reference in New Issue