HBASE-7405 Enforce PB ser/de for Aggregate protocol and associated ColumnInterpreter user code bits (Devaraj Das)

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1429278 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Zhihong Yu 2013-01-05 13:22:04 +00:00
parent 94bd157d21
commit c8138e1c3e
8 changed files with 1012 additions and 165 deletions

View File

@ -11303,6 +11303,685 @@ public final class HBaseProtos {
// @@protoc_insertion_point(class_scope:NameInt64Pair) // @@protoc_insertion_point(class_scope:NameInt64Pair)
} }
public interface EmptyMsgOrBuilder
extends com.google.protobuf.MessageOrBuilder {
}
public static final class EmptyMsg extends
com.google.protobuf.GeneratedMessage
implements EmptyMsgOrBuilder {
// Use EmptyMsg.newBuilder() to construct.
private EmptyMsg(Builder builder) {
super(builder);
}
private EmptyMsg(boolean noInit) {}
private static final EmptyMsg defaultInstance;
public static EmptyMsg getDefaultInstance() {
return defaultInstance;
}
public EmptyMsg getDefaultInstanceForType() {
return defaultInstance;
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_EmptyMsg_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_EmptyMsg_fieldAccessorTable;
}
private void initFields() {
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg)) {
return super.equals(obj);
}
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg) obj;
boolean result = true;
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
@java.lang.Override
public int hashCode() {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
return hash;
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseFrom(java.io.InputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsgOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_EmptyMsg_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_EmptyMsg_fieldAccessorTable;
}
// Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg.getDescriptor();
}
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg getDefaultInstanceForType() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg.getDefaultInstance();
}
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg build() {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg buildParsed()
throws com.google.protobuf.InvalidProtocolBufferException {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(
result).asInvalidProtocolBufferException();
}
return result;
}
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg buildPartial() {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg(this);
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg) {
return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg other) {
if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder(
this.getUnknownFields());
while (true) {
int tag = input.readTag();
switch (tag) {
case 0:
this.setUnknownFields(unknownFields.build());
onChanged();
return this;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
this.setUnknownFields(unknownFields.build());
onChanged();
return this;
}
break;
}
}
}
}
// @@protoc_insertion_point(builder_scope:EmptyMsg)
}
static {
defaultInstance = new EmptyMsg(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:EmptyMsg)
}
public interface LongMsgOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required int64 longMsg = 1;
boolean hasLongMsg();
long getLongMsg();
}
public static final class LongMsg extends
com.google.protobuf.GeneratedMessage
implements LongMsgOrBuilder {
// Use LongMsg.newBuilder() to construct.
private LongMsg(Builder builder) {
super(builder);
}
private LongMsg(boolean noInit) {}
private static final LongMsg defaultInstance;
public static LongMsg getDefaultInstance() {
return defaultInstance;
}
public LongMsg getDefaultInstanceForType() {
return defaultInstance;
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_LongMsg_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_LongMsg_fieldAccessorTable;
}
private int bitField0_;
// required int64 longMsg = 1;
public static final int LONGMSG_FIELD_NUMBER = 1;
private long longMsg_;
public boolean hasLongMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
public long getLongMsg() {
return longMsg_;
}
private void initFields() {
longMsg_ = 0L;
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasLongMsg()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeInt64(1, longMsg_);
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeInt64Size(1, longMsg_);
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg)) {
return super.equals(obj);
}
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg other = (org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg) obj;
boolean result = true;
result = result && (hasLongMsg() == other.hasLongMsg());
if (hasLongMsg()) {
result = result && (getLongMsg()
== other.getLongMsg());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
@java.lang.Override
public int hashCode() {
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasLongMsg()) {
hash = (37 * hash) + LONGMSG_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getLongMsg());
}
hash = (29 * hash) + getUnknownFields().hashCode();
return hash;
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data).buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return newBuilder().mergeFrom(data, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseFrom(java.io.InputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
Builder builder = newBuilder();
if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
return builder.buildParsed();
} else {
return null;
}
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return newBuilder().mergeFrom(input).buildParsed();
}
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return newBuilder().mergeFrom(input, extensionRegistry)
.buildParsed();
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder<Builder>
implements org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsgOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_LongMsg_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.internal_static_LongMsg_fieldAccessorTable;
}
// Construct using org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
longMsg_ = 0L;
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg.getDescriptor();
}
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg getDefaultInstanceForType() {
return org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg.getDefaultInstance();
}
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg build() {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg buildParsed()
throws com.google.protobuf.InvalidProtocolBufferException {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(
result).asInvalidProtocolBufferException();
}
return result;
}
public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg buildPartial() {
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg result = new org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.longMsg_ = longMsg_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg) {
return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg other) {
if (other == org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg.getDefaultInstance()) return this;
if (other.hasLongMsg()) {
setLongMsg(other.getLongMsg());
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasLongMsg()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder(
this.getUnknownFields());
while (true) {
int tag = input.readTag();
switch (tag) {
case 0:
this.setUnknownFields(unknownFields.build());
onChanged();
return this;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
this.setUnknownFields(unknownFields.build());
onChanged();
return this;
}
break;
}
case 8: {
bitField0_ |= 0x00000001;
longMsg_ = input.readInt64();
break;
}
}
}
}
private int bitField0_;
// required int64 longMsg = 1;
private long longMsg_ ;
public boolean hasLongMsg() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
public long getLongMsg() {
return longMsg_;
}
public Builder setLongMsg(long value) {
bitField0_ |= 0x00000001;
longMsg_ = value;
onChanged();
return this;
}
public Builder clearLongMsg() {
bitField0_ = (bitField0_ & ~0x00000001);
longMsg_ = 0L;
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:LongMsg)
}
static {
defaultInstance = new LongMsg(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:LongMsg)
}
private static com.google.protobuf.Descriptors.Descriptor private static com.google.protobuf.Descriptors.Descriptor
internal_static_TableSchema_descriptor; internal_static_TableSchema_descriptor;
private static private static
@ -11388,6 +12067,16 @@ public final class HBaseProtos {
private static private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_NameInt64Pair_fieldAccessorTable; internal_static_NameInt64Pair_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_EmptyMsg_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_EmptyMsg_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_LongMsg_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_LongMsg_fieldAccessorTable;
public static com.google.protobuf.Descriptors.FileDescriptor public static com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() { getDescriptor() {
@ -11440,13 +12129,14 @@ public final class HBaseProtos {
"value\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004name\030\001 \002" + "value\030\002 \002(\t\",\n\rNameBytesPair\022\014\n\004name\030\001 \002" +
"(\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPair\022\r\n\005f" + "(\t\022\r\n\005value\030\002 \001(\014\"/\n\016BytesBytesPair\022\r\n\005f" +
"irst\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNameInt64P" + "irst\030\001 \002(\014\022\016\n\006second\030\002 \002(\014\",\n\rNameInt64P" +
"air\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003*r\n\013Comp" + "air\022\014\n\004name\030\001 \001(\t\022\r\n\005value\030\002 \001(\003\"\n\n\010Empt" +
"yMsg\"\032\n\007LongMsg\022\017\n\007longMsg\030\001 \002(\003*r\n\013Comp" +
"areType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005" + "areType\022\010\n\004LESS\020\000\022\021\n\rLESS_OR_EQUAL\020\001\022\t\n\005" +
"EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQU" + "EQUAL\020\002\022\r\n\tNOT_EQUAL\020\003\022\024\n\020GREATER_OR_EQU" +
"AL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006*_\n\007KeyType\022" + "AL\020\004\022\013\n\007GREATER\020\005\022\t\n\005NO_OP\020\006*_\n\007KeyType\022" +
"\013\n\007MINIMUM\020\000\022\007\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDEL" + "\013\n\007MINIMUM\020\000\022\007\n\003PUT\020\004\022\n\n\006DELETE\020\010\022\021\n\rDEL" +
"ETE_COLUMN\020\014\022\021\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIM" + "ETE_COLUMN\020\014\022\021\n\rDELETE_FAMILY\020\016\022\014\n\007MAXIM",
"UM\020\377\001B>\n*org.apache.hadoop.hbase.protobu", "UM\020\377\001B>\n*org.apache.hadoop.hbase.protobu" +
"f.generatedB\013HBaseProtosH\001\240\001\001" "f.generatedB\013HBaseProtosH\001\240\001\001"
}; };
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
@ -11590,6 +12280,22 @@ public final class HBaseProtos {
new java.lang.String[] { "Name", "Value", }, new java.lang.String[] { "Name", "Value", },
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameInt64Pair.class, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameInt64Pair.class,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameInt64Pair.Builder.class); org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameInt64Pair.Builder.class);
internal_static_EmptyMsg_descriptor =
getDescriptor().getMessageTypes().get(15);
internal_static_EmptyMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_EmptyMsg_descriptor,
new java.lang.String[] { },
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg.class,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg.Builder.class);
internal_static_LongMsg_descriptor =
getDescriptor().getMessageTypes().get(16);
internal_static_LongMsg_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_LongMsg_descriptor,
new java.lang.String[] { "LongMsg", },
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg.class,
org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg.Builder.class);
return null; return null;
} }
}; };

View File

@ -268,3 +268,10 @@ message NameInt64Pair {
optional string name = 1; optional string name = 1;
optional int64 value = 2; optional int64 value = 2;
} }
message EmptyMsg {
}
message LongMsg {
required int64 longMsg = 1;
}

View File

@ -20,6 +20,10 @@
package org.apache.hadoop.hbase.client.coprocessor; package org.apache.hadoop.hbase.client.coprocessor;
import java.io.IOException; import java.io.IOException;
import java.lang.reflect.InvocationTargetException;
import java.lang.reflect.Method;
import java.lang.reflect.ParameterizedType;
import java.lang.reflect.Type;
import java.nio.ByteBuffer; import java.nio.ByteBuffer;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.List; import java.util.List;
@ -51,6 +55,7 @@ import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.hbase.util.Pair;
import com.google.protobuf.ByteString; import com.google.protobuf.ByteString;
import com.google.protobuf.Message;
/** /**
* This client class is for invoking the aggregate functions deployed on the * This client class is for invoking the aggregate functions deployed on the
@ -98,7 +103,8 @@ public class AggregationClient {
* The caller is supposed to handle the exception as they are thrown * The caller is supposed to handle the exception as they are thrown
* & propagated to it. * & propagated to it.
*/ */
public <R, S> R max(final byte[] tableName, final ColumnInterpreter<R, S> ci, public <R, S, P extends Message, Q extends Message, T extends Message>
R max(final byte[] tableName, final ColumnInterpreter<R, S, P, Q, T> ci,
final Scan scan) throws Throwable { final Scan scan) throws Throwable {
final AggregateArgument requestArg = validateArgAndGetPB(scan, ci); final AggregateArgument requestArg = validateArgAndGetPB(scan, ci);
class MaxCallBack implements Batch.Callback<R> { class MaxCallBack implements Batch.Callback<R> {
@ -130,9 +136,9 @@ public class AggregationClient {
throw controller.getFailedOn(); throw controller.getFailedOn();
} }
if (response.getFirstPartCount() > 0) { if (response.getFirstPartCount() > 0) {
return ci.castToCellType( ByteString b = response.getFirstPart(0);
ci.parseResponseAsPromotedType( Q q = ProtobufUtil.getParsedGenericInstance(ci.getClass(), 3, b);
getBytesFromResponse(response.getFirstPart(0)))); return ci.getCellValueFromProto(q);
} }
return null; return null;
} }
@ -168,7 +174,8 @@ public class AggregationClient {
* @return min val <R> * @return min val <R>
* @throws Throwable * @throws Throwable
*/ */
public <R, S> R min(final byte[] tableName, final ColumnInterpreter<R, S> ci, public <R, S, P extends Message, Q extends Message, T extends Message>
R min(final byte[] tableName, final ColumnInterpreter<R, S, P, Q, T> ci,
final Scan scan) throws Throwable { final Scan scan) throws Throwable {
final AggregateArgument requestArg = validateArgAndGetPB(scan, ci); final AggregateArgument requestArg = validateArgAndGetPB(scan, ci);
class MinCallBack implements Batch.Callback<R> { class MinCallBack implements Batch.Callback<R> {
@ -202,9 +209,9 @@ public class AggregationClient {
throw controller.getFailedOn(); throw controller.getFailedOn();
} }
if (response.getFirstPartCount() > 0) { if (response.getFirstPartCount() > 0) {
return ci.castToCellType( ByteString b = response.getFirstPart(0);
ci.parseResponseAsPromotedType( Q q = ProtobufUtil.getParsedGenericInstance(ci.getClass(), 3, b);
getBytesFromResponse(response.getFirstPart(0)))); return ci.getCellValueFromProto(q);
} }
return null; return null;
} }
@ -231,8 +238,9 @@ public class AggregationClient {
* @return <R, S> * @return <R, S>
* @throws Throwable * @throws Throwable
*/ */
public <R, S> long rowCount(final byte[] tableName, public <R, S, P extends Message, Q extends Message, T extends Message>
final ColumnInterpreter<R, S> ci, final Scan scan) throws Throwable { long rowCount(final byte[] tableName,
final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan) throws Throwable {
final AggregateArgument requestArg = validateArgAndGetPB(scan, ci); final AggregateArgument requestArg = validateArgAndGetPB(scan, ci);
class RowNumCallback implements Batch.Callback<Long> { class RowNumCallback implements Batch.Callback<Long> {
private final AtomicLong rowCountL = new AtomicLong(0); private final AtomicLong rowCountL = new AtomicLong(0);
@ -285,7 +293,8 @@ public class AggregationClient {
* @return sum <S> * @return sum <S>
* @throws Throwable * @throws Throwable
*/ */
public <R, S> S sum(final byte[] tableName, final ColumnInterpreter<R, S> ci, public <R, S, P extends Message, Q extends Message, T extends Message>
S sum(final byte[] tableName, final ColumnInterpreter<R, S, P, Q, T> ci,
final Scan scan) throws Throwable { final Scan scan) throws Throwable {
final AggregateArgument requestArg = validateArgAndGetPB(scan, ci); final AggregateArgument requestArg = validateArgAndGetPB(scan, ci);
@ -320,8 +329,10 @@ public class AggregationClient {
if (response.getFirstPartCount() == 0) { if (response.getFirstPartCount() == 0) {
return null; return null;
} }
return ci.parseResponseAsPromotedType( ByteString b = response.getFirstPart(0);
getBytesFromResponse(response.getFirstPart(0))); T t = ProtobufUtil.getParsedGenericInstance(ci.getClass(), 4, b);
S s = ci.getPromotedValueFromProto(t);
return s;
} }
}, sumCallBack); }, sumCallBack);
} finally { } finally {
@ -340,8 +351,9 @@ public class AggregationClient {
* @param scan * @param scan
* @throws Throwable * @throws Throwable
*/ */
private <R, S> Pair<S, Long> getAvgArgs(final byte[] tableName, private <R, S, P extends Message, Q extends Message, T extends Message>
final ColumnInterpreter<R, S> ci, final Scan scan) throws Throwable { Pair<S, Long> getAvgArgs(final byte[] tableName,
final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan) throws Throwable {
final AggregateArgument requestArg = validateArgAndGetPB(scan, ci); final AggregateArgument requestArg = validateArgAndGetPB(scan, ci);
class AvgCallBack implements Batch.Callback<Pair<S, Long>> { class AvgCallBack implements Batch.Callback<Pair<S, Long>> {
S sum = null; S sum = null;
@ -379,8 +391,10 @@ public class AggregationClient {
if (response.getFirstPartCount() == 0) { if (response.getFirstPartCount() == 0) {
return pair; return pair;
} }
pair.setFirst(ci.parseResponseAsPromotedType( ByteString b = response.getFirstPart(0);
getBytesFromResponse(response.getFirstPart(0)))); T t = ProtobufUtil.getParsedGenericInstance(ci.getClass(), 4, b);
S s = ci.getPromotedValueFromProto(t);
pair.setFirst(s);
ByteBuffer bb = ByteBuffer.allocate(8).put( ByteBuffer bb = ByteBuffer.allocate(8).put(
getBytesFromResponse(response.getSecondPart())); getBytesFromResponse(response.getSecondPart()));
bb.rewind(); bb.rewind();
@ -408,8 +422,9 @@ public class AggregationClient {
* @return <R, S> * @return <R, S>
* @throws Throwable * @throws Throwable
*/ */
public <R, S> double avg(final byte[] tableName, public <R, S, P extends Message, Q extends Message, T extends Message>
final ColumnInterpreter<R, S> ci, Scan scan) throws Throwable { double avg(final byte[] tableName,
final ColumnInterpreter<R, S, P, Q, T> ci, Scan scan) throws Throwable {
Pair<S, Long> p = getAvgArgs(tableName, ci, scan); Pair<S, Long> p = getAvgArgs(tableName, ci, scan);
return ci.divideForAvg(p.getFirst(), p.getSecond()); return ci.divideForAvg(p.getFirst(), p.getSecond());
} }
@ -425,8 +440,9 @@ public class AggregationClient {
* @return * @return
* @throws Throwable * @throws Throwable
*/ */
private <R, S> Pair<List<S>, Long> getStdArgs(final byte[] tableName, private <R, S, P extends Message, Q extends Message, T extends Message>
final ColumnInterpreter<R, S> ci, final Scan scan) throws Throwable { Pair<List<S>, Long> getStdArgs(final byte[] tableName,
final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan) throws Throwable {
final AggregateArgument requestArg = validateArgAndGetPB(scan, ci); final AggregateArgument requestArg = validateArgAndGetPB(scan, ci);
class StdCallback implements Batch.Callback<Pair<List<S>, Long>> { class StdCallback implements Batch.Callback<Pair<List<S>, Long>> {
long rowCountVal = 0l; long rowCountVal = 0l;
@ -474,8 +490,10 @@ public class AggregationClient {
} }
List<S> list = new ArrayList<S>(); List<S> list = new ArrayList<S>();
for (int i = 0; i < response.getFirstPartCount(); i++) { for (int i = 0; i < response.getFirstPartCount(); i++) {
list.add(ci.parseResponseAsPromotedType( ByteString b = response.getFirstPart(i);
getBytesFromResponse(response.getFirstPart(i)))); T t = ProtobufUtil.getParsedGenericInstance(ci.getClass(), 4, b);
S s = ci.getPromotedValueFromProto(t);
list.add(s);
} }
pair.setFirst(list); pair.setFirst(list);
ByteBuffer bb = ByteBuffer.allocate(8).put( ByteBuffer bb = ByteBuffer.allocate(8).put(
@ -505,7 +523,8 @@ public class AggregationClient {
* @return <R, S> * @return <R, S>
* @throws Throwable * @throws Throwable
*/ */
public <R, S> double std(final byte[] tableName, ColumnInterpreter<R, S> ci, public <R, S, P extends Message, Q extends Message, T extends Message>
double std(final byte[] tableName, ColumnInterpreter<R, S, P, Q, T> ci,
Scan scan) throws Throwable { Scan scan) throws Throwable {
Pair<List<S>, Long> p = getStdArgs(tableName, ci, scan); Pair<List<S>, Long> p = getStdArgs(tableName, ci, scan);
double res = 0d; double res = 0d;
@ -528,9 +547,10 @@ public class AggregationClient {
* (sum of values, sum of weights) for all the regions chosen * (sum of values, sum of weights) for all the regions chosen
* @throws Throwable * @throws Throwable
*/ */
private <R, S> Pair<NavigableMap<byte[], List<S>>, List<S>> private <R, S, P extends Message, Q extends Message, T extends Message>
Pair<NavigableMap<byte[], List<S>>, List<S>>
getMedianArgs(final byte[] tableName, getMedianArgs(final byte[] tableName,
final ColumnInterpreter<R, S> ci, final Scan scan) throws Throwable { final ColumnInterpreter<R, S, P, Q, T> ci, final Scan scan) throws Throwable {
final AggregateArgument requestArg = validateArgAndGetPB(scan, ci); final AggregateArgument requestArg = validateArgAndGetPB(scan, ci);
final NavigableMap<byte[], List<S>> map = final NavigableMap<byte[], List<S>> map =
new TreeMap<byte[], List<S>>(Bytes.BYTES_COMPARATOR); new TreeMap<byte[], List<S>>(Bytes.BYTES_COMPARATOR);
@ -572,8 +592,10 @@ public class AggregationClient {
List<S> list = new ArrayList<S>(); List<S> list = new ArrayList<S>();
for (int i = 0; i < response.getFirstPartCount(); i++) { for (int i = 0; i < response.getFirstPartCount(); i++) {
list.add(ci.parseResponseAsPromotedType( ByteString b = response.getFirstPart(i);
getBytesFromResponse(response.getFirstPart(i)))); T t = ProtobufUtil.getParsedGenericInstance(ci.getClass(), 4, b);
S s = ci.getPromotedValueFromProto(t);
list.add(s);
} }
return list; return list;
} }
@ -597,7 +619,8 @@ public class AggregationClient {
* @return R the median * @return R the median
* @throws Throwable * @throws Throwable
*/ */
public <R, S> R median(final byte[] tableName, ColumnInterpreter<R, S> ci, public <R, S, P extends Message, Q extends Message, T extends Message>
R median(final byte[] tableName, ColumnInterpreter<R, S, P, Q, T> ci,
Scan scan) throws Throwable { Scan scan) throws Throwable {
Pair<NavigableMap<byte[], List<S>>, List<S>> p = getMedianArgs(tableName, ci, scan); Pair<NavigableMap<byte[], List<S>>, List<S>> p = getMedianArgs(tableName, ci, scan);
byte[] startRow = null; byte[] startRow = null;
@ -672,16 +695,17 @@ public class AggregationClient {
return null; return null;
} }
<R,S>AggregateArgument validateArgAndGetPB(Scan scan, ColumnInterpreter<R,S> ci) <R, S, P extends Message, Q extends Message, T extends Message> AggregateArgument
validateArgAndGetPB(Scan scan, ColumnInterpreter<R,S,P,Q,T> ci)
throws IOException { throws IOException {
validateParameters(scan); validateParameters(scan);
final AggregateArgument.Builder requestBuilder = final AggregateArgument.Builder requestBuilder =
AggregateArgument.newBuilder(); AggregateArgument.newBuilder();
requestBuilder.setInterpreterClassName(ci.getClass().getCanonicalName()); requestBuilder.setInterpreterClassName(ci.getClass().getCanonicalName());
ByteString columnInterpreterSpecificData = null; P columnInterpreterSpecificData = null;
if ((columnInterpreterSpecificData = ci.columnInterpreterSpecificData()) if ((columnInterpreterSpecificData = ci.getRequestData())
!= null) { != null) {
requestBuilder.setInterpreterSpecificBytes(columnInterpreterSpecificData); requestBuilder.setInterpreterSpecificBytes(columnInterpreterSpecificData.toByteString());
} }
requestBuilder.setScan(ProtobufUtil.toScan(scan)); requestBuilder.setScan(ProtobufUtil.toScan(scan));
return requestBuilder.build(); return requestBuilder.build();

View File

@ -19,16 +19,15 @@
package org.apache.hadoop.hbase.client.coprocessor; package org.apache.hadoop.hbase.client.coprocessor;
import java.io.IOException; import java.io.IOException;
import java.nio.ByteBuffer;
import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.coprocessor.ColumnInterpreter; import org.apache.hadoop.hbase.coprocessor.ColumnInterpreter;
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg;
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
import com.google.protobuf.ByteString;
/** /**
* a concrete column interpreter implementation. The cell value is a Long value * a concrete column interpreter implementation. The cell value is a Long value
* and its promoted data type is also a Long value. For computing aggregation * and its promoted data type is also a Long value. For computing aggregation
@ -39,7 +38,8 @@ import com.google.protobuf.ByteString;
*/ */
@InterfaceAudience.Public @InterfaceAudience.Public
@InterfaceStability.Evolving @InterfaceStability.Evolving
public class LongColumnInterpreter implements ColumnInterpreter<Long, Long> { public class LongColumnInterpreter extends ColumnInterpreter<Long, Long,
EmptyMsg, LongMsg, LongMsg> {
public Long getValue(byte[] colFamily, byte[] colQualifier, KeyValue kv) public Long getValue(byte[] colFamily, byte[] colQualifier, KeyValue kv)
throws IOException { throws IOException {
@ -97,45 +97,40 @@ public class LongColumnInterpreter implements ColumnInterpreter<Long, Long> {
return o; return o;
} }
@Override
public Long parseResponseAsPromotedType(byte[] response) {
ByteBuffer b = ByteBuffer.allocate(8).put(response);
b.rewind();
long l = b.getLong();
return l;
}
@Override @Override
public Long castToCellType(Long l) { public Long castToCellType(Long l) {
return l; return l;
} }
@Override @Override
public ByteString columnInterpreterSpecificData() { public EmptyMsg getRequestData() {
// nothing return EmptyMsg.getDefaultInstance();
return null;
} }
@Override @Override
public void initialize(ByteString bytes) { public void initialize(EmptyMsg msg) {
//nothing //nothing
} }
@Override @Override
public ByteString getProtoForCellType(Long t) { public LongMsg getProtoForCellType(Long t) {
return getProtoForPromotedOrCellType(t); LongMsg.Builder builder = LongMsg.newBuilder();
return builder.setLongMsg(t).build();
} }
@Override @Override
public ByteString getProtoForPromotedType(Long s) { public LongMsg getProtoForPromotedType(Long s) {
return getProtoForPromotedOrCellType(s); LongMsg.Builder builder = LongMsg.newBuilder();
return builder.setLongMsg(s).build();
} }
private ByteString getProtoForPromotedOrCellType(Long s) { @Override
ByteBuffer bb = ByteBuffer.allocate(8).putLong(s); public Long getPromotedValueFromProto(LongMsg r) {
bb.rewind(); return r.getLongMsg();
ByteString bs = ByteString.copyFrom(bb); }
return bs;
@Override
public Long getCellValueFromProto(LongMsg q) {
return q.getLongMsg();
} }
} }

View File

@ -19,6 +19,10 @@
package org.apache.hadoop.hbase.coprocessor; package org.apache.hadoop.hbase.coprocessor;
import java.io.IOException; import java.io.IOException;
import java.lang.reflect.InvocationTargetException;
import java.lang.reflect.Method;
import java.lang.reflect.ParameterizedType;
import java.lang.reflect.Type;
import java.nio.ByteBuffer; import java.nio.ByteBuffer;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.List; import java.util.List;
@ -42,6 +46,7 @@ import org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateServi
import org.apache.hadoop.hbase.regionserver.InternalScanner; import org.apache.hadoop.hbase.regionserver.InternalScanner;
import com.google.protobuf.ByteString; import com.google.protobuf.ByteString;
import com.google.protobuf.Message;
import com.google.protobuf.RpcCallback; import com.google.protobuf.RpcCallback;
import com.google.protobuf.RpcController; import com.google.protobuf.RpcController;
import com.google.protobuf.Service; import com.google.protobuf.Service;
@ -49,13 +54,19 @@ import com.google.protobuf.Service;
/** /**
* A concrete AggregateProtocol implementation. Its system level coprocessor * A concrete AggregateProtocol implementation. Its system level coprocessor
* that computes the aggregate function at a region level. * that computes the aggregate function at a region level.
* @param <T> * {@link ColumnInterpreter} is used to interpret column value. This class is
* @param <S> * parameterized with the following (these are the types with which the {@link ColumnInterpreter}
* is parameterized, and for more description on these, refer to {@link ColumnInterpreter}):
* @param <T> Cell value data type
* @param <S> Promoted data type
* @param <P> PB message that is used to transport initializer specific bytes
* @param <Q> PB message that is used to transport Cell (<T>) instance
* @param <R> PB message that is used to transport Promoted (<S>) instance
*/ */
@InterfaceAudience.Public @InterfaceAudience.Public
@InterfaceStability.Evolving @InterfaceStability.Evolving
public class AggregateImplementation<T, S> extends AggregateService implements public class AggregateImplementation<T, S, P extends Message, Q extends Message, R extends Message>
CoprocessorService, Coprocessor { extends AggregateService implements CoprocessorService, Coprocessor {
protected static Log log = LogFactory.getLog(AggregateImplementation.class); protected static Log log = LogFactory.getLog(AggregateImplementation.class);
private RegionCoprocessorEnvironment env; private RegionCoprocessorEnvironment env;
@ -73,7 +84,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
AggregateResponse response = null; AggregateResponse response = null;
T max = null; T max = null;
try { try {
ColumnInterpreter<T, S> ci = constructColumnInterpreterFromRequest(request); ColumnInterpreter<T, S, P, Q, R> ci = constructColumnInterpreterFromRequest(request);
T temp; T temp;
Scan scan = ProtobufUtil.toScan(request.getScan()); Scan scan = ProtobufUtil.toScan(request.getScan());
scanner = env.getRegion().getScanner(scan); scanner = env.getRegion().getScanner(scan);
@ -96,7 +107,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
} while (hasMoreRows); } while (hasMoreRows);
if (max != null) { if (max != null) {
AggregateResponse.Builder builder = AggregateResponse.newBuilder(); AggregateResponse.Builder builder = AggregateResponse.newBuilder();
builder.addFirstPart(ci.getProtoForCellType(max)); builder.addFirstPart(ci.getProtoForCellType(max).toByteString());
response = builder.build(); response = builder.build();
} }
} catch (IOException e) { } catch (IOException e) {
@ -127,7 +138,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
InternalScanner scanner = null; InternalScanner scanner = null;
T min = null; T min = null;
try { try {
ColumnInterpreter<T, S> ci = constructColumnInterpreterFromRequest(request); ColumnInterpreter<T, S, P, Q, R> ci = constructColumnInterpreterFromRequest(request);
T temp; T temp;
Scan scan = ProtobufUtil.toScan(request.getScan()); Scan scan = ProtobufUtil.toScan(request.getScan());
scanner = env.getRegion().getScanner(scan); scanner = env.getRegion().getScanner(scan);
@ -149,7 +160,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
} while (hasMoreRows); } while (hasMoreRows);
if (min != null) { if (min != null) {
response = AggregateResponse.newBuilder().addFirstPart( response = AggregateResponse.newBuilder().addFirstPart(
ci.getProtoForCellType(min)).build(); ci.getProtoForCellType(min).toByteString()).build();
} }
} catch (IOException e) { } catch (IOException e) {
ResponseConverter.setControllerException(controller, e); ResponseConverter.setControllerException(controller, e);
@ -179,7 +190,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
InternalScanner scanner = null; InternalScanner scanner = null;
long sum = 0l; long sum = 0l;
try { try {
ColumnInterpreter<T, S> ci = constructColumnInterpreterFromRequest(request); ColumnInterpreter<T, S, P, Q, R> ci = constructColumnInterpreterFromRequest(request);
S sumVal = null; S sumVal = null;
T temp; T temp;
Scan scan = ProtobufUtil.toScan(request.getScan()); Scan scan = ProtobufUtil.toScan(request.getScan());
@ -203,7 +214,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
} while (hasMoreRows); } while (hasMoreRows);
if (sumVal != null) { if (sumVal != null) {
response = AggregateResponse.newBuilder().addFirstPart( response = AggregateResponse.newBuilder().addFirstPart(
ci.getProtoForPromotedType(sumVal)).build(); ci.getProtoForPromotedType(sumVal).toByteString()).build();
} }
} catch (IOException e) { } catch (IOException e) {
ResponseConverter.setControllerException(controller, e); ResponseConverter.setControllerException(controller, e);
@ -287,7 +298,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
AggregateResponse response = null; AggregateResponse response = null;
InternalScanner scanner = null; InternalScanner scanner = null;
try { try {
ColumnInterpreter<T, S> ci = constructColumnInterpreterFromRequest(request); ColumnInterpreter<T, S, P, Q, R> ci = constructColumnInterpreterFromRequest(request);
S sumVal = null; S sumVal = null;
Long rowCountVal = 0l; Long rowCountVal = 0l;
Scan scan = ProtobufUtil.toScan(request.getScan()); Scan scan = ProtobufUtil.toScan(request.getScan());
@ -311,7 +322,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
rowCountVal++; rowCountVal++;
} while (hasMoreRows); } while (hasMoreRows);
if (sumVal != null) { if (sumVal != null) {
ByteString first = ci.getProtoForPromotedType(sumVal); ByteString first = ci.getProtoForPromotedType(sumVal).toByteString();
AggregateResponse.Builder pair = AggregateResponse.newBuilder(); AggregateResponse.Builder pair = AggregateResponse.newBuilder();
pair.addFirstPart(first); pair.addFirstPart(first);
ByteBuffer bb = ByteBuffer.allocate(8).putLong(rowCountVal); ByteBuffer bb = ByteBuffer.allocate(8).putLong(rowCountVal);
@ -346,7 +357,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
InternalScanner scanner = null; InternalScanner scanner = null;
AggregateResponse response = null; AggregateResponse response = null;
try { try {
ColumnInterpreter<T, S> ci = constructColumnInterpreterFromRequest(request); ColumnInterpreter<T, S, P, Q, R> ci = constructColumnInterpreterFromRequest(request);
S sumVal = null, sumSqVal = null, tempVal = null; S sumVal = null, sumSqVal = null, tempVal = null;
long rowCountVal = 0l; long rowCountVal = 0l;
Scan scan = ProtobufUtil.toScan(request.getScan()); Scan scan = ProtobufUtil.toScan(request.getScan());
@ -374,8 +385,8 @@ public class AggregateImplementation<T, S> extends AggregateService implements
rowCountVal++; rowCountVal++;
} while (hasMoreRows); } while (hasMoreRows);
if (sumVal != null) { if (sumVal != null) {
ByteString first_sumVal = ci.getProtoForPromotedType(sumVal); ByteString first_sumVal = ci.getProtoForPromotedType(sumVal).toByteString();
ByteString first_sumSqVal = ci.getProtoForPromotedType(sumSqVal); ByteString first_sumSqVal = ci.getProtoForPromotedType(sumSqVal).toByteString();
AggregateResponse.Builder pair = AggregateResponse.newBuilder(); AggregateResponse.Builder pair = AggregateResponse.newBuilder();
pair.addFirstPart(first_sumVal); pair.addFirstPart(first_sumVal);
pair.addFirstPart(first_sumSqVal); pair.addFirstPart(first_sumSqVal);
@ -410,7 +421,7 @@ public class AggregateImplementation<T, S> extends AggregateService implements
AggregateResponse response = null; AggregateResponse response = null;
InternalScanner scanner = null; InternalScanner scanner = null;
try { try {
ColumnInterpreter<T, S> ci = constructColumnInterpreterFromRequest(request); ColumnInterpreter<T, S, P, Q, R> ci = constructColumnInterpreterFromRequest(request);
S sumVal = null, sumWeights = null, tempVal = null, tempWeight = null; S sumVal = null, sumWeights = null, tempVal = null, tempWeight = null;
Scan scan = ProtobufUtil.toScan(request.getScan()); Scan scan = ProtobufUtil.toScan(request.getScan());
scanner = env.getRegion().getScanner(scan); scanner = env.getRegion().getScanner(scan);
@ -442,9 +453,9 @@ public class AggregateImplementation<T, S> extends AggregateService implements
sumVal = ci.add(sumVal, tempVal); sumVal = ci.add(sumVal, tempVal);
sumWeights = ci.add(sumWeights, tempWeight); sumWeights = ci.add(sumWeights, tempWeight);
} while (hasMoreRows); } while (hasMoreRows);
ByteString first_sumVal = ci.getProtoForPromotedType(sumVal); ByteString first_sumVal = ci.getProtoForPromotedType(sumVal).toByteString();
S s = sumWeights == null ? ci.castToReturnType(ci.getMinValue()) : sumWeights; S s = sumWeights == null ? ci.castToReturnType(ci.getMinValue()) : sumWeights;
ByteString first_sumWeights = ci.getProtoForPromotedType(s); ByteString first_sumWeights = ci.getProtoForPromotedType(s).toByteString();
AggregateResponse.Builder pair = AggregateResponse.newBuilder(); AggregateResponse.Builder pair = AggregateResponse.newBuilder();
pair.addFirstPart(first_sumVal); pair.addFirstPart(first_sumVal);
pair.addFirstPart(first_sumWeights); pair.addFirstPart(first_sumWeights);
@ -462,15 +473,17 @@ public class AggregateImplementation<T, S> extends AggregateService implements
} }
@SuppressWarnings("unchecked") @SuppressWarnings("unchecked")
ColumnInterpreter<T,S> constructColumnInterpreterFromRequest( ColumnInterpreter<T,S,P,Q,R> constructColumnInterpreterFromRequest(
AggregateArgument request) throws IOException { AggregateArgument request) throws IOException {
String className = request.getInterpreterClassName(); String className = request.getInterpreterClassName();
Class<?> cls; Class<?> cls;
try { try {
cls = Class.forName(className); cls = Class.forName(className);
ColumnInterpreter<T,S> ci = (ColumnInterpreter<T, S>) cls.newInstance(); ColumnInterpreter<T,S,P,Q,R> ci = (ColumnInterpreter<T, S, P, Q, R>) cls.newInstance();
if (request.hasInterpreterSpecificBytes()) { if (request.hasInterpreterSpecificBytes()) {
ci.initialize(request.getInterpreterSpecificBytes()); ByteString b = request.getInterpreterSpecificBytes();
P initMsg = ProtobufUtil.getParsedGenericInstance(ci.getClass(), 2, b);
ci.initialize(initMsg);
} }
return ci; return ci;
} catch (ClassNotFoundException e) { } catch (ClassNotFoundException e) {

View File

@ -24,9 +24,11 @@ import java.io.IOException;
import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.coprocessor.AggregationClient;
import org.apache.hadoop.hbase.client.coprocessor.LongColumnInterpreter; import org.apache.hadoop.hbase.client.coprocessor.LongColumnInterpreter;
import com.google.protobuf.ByteString; import com.google.protobuf.ByteString;
import com.google.protobuf.Message;
/** /**
* Defines how value for specific column is interpreted and provides utility * Defines how value for specific column is interpreted and provides utility
@ -35,7 +37,8 @@ import com.google.protobuf.ByteString;
* handle null case gracefully. Refer to {@link LongColumnInterpreter} for an * handle null case gracefully. Refer to {@link LongColumnInterpreter} for an
* example. * example.
* <p> * <p>
* Takes two generic parameters. The cell value type of the interpreter is <T>. * Takes two generic parameters and three Message parameters.
* The cell value type of the interpreter is <T>.
* During some computations like sum, average, the return type can be different * During some computations like sum, average, the return type can be different
* than the cell value data type, for eg, sum of int cell values might overflow * than the cell value data type, for eg, sum of int cell values might overflow
* in case of a int result, we should use Long for its result. Therefore, this * in case of a int result, we should use Long for its result. Therefore, this
@ -44,12 +47,19 @@ import com.google.protobuf.ByteString;
* <S>. There is a conversion method * <S>. There is a conversion method
* {@link ColumnInterpreter#castToReturnType(Object)} which takes a <T> type and * {@link ColumnInterpreter#castToReturnType(Object)} which takes a <T> type and
* returns a <S> type. * returns a <S> type.
* The {@link AggregateImplementation} uses PB messages to initialize the
* user's ColumnInterpreter implementation, and for sending the responses
* back to {@link AggregationClient}.
* @param <T> Cell value data type * @param <T> Cell value data type
* @param <S> Promoted data type * @param <S> Promoted data type
* @param <P> PB message that is used to transport initializer specific bytes
* @param <Q> PB message that is used to transport Cell (<T>) instance
* @param <R> PB message that is used to transport Promoted (<S>) instance
*/ */
@InterfaceAudience.Public @InterfaceAudience.Public
@InterfaceStability.Evolving @InterfaceStability.Evolving
public interface ColumnInterpreter<T, S> { public abstract class ColumnInterpreter<T, S, P extends Message,
Q extends Message, R extends Message> {
/** /**
* @param colFamily * @param colFamily
@ -58,7 +68,7 @@ public interface ColumnInterpreter<T, S> {
* @return value of type T * @return value of type T
* @throws IOException * @throws IOException
*/ */
T getValue(byte[] colFamily, byte[] colQualifier, KeyValue kv) public abstract T getValue(byte[] colFamily, byte[] colQualifier, KeyValue kv)
throws IOException; throws IOException;
/** /**
@ -67,36 +77,36 @@ public interface ColumnInterpreter<T, S> {
* @return sum or non null value among (if either of them is null); otherwise * @return sum or non null value among (if either of them is null); otherwise
* returns a null. * returns a null.
*/ */
public S add(S l1, S l2); public abstract S add(S l1, S l2);
/** /**
* returns the maximum value for this type T * returns the maximum value for this type T
* @return max * @return max
*/ */
T getMaxValue(); public abstract T getMaxValue();
T getMinValue(); public abstract T getMinValue();
/** /**
* @param o1 * @param o1
* @param o2 * @param o2
* @return multiplication * @return multiplication
*/ */
S multiply(S o1, S o2); public abstract S multiply(S o1, S o2);
/** /**
* @param o * @param o
* @return increment * @return increment
*/ */
S increment(S o); public abstract S increment(S o);
/** /**
* provides casting opportunity between the data types. * provides casting opportunity between the data types.
* @param o * @param o
* @return cast * @return cast
*/ */
S castToReturnType(T o); public abstract S castToReturnType(T o);
/** /**
* This takes care if either of arguments are null. returns 0 if they are * This takes care if either of arguments are null. returns 0 if they are
@ -105,7 +115,7 @@ public interface ColumnInterpreter<T, S> {
* <li>>0 if l1 > l2 or l1 is not null and l2 is null. * <li>>0 if l1 > l2 or l1 is not null and l2 is null.
* <li>< 0 if l1 < l2 or l1 is null and l2 is not null. * <li>< 0 if l1 < l2 or l1 is null and l2 is not null.
*/ */
int compare(final T l1, final T l2); public abstract int compare(final T l1, final T l2);
/** /**
* used for computing average of <S> data values. Not providing the divide * used for computing average of <S> data values. Not providing the divide
@ -114,45 +124,52 @@ public interface ColumnInterpreter<T, S> {
* @param l * @param l
* @return Average * @return Average
*/ */
double divideForAvg(S o, Long l); public abstract double divideForAvg(S o, Long l);
/** /**
* This method should return any additional data that is needed on the * This method should return any additional data that is needed on the
* server side to construct the ColumnInterpreter. The server * server side to construct the ColumnInterpreter. The server
* will pass this to the {@link #initialize(ByteString)} * will pass this to the {@link #initialize}
* method. If there is no ColumnInterpreter specific data (for e.g., * method. If there is no ColumnInterpreter specific data (for e.g.,
* {@link LongColumnInterpreter}) then null should be returned. * {@link LongColumnInterpreter}) then null should be returned.
* @return the PB message * @return the PB message
*/ */
ByteString columnInterpreterSpecificData(); public abstract P getRequestData();
/**
* Return the PB for type T
* @param t
* @return PB-message
*/
ByteString getProtoForCellType(T t);
/**
* Return the PB for type S
* @param s
* @return PB-message
*/
ByteString getProtoForPromotedType(S s);
/** /**
* This method should initialize any field(s) of the ColumnInterpreter with * This method should initialize any field(s) of the ColumnInterpreter with
* a parsing of the passed message bytes (used on the server side). * a parsing of the passed message bytes (used on the server side).
* @param bytes * @param msg
*/ */
void initialize(ByteString bytes); public abstract void initialize(P msg);
/** /**
* Converts the bytes in the server's response to the expected type S * This method gets the PB message corresponding to the cell type
* @param response * @param t
* @return response of type S constructed from the message * @return the PB message for the cell-type instance
*/ */
S parseResponseAsPromotedType(byte[] response); public abstract Q getProtoForCellType(T t);
/**
* This method gets the PB message corresponding to the cell type
* @param q
* @return the cell-type instance from the PB message
*/
public abstract T getCellValueFromProto(Q q);
/**
* This method gets the PB message corresponding to the promoted type
* @param s
* @return the PB message for the promoted-type instance
*/
public abstract R getProtoForPromotedType(S s);
/**
* This method gets the promoted type from the proto message
* @param r
* @return the promoted-type instance from the PB message
*/
public abstract S getPromotedValueFromProto(R r);
/** /**
* The response message comes as type S. This will convert/cast it to T. * The response message comes as type S. This will convert/cast it to T.
@ -160,5 +177,5 @@ public interface ColumnInterpreter<T, S> {
* @param response * @param response
* @return cast * @return cast
*/ */
T castToCellType(S response); public abstract T castToCellType(S response);
} }

View File

@ -22,7 +22,10 @@ import static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpeci
import java.io.ByteArrayOutputStream; import java.io.ByteArrayOutputStream;
import java.io.IOException; import java.io.IOException;
import java.lang.reflect.Constructor; import java.lang.reflect.Constructor;
import java.lang.reflect.InvocationTargetException;
import java.lang.reflect.Method; import java.lang.reflect.Method;
import java.lang.reflect.ParameterizedType;
import java.lang.reflect.Type;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Collection; import java.util.Collection;
import java.util.HashMap; import java.util.HashMap;
@ -1836,4 +1839,39 @@ public final class ProtobufUtil {
KeyValue.Type.codeToType((byte)kv.getKeyType().getNumber()), KeyValue.Type.codeToType((byte)kv.getKeyType().getNumber()),
kv.getValue().toByteArray()); kv.getValue().toByteArray());
} }
/**
* Get an instance of the argument type declared in a class's signature. The
* argument type is assumed to be a PB Message subclass, and the instance is
* created using parseFrom method on the passed ByteString.
* @param runtimeClass the runtime type of the class
* @param position the position of the argument in the class declaration
* @param b the ByteString which should be parsed to get the instance created
* @return the instance
* @throws IOException
*/
@SuppressWarnings("unchecked")
public static <T extends Message>
T getParsedGenericInstance(Class<?> runtimeClass, int position, ByteString b)
throws IOException {
Type type = runtimeClass.getGenericSuperclass();
Type argType = ((ParameterizedType)type).getActualTypeArguments()[position];
Class<T> classType = (Class<T>)argType;
T inst;
try {
Method m = classType.getMethod("parseFrom", ByteString.class);
inst = (T)m.invoke(null, b);
return inst;
} catch (SecurityException e) {
throw new IOException(e);
} catch (NoSuchMethodException e) {
throw new IOException(e);
} catch (IllegalArgumentException e) {
throw new IOException(e);
} catch (InvocationTargetException e) {
throw new IOException(e);
} catch (IllegalAccessException e) {
throw new IOException(e);
}
}
} }

View File

@ -31,6 +31,8 @@ import org.apache.hadoop.hbase.client.coprocessor.AggregationClient;
import org.apache.hadoop.hbase.client.coprocessor.LongColumnInterpreter; import org.apache.hadoop.hbase.client.coprocessor.LongColumnInterpreter;
import org.apache.hadoop.hbase.filter.Filter; import org.apache.hadoop.hbase.filter.Filter;
import org.apache.hadoop.hbase.filter.PrefixFilter; import org.apache.hadoop.hbase.filter.PrefixFilter;
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.EmptyMsg;
import org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LongMsg;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
import org.junit.AfterClass; import org.junit.AfterClass;
import org.junit.BeforeClass; import org.junit.BeforeClass;
@ -131,7 +133,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addColumn(TEST_FAMILY,TEST_QUALIFIER); scan.addColumn(TEST_FAMILY,TEST_QUALIFIER);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long median = aClient.median(TEST_TABLE, ci, long median = aClient.median(TEST_TABLE, ci,
scan); scan);
assertEquals(8L, median); assertEquals(8L, median);
@ -153,7 +156,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY, TEST_QUALIFIER); scan.addColumn(TEST_FAMILY, TEST_QUALIFIER);
scan.setStartRow(ROWS[2]); scan.setStartRow(ROWS[2]);
scan.setStopRow(ROWS[14]); scan.setStopRow(ROWS[14]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long rowCount = aClient.rowCount(TEST_TABLE, ci, scan); long rowCount = aClient.rowCount(TEST_TABLE, ci, scan);
assertEquals(12, rowCount); assertEquals(12, rowCount);
} }
@ -168,7 +172,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addColumn(TEST_FAMILY, TEST_QUALIFIER); scan.addColumn(TEST_FAMILY, TEST_QUALIFIER);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long rowCount = aClient.rowCount(TEST_TABLE, ci, long rowCount = aClient.rowCount(TEST_TABLE, ci,
scan); scan);
assertEquals(ROWSIZE, rowCount); assertEquals(ROWSIZE, rowCount);
@ -187,7 +192,8 @@ public class TestAggregateProtocol {
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[2]); scan.setStopRow(ROWS[2]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long rowCount = -1; long rowCount = -1;
try { try {
rowCount = aClient.rowCount(TEST_TABLE, ci, scan); rowCount = aClient.rowCount(TEST_TABLE, ci, scan);
@ -211,7 +217,8 @@ public class TestAggregateProtocol {
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[5]); scan.setStopRow(ROWS[5]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long rowCount = -1; long rowCount = -1;
try { try {
rowCount = aClient.rowCount(TEST_TABLE, ci, scan); rowCount = aClient.rowCount(TEST_TABLE, ci, scan);
@ -230,7 +237,8 @@ public class TestAggregateProtocol {
Scan scan = new Scan(); Scan scan = new Scan();
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[15]); scan.setStopRow(ROWS[15]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long rowCount = -1; long rowCount = -1;
try { try {
rowCount = aClient.rowCount(TEST_TABLE, ci, scan); rowCount = aClient.rowCount(TEST_TABLE, ci, scan);
@ -245,7 +253,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long rowCount = aClient.rowCount(TEST_TABLE, ci, long rowCount = aClient.rowCount(TEST_TABLE, ci,
scan); scan);
assertEquals(20, rowCount); assertEquals(20, rowCount);
@ -256,7 +265,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addColumn(TEST_FAMILY, TEST_QUALIFIER); scan.addColumn(TEST_FAMILY, TEST_QUALIFIER);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Filter f = new PrefixFilter(Bytes.toBytes("foo:bar")); Filter f = new PrefixFilter(Bytes.toBytes("foo:bar"));
scan.setFilter(f); scan.setFilter(f);
long rowCount = aClient.rowCount(TEST_TABLE, ci, long rowCount = aClient.rowCount(TEST_TABLE, ci,
@ -277,7 +287,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addColumn(TEST_FAMILY, TEST_QUALIFIER); scan.addColumn(TEST_FAMILY, TEST_QUALIFIER);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long maximum = aClient.max(TEST_TABLE, ci, scan); long maximum = aClient.max(TEST_TABLE, ci, scan);
assertEquals(19, maximum); assertEquals(19, maximum);
} }
@ -292,7 +303,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY, TEST_QUALIFIER); scan.addColumn(TEST_FAMILY, TEST_QUALIFIER);
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[15]); scan.setStopRow(ROWS[15]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long max = aClient.max(TEST_TABLE, ci, scan); long max = aClient.max(TEST_TABLE, ci, scan);
assertEquals(14, max); assertEquals(14, max);
} }
@ -302,7 +314,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long maximum = aClient.max(TEST_TABLE, ci, scan); long maximum = aClient.max(TEST_TABLE, ci, scan);
assertEquals(190, maximum); assertEquals(190, maximum);
} }
@ -314,7 +327,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[7]); scan.setStopRow(ROWS[7]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long max = aClient.max(TEST_TABLE, ci, scan); long max = aClient.max(TEST_TABLE, ci, scan);
assertEquals(60, max); assertEquals(60, max);
} }
@ -322,7 +336,8 @@ public class TestAggregateProtocol {
@Test @Test
public void testMaxWithValidRangeWithNullCF() { public void testMaxWithValidRangeWithNullCF() {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Scan scan = new Scan(); Scan scan = new Scan();
Long max = null; Long max = null;
try { try {
@ -337,7 +352,8 @@ public class TestAggregateProtocol {
@Test @Test
public void testMaxWithInvalidRange() { public void testMaxWithInvalidRange() {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Scan scan = new Scan(); Scan scan = new Scan();
scan.setStartRow(ROWS[4]); scan.setStartRow(ROWS[4]);
scan.setStopRow(ROWS[2]); scan.setStopRow(ROWS[2]);
@ -360,7 +376,8 @@ public class TestAggregateProtocol {
scan.setStopRow(ROWS[4]); scan.setStopRow(ROWS[4]);
try { try {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
max = aClient.max(TEST_TABLE, ci, scan); max = aClient.max(TEST_TABLE, ci, scan);
} catch (Exception e) { } catch (Exception e) {
max = 0; max = 0;
@ -376,7 +393,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY, TEST_QUALIFIER); scan.addColumn(TEST_FAMILY, TEST_QUALIFIER);
Filter f = new PrefixFilter(Bytes.toBytes("foo:bar")); Filter f = new PrefixFilter(Bytes.toBytes("foo:bar"));
scan.setFilter(f); scan.setFilter(f);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
max = aClient.max(TEST_TABLE, ci, scan); max = aClient.max(TEST_TABLE, ci, scan);
assertEquals(null, max); assertEquals(null, max);
} }
@ -395,7 +413,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY, TEST_QUALIFIER); scan.addColumn(TEST_FAMILY, TEST_QUALIFIER);
scan.setStartRow(HConstants.EMPTY_START_ROW); scan.setStartRow(HConstants.EMPTY_START_ROW);
scan.setStopRow(HConstants.EMPTY_END_ROW); scan.setStopRow(HConstants.EMPTY_END_ROW);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Long min = aClient.min(TEST_TABLE, ci, Long min = aClient.min(TEST_TABLE, ci,
scan); scan);
assertEquals(0l, min.longValue()); assertEquals(0l, min.longValue());
@ -411,7 +430,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY, TEST_QUALIFIER); scan.addColumn(TEST_FAMILY, TEST_QUALIFIER);
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[15]); scan.setStopRow(ROWS[15]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long min = aClient.min(TEST_TABLE, ci, scan); long min = aClient.min(TEST_TABLE, ci, scan);
assertEquals(5, min); assertEquals(5, min);
} }
@ -423,7 +443,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(HConstants.EMPTY_START_ROW); scan.setStartRow(HConstants.EMPTY_START_ROW);
scan.setStopRow(HConstants.EMPTY_END_ROW); scan.setStopRow(HConstants.EMPTY_END_ROW);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long min = aClient.min(TEST_TABLE, ci, long min = aClient.min(TEST_TABLE, ci,
scan); scan);
assertEquals(0, min); assertEquals(0, min);
@ -436,7 +457,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[7]); scan.setStopRow(ROWS[7]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long min = aClient.min(TEST_TABLE, ci, scan); long min = aClient.min(TEST_TABLE, ci, scan);
assertEquals(6, min); assertEquals(6, min);
} }
@ -447,7 +469,8 @@ public class TestAggregateProtocol {
Scan scan = new Scan(); Scan scan = new Scan();
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[15]); scan.setStopRow(ROWS[15]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Long min = null; Long min = null;
try { try {
min = aClient.min(TEST_TABLE, ci, scan); min = aClient.min(TEST_TABLE, ci, scan);
@ -465,7 +488,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(ROWS[4]); scan.setStartRow(ROWS[4]);
scan.setStopRow(ROWS[2]); scan.setStopRow(ROWS[2]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
try { try {
min = aClient.min(TEST_TABLE, ci, scan); min = aClient.min(TEST_TABLE, ci, scan);
} catch (Throwable e) { } catch (Throwable e) {
@ -480,7 +504,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[6]); scan.setStopRow(ROWS[6]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Long min = null; Long min = null;
try { try {
min = aClient.min(TEST_TABLE, ci, scan); min = aClient.min(TEST_TABLE, ci, scan);
@ -496,7 +521,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY, TEST_QUALIFIER); scan.addColumn(TEST_FAMILY, TEST_QUALIFIER);
Filter f = new PrefixFilter(Bytes.toBytes("foo:bar")); Filter f = new PrefixFilter(Bytes.toBytes("foo:bar"));
scan.setFilter(f); scan.setFilter(f);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Long min = null; Long min = null;
min = aClient.min(TEST_TABLE, ci, scan); min = aClient.min(TEST_TABLE, ci, scan);
assertEquals(null, min); assertEquals(null, min);
@ -513,7 +539,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addColumn(TEST_FAMILY,TEST_QUALIFIER); scan.addColumn(TEST_FAMILY,TEST_QUALIFIER);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long sum = aClient.sum(TEST_TABLE, ci, long sum = aClient.sum(TEST_TABLE, ci,
scan); scan);
assertEquals(190, sum); assertEquals(190, sum);
@ -529,7 +556,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY,TEST_QUALIFIER); scan.addColumn(TEST_FAMILY,TEST_QUALIFIER);
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[15]); scan.setStopRow(ROWS[15]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long sum = aClient.sum(TEST_TABLE, ci, scan); long sum = aClient.sum(TEST_TABLE, ci, scan);
assertEquals(95, sum); assertEquals(95, sum);
} }
@ -539,7 +567,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long sum = aClient.sum(TEST_TABLE, ci, long sum = aClient.sum(TEST_TABLE, ci,
scan); scan);
assertEquals(190 + 1900, sum); assertEquals(190 + 1900, sum);
@ -552,7 +581,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[7]); scan.setStopRow(ROWS[7]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
long sum = aClient.sum(TEST_TABLE, ci, scan); long sum = aClient.sum(TEST_TABLE, ci, scan);
assertEquals(6 + 60, sum); assertEquals(6 + 60, sum);
} }
@ -563,7 +593,8 @@ public class TestAggregateProtocol {
Scan scan = new Scan(); Scan scan = new Scan();
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[7]); scan.setStopRow(ROWS[7]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Long sum = null; Long sum = null;
try { try {
sum = aClient.sum(TEST_TABLE, ci, scan); sum = aClient.sum(TEST_TABLE, ci, scan);
@ -580,7 +611,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[2]); scan.setStopRow(ROWS[2]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Long sum = null; Long sum = null;
try { try {
sum = aClient.sum(TEST_TABLE, ci, scan); sum = aClient.sum(TEST_TABLE, ci, scan);
@ -596,7 +628,8 @@ public class TestAggregateProtocol {
Scan scan = new Scan(); Scan scan = new Scan();
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setFilter(f); scan.setFilter(f);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Long sum = null; Long sum = null;
sum = aClient.sum(TEST_TABLE, ci, scan); sum = aClient.sum(TEST_TABLE, ci, scan);
assertEquals(null, sum); assertEquals(null, sum);
@ -613,7 +646,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addColumn(TEST_FAMILY,TEST_QUALIFIER); scan.addColumn(TEST_FAMILY,TEST_QUALIFIER);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
double avg = aClient.avg(TEST_TABLE, ci, double avg = aClient.avg(TEST_TABLE, ci,
scan); scan);
assertEquals(9.5, avg, 0); assertEquals(9.5, avg, 0);
@ -629,7 +663,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY,TEST_QUALIFIER); scan.addColumn(TEST_FAMILY,TEST_QUALIFIER);
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[15]); scan.setStopRow(ROWS[15]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
double avg = aClient.avg(TEST_TABLE, ci, scan); double avg = aClient.avg(TEST_TABLE, ci, scan);
assertEquals(9.5, avg, 0); assertEquals(9.5, avg, 0);
} }
@ -639,7 +674,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
double avg = aClient.avg(TEST_TABLE, ci, double avg = aClient.avg(TEST_TABLE, ci,
scan); scan);
assertEquals(104.5, avg, 0); assertEquals(104.5, avg, 0);
@ -652,7 +688,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[7]); scan.setStopRow(ROWS[7]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
double avg = aClient.avg(TEST_TABLE, ci, scan); double avg = aClient.avg(TEST_TABLE, ci, scan);
assertEquals(6 + 60, avg, 0); assertEquals(6 + 60, avg, 0);
} }
@ -661,7 +698,8 @@ public class TestAggregateProtocol {
public void testAvgWithValidRangeWithNullCF() { public void testAvgWithValidRangeWithNullCF() {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Double avg = null; Double avg = null;
try { try {
avg = aClient.avg(TEST_TABLE, ci, scan); avg = aClient.avg(TEST_TABLE, ci, scan);
@ -678,7 +716,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY,TEST_QUALIFIER); scan.addColumn(TEST_FAMILY,TEST_QUALIFIER);
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[1]); scan.setStopRow(ROWS[1]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Double avg = null; Double avg = null;
try { try {
avg = aClient.avg(TEST_TABLE, ci, scan); avg = aClient.avg(TEST_TABLE, ci, scan);
@ -694,7 +733,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY,TEST_QUALIFIER); scan.addColumn(TEST_FAMILY,TEST_QUALIFIER);
Filter f = new PrefixFilter(Bytes.toBytes("foo:bar")); Filter f = new PrefixFilter(Bytes.toBytes("foo:bar"));
scan.setFilter(f); scan.setFilter(f);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Double avg = null; Double avg = null;
avg = aClient.avg(TEST_TABLE, ci, scan); avg = aClient.avg(TEST_TABLE, ci, scan);
assertEquals(Double.NaN, avg, 0); assertEquals(Double.NaN, avg, 0);
@ -711,7 +751,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addColumn(TEST_FAMILY,TEST_QUALIFIER); scan.addColumn(TEST_FAMILY,TEST_QUALIFIER);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
double std = aClient.std(TEST_TABLE, ci, double std = aClient.std(TEST_TABLE, ci,
scan); scan);
assertEquals(5.766, std, 0.05d); assertEquals(5.766, std, 0.05d);
@ -727,7 +768,8 @@ public class TestAggregateProtocol {
scan.addColumn(TEST_FAMILY,TEST_QUALIFIER); scan.addColumn(TEST_FAMILY,TEST_QUALIFIER);
scan.setStartRow(ROWS[5]); scan.setStartRow(ROWS[5]);
scan.setStopRow(ROWS[15]); scan.setStopRow(ROWS[15]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
double std = aClient.std(TEST_TABLE, ci, scan); double std = aClient.std(TEST_TABLE, ci, scan);
assertEquals(2.87, std, 0.05d); assertEquals(2.87, std, 0.05d);
} }
@ -737,7 +779,8 @@ public class TestAggregateProtocol {
AggregationClient aClient = new AggregationClient(conf); AggregationClient aClient = new AggregationClient(conf);
Scan scan = new Scan(); Scan scan = new Scan();
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
double std = aClient.std(TEST_TABLE, ci, double std = aClient.std(TEST_TABLE, ci,
scan); scan);
assertEquals(63.42, std, 0.05d); assertEquals(63.42, std, 0.05d);
@ -750,7 +793,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[7]); scan.setStopRow(ROWS[7]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
double std = aClient.std(TEST_TABLE, ci, scan); double std = aClient.std(TEST_TABLE, ci, scan);
assertEquals(0, std, 0); assertEquals(0, std, 0);
} }
@ -761,7 +805,8 @@ public class TestAggregateProtocol {
Scan scan = new Scan(); Scan scan = new Scan();
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[17]); scan.setStopRow(ROWS[17]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Double std = null; Double std = null;
try { try {
std = aClient.std(TEST_TABLE, ci, scan); std = aClient.std(TEST_TABLE, ci, scan);
@ -778,7 +823,8 @@ public class TestAggregateProtocol {
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setStartRow(ROWS[6]); scan.setStartRow(ROWS[6]);
scan.setStopRow(ROWS[1]); scan.setStopRow(ROWS[1]);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Double std = null; Double std = null;
try { try {
std = aClient.std(TEST_TABLE, ci, scan); std = aClient.std(TEST_TABLE, ci, scan);
@ -794,7 +840,8 @@ public class TestAggregateProtocol {
Scan scan = new Scan(); Scan scan = new Scan();
scan.addFamily(TEST_FAMILY); scan.addFamily(TEST_FAMILY);
scan.setFilter(f); scan.setFilter(f);
final ColumnInterpreter<Long, Long> ci = new LongColumnInterpreter(); final ColumnInterpreter<Long, Long, EmptyMsg, LongMsg, LongMsg> ci =
new LongColumnInterpreter();
Double std = null; Double std = null;
std = aClient.std(TEST_TABLE, ci, scan); std = aClient.std(TEST_TABLE, ci, scan);
assertEquals(Double.NaN, std, 0); assertEquals(Double.NaN, std, 0);