HBASE-1714

git-svn-id: https://svn.apache.org/repos/asf/hadoop/hbase/trunk@798334 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Ryan Rawson 2009-07-27 23:48:22 +00:00
parent ace2be6b00
commit a6c6f77612
17 changed files with 981 additions and 112 deletions

View File

@ -521,6 +521,7 @@ Release 0.20.0 - Unreleased
(Tim Sell via Stack)
HBASE-1683 OOME on master splitting logs; stuck, won't go down
HBASE-1704 Better zk error when failed connect
HBASE-1714 Thrift server: prefix scan API
OPTIMIZATIONS
HBASE-1412 Change values for delete column and column family in KeyValue

View File

@ -228,6 +228,11 @@ public class Delete implements Writable {
familyMap.put(family, list);
return this;
}
public void deleteColumns(byte [] column) {
byte [][] parts = KeyValue.parseColumn(column);
this.deleteColumns(parts[0], parts[1]);
}
/**
* Delete the latest version of the specified column, given in

View File

@ -39,6 +39,7 @@ import org.apache.hadoop.hbase.io.ImmutableBytesWritable;
import org.apache.hadoop.hbase.io.RowResult;
import org.apache.hadoop.hbase.regionserver.HRegion;
import org.apache.hadoop.hbase.util.Writables;
import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.mapred.InputFormat;
import org.apache.hadoop.mapred.InputSplit;
import org.apache.hadoop.mapred.JobConf;
@ -106,21 +107,22 @@ implements InputFormat<ImmutableBytesWritable, RowResult> {
public void restart(byte[] firstRow) throws IOException {
if ((endRow != null) && (endRow.length > 0)) {
if (trrRowFilter != null) {
final Set<RowFilterInterface> rowFiltersSet =
new HashSet<RowFilterInterface>();
rowFiltersSet.add(new WhileMatchRowFilter(new StopRowFilter(endRow)));
rowFiltersSet.add(trrRowFilter);
Scan scan = new Scan(startRow);
Scan scan = new Scan(firstRow, endRow);
scan.addColumns(trrInputColumns);
// scan.setFilter(new RowFilterSet(RowFilterSet.Operator.MUST_PASS_ALL,
// rowFiltersSet));
scan.setOldFilter(trrRowFilter);
this.scanner = this.htable.getScanner(scan);
} else {
LOG.debug("TIFB.restart, firstRow: " +
Bytes.toStringBinary(firstRow) + ", endRow: " +
Bytes.toStringBinary(endRow));
Scan scan = new Scan(firstRow, endRow);
scan.addColumns(trrInputColumns);
this.scanner = this.htable.getScanner(scan);
}
} else {
LOG.debug("TIFB.restart, firstRow: " +
Bytes.toStringBinary(firstRow) + ", no endRow");
Scan scan = new Scan(firstRow);
scan.addColumns(trrInputColumns);
// scan.setFilter(trrRowFilter);

View File

@ -1,3 +1,23 @@
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.hbase.regionserver;
import org.apache.hadoop.hbase.KeyValue;
@ -69,7 +89,6 @@ public class DeleteCompare {
if(res > 0) {
return DeleteCode.DONE;
} else if(res < 0){
System.out.println("SKIPPING ROW");
return DeleteCode.SKIP;
}
@ -113,7 +132,6 @@ public class DeleteCompare {
}
return DeleteCode.DONE;
} else {
System.out.println("SKIPPING TS");
return DeleteCode.SKIP;
}
}

View File

@ -74,12 +74,11 @@ struct ColumnDescriptor {
2:i32 maxVersions = 3,
3:string compression = "NONE",
4:bool inMemory = 0,
5:i32 maxValueLength = 2147483647,
6:string bloomFilterType = "NONE",
7:i32 bloomFilterVectorSize = 0,
8:i32 bloomFilterNbHashes = 0,
9:bool blockCacheEnabled = 0,
10:i32 timeToLive = -1
5:string bloomFilterType = "NONE",
6:i32 bloomFilterVectorSize = 0,
7:i32 bloomFilterNbHashes = 0,
8:bool blockCacheEnabled = 0,
9:i32 timeToLive = -1
}
/**
@ -463,6 +462,20 @@ service Hbase {
4:list<Text> columns)
throws (1:IOError io)
/**
* Open a scanner for a given prefix. That is all rows will have the specified
* prefix. No other rows will be returned.
*
* @param tableName name of table
* @param startAndPrefix the prefix (and thus start row) of the keys you want
* @param columns the columns you want returned
* @return scanner id to use with other scanner calls
*/
ScannerID scannerOpenWithPrefix(1:Text tableName,
2:Text startAndPrefix,
3:list<Text> columns)
throws (1:IOError io)
/**
* Get a scanner on the current table starting at the specified row and
* ending at the last row in the table. Return the specified columns.

View File

@ -35,6 +35,9 @@ import org.apache.hadoop.hbase.HServerAddress;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.MasterNotRunningException;
import org.apache.hadoop.hbase.filter.Filter;
import org.apache.hadoop.hbase.filter.WhileMatchFilter;
import org.apache.hadoop.hbase.filter.PrefixFilter;
import org.apache.hadoop.hbase.client.Delete;
import org.apache.hadoop.hbase.client.Get;
import org.apache.hadoop.hbase.client.HBaseAdmin;
@ -603,7 +606,24 @@ public class ThriftServer {
throw new IOError(e.getMessage());
}
}
@Override
public int scannerOpenWithPrefix(byte[] tableName, byte[] startAndPrefix, List<byte[]> columns) throws IOError, TException {
try {
HTable table = getTable(tableName);
byte [][] columnsArray = null;
columnsArray = columns.toArray(new byte[0][]);
Scan scan = new Scan(startAndPrefix);
scan.addColumns(columnsArray);
Filter f = new WhileMatchFilter(
new PrefixFilter(startAndPrefix));
scan.setFilter(f);
return addScanner(table.getScanner(scan));
} catch (IOException e) {
throw new IOError(e.getMessage());
}
}
public int scannerOpenTs(byte[] tableName, byte[] startRow,
List<byte[]> columns, long timestamp) throws IOError, TException {
try {

View File

@ -1,4 +1,6 @@
/**
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
@ -15,11 +17,7 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Autogenerated by Thrift
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
package org.apache.hadoop.hbase.thrift.generated;
import java.util.List;

View File

@ -1,4 +1,6 @@
/**
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
@ -15,11 +17,6 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Autogenerated by Thrift
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
package org.apache.hadoop.hbase.thrift.generated;
import java.util.List;

View File

@ -1,4 +1,6 @@
/**
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
@ -15,11 +17,6 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Autogenerated by Thrift
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
package org.apache.hadoop.hbase.thrift.generated;
import java.util.List;
@ -129,7 +126,6 @@ public class ColumnDescriptor implements TBase, java.io.Serializable, Cloneable
int maxVersions,
String compression,
boolean inMemory,
int maxValueLength,
String bloomFilterType,
int bloomFilterVectorSize,
int bloomFilterNbHashes,

File diff suppressed because it is too large Load Diff

View File

@ -1,4 +1,6 @@
/**
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
@ -15,11 +17,6 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Autogenerated by Thrift
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
package org.apache.hadoop.hbase.thrift.generated;
import java.util.List;

View File

@ -1,4 +1,6 @@
/**
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
@ -15,11 +17,6 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Autogenerated by Thrift
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
package org.apache.hadoop.hbase.thrift.generated;
import java.util.List;

View File

@ -1,4 +1,6 @@
/**
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
@ -15,11 +17,6 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Autogenerated by Thrift
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
package org.apache.hadoop.hbase.thrift.generated;
import java.util.List;

View File

@ -1,4 +1,6 @@
/**
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
@ -15,11 +17,6 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Autogenerated by Thrift
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
package org.apache.hadoop.hbase.thrift.generated;
import java.util.List;

View File

@ -1,4 +1,6 @@
/**
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
@ -15,11 +17,6 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Autogenerated by Thrift
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
package org.apache.hadoop.hbase.thrift.generated;
import java.util.List;

View File

@ -1,4 +1,6 @@
/**
/*
* Copyright 2009 The Apache Software Foundation
*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
@ -15,11 +17,6 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/**
* Autogenerated by Thrift
*
* DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
*/
package org.apache.hadoop.hbase.thrift.generated;
import java.util.List;

View File

@ -322,7 +322,7 @@ public class TestThriftServer extends HBaseClusterTestCase {
// A slightly customized ColumnDescriptor (only 2 versions)
ColumnDescriptor cDescB = new ColumnDescriptor(columnBname, 2, "NONE",
false, 2147483647, "NONE", 0, 0, false, -1);
false, "NONE", 0, 0, false, -1);
cDescriptors.add(cDescB);
return cDescriptors;