HBASE-18983 fixes from update error-prone to 2.1.1
This commit is contained in:
parent
d592b29619
commit
33ae6dce42
|
@ -27,6 +27,7 @@ import java.io.IOException;
|
||||||
import java.io.InterruptedIOException;
|
import java.io.InterruptedIOException;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Arrays;
|
import java.util.Arrays;
|
||||||
|
import java.util.Collections;
|
||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
import java.util.Iterator;
|
import java.util.Iterator;
|
||||||
import java.util.LinkedList;
|
import java.util.LinkedList;
|
||||||
|
@ -746,12 +747,12 @@ public class TestAsyncProcess {
|
||||||
puts.add(createPut(1, true));
|
puts.add(createPut(1, true));
|
||||||
|
|
||||||
for (int i = 0; i != controller.maxConcurrentTasksPerRegion; ++i) {
|
for (int i = 0; i != controller.maxConcurrentTasksPerRegion; ++i) {
|
||||||
ap.incTaskCounters(Arrays.asList(hri1.getRegionName()), sn);
|
ap.incTaskCounters(Collections.singleton(hri1.getRegionName()), sn);
|
||||||
}
|
}
|
||||||
ap.submit(null, DUMMY_TABLE, puts, false, null, false);
|
ap.submit(null, DUMMY_TABLE, puts, false, null, false);
|
||||||
Assert.assertEquals(puts.size(), 1);
|
Assert.assertEquals(puts.size(), 1);
|
||||||
|
|
||||||
ap.decTaskCounters(Arrays.asList(hri1.getRegionName()), sn);
|
ap.decTaskCounters(Collections.singleton(hri1.getRegionName()), sn);
|
||||||
ap.submit(null, DUMMY_TABLE, puts, false, null, false);
|
ap.submit(null, DUMMY_TABLE, puts, false, null, false);
|
||||||
Assert.assertEquals(0, puts.size());
|
Assert.assertEquals(0, puts.size());
|
||||||
if (defaultClazz != null) {
|
if (defaultClazz != null) {
|
||||||
|
@ -966,7 +967,7 @@ public class TestAsyncProcess {
|
||||||
|
|
||||||
|
|
||||||
for (int i = 0; i < 1000; i++) {
|
for (int i = 0; i < 1000; i++) {
|
||||||
ap.incTaskCounters(Arrays.asList("dummy".getBytes()), sn);
|
ap.incTaskCounters(Collections.singleton("dummy".getBytes()), sn);
|
||||||
}
|
}
|
||||||
|
|
||||||
final Thread myThread = Thread.currentThread();
|
final Thread myThread = Thread.currentThread();
|
||||||
|
@ -997,7 +998,7 @@ public class TestAsyncProcess {
|
||||||
public void run() {
|
public void run() {
|
||||||
Threads.sleep(sleepTime);
|
Threads.sleep(sleepTime);
|
||||||
while (controller.tasksInProgress.get() > 0) {
|
while (controller.tasksInProgress.get() > 0) {
|
||||||
ap.decTaskCounters(Arrays.asList("dummy".getBytes()), sn);
|
ap.decTaskCounters(Collections.singleton("dummy".getBytes()), sn);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
@ -1119,6 +1120,7 @@ public class TestAsyncProcess {
|
||||||
Assert.assertEquals("the put should not been inserted.", 0, mutator.size());
|
Assert.assertEquals("the put should not been inserted.", 0, mutator.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@SuppressWarnings("SelfComparison")
|
||||||
@Test
|
@Test
|
||||||
public void testAction() {
|
public void testAction() {
|
||||||
Action action_0 = new Action(new Put(Bytes.toBytes("abc")), 10);
|
Action action_0 = new Action(new Put(Bytes.toBytes("abc")), 10);
|
||||||
|
|
|
@ -96,6 +96,7 @@ public class TestZKUtil {
|
||||||
Assert.assertTrue(aclList.contains(new ACL(Perms.ALL, new Id("sasl", "user6"))));
|
Assert.assertTrue(aclList.contains(new ACL(Perms.ALL, new Id("sasl", "user6"))));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test(expected = KeeperException.SystemErrorException.class)
|
||||||
public void testInterruptedDuringAction()
|
public void testInterruptedDuringAction()
|
||||||
throws ZooKeeperConnectionException, IOException, KeeperException, InterruptedException {
|
throws ZooKeeperConnectionException, IOException, KeeperException, InterruptedException {
|
||||||
final RecoverableZooKeeper recoverableZk = Mockito.mock(RecoverableZooKeeper.class);
|
final RecoverableZooKeeper recoverableZk = Mockito.mock(RecoverableZooKeeper.class);
|
||||||
|
@ -107,12 +108,6 @@ public class TestZKUtil {
|
||||||
};
|
};
|
||||||
Mockito.doThrow(new InterruptedException()).when(recoverableZk)
|
Mockito.doThrow(new InterruptedException()).when(recoverableZk)
|
||||||
.getChildren(zkw.znodePaths.baseZNode, null);
|
.getChildren(zkw.znodePaths.baseZNode, null);
|
||||||
try {
|
ZKUtil.listChildrenNoWatch(zkw, zkw.znodePaths.baseZNode);
|
||||||
ZKUtil.listChildrenNoWatch(zkw, zkw.znodePaths.baseZNode);
|
|
||||||
} catch (KeeperException.SystemErrorException e) {
|
|
||||||
// expected
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
Assert.fail("Should have thrown KeeperException but not");
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -174,7 +174,7 @@ public abstract class RemoteProcedureDispatcher<TEnv, TRemote extends Comparable
|
||||||
}
|
}
|
||||||
node.add(rp);
|
node.add(rp);
|
||||||
// Check our node still in the map; could have been removed by #removeNode.
|
// Check our node still in the map; could have been removed by #removeNode.
|
||||||
return nodeMap.contains(node);
|
return nodeMap.containsValue(node);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -75,14 +75,13 @@ public class SplitLogWorker implements Runnable {
|
||||||
this.server = server;
|
this.server = server;
|
||||||
this.conf = conf;
|
this.conf = conf;
|
||||||
this.coordination = hserver.getCoordinatedStateManager().getSplitLogWorkerCoordination();
|
this.coordination = hserver.getCoordinatedStateManager().getSplitLogWorkerCoordination();
|
||||||
this.server = server;
|
|
||||||
coordination.init(server, conf, splitTaskExecutor, this);
|
coordination.init(server, conf, splitTaskExecutor, this);
|
||||||
}
|
}
|
||||||
|
|
||||||
public SplitLogWorker(final Server hserver, final Configuration conf,
|
public SplitLogWorker(final Server hserver, final Configuration conf,
|
||||||
final RegionServerServices server, final LastSequenceId sequenceIdChecker,
|
final RegionServerServices server, final LastSequenceId sequenceIdChecker,
|
||||||
final WALFactory factory) {
|
final WALFactory factory) {
|
||||||
this(server, conf, server, new TaskExecutor() {
|
this(hserver, conf, server, new TaskExecutor() {
|
||||||
@Override
|
@Override
|
||||||
public Status exec(String filename, RecoveryMode mode, CancelableProgressable p) {
|
public Status exec(String filename, RecoveryMode mode, CancelableProgressable p) {
|
||||||
Path walDir;
|
Path walDir;
|
||||||
|
|
|
@ -1,61 +0,0 @@
|
||||||
/**
|
|
||||||
* Licensed to the Apache Software Foundation (ASF) under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. The ASF licenses this file
|
|
||||||
* to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.apache.hadoop.hbase;
|
|
||||||
|
|
||||||
import junit.framework.TestCase;
|
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.testclassification.MiscTests;
|
|
||||||
import org.apache.hadoop.hbase.testclassification.SmallTests;
|
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
|
||||||
import org.junit.experimental.categories.Category;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Test comparing HBase objects.
|
|
||||||
*/
|
|
||||||
@Category({MiscTests.class, SmallTests.class})
|
|
||||||
public class TestCompare extends TestCase {
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Sort of HRegionInfo.
|
|
||||||
*/
|
|
||||||
public void testHRegionInfo() {
|
|
||||||
HRegionInfo a = new HRegionInfo(TableName.valueOf("a"), null, null);
|
|
||||||
HRegionInfo b = new HRegionInfo(TableName.valueOf("b"), null, null);
|
|
||||||
assertTrue(a.compareTo(b) != 0);
|
|
||||||
HTableDescriptor t = new HTableDescriptor(TableName.valueOf("t"));
|
|
||||||
byte [] midway = Bytes.toBytes("midway");
|
|
||||||
a = new HRegionInfo(t.getTableName(), null, midway);
|
|
||||||
b = new HRegionInfo(t.getTableName(), midway, null);
|
|
||||||
assertTrue(a.compareTo(b) < 0);
|
|
||||||
assertTrue(b.compareTo(a) > 0);
|
|
||||||
assertEquals(a, a);
|
|
||||||
assertTrue(a.compareTo(a) == 0);
|
|
||||||
a = new HRegionInfo(t.getTableName(), Bytes.toBytes("a"), Bytes.toBytes("d"));
|
|
||||||
b = new HRegionInfo(t.getTableName(), Bytes.toBytes("e"), Bytes.toBytes("g"));
|
|
||||||
assertTrue(a.compareTo(b) < 0);
|
|
||||||
a = new HRegionInfo(t.getTableName(), Bytes.toBytes("aaaa"), Bytes.toBytes("dddd"));
|
|
||||||
b = new HRegionInfo(t.getTableName(), Bytes.toBytes("e"), Bytes.toBytes("g"));
|
|
||||||
assertTrue(a.compareTo(b) < 0);
|
|
||||||
a = new HRegionInfo(t.getTableName(), Bytes.toBytes("aaaa"), Bytes.toBytes("dddd"));
|
|
||||||
b = new HRegionInfo(t.getTableName(), Bytes.toBytes("aaaa"), Bytes.toBytes("eeee"));
|
|
||||||
assertTrue(a.compareTo(b) < 0);
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
|
@ -63,6 +63,7 @@ public class TestHRegionLocation {
|
||||||
System.out.println(hrl1.toString());
|
System.out.println(hrl1.toString());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@SuppressWarnings("SelfComparison")
|
||||||
@Test
|
@Test
|
||||||
public void testCompareTo() {
|
public void testCompareTo() {
|
||||||
ServerName hsa1 = ServerName.valueOf("localhost", 1234, -1L);
|
ServerName hsa1 = ServerName.valueOf("localhost", 1234, -1L);
|
||||||
|
@ -71,8 +72,8 @@ public class TestHRegionLocation {
|
||||||
ServerName hsa2 = ServerName.valueOf("localhost", 1235, -1L);
|
ServerName hsa2 = ServerName.valueOf("localhost", 1235, -1L);
|
||||||
HRegionLocation hsl2 =
|
HRegionLocation hsl2 =
|
||||||
new HRegionLocation(HRegionInfo.FIRST_META_REGIONINFO, hsa2);
|
new HRegionLocation(HRegionInfo.FIRST_META_REGIONINFO, hsa2);
|
||||||
assertTrue(hsl1.compareTo(hsl1) == 0);
|
assertEquals(0, hsl1.compareTo(hsl1));
|
||||||
assertTrue(hsl2.compareTo(hsl2) == 0);
|
assertEquals(0, hsl2.compareTo(hsl2));
|
||||||
int compare1 = hsl1.compareTo(hsl2);
|
int compare1 = hsl1.compareTo(hsl2);
|
||||||
int compare2 = hsl2.compareTo(hsl1);
|
int compare2 = hsl2.compareTo(hsl1);
|
||||||
assertTrue((compare1 > 0)? compare2 < 0: compare2 > 0);
|
assertTrue((compare1 > 0)? compare2 < 0: compare2 > 0);
|
||||||
|
|
|
@ -124,7 +124,7 @@ public class TestIPv6NIOServerSocketChannel {
|
||||||
//On Windows JDK6, we will get expected exception:
|
//On Windows JDK6, we will get expected exception:
|
||||||
//java.net.SocketException: Address family not supported by protocol family
|
//java.net.SocketException: Address family not supported by protocol family
|
||||||
//or java.net.SocketException: Protocol family not supported
|
//or java.net.SocketException: Protocol family not supported
|
||||||
Assert.assertFalse(ex.getClass().isInstance(BindException.class));
|
Assert.assertFalse(ex instanceof BindException);
|
||||||
Assert.assertTrue(ex.getMessage().toLowerCase(Locale.ROOT).contains("protocol family"));
|
Assert.assertTrue(ex.getMessage().toLowerCase(Locale.ROOT).contains("protocol family"));
|
||||||
LOG.info("Received expected exception:");
|
LOG.info("Received expected exception:");
|
||||||
LOG.info(ex);
|
LOG.info(ex);
|
||||||
|
|
|
@ -217,6 +217,7 @@ public class TestScannersFromClientSide2 {
|
||||||
testScan(456, false, 678, false, 200);
|
testScan(456, false, 678, false, 200);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
public void testReversedScanWithLimit() throws Exception {
|
public void testReversedScanWithLimit() throws Exception {
|
||||||
testReversedScan(998, true, 1, false, 900); // from last region to first region
|
testReversedScan(998, true, 1, false, 900); // from last region to first region
|
||||||
testReversedScan(543, true, 321, true, 100);
|
testReversedScan(543, true, 321, true, 100);
|
||||||
|
|
|
@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver;
|
||||||
import static org.junit.Assert.assertArrayEquals;
|
import static org.junit.Assert.assertArrayEquals;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
import static org.junit.Assert.assertFalse;
|
import static org.junit.Assert.assertFalse;
|
||||||
|
import static org.junit.Assert.assertNotEquals;
|
||||||
import static org.junit.Assert.assertTrue;
|
import static org.junit.Assert.assertTrue;
|
||||||
import static org.junit.Assert.fail;
|
import static org.junit.Assert.fail;
|
||||||
|
|
||||||
|
@ -169,8 +170,30 @@ public class TestHRegionInfo {
|
||||||
HRegionInfo newer = new HRegionInfo(tableName, empty, empty, false, 1L);
|
HRegionInfo newer = new HRegionInfo(tableName, empty, empty, false, 1L);
|
||||||
assertTrue(older.compareTo(newer) < 0);
|
assertTrue(older.compareTo(newer) < 0);
|
||||||
assertTrue(newer.compareTo(older) > 0);
|
assertTrue(newer.compareTo(older) > 0);
|
||||||
assertTrue(older.compareTo(older) == 0);
|
assertEquals(0, older.compareTo(older));
|
||||||
assertTrue(newer.compareTo(newer) == 0);
|
assertEquals(0, newer.compareTo(newer));
|
||||||
|
|
||||||
|
HRegionInfo a = new HRegionInfo(TableName.valueOf("a"), null, null);
|
||||||
|
HRegionInfo b = new HRegionInfo(TableName.valueOf("b"), null, null);
|
||||||
|
assertNotEquals(0, a.compareTo(b));
|
||||||
|
HTableDescriptor t = new HTableDescriptor(TableName.valueOf("t"));
|
||||||
|
byte [] midway = Bytes.toBytes("midway");
|
||||||
|
a = new HRegionInfo(t.getTableName(), null, midway);
|
||||||
|
b = new HRegionInfo(t.getTableName(), midway, null);
|
||||||
|
assertTrue(a.compareTo(b) < 0);
|
||||||
|
assertTrue(b.compareTo(a) > 0);
|
||||||
|
assertEquals(a, a);
|
||||||
|
assertEquals(0, a.compareTo(a));
|
||||||
|
a = new HRegionInfo(t.getTableName(), Bytes.toBytes("a"), Bytes.toBytes("d"));
|
||||||
|
b = new HRegionInfo(t.getTableName(), Bytes.toBytes("e"), Bytes.toBytes("g"));
|
||||||
|
assertTrue(a.compareTo(b) < 0);
|
||||||
|
a = new HRegionInfo(t.getTableName(), Bytes.toBytes("aaaa"), Bytes.toBytes("dddd"));
|
||||||
|
b = new HRegionInfo(t.getTableName(), Bytes.toBytes("e"), Bytes.toBytes("g"));
|
||||||
|
assertTrue(a.compareTo(b) < 0);
|
||||||
|
a = new HRegionInfo(t.getTableName(), Bytes.toBytes("aaaa"), Bytes.toBytes("dddd"));
|
||||||
|
b = new HRegionInfo(t.getTableName(), Bytes.toBytes("aaaa"), Bytes.toBytes("eeee"));
|
||||||
|
assertTrue(a.compareTo(b) < 0);
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
@ -325,7 +348,7 @@ public class TestHRegionInfo {
|
||||||
if (i != 1) {
|
if (i != 1) {
|
||||||
Assert.assertArrayEquals(regionNameParts[i], modifiedRegionNameParts[i]);
|
Assert.assertArrayEquals(regionNameParts[i], modifiedRegionNameParts[i]);
|
||||||
} else {
|
} else {
|
||||||
Assert.assertNotEquals(regionNameParts[i][0], modifiedRegionNameParts[i][0]);
|
assertNotEquals(regionNameParts[i][0], modifiedRegionNameParts[i][0]);
|
||||||
Assert.assertArrayEquals(modifiedRegionNameParts[1],
|
Assert.assertArrayEquals(modifiedRegionNameParts[1],
|
||||||
HRegionInfo.getStartKeyForDisplay(h, conf));
|
HRegionInfo.getStartKeyForDisplay(h, conf));
|
||||||
}
|
}
|
||||||
|
|
|
@ -129,7 +129,7 @@ public class TestJoinedScanners {
|
||||||
puts.clear();
|
puts.clear();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (puts.size() >= 0) {
|
if (!puts.isEmpty()) {
|
||||||
ht.put(puts);
|
ht.put(puts);
|
||||||
puts.clear();
|
puts.clear();
|
||||||
}
|
}
|
||||||
|
|
|
@ -58,6 +58,7 @@ import org.apache.hadoop.hbase.testclassification.RegionServerTests;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
import org.apache.hadoop.hbase.util.EnvironmentEdge;
|
import org.apache.hadoop.hbase.util.EnvironmentEdge;
|
||||||
import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper;
|
import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper;
|
||||||
|
import org.junit.Ignore;
|
||||||
import org.junit.Rule;
|
import org.junit.Rule;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.junit.experimental.categories.Category;
|
import org.junit.experimental.categories.Category;
|
||||||
|
@ -599,6 +600,7 @@ public class TestStoreScanner {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
public void testDeleteVersionMaskingMultiplePuts() throws IOException {
|
public void testDeleteVersionMaskingMultiplePuts() throws IOException {
|
||||||
long now = System.currentTimeMillis();
|
long now = System.currentTimeMillis();
|
||||||
KeyValue [] kvs1 = new KeyValue[] {
|
KeyValue [] kvs1 = new KeyValue[] {
|
||||||
|
@ -623,6 +625,8 @@ public class TestStoreScanner {
|
||||||
assertEquals(kvs2[1], results.get(0));
|
assertEquals(kvs2[1], results.get(0));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
public void testDeleteVersionsMixedAndMultipleVersionReturn() throws IOException {
|
public void testDeleteVersionsMixedAndMultipleVersionReturn() throws IOException {
|
||||||
long now = System.currentTimeMillis();
|
long now = System.currentTimeMillis();
|
||||||
KeyValue [] kvs1 = new KeyValue[] {
|
KeyValue [] kvs1 = new KeyValue[] {
|
||||||
|
@ -834,10 +838,8 @@ public class TestStoreScanner {
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
/**
|
@Test @Ignore("this fails, since we don't handle deletions, etc, in peek")
|
||||||
* TODO this fails, since we don't handle deletions, etc, in peek
|
public void testPeek() throws Exception {
|
||||||
*/
|
|
||||||
public void SKIP_testPeek() throws Exception {
|
|
||||||
KeyValue[] kvs = new KeyValue [] {
|
KeyValue[] kvs = new KeyValue [] {
|
||||||
create("R1", "cf", "a", 1, KeyValue.Type.Put, "dont-care"),
|
create("R1", "cf", "a", 1, KeyValue.Type.Put, "dont-care"),
|
||||||
create("R1", "cf", "a", 1, KeyValue.Type.Delete, "dont-care"),
|
create("R1", "cf", "a", 1, KeyValue.Type.Delete, "dont-care"),
|
||||||
|
|
|
@ -167,7 +167,7 @@ public class MultiThreadedUpdater extends MultiThreadedWriterBase {
|
||||||
get = dataGenerator.beforeGet(rowKeyBase, get);
|
get = dataGenerator.beforeGet(rowKeyBase, get);
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
// Ideally wont happen
|
// Ideally wont happen
|
||||||
LOG.warn("Failed to modify the get from the load generator = [" + get.getRow()
|
LOG.warn("Failed to modify the get from the load generator = [" + Bytes.toString(get.getRow())
|
||||||
+ "], column family = [" + Bytes.toString(cf) + "]", e);
|
+ "], column family = [" + Bytes.toString(cf) + "]", e);
|
||||||
}
|
}
|
||||||
Result result = getRow(get, rowKeyBase, cf);
|
Result result = getRow(get, rowKeyBase, cf);
|
||||||
|
@ -234,7 +234,7 @@ public class MultiThreadedUpdater extends MultiThreadedWriterBase {
|
||||||
if (isBatchUpdate) {
|
if (isBatchUpdate) {
|
||||||
if (verbose) {
|
if (verbose) {
|
||||||
LOG.debug("Preparing increment and append for key = ["
|
LOG.debug("Preparing increment and append for key = ["
|
||||||
+ rowKey + "], " + columnCount + " columns");
|
+ Bytes.toString(rowKey) + "], " + columnCount + " columns");
|
||||||
}
|
}
|
||||||
mutate(table, inc, rowKeyBase);
|
mutate(table, inc, rowKeyBase);
|
||||||
mutate(table, app, rowKeyBase);
|
mutate(table, app, rowKeyBase);
|
||||||
|
@ -267,7 +267,7 @@ public class MultiThreadedUpdater extends MultiThreadedWriterBase {
|
||||||
result = table.get(get);
|
result = table.get(get);
|
||||||
} catch (IOException ie) {
|
} catch (IOException ie) {
|
||||||
LOG.warn(
|
LOG.warn(
|
||||||
"Failed to get the row for key = [" + get.getRow() + "], column family = ["
|
"Failed to get the row for key = [" + Bytes.toString(get.getRow()) + "], column family = ["
|
||||||
+ Bytes.toString(cf) + "]", ie);
|
+ Bytes.toString(cf) + "]", ie);
|
||||||
}
|
}
|
||||||
return result;
|
return result;
|
||||||
|
|
|
@ -123,7 +123,7 @@ public class MultiThreadedWriter extends MultiThreadedWriterBase {
|
||||||
}
|
}
|
||||||
if (isMultiPut) {
|
if (isMultiPut) {
|
||||||
if (verbose) {
|
if (verbose) {
|
||||||
LOG.debug("Preparing put for key = [" + rowKey + "], " + columnCount + " columns");
|
LOG.debug("Preparing put for key = [" + Bytes.toString(rowKey) + "], " + columnCount + " columns");
|
||||||
}
|
}
|
||||||
insert(table, put, rowKeyBase);
|
insert(table, put, rowKeyBase);
|
||||||
numCols.addAndGet(columnCount);
|
numCols.addAndGet(columnCount);
|
||||||
|
|
Loading…
Reference in New Issue