From d0e4912f678abd2efe497ff8a52eae115cc74d8a Mon Sep 17 00:00:00 2001 From: Josh Elser Date: Fri, 15 Feb 2019 12:10:04 -0500 Subject: [PATCH] Revert "HBASE-21782 Use a call that defaults to UTF-8 charset for string to byte encoding" This reverts commit 3d4e1d57e57edd9c23e83048e5d376bff2270859. Fixing incorrect Jira id. --- .../hbase/backup/impl/BackupSystemTable.java | 30 +- .../impl/IncrementalTableBackupClient.java | 6 +- .../master/LogRollMasterProcedureManager.java | 3 +- .../hbase/backup/TestBackupDeleteRestore.java | 3 +- .../hbase/backup/TestBackupHFileCleaner.java | 3 +- .../hadoop/hbase/rest/RESTDemoClient.java | 6 +- .../hadoop/hbase/thrift2/DemoClient.java | 13 +- .../apache/hadoop/hbase/http/HtmlQuoting.java | 16 +- .../hbase/IntegrationTestLazyCfLoading.java | 2 +- .../apache/hadoop/hbase/mapreduce/Import.java | 2 +- .../hadoop/hbase/mapreduce/ImportTsv.java | 8 +- .../hbase/TestPerformanceEvaluation.java | 4 +- .../hbase/mapred/TestGroupingTableMap.java | 24 +- .../hadoop/hbase/mapred/TestSplitTable.java | 4 +- .../hbase/mapred/TestTableInputFormat.java | 30 +- .../hadoop/hbase/mapreduce/TestCopyTable.java | 2 +- .../hbase/mapreduce/TestImportExport.java | 2 +- .../mapreduce/TestRegionSizeCalculator.java | 20 +- .../hbase/mapreduce/TestTableInputFormat.java | 30 +- .../hbase/mapreduce/TestTableSplit.java | 33 +- .../TestVerifyReplicationCrossDiffHdfs.java | 8 +- .../hadoop/hbase/TestSerialization.java | 22 +- .../hadoop/hbase/client/TestAdmin1.java | 18 +- .../hbase/client/TestAsyncTableAdminApi.java | 8 +- .../client/TestConnectionImplementation.java | 12 +- .../hadoop/hbase/client/TestFastFail.java | 2 +- .../client/TestGetScanPartialResult.java | 3 +- .../hbase/client/TestMetaWithReplicas.java | 6 +- .../hbase/client/TestReplicaWithCluster.java | 2 +- .../hbase/client/TestReplicasClient.java | 18 +- .../TestRegionObserverInterface.java | 4 +- .../hbase/filter/TestFilterWrapper.java | 2 +- .../TestFuzzyRowAndColumnRangeFilter.java | 4 +- .../filter/TestFuzzyRowFilterEndToEnd.java | 16 +- .../hbase/filter/TestScanRowPrefix.java | 3 +- .../hbase/io/encoding/TestEncodedSeekers.java | 4 +- .../hbase/io/hfile/RandomKeyValueUtil.java | 13 +- .../hadoop/hbase/io/hfile/TestHFile.java | 9 +- .../hbase/io/hfile/TestHFileBlockIndex.java | 2 +- .../hbase/io/hfile/TestHFileEncryption.java | 3 +- .../hadoop/hbase/io/hfile/TestHFileSeek.java | 5 +- .../master/balancer/TestBaseLoadBalancer.java | 21 +- .../hbase/master/locking/TestLockManager.java | 4 +- .../MasterProcedureTestingUtility.java | 6 +- .../procedure/SimpleRSProcedureManager.java | 3 +- .../hbase/procedure/TestProcedureManager.java | 3 +- .../procedure/TestZKProcedureControllers.java | 3 +- .../regionserver/TestCompoundBloomFilter.java | 4 +- .../regionserver/TestMultiLogThreshold.java | 2 +- .../TestRegionReplicasWithModifyTable.java | 1 + .../TestRegionServerReadRequestMetrics.java | 20 +- .../TestSplitTransactionOnCluster.java | 22 +- .../TestWALMonotonicallyIncreasingSeqId.java | 4 +- .../throttle/TestStoreHotnessProtector.java | 3 +- ...ionReplicaReplicationEndpointNoMaster.java | 3 +- .../access/TestWithDisabledAuthorization.java | 4 +- .../hbase/tool/TestLoadIncrementalHFiles.java | 24 +- .../hbase/util/TestBloomFilterChunk.java | 2 +- .../hadoop/hbase/util/TestRegionMover.java | 2 +- .../hadoop/hbase/util/TestRegionSplitter.java | 472 +++++++++--------- .../hbase/zookeeper/TestZooKeeperACL.java | 37 +- .../hbase/zookeeper/MiniZooKeeperCluster.java | 6 +- .../zookeeper/TestRecoverableZooKeeper.java | 4 +- .../hadoop/hbase/zookeeper/TestZKUtil.java | 4 +- 64 files changed, 511 insertions(+), 548 deletions(-) diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.java index 89945cba5a3..94ccfe56e60 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/BackupSystemTable.java @@ -141,12 +141,12 @@ public final class BackupSystemTable implements Closeable { /** * Stores backup sessions (contexts) */ - final static byte[] SESSIONS_FAMILY = Bytes.toBytes("session"); + final static byte[] SESSIONS_FAMILY = "session".getBytes(); /** * Stores other meta */ - final static byte[] META_FAMILY = Bytes.toBytes("meta"); - final static byte[] BULK_LOAD_FAMILY = Bytes.toBytes("bulk"); + final static byte[] META_FAMILY = "meta".getBytes(); + final static byte[] BULK_LOAD_FAMILY = "bulk".getBytes(); /** * Connection to HBase cluster, shared among all instances */ @@ -154,20 +154,20 @@ public final class BackupSystemTable implements Closeable { private final static String BACKUP_INFO_PREFIX = "session:"; private final static String START_CODE_ROW = "startcode:"; - private final static byte[] ACTIVE_SESSION_ROW = Bytes.toBytes("activesession:"); - private final static byte[] ACTIVE_SESSION_COL = Bytes.toBytes("c"); + private final static byte[] ACTIVE_SESSION_ROW = "activesession:".getBytes(); + private final static byte[] ACTIVE_SESSION_COL = "c".getBytes(); - private final static byte[] ACTIVE_SESSION_YES = Bytes.toBytes("yes"); - private final static byte[] ACTIVE_SESSION_NO = Bytes.toBytes("no"); + private final static byte[] ACTIVE_SESSION_YES = "yes".getBytes(); + private final static byte[] ACTIVE_SESSION_NO = "no".getBytes(); private final static String INCR_BACKUP_SET = "incrbackupset:"; private final static String TABLE_RS_LOG_MAP_PREFIX = "trslm:"; private final static String RS_LOG_TS_PREFIX = "rslogts:"; private final static String BULK_LOAD_PREFIX = "bulk:"; - private final static byte[] BULK_LOAD_PREFIX_BYTES = Bytes.toBytes(BULK_LOAD_PREFIX); - private final static byte[] DELETE_OP_ROW = Bytes.toBytes("delete_op_row"); - private final static byte[] MERGE_OP_ROW = Bytes.toBytes("merge_op_row"); + private final static byte[] BULK_LOAD_PREFIX_BYTES = BULK_LOAD_PREFIX.getBytes(); + private final static byte[] DELETE_OP_ROW = "delete_op_row".getBytes(); + private final static byte[] MERGE_OP_ROW = "merge_op_row".getBytes(); final static byte[] TBL_COL = Bytes.toBytes("tbl"); final static byte[] FAM_COL = Bytes.toBytes("fam"); @@ -1615,7 +1615,7 @@ public final class BackupSystemTable implements Closeable { Bytes.toString(region), BLK_LD_DELIM, filename)); put.addColumn(BackupSystemTable.META_FAMILY, TBL_COL, table.getName()); put.addColumn(BackupSystemTable.META_FAMILY, FAM_COL, entry.getKey()); - put.addColumn(BackupSystemTable.META_FAMILY, PATH_COL, Bytes.toBytes(file)); + put.addColumn(BackupSystemTable.META_FAMILY, PATH_COL, file.getBytes()); put.addColumn(BackupSystemTable.META_FAMILY, STATE_COL, BL_COMMIT); puts.add(put); LOG.debug( @@ -1695,7 +1695,7 @@ public final class BackupSystemTable implements Closeable { Bytes.toString(region), BLK_LD_DELIM, filename)); put.addColumn(BackupSystemTable.META_FAMILY, TBL_COL, table.getName()); put.addColumn(BackupSystemTable.META_FAMILY, FAM_COL, family); - put.addColumn(BackupSystemTable.META_FAMILY, PATH_COL, Bytes.toBytes(file)); + put.addColumn(BackupSystemTable.META_FAMILY, PATH_COL, file.getBytes()); put.addColumn(BackupSystemTable.META_FAMILY, STATE_COL, BL_PREPARE); puts.add(put); LOG.debug("writing raw bulk path " + file + " for " + table + " " + Bytes.toString(region)); @@ -1902,7 +1902,7 @@ public final class BackupSystemTable implements Closeable { Put put = new Put(rowkey(BULK_LOAD_PREFIX, backupId + BLK_LD_DELIM + ts + BLK_LD_DELIM + idx)); put.addColumn(BackupSystemTable.META_FAMILY, TBL_COL, tn.getName()); put.addColumn(BackupSystemTable.META_FAMILY, FAM_COL, fam); - put.addColumn(BackupSystemTable.META_FAMILY, PATH_COL, Bytes.toBytes(p)); + put.addColumn(BackupSystemTable.META_FAMILY, PATH_COL, p.getBytes()); return put; } @@ -2006,7 +2006,7 @@ public final class BackupSystemTable implements Closeable { } private byte[] convertToByteArray(String[] tables) { - return Bytes.toBytes(StringUtils.join(tables, ",")); + return StringUtils.join(tables, ",").getBytes(); } /** @@ -2037,6 +2037,6 @@ public final class BackupSystemTable implements Closeable { for (String ss : other) { sb.append(ss); } - return Bytes.toBytes(sb.toString()); + return sb.toString().getBytes(); } } diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.java index 1e4b266b7ad..43824d78d03 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/impl/IncrementalTableBackupClient.java @@ -155,11 +155,11 @@ public class IncrementalTableBackupClient extends TableBackupClient { String fam = famEntry.getKey(); Path famDir = new Path(regionDir, fam); List files; - if (!mapForSrc[srcIdx].containsKey(Bytes.toBytes(fam))) { + if (!mapForSrc[srcIdx].containsKey(fam.getBytes())) { files = new ArrayList<>(); - mapForSrc[srcIdx].put(Bytes.toBytes(fam), files); + mapForSrc[srcIdx].put(fam.getBytes(), files); } else { - files = mapForSrc[srcIdx].get(Bytes.toBytes(fam)); + files = mapForSrc[srcIdx].get(fam.getBytes()); } Path archiveDir = HFileArchiveUtil.getStoreArchivePath(conf, srcTable, regionName, fam); String tblName = srcTable.getQualifierAsString(); diff --git a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/master/LogRollMasterProcedureManager.java b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/master/LogRollMasterProcedureManager.java index 32e3e23fdaf..a3d5ce2d91f 100644 --- a/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/master/LogRollMasterProcedureManager.java +++ b/hbase-backup/src/main/java/org/apache/hadoop/hbase/backup/master/LogRollMasterProcedureManager.java @@ -40,7 +40,6 @@ import org.apache.hadoop.hbase.procedure.RegionServerProcedureManager; import org.apache.hadoop.hbase.procedure.ZKProcedureCoordinationManager; import org.apache.hadoop.hbase.security.User; import org.apache.hadoop.hbase.security.access.AccessChecker; -import org.apache.hadoop.hbase.util.Bytes; import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -131,7 +130,7 @@ public class LogRollMasterProcedureManager extends MasterProcedureManager { byte[] data = new byte[0]; if (conf.size() > 0) { // Get backup root path - data = Bytes.toBytes(conf.get(0).getValue()); + data = conf.get(0).getValue().getBytes(); } Procedure proc = coordinator.startProcedure(monitor, desc.getInstance(), data, servers); if (proc == null) { diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDeleteRestore.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDeleteRestore.java index 74176e390c6..a557b9b91b9 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDeleteRestore.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupDeleteRestore.java @@ -28,7 +28,6 @@ import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.testclassification.MediumTests; -import org.apache.hadoop.hbase.util.Bytes; import org.junit.ClassRule; import org.junit.Test; import org.junit.experimental.categories.Category; @@ -64,7 +63,7 @@ public class TestBackupDeleteRestore extends TestBackupBase { HBaseAdmin hba = TEST_UTIL.getHBaseAdmin(); // delete row try (Table table = TEST_UTIL.getConnection().getTable(table1)) { - Delete delete = new Delete(Bytes.toBytes("row0")); + Delete delete = new Delete("row0".getBytes()); table.delete(delete); hba.flush(table1); } diff --git a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupHFileCleaner.java b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupHFileCleaner.java index 1680dc643d8..8d22f2753f4 100644 --- a/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupHFileCleaner.java +++ b/hbase-backup/src/test/java/org/apache/hadoop/hbase/backup/TestBackupHFileCleaner.java @@ -37,7 +37,6 @@ import org.apache.hadoop.hbase.client.Connection; import org.apache.hadoop.hbase.client.ConnectionFactory; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.SmallTests; -import org.apache.hadoop.hbase.util.Bytes; import org.junit.After; import org.junit.AfterClass; import org.junit.Before; @@ -133,7 +132,7 @@ public class TestBackupHFileCleaner { sTableList.add(tableName); Map>[] maps = new Map[1]; maps[0] = new HashMap<>(); - maps[0].put(Bytes.toBytes(famName), list); + maps[0].put(famName.getBytes(), list); sysTbl.writeBulkLoadedFiles(sTableList, maps, "1"); } diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/rest/RESTDemoClient.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/rest/RESTDemoClient.java index 2e462e543ce..19fae470ce0 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/rest/RESTDemoClient.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/rest/RESTDemoClient.java @@ -87,12 +87,12 @@ public class RESTDemoClient { try (RemoteHTable remoteTable = new RemoteHTable(restClient, conf, "example")) { // Write data to the table String rowKey = "row1"; - Put p = new Put(Bytes.toBytes(rowKey)); - p.addColumn(Bytes.toBytes("family1"), Bytes.toBytes("qualifier1"), Bytes.toBytes("value1")); + Put p = new Put(rowKey.getBytes()); + p.addColumn("family1".getBytes(), "qualifier1".getBytes(), "value1".getBytes()); remoteTable.put(p); // Get the data from the table - Get g = new Get(Bytes.toBytes(rowKey)); + Get g = new Get(rowKey.getBytes()); Result result = remoteTable.get(g); Preconditions.checkArgument(result != null, diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift2/DemoClient.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift2/DemoClient.java index f8d4f64f587..7fbe2aa72c9 100644 --- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift2/DemoClient.java +++ b/hbase-examples/src/main/java/org/apache/hadoop/hbase/thrift2/DemoClient.java @@ -35,7 +35,6 @@ import org.apache.hadoop.hbase.thrift2.generated.TGet; import org.apache.hadoop.hbase.thrift2.generated.THBaseService; import org.apache.hadoop.hbase.thrift2.generated.TPut; import org.apache.hadoop.hbase.thrift2.generated.TResult; -import org.apache.hadoop.hbase.util.Bytes; import org.apache.thrift.protocol.TBinaryProtocol; import org.apache.thrift.protocol.TProtocol; import org.apache.thrift.transport.TFramedTransport; @@ -117,15 +116,15 @@ public class DemoClient { // open the transport transport.open(); - ByteBuffer table = ByteBuffer.wrap(Bytes.toBytes("example")); + ByteBuffer table = ByteBuffer.wrap("example".getBytes()); TPut put = new TPut(); - put.setRow(Bytes.toBytes("row1")); + put.setRow("row1".getBytes()); TColumnValue columnValue = new TColumnValue(); - columnValue.setFamily(Bytes.toBytes("family1")); - columnValue.setQualifier(Bytes.toBytes("qualifier1")); - columnValue.setValue(Bytes.toBytes("value1")); + columnValue.setFamily("family1".getBytes()); + columnValue.setQualifier("qualifier1".getBytes()); + columnValue.setValue("value1".getBytes()); List columnValues = new ArrayList<>(1); columnValues.add(columnValue); put.setColumnValues(columnValues); @@ -133,7 +132,7 @@ public class DemoClient { client.put(table, put); TGet get = new TGet(); - get.setRow(Bytes.toBytes("row1")); + get.setRow("row1".getBytes()); TResult result = client.get(table, get); diff --git a/hbase-http/src/main/java/org/apache/hadoop/hbase/http/HtmlQuoting.java b/hbase-http/src/main/java/org/apache/hadoop/hbase/http/HtmlQuoting.java index 6d1bfbce1b6..e2eebb2519e 100644 --- a/hbase-http/src/main/java/org/apache/hadoop/hbase/http/HtmlQuoting.java +++ b/hbase-http/src/main/java/org/apache/hadoop/hbase/http/HtmlQuoting.java @@ -20,8 +20,6 @@ package org.apache.hadoop.hbase.http; import java.io.ByteArrayOutputStream; import java.io.IOException; import java.io.OutputStream; - -import org.apache.hadoop.hbase.util.Bytes; import org.apache.yetus.audience.InterfaceAudience; /** @@ -29,11 +27,11 @@ import org.apache.yetus.audience.InterfaceAudience; */ @InterfaceAudience.Private public final class HtmlQuoting { - private static final byte[] ampBytes = Bytes.toBytes("&"); - private static final byte[] aposBytes = Bytes.toBytes("'"); - private static final byte[] gtBytes = Bytes.toBytes(">"); - private static final byte[] ltBytes = Bytes.toBytes("<"); - private static final byte[] quotBytes = Bytes.toBytes("""); + private static final byte[] ampBytes = "&".getBytes(); + private static final byte[] aposBytes = "'".getBytes(); + private static final byte[] gtBytes = ">".getBytes(); + private static final byte[] ltBytes = "<".getBytes(); + private static final byte[] quotBytes = """.getBytes(); /** * Does the given string need to be quoted? @@ -71,7 +69,7 @@ public final class HtmlQuoting { if (str == null) { return false; } - byte[] bytes = Bytes.toBytes(str); + byte[] bytes = str.getBytes(); return needsQuoting(bytes, 0 , bytes.length); } @@ -106,7 +104,7 @@ public final class HtmlQuoting { if (item == null) { return null; } - byte[] bytes = Bytes.toBytes(item); + byte[] bytes = item.getBytes(); if (needsQuoting(bytes, 0, bytes.length)) { ByteArrayOutputStream buffer = new ByteArrayOutputStream(); try { diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java index 128767b4b29..c2966a34635 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java +++ b/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestLazyCfLoading.java @@ -112,7 +112,7 @@ public class IntegrationTestLazyCfLoading { @Override public byte[] getDeterministicUniqueKey(long keyBase) { - return Bytes.toBytes(LoadTestKVGenerator.md5PrefixedKey(keyBase)); + return LoadTestKVGenerator.md5PrefixedKey(keyBase).getBytes(); } @Override diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java index 239a12bdc68..7805bc5dff2 100644 --- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java +++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java @@ -545,7 +545,7 @@ public class Import extends Configured implements Tool { if(srcAndDest.length != 2) { continue; } - cfRenameMap.put(Bytes.toBytes(srcAndDest[0]), Bytes.toBytes(srcAndDest[1])); + cfRenameMap.put(srcAndDest[0].getBytes(), srcAndDest[1].getBytes()); } } return cfRenameMap; diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java index 0fd9483cce6..58d9184aa5f 100644 --- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java +++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java @@ -194,11 +194,11 @@ public class ImportTsv extends Configured implements Tool { } String[] parts = str.split(":", 2); if (parts.length == 1) { - families[i] = Bytes.toBytes(str); + families[i] = str.getBytes(); qualifiers[i] = HConstants.EMPTY_BYTE_ARRAY; } else { - families[i] = Bytes.toBytes(parts[0]); - qualifiers[i] = Bytes.toBytes(parts[1]); + families[i] = parts[0].getBytes(); + qualifiers[i] = parts[1].getBytes(); } } } @@ -471,7 +471,7 @@ public class ImportTsv extends Configured implements Tool { String actualSeparator = conf.get(SEPARATOR_CONF_KEY); if (actualSeparator != null) { conf.set(SEPARATOR_CONF_KEY, - Bytes.toString(Base64.getEncoder().encode(Bytes.toBytes(actualSeparator)))); + Bytes.toString(Base64.getEncoder().encode(actualSeparator.getBytes()))); } // See if a non-default Mapper was set diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/TestPerformanceEvaluation.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/TestPerformanceEvaluation.java index dfcf2d63b2b..2e1ecd01e08 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/TestPerformanceEvaluation.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/TestPerformanceEvaluation.java @@ -31,7 +31,6 @@ import java.io.IOException; import java.io.InputStreamReader; import java.lang.reflect.Constructor; import java.lang.reflect.InvocationTargetException; -import java.nio.charset.StandardCharsets; import java.util.LinkedList; import java.util.NoSuchElementException; import java.util.Queue; @@ -105,8 +104,7 @@ public class TestPerformanceEvaluation { try { dis.readFully(content); BufferedReader br = - new BufferedReader(new InputStreamReader( - new ByteArrayInputStream(content), StandardCharsets.UTF_8)); + new BufferedReader(new InputStreamReader(new ByteArrayInputStream(content))); int count = 0; while (br.readLine() != null) { count++; diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestGroupingTableMap.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestGroupingTableMap.java index 76e3c73e2d5..12db348ba8b 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestGroupingTableMap.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestGroupingTableMap.java @@ -72,12 +72,9 @@ public class TestGroupingTableMap { byte[] row = {}; List keyValues = ImmutableList.of( - new KeyValue(row, Bytes.toBytes("familyA"), Bytes.toBytes("qualifierA"), - Bytes.toBytes("1111")), - new KeyValue(row, Bytes.toBytes("familyA"), Bytes.toBytes("qualifierA"), - Bytes.toBytes("2222")), - new KeyValue(row, Bytes.toBytes("familyB"), Bytes.toBytes("qualifierB"), - Bytes.toBytes("3333"))); + new KeyValue(row, "familyA".getBytes(), "qualifierA".getBytes(), Bytes.toBytes("1111")), + new KeyValue(row, "familyA".getBytes(), "qualifierA".getBytes(), Bytes.toBytes("2222")), + new KeyValue(row, "familyB".getBytes(), "qualifierB".getBytes(), Bytes.toBytes("3333"))); when(result.listCells()).thenReturn(keyValues); OutputCollector outputCollectorMock = mock(OutputCollector.class); @@ -105,12 +102,9 @@ public class TestGroupingTableMap { byte[] row = {}; List keyValues = ImmutableList.of( - new KeyValue(row, Bytes.toBytes("familyA"), Bytes.toBytes("qualifierA"), - Bytes.toBytes("1111")), - new KeyValue(row, Bytes.toBytes("familyB"), Bytes.toBytes("qualifierB"), - Bytes.toBytes("2222")), - new KeyValue(row, Bytes.toBytes("familyC"), Bytes.toBytes("qualifierC"), - Bytes.toBytes("3333"))); + new KeyValue(row, "familyA".getBytes(), "qualifierA".getBytes(), Bytes.toBytes("1111")), + new KeyValue(row, "familyB".getBytes(), "qualifierB".getBytes(), Bytes.toBytes("2222")), + new KeyValue(row, "familyC".getBytes(), "qualifierC".getBytes(), Bytes.toBytes("3333"))); when(result.listCells()).thenReturn(keyValues); OutputCollector outputCollectorMock = mock(OutputCollector.class); @@ -143,10 +137,8 @@ public class TestGroupingTableMap { final byte[] secondPartKeyValue = Bytes.toBytes("35245142671437"); byte[] row = {}; List cells = ImmutableList.of( - new KeyValue(row, Bytes.toBytes("familyA"), Bytes.toBytes("qualifierA"), - firstPartKeyValue), - new KeyValue(row, Bytes.toBytes("familyB"), Bytes.toBytes("qualifierB"), - secondPartKeyValue)); + new KeyValue(row, "familyA".getBytes(), "qualifierA".getBytes(), firstPartKeyValue), + new KeyValue(row, "familyB".getBytes(), "qualifierB".getBytes(), secondPartKeyValue)); when(result.listCells()).thenReturn(cells); final AtomicBoolean outputCollected = new AtomicBoolean(); diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestSplitTable.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestSplitTable.java index bf46a7ac6d8..7b097d264ce 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestSplitTable.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestSplitTable.java @@ -106,8 +106,8 @@ public class TestSplitTable { @SuppressWarnings("deprecation") public void testToString() { TableSplit split = - new TableSplit(TableName.valueOf(name.getMethodName()), Bytes.toBytes("row-start"), - Bytes.toBytes("row-end"), "location"); + new TableSplit(TableName.valueOf(name.getMethodName()), "row-start".getBytes(), "row-end".getBytes(), + "location"); String str = "HBase table split(table name: " + name.getMethodName() + ", start row: row-start, " + "end row: row-end, region location: location)"; diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestTableInputFormat.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestTableInputFormat.java index e93d36d4b25..2ef8351a09b 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestTableInputFormat.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapred/TestTableInputFormat.java @@ -124,14 +124,14 @@ public class TestTableInputFormat { */ public static Table createTable(byte[] tableName, byte[][] families) throws IOException { Table table = UTIL.createTable(TableName.valueOf(tableName), families); - Put p = new Put(Bytes.toBytes("aaa")); + Put p = new Put("aaa".getBytes()); for (byte[] family : families) { - p.addColumn(family, null, Bytes.toBytes("value aaa")); + p.addColumn(family, null, "value aaa".getBytes()); } table.put(p); - p = new Put(Bytes.toBytes("bbb")); + p = new Put("bbb".getBytes()); for (byte[] family : families) { - p.addColumn(family, null, Bytes.toBytes("value bbb")); + p.addColumn(family, null, "value bbb".getBytes()); } table.put(p); return table; @@ -165,8 +165,8 @@ public class TestTableInputFormat { static void runTestMapred(Table table) throws IOException { org.apache.hadoop.hbase.mapred.TableRecordReader trr = new org.apache.hadoop.hbase.mapred.TableRecordReader(); - trr.setStartRow(Bytes.toBytes("aaa")); - trr.setEndRow(Bytes.toBytes("zzz")); + trr.setStartRow("aaa".getBytes()); + trr.setEndRow("zzz".getBytes()); trr.setHTable(table); trr.setInputColumns(columns); @@ -176,11 +176,11 @@ public class TestTableInputFormat { boolean more = trr.next(key, r); assertTrue(more); - checkResult(r, key, Bytes.toBytes("aaa"), Bytes.toBytes("value aaa")); + checkResult(r, key, "aaa".getBytes(), "value aaa".getBytes()); more = trr.next(key, r); assertTrue(more); - checkResult(r, key, Bytes.toBytes("bbb"), Bytes.toBytes("value bbb")); + checkResult(r, key, "bbb".getBytes(), "value bbb".getBytes()); // no more data more = trr.next(key, r); @@ -204,7 +204,7 @@ public class TestTableInputFormat { if (cnt++ < failCnt) { // create mock ResultScanner that always fails. Scan scan = mock(Scan.class); - doReturn(Bytes.toBytes("bogus")).when(scan).getStartRow(); // avoid npe + doReturn("bogus".getBytes()).when(scan).getStartRow(); // avoid npe ResultScanner scanner = mock(ResultScanner.class); // simulate TimeoutException / IOException doThrow(new IOException("Injected exception")).when(scanner).next(); @@ -239,7 +239,7 @@ public class TestTableInputFormat { if (cnt++ < failCnt) { // create mock ResultScanner that always fails. Scan scan = mock(Scan.class); - doReturn(Bytes.toBytes("bogus")).when(scan).getStartRow(); // avoid npe + doReturn("bogus".getBytes()).when(scan).getStartRow(); // avoid npe ResultScanner scanner = mock(ResultScanner.class); invocation.callRealMethod(); // simulate NotServingRegionException @@ -266,7 +266,7 @@ public class TestTableInputFormat { */ @Test public void testTableRecordReader() throws IOException { - Table table = createTable(Bytes.toBytes("table1")); + Table table = createTable("table1".getBytes()); runTestMapred(table); } @@ -277,7 +277,7 @@ public class TestTableInputFormat { */ @Test public void testTableRecordReaderScannerFail() throws IOException { - Table htable = createIOEScannerTable(Bytes.toBytes("table2"), 1); + Table htable = createIOEScannerTable("table2".getBytes(), 1); runTestMapred(htable); } @@ -288,7 +288,7 @@ public class TestTableInputFormat { */ @Test(expected = IOException.class) public void testTableRecordReaderScannerFailTwice() throws IOException { - Table htable = createIOEScannerTable(Bytes.toBytes("table3"), 2); + Table htable = createIOEScannerTable("table3".getBytes(), 2); runTestMapred(htable); } @@ -299,7 +299,7 @@ public class TestTableInputFormat { */ @Test public void testTableRecordReaderScannerTimeout() throws IOException { - Table htable = createDNRIOEScannerTable(Bytes.toBytes("table4"), 1); + Table htable = createDNRIOEScannerTable("table4".getBytes(), 1); runTestMapred(htable); } @@ -310,7 +310,7 @@ public class TestTableInputFormat { */ @Test(expected = org.apache.hadoop.hbase.NotServingRegionException.class) public void testTableRecordReaderScannerTimeoutTwice() throws IOException { - Table htable = createDNRIOEScannerTable(Bytes.toBytes("table5"), 2); + Table htable = createDNRIOEScannerTable("table5".getBytes(), 2); runTestMapred(htable); } diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java index 1b94ca8a75a..5591e5fa4f6 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestCopyTable.java @@ -301,7 +301,7 @@ public class TestCopyTable { assertNotNull(t2.get(new Get(ROW1)).getRow()); Result res = t2.get(new Get(ROW1)); byte[] b1 = res.getValue(FAMILY_B, QUALIFIER); - assertEquals("Data13", Bytes.toString(b1)); + assertEquals("Data13", new String(b1)); assertNotNull(t2.get(new Get(ROW2)).getRow()); res = t2.get(new Get(ROW2)); b1 = res.getValue(FAMILY_A, QUALIFIER); diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java index 9c4da4b8910..8a189c5ac3b 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestImportExport.java @@ -415,7 +415,7 @@ public class TestImportExport { //Add second version of QUAL p = new Put(ROW1); - p.addColumn(FAMILYA, QUAL, now + 5, Bytes.toBytes("s")); + p.addColumn(FAMILYA, QUAL, now + 5, "s".getBytes()); exportT.put(p); //Add second Delete family marker diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRegionSizeCalculator.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRegionSizeCalculator.java index c1562cc7414..85d0f1c8ddd 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRegionSizeCalculator.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestRegionSizeCalculator.java @@ -36,7 +36,6 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.testclassification.MiscTests; import org.apache.hadoop.hbase.testclassification.SmallTests; -import org.apache.hadoop.hbase.util.Bytes; import org.junit.ClassRule; import org.junit.Test; import org.junit.experimental.categories.Category; @@ -67,11 +66,11 @@ public class TestRegionSizeCalculator { RegionSizeCalculator calculator = new RegionSizeCalculator(regionLocator, admin); - assertEquals(123 * megabyte, calculator.getRegionSize(Bytes.toBytes("region1"))); - assertEquals(54321 * megabyte, calculator.getRegionSize(Bytes.toBytes("region2"))); - assertEquals(1232 * megabyte, calculator.getRegionSize(Bytes.toBytes("region3"))); + assertEquals(123 * megabyte, calculator.getRegionSize("region1".getBytes())); + assertEquals(54321 * megabyte, calculator.getRegionSize("region2".getBytes())); + assertEquals(1232 * megabyte, calculator.getRegionSize("region3".getBytes())); // if regionCalculator does not know about a region, it should return 0 - assertEquals(0 * megabyte, calculator.getRegionSize(Bytes.toBytes("otherTableRegion"))); + assertEquals(0 * megabyte, calculator.getRegionSize("otherTableRegion".getBytes())); assertEquals(3, calculator.getRegionSizeMap().size()); } @@ -92,8 +91,7 @@ public class TestRegionSizeCalculator { RegionSizeCalculator calculator = new RegionSizeCalculator(regionLocator, admin); - assertEquals(((long) Integer.MAX_VALUE) * megabyte, - calculator.getRegionSize(Bytes.toBytes("largeRegion"))); + assertEquals(((long) Integer.MAX_VALUE) * megabyte, calculator.getRegionSize("largeRegion".getBytes())); } /** When calculator is disabled, it should return 0 for each request.*/ @@ -108,12 +106,12 @@ public class TestRegionSizeCalculator { //first request on enabled calculator RegionSizeCalculator calculator = new RegionSizeCalculator(table, admin); - assertEquals(999 * megabyte, calculator.getRegionSize(Bytes.toBytes(regionName))); + assertEquals(999 * megabyte, calculator.getRegionSize(regionName.getBytes())); //then disabled calculator. configuration.setBoolean(RegionSizeCalculator.ENABLE_REGIONSIZECALCULATOR, false); RegionSizeCalculator disabledCalculator = new RegionSizeCalculator(table, admin); - assertEquals(0 * megabyte, disabledCalculator.getRegionSize(Bytes.toBytes(regionName))); + assertEquals(0 * megabyte, disabledCalculator.getRegionSize(regionName.getBytes())); assertEquals(0, disabledCalculator.getRegionSizeMap().size()); } @@ -129,7 +127,7 @@ public class TestRegionSizeCalculator { for (String regionName : regionNames) { HRegionInfo info = Mockito.mock(HRegionInfo.class); - when(info.getRegionName()).thenReturn(Bytes.toBytes(regionName)); + when(info.getRegionName()).thenReturn(regionName.getBytes()); regionLocations.add(new HRegionLocation(info, sn)); } @@ -158,7 +156,7 @@ public class TestRegionSizeCalculator { * */ private RegionMetrics mockRegion(String regionName, int fileSizeMb) { RegionMetrics region = Mockito.mock(RegionMetrics.class); - when(region.getRegionName()).thenReturn(Bytes.toBytes(regionName)); + when(region.getRegionName()).thenReturn(regionName.getBytes()); when(region.getNameAsString()).thenReturn(regionName); when(region.getStoreFileSize()).thenReturn(new Size(fileSizeMb, Size.Unit.MEGABYTE)); return region; diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormat.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormat.java index 0697e726cc9..eae860627b6 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormat.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableInputFormat.java @@ -121,14 +121,14 @@ public class TestTableInputFormat { */ public static Table createTable(byte[] tableName, byte[][] families) throws IOException { Table table = UTIL.createTable(TableName.valueOf(tableName), families); - Put p = new Put(Bytes.toBytes("aaa")); + Put p = new Put("aaa".getBytes()); for (byte[] family : families) { - p.addColumn(family, null, Bytes.toBytes("value aaa")); + p.addColumn(family, null, "value aaa".getBytes()); } table.put(p); - p = new Put(Bytes.toBytes("bbb")); + p = new Put("bbb".getBytes()); for (byte[] family : families) { - p.addColumn(family, null, Bytes.toBytes("value bbb")); + p.addColumn(family, null, "value bbb".getBytes()); } table.put(p); return table; @@ -165,8 +165,8 @@ public class TestTableInputFormat { org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl trr = new org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl(); Scan s = new Scan(); - s.setStartRow(Bytes.toBytes("aaa")); - s.setStopRow(Bytes.toBytes("zzz")); + s.setStartRow("aaa".getBytes()); + s.setStopRow("zzz".getBytes()); s.addFamily(FAMILY); trr.setScan(s); trr.setHTable(table); @@ -179,13 +179,13 @@ public class TestTableInputFormat { assertTrue(more); key = trr.getCurrentKey(); r = trr.getCurrentValue(); - checkResult(r, key, Bytes.toBytes("aaa"), Bytes.toBytes("value aaa")); + checkResult(r, key, "aaa".getBytes(), "value aaa".getBytes()); more = trr.nextKeyValue(); assertTrue(more); key = trr.getCurrentKey(); r = trr.getCurrentValue(); - checkResult(r, key, Bytes.toBytes("bbb"), Bytes.toBytes("value bbb")); + checkResult(r, key, "bbb".getBytes(), "value bbb".getBytes()); // no more data more = trr.nextKeyValue(); @@ -209,7 +209,7 @@ public class TestTableInputFormat { if (cnt++ < failCnt) { // create mock ResultScanner that always fails. Scan scan = mock(Scan.class); - doReturn(Bytes.toBytes("bogus")).when(scan).getStartRow(); // avoid npe + doReturn("bogus".getBytes()).when(scan).getStartRow(); // avoid npe ResultScanner scanner = mock(ResultScanner.class); // simulate TimeoutException / IOException doThrow(new IOException("Injected exception")).when(scanner).next(); @@ -244,7 +244,7 @@ public class TestTableInputFormat { if (cnt++ < failCnt) { // create mock ResultScanner that always fails. Scan scan = mock(Scan.class); - doReturn(Bytes.toBytes("bogus")).when(scan).getStartRow(); // avoid npe + doReturn("bogus".getBytes()).when(scan).getStartRow(); // avoid npe ResultScanner scanner = mock(ResultScanner.class); invocation.callRealMethod(); // simulate NotServingRegionException @@ -273,7 +273,7 @@ public class TestTableInputFormat { @Test public void testTableRecordReaderMapreduce() throws IOException, InterruptedException { - Table table = createTable(Bytes.toBytes("table1-mr")); + Table table = createTable("table1-mr".getBytes()); runTestMapreduce(table); } @@ -286,7 +286,7 @@ public class TestTableInputFormat { @Test public void testTableRecordReaderScannerFailMapreduce() throws IOException, InterruptedException { - Table htable = createIOEScannerTable(Bytes.toBytes("table2-mr"), 1); + Table htable = createIOEScannerTable("table2-mr".getBytes(), 1); runTestMapreduce(htable); } @@ -299,7 +299,7 @@ public class TestTableInputFormat { @Test(expected = IOException.class) public void testTableRecordReaderScannerFailMapreduceTwice() throws IOException, InterruptedException { - Table htable = createIOEScannerTable(Bytes.toBytes("table3-mr"), 2); + Table htable = createIOEScannerTable("table3-mr".getBytes(), 2); runTestMapreduce(htable); } @@ -312,7 +312,7 @@ public class TestTableInputFormat { @Test public void testTableRecordReaderScannerTimeoutMapreduce() throws IOException, InterruptedException { - Table htable = createDNRIOEScannerTable(Bytes.toBytes("table4-mr"), 1); + Table htable = createDNRIOEScannerTable("table4-mr".getBytes(), 1); runTestMapreduce(htable); } @@ -325,7 +325,7 @@ public class TestTableInputFormat { @Test(expected = org.apache.hadoop.hbase.NotServingRegionException.class) public void testTableRecordReaderScannerTimeoutMapreduceTwice() throws IOException, InterruptedException { - Table htable = createDNRIOEScannerTable(Bytes.toBytes("table5-mr"), 2); + Table htable = createDNRIOEScannerTable("table5-mr".getBytes(), 2); runTestMapreduce(htable); } diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableSplit.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableSplit.java index 4d754fa4fdb..9be163448c8 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableSplit.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/mapreduce/TestTableSplit.java @@ -26,7 +26,6 @@ import org.apache.hadoop.hbase.HBaseClassTestRule; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.testclassification.MapReduceTests; import org.apache.hadoop.hbase.testclassification.SmallTests; -import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.util.ReflectionUtils; import org.junit.Assert; import org.junit.ClassRule; @@ -48,11 +47,11 @@ public class TestTableSplit { @Test public void testHashCode() { TableSplit split1 = new TableSplit(TableName.valueOf(name.getMethodName()), - Bytes.toBytes("row-start"), - Bytes.toBytes("row-end"), "location"); + "row-start".getBytes(), + "row-end".getBytes(), "location"); TableSplit split2 = new TableSplit(TableName.valueOf(name.getMethodName()), - Bytes.toBytes("row-start"), - Bytes.toBytes("row-end"), "location"); + "row-start".getBytes(), + "row-end".getBytes(), "location"); assertEquals (split1, split2); assertTrue (split1.hashCode() == split2.hashCode()); HashSet set = new HashSet<>(2); @@ -67,11 +66,11 @@ public class TestTableSplit { @Test public void testHashCode_length() { TableSplit split1 = new TableSplit(TableName.valueOf(name.getMethodName()), - Bytes.toBytes("row-start"), - Bytes.toBytes("row-end"), "location", 1984); + "row-start".getBytes(), + "row-end".getBytes(), "location", 1984); TableSplit split2 = new TableSplit(TableName.valueOf(name.getMethodName()), - Bytes.toBytes("row-start"), - Bytes.toBytes("row-end"), "location", 1982); + "row-start".getBytes(), + "row-end".getBytes(), "location", 1982); assertEquals (split1, split2); assertTrue (split1.hashCode() == split2.hashCode()); @@ -87,12 +86,12 @@ public class TestTableSplit { @Test public void testLengthIsSerialized() throws Exception { TableSplit split1 = new TableSplit(TableName.valueOf(name.getMethodName()), - Bytes.toBytes("row-start"), - Bytes.toBytes("row-end"), "location", 666); + "row-start".getBytes(), + "row-end".getBytes(), "location", 666); TableSplit deserialized = new TableSplit(TableName.valueOf(name.getMethodName()), - Bytes.toBytes("row-start2"), - Bytes.toBytes("row-end2"), "location1"); + "row-start2".getBytes(), + "row-end2".getBytes(), "location1"); ReflectionUtils.copy(new Configuration(), split1, deserialized); Assert.assertEquals(666, deserialized.getLength()); @@ -101,8 +100,8 @@ public class TestTableSplit { @Test public void testToString() { TableSplit split = - new TableSplit(TableName.valueOf(name.getMethodName()), Bytes.toBytes("row-start"), - Bytes.toBytes("row-end"), "location"); + new TableSplit(TableName.valueOf(name.getMethodName()), "row-start".getBytes(), "row-end".getBytes(), + "location"); String str = "HBase table split(table name: " + name.getMethodName() + ", scan: , start row: row-start, " + "end row: row-end, region location: location, " @@ -110,8 +109,8 @@ public class TestTableSplit { Assert.assertEquals(str, split.toString()); split = - new TableSplit(TableName.valueOf(name.getMethodName()), null, Bytes.toBytes("row-start"), - Bytes.toBytes("row-end"), "location", "encoded-region-name", 1000L); + new TableSplit(TableName.valueOf(name.getMethodName()), null, "row-start".getBytes(), + "row-end".getBytes(), "location", "encoded-region-name", 1000L); str = "HBase table split(table name: " + name.getMethodName() + ", scan: , start row: row-start, " + "end row: row-end, region location: location, " diff --git a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/replication/TestVerifyReplicationCrossDiffHdfs.java b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/replication/TestVerifyReplicationCrossDiffHdfs.java index aaa5b6ca4c6..a07e0a82aed 100644 --- a/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/replication/TestVerifyReplicationCrossDiffHdfs.java +++ b/hbase-mapreduce/src/test/java/org/apache/hadoop/hbase/replication/TestVerifyReplicationCrossDiffHdfs.java @@ -168,15 +168,15 @@ public class TestVerifyReplicationCrossDiffHdfs { Path rootDir = FSUtils.getRootDir(conf1); FileSystem fs = rootDir.getFileSystem(conf1); String sourceSnapshotName = "sourceSnapshot-" + System.currentTimeMillis(); - SnapshotTestingUtils.createSnapshotAndValidate(util1.getAdmin(), TABLE_NAME, - Bytes.toString(FAMILY), sourceSnapshotName, rootDir, fs, true); + SnapshotTestingUtils.createSnapshotAndValidate(util1.getAdmin(), TABLE_NAME, new String(FAMILY), + sourceSnapshotName, rootDir, fs, true); // Take target snapshot Path peerRootDir = FSUtils.getRootDir(conf2); FileSystem peerFs = peerRootDir.getFileSystem(conf2); String peerSnapshotName = "peerSnapshot-" + System.currentTimeMillis(); - SnapshotTestingUtils.createSnapshotAndValidate(util2.getAdmin(), TABLE_NAME, - Bytes.toString(FAMILY), peerSnapshotName, peerRootDir, peerFs, true); + SnapshotTestingUtils.createSnapshotAndValidate(util2.getAdmin(), TABLE_NAME, new String(FAMILY), + peerSnapshotName, peerRootDir, peerFs, true); String peerFSAddress = peerFs.getUri().toString(); String temPath1 = new Path(fs.getUri().toString(), "/tmp1").toString(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestSerialization.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestSerialization.java index 24c9d78fa8a..9f0cf6ba18a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/TestSerialization.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/TestSerialization.java @@ -61,11 +61,11 @@ public class TestSerialization { @Test public void testKeyValue() throws Exception { final String name = "testKeyValue2"; - byte[] row = Bytes.toBytes(name); - byte[] fam = Bytes.toBytes("fam"); - byte[] qf = Bytes.toBytes("qf"); + byte[] row = name.getBytes(); + byte[] fam = "fam".getBytes(); + byte[] qf = "qf".getBytes(); long ts = System.currentTimeMillis(); - byte[] val = Bytes.toBytes("val"); + byte[] val = "val".getBytes(); KeyValue kv = new KeyValue(row, fam, qf, ts, val); ByteArrayOutputStream baos = new ByteArrayOutputStream(); DataOutputStream dos = new DataOutputStream(baos); @@ -290,9 +290,9 @@ public class TestSerialization { */ @Test public void testGet() throws Exception{ - byte[] row = Bytes.toBytes("row"); - byte[] fam = Bytes.toBytes("fam"); - byte[] qf1 = Bytes.toBytes("qf1"); + byte[] row = "row".getBytes(); + byte[] fam = "fam".getBytes(); + byte[] qf1 = "qf1".getBytes(); long ts = System.currentTimeMillis(); int maxVersions = 2; @@ -329,10 +329,10 @@ public class TestSerialization { @Test public void testScan() throws Exception { - byte[] startRow = Bytes.toBytes("startRow"); - byte[] stopRow = Bytes.toBytes("stopRow"); - byte[] fam = Bytes.toBytes("fam"); - byte[] qf1 = Bytes.toBytes("qf1"); + byte[] startRow = "startRow".getBytes(); + byte[] stopRow = "stopRow".getBytes(); + byte[] fam = "fam".getBytes(); + byte[] qf1 = "qf1".getBytes(); long ts = System.currentTimeMillis(); int maxVersions = 2; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java index 376aa92cb6e..dfc3a2cd713 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAdmin1.java @@ -686,7 +686,7 @@ public class TestAdmin1 { TableName TABLE_3 = TableName.valueOf(tableName.getNameAsString() + "_3"); desc = new HTableDescriptor(TABLE_3); desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY)); - admin.createTable(desc, Bytes.toBytes("a"), Bytes.toBytes("z"), 3); + admin.createTable(desc, "a".getBytes(), "z".getBytes(), 3); try (RegionLocator l = TEST_UTIL.getConnection().getRegionLocator(TABLE_3)) { regions = l.getAllRegionLocations(); assertEquals("Table should have only 3 region", 3, regions.size()); @@ -696,7 +696,7 @@ public class TestAdmin1 { desc = new HTableDescriptor(TABLE_4); desc.addFamily(new HColumnDescriptor(HConstants.CATALOG_FAMILY)); try { - admin.createTable(desc, Bytes.toBytes("a"), Bytes.toBytes("z"), 2); + admin.createTable(desc, "a".getBytes(), "z".getBytes(), 2); fail("Should not be able to create a table with only 2 regions using this API."); } catch (IllegalArgumentException eae) { // Expected @@ -921,9 +921,9 @@ public class TestAdmin1 { public void testCreateTableWithEmptyRowInTheSplitKeys() throws IOException{ final byte[] tableName = Bytes.toBytes(name.getMethodName()); byte[][] splitKeys = new byte[3][]; - splitKeys[0] = Bytes.toBytes("region1"); + splitKeys[0] = "region1".getBytes(); splitKeys[1] = HConstants.EMPTY_BYTE_ARRAY; - splitKeys[2] = Bytes.toBytes("region2"); + splitKeys[2] = "region2".getBytes(); HTableDescriptor desc = new HTableDescriptor(TableName.valueOf(tableName)); desc.addFamily(new HColumnDescriptor("col")); try { @@ -1181,7 +1181,7 @@ public class TestAdmin1 { final TableName tableName = TableName.valueOf(name.getMethodName()); HTableDescriptor desc = new HTableDescriptor(tableName); desc.setRegionReplication(3); - byte[] cf = Bytes.toBytes("f"); + byte[] cf = "f".getBytes(); HColumnDescriptor hcd = new HColumnDescriptor(cf); desc.addFamily(hcd); byte[][] splitRows = new byte[2][]; @@ -1196,15 +1196,15 @@ public class TestAdmin1 { // write some data to the table Table ht = TEST_UTIL.getConnection().getTable(tableName); List puts = new ArrayList<>(); - byte[] qualifier = Bytes.toBytes("c"); + byte[] qualifier = "c".getBytes(); Put put = new Put(new byte[]{(byte)'1'}); - put.addColumn(cf, qualifier, Bytes.toBytes("100")); + put.addColumn(cf, qualifier, "100".getBytes()); puts.add(put); put = new Put(new byte[]{(byte)'6'}); - put.addColumn(cf, qualifier, Bytes.toBytes("100")); + put.addColumn(cf, qualifier, "100".getBytes()); puts.add(put); put = new Put(new byte[]{(byte)'8'}); - put.addColumn(cf, qualifier, Bytes.toBytes("100")); + put.addColumn(cf, qualifier, "100".getBytes()); puts.add(put); ht.put(puts); ht.close(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAsyncTableAdminApi.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAsyncTableAdminApi.java index 990eed424d3..1750926785b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAsyncTableAdminApi.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestAsyncTableAdminApi.java @@ -100,7 +100,7 @@ public class TestAsyncTableAdminApi extends TestAsyncAdminBase { final TableName tableName3 = TableName.valueOf(tableName.getNameAsString() + "_3"); TableDescriptorBuilder builder = TableDescriptorBuilder.newBuilder(tableName3); builder.setColumnFamily(ColumnFamilyDescriptorBuilder.of(FAMILY)); - admin.createTable(builder.build(), Bytes.toBytes("a"), Bytes.toBytes("z"), 3).join(); + admin.createTable(builder.build(), "a".getBytes(), "z".getBytes(), 3).join(); regionLocations = AsyncMetaTableAccessor.getTableHRegionLocations(metaTable, Optional.of(tableName3)).get(); assertEquals("Table should have only 3 region", 3, regionLocations.size()); @@ -109,7 +109,7 @@ public class TestAsyncTableAdminApi extends TestAsyncAdminBase { builder = TableDescriptorBuilder.newBuilder(tableName4); builder.setColumnFamily(ColumnFamilyDescriptorBuilder.of(FAMILY)); try { - admin.createTable(builder.build(), Bytes.toBytes("a"), Bytes.toBytes("z"), 2).join(); + admin.createTable(builder.build(), "a".getBytes(), "z".getBytes(), 2).join(); fail("Should not be able to create a table with only 2 regions using this API."); } catch (CompletionException e) { assertTrue(e.getCause() instanceof IllegalArgumentException); @@ -309,9 +309,9 @@ public class TestAsyncTableAdminApi extends TestAsyncAdminBase { @Test public void testCreateTableWithEmptyRowInTheSplitKeys() throws Exception { byte[][] splitKeys = new byte[3][]; - splitKeys[0] = Bytes.toBytes("region1"); + splitKeys[0] = "region1".getBytes(); splitKeys[1] = HConstants.EMPTY_BYTE_ARRAY; - splitKeys[2] = Bytes.toBytes("region2"); + splitKeys[2] = "region2".getBytes(); try { createTableWithDefaultConf(tableName, splitKeys); fail("Test case should fail as empty split key is passed."); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestConnectionImplementation.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestConnectionImplementation.java index 9523a615a14..4d9f39bebe0 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestConnectionImplementation.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestConnectionImplementation.java @@ -218,8 +218,8 @@ public class TestConnectionImplementation { // dead servers is broke" public void testClusterStatus() throws Exception { final TableName tableName = TableName.valueOf(name.getMethodName()); - byte[] cf = Bytes.toBytes("cf"); - byte[] rk = Bytes.toBytes("rk1"); + byte[] cf = "cf".getBytes(); + byte[] rk = "rk1".getBytes(); JVMClusterUtil.RegionServerThread rs = TEST_UTIL.getHBaseCluster().startRegionServer(); rs.waitForServerOnline(); @@ -242,7 +242,7 @@ public class TestConnectionImplementation { } Put p1 = new Put(rk); - p1.addColumn(cf, Bytes.toBytes("qual"), Bytes.toBytes("val")); + p1.addColumn(cf, "qual".getBytes(), "val".getBytes()); t.put(p1); rs.getRegionServer().abort("I'm dead"); @@ -606,7 +606,7 @@ public class TestConnectionImplementation { LOG.info("Move starting region="+toMove.getRegionInfo().getRegionNameAsString()); TEST_UTIL.getAdmin().move( toMove.getRegionInfo().getEncodedNameAsBytes(), - Bytes.toBytes(destServerName.getServerName()) + destServerName.getServerName().getBytes() ); while (destServer.getOnlineRegion(regionName) == null || @@ -672,7 +672,7 @@ public class TestConnectionImplementation { LOG.info("Move starting region=" + toMove.getRegionInfo().getRegionNameAsString()); TEST_UTIL.getAdmin().move( toMove.getRegionInfo().getEncodedNameAsBytes(), - Bytes.toBytes(curServer.getServerName().getServerName()) + curServer.getServerName().getServerName().getBytes() ); while (curServer.getOnlineRegion(regionName) == null || @@ -930,7 +930,7 @@ public class TestConnectionImplementation { LOG.info("Move starting region=" + toMove.getRegionInfo().getRegionNameAsString()); TEST_UTIL.getAdmin().move( toMove.getRegionInfo().getEncodedNameAsBytes(), - Bytes.toBytes(destServerName.getServerName()) + destServerName.getServerName().getBytes() ); while (destServer.getOnlineRegion(regionName) == null || diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java index 9e65c57f483..5d088b3257b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFastFail.java @@ -367,7 +367,7 @@ public class TestFastFail { } private byte[] longToByteArrayKey(long rowKey) { - return Bytes.toBytes(LoadTestKVGenerator.md5PrefixedKey(rowKey)); + return LoadTestKVGenerator.md5PrefixedKey(rowKey).getBytes(); } public static class CallQueueTooBigPffeInterceptor extends diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestGetScanPartialResult.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestGetScanPartialResult.java index 63976e0169a..e9295c4fda0 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestGetScanPartialResult.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestGetScanPartialResult.java @@ -26,7 +26,6 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.MediumTests; -import org.apache.hadoop.hbase.util.Bytes; import org.junit.AfterClass; import org.junit.BeforeClass; import org.junit.ClassRule; @@ -78,7 +77,7 @@ public class TestGetScanPartialResult { byte[] val = makeLargeValue(VALUE_SIZE); Put p = new Put(ROW); for (int i = 0; i < NUM_COLUMNS; i++) { - p.addColumn(CF, Bytes.toBytes(Integer.toString(i)), val); + p.addColumn(CF, Integer.toString(i).getBytes(), val); } t.put(p); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaWithReplicas.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaWithReplicas.java index 6c633a2716d..3b14b7f0a53 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaWithReplicas.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMetaWithReplicas.java @@ -258,10 +258,10 @@ public class TestMetaWithReplicas { LOG.info("Running GETs"); Get get = null; Result r = null; - byte[] row = Bytes.toBytes("test"); + byte[] row = "test".getBytes(); try (Table htable = c.getTable(TABLE)) { Put put = new Put(row); - put.addColumn(Bytes.toBytes("foo"), row, row); + put.addColumn("foo".getBytes(), row, row); BufferedMutator m = c.getBufferedMutator(TABLE); m.mutate(put); m.flush(); @@ -296,7 +296,7 @@ public class TestMetaWithReplicas { TEST_UTIL.getAdmin().deleteTable(tableName); } try (Table htable = TEST_UTIL.createTable(tableName, FAMILIES)) { - byte[] row = Bytes.toBytes("test"); + byte[] row = "test".getBytes(); ConnectionImplementation c = ((ConnectionImplementation) TEST_UTIL.getConnection()); // check that metalookup pool would get created c.relocateRegion(tableName, row); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java index a42b26d857c..59920d6eee3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicaWithCluster.java @@ -74,7 +74,7 @@ public class TestReplicaWithCluster { private static final Logger LOG = LoggerFactory.getLogger(TestReplicaWithCluster.class); private static final int NB_SERVERS = 3; - private static final byte[] row = Bytes.toBytes(TestReplicaWithCluster.class.getName()); + private static final byte[] row = TestReplicaWithCluster.class.getName().getBytes(); private static final HBaseTestingUtility HTU = new HBaseTestingUtility(); // second minicluster used in testing of replication diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java index 4a72410200b..311f65185ed 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestReplicasClient.java @@ -87,7 +87,7 @@ public class TestReplicasClient { private static final int NB_SERVERS = 1; private static Table table = null; - private static final byte[] row = Bytes.toBytes(TestReplicasClient.class.getName()); + private static final byte[] row = TestReplicasClient.class.getName().getBytes(); private static HRegionInfo hriPrimary; private static HRegionInfo hriSecondary; @@ -313,7 +313,7 @@ public class TestReplicasClient { @Test public void testUseRegionWithoutReplica() throws Exception { - byte[] b1 = Bytes.toBytes("testUseRegionWithoutReplica"); + byte[] b1 = "testUseRegionWithoutReplica".getBytes(); openRegion(hriSecondary); SlowMeCopro.getPrimaryCdl().set(new CountDownLatch(0)); try { @@ -327,7 +327,7 @@ public class TestReplicasClient { @Test public void testLocations() throws Exception { - byte[] b1 = Bytes.toBytes("testLocations"); + byte[] b1 = "testLocations".getBytes(); openRegion(hriSecondary); ClusterConnection hc = (ClusterConnection) HTU.getAdmin().getConnection(); @@ -352,7 +352,7 @@ public class TestReplicasClient { @Test public void testGetNoResultNoStaleRegionWithReplica() throws Exception { - byte[] b1 = Bytes.toBytes("testGetNoResultNoStaleRegionWithReplica"); + byte[] b1 = "testGetNoResultNoStaleRegionWithReplica".getBytes(); openRegion(hriSecondary); try { @@ -368,7 +368,7 @@ public class TestReplicasClient { @Test public void testGetNoResultStaleRegionWithReplica() throws Exception { - byte[] b1 = Bytes.toBytes("testGetNoResultStaleRegionWithReplica"); + byte[] b1 = "testGetNoResultStaleRegionWithReplica".getBytes(); openRegion(hriSecondary); SlowMeCopro.getPrimaryCdl().set(new CountDownLatch(1)); @@ -385,7 +385,7 @@ public class TestReplicasClient { @Test public void testGetNoResultNotStaleSleepRegionWithReplica() throws Exception { - byte[] b1 = Bytes.toBytes("testGetNoResultNotStaleSleepRegionWithReplica"); + byte[] b1 = "testGetNoResultNotStaleSleepRegionWithReplica".getBytes(); openRegion(hriSecondary); try { @@ -461,7 +461,7 @@ public class TestReplicasClient { @Test public void testUseRegionWithReplica() throws Exception { - byte[] b1 = Bytes.toBytes("testUseRegionWithReplica"); + byte[] b1 = "testUseRegionWithReplica".getBytes(); openRegion(hriSecondary); try { @@ -554,7 +554,7 @@ public class TestReplicasClient { @Test public void testHedgedRead() throws Exception { - byte[] b1 = Bytes.toBytes("testHedgedRead"); + byte[] b1 = "testHedgedRead".getBytes(); openRegion(hriSecondary); try { @@ -769,7 +769,7 @@ public class TestReplicasClient { for (int col = 0; col < NUMCOLS; col++) { Put p = new Put(b1); String qualifier = "qualifer" + col; - KeyValue kv = new KeyValue(b1, f, Bytes.toBytes(qualifier)); + KeyValue kv = new KeyValue(b1, f, qualifier.getBytes()); p.add(kv); table.put(p); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java index b222cc9c555..ef3fa10f0ee 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRegionObserverInterface.java @@ -589,7 +589,7 @@ public class TestRegionObserverInterface { ServerName sn2 = rs1.getRegionServer().getServerName(); String regEN = locator.getAllRegionLocations().get(0).getRegionInfo().getEncodedName(); - util.getAdmin().move(Bytes.toBytes(regEN), Bytes.toBytes(sn2.getServerName())); + util.getAdmin().move(regEN.getBytes(), sn2.getServerName().getBytes()); while (!sn2.equals(locator.getAllRegionLocations().get(0).getServerName())) { Thread.sleep(100); } @@ -639,7 +639,7 @@ public class TestRegionObserverInterface { ServerName sn2 = rs1.getRegionServer().getServerName(); String regEN = locator.getAllRegionLocations().get(0).getRegionInfo().getEncodedName(); - util.getAdmin().move(Bytes.toBytes(regEN), Bytes.toBytes(sn2.getServerName())); + util.getAdmin().move(regEN.getBytes(), sn2.getServerName().getBytes()); while (!sn2.equals(locator.getAllRegionLocations().get(0).getServerName())) { Thread.sleep(100); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java index 05df8b0c18b..85e02cb257a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFilterWrapper.java @@ -99,7 +99,7 @@ public class TestFilterWrapper { for (Cell kv : result.listCells()) { LOG.debug(kv_number + ". kv: " + kv); kv_number++; - assertEquals("Returned row is not correct", Bytes.toString(CellUtil.cloneRow(kv)), + assertEquals("Returned row is not correct", new String(CellUtil.cloneRow(kv)), "row" + ( row_number + 1 )); } } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowAndColumnRangeFilter.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowAndColumnRangeFilter.java index dcd9cebf307..a612290208b 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowAndColumnRangeFilter.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowAndColumnRangeFilter.java @@ -125,7 +125,7 @@ public class TestFuzzyRowAndColumnRangeFilter { Put p = new Put(rk); p.setDurability(Durability.SKIP_WAL); - p.addColumn(Bytes.toBytes(cf), cq, Bytes.toBytes(c)); + p.addColumn(cf.getBytes(), cq, Bytes.toBytes(c)); ht.put(p); LOG.info("Inserting: rk: " + Bytes.toStringBinary(rk) + " cq: " + Bytes.toStringBinary(cq)); @@ -167,7 +167,7 @@ public class TestFuzzyRowAndColumnRangeFilter { private void runScanner(Table hTable, int expectedSize, Filter... filters) throws IOException { String cf = "f"; Scan scan = new Scan(); - scan.addFamily(Bytes.toBytes(cf)); + scan.addFamily(cf.getBytes()); FilterList filterList = new FilterList(filters); scan.setFilter(filterList); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.java index ca538aa4737..cf326311ab0 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestFuzzyRowFilterEndToEnd.java @@ -136,7 +136,7 @@ public class TestFuzzyRowFilterEndToEnd { for (int i = 0; i < rows.length; i++) { Put p = new Put(Bytes.toBytesBinary(rows[i])); - p.addColumn(Bytes.toBytes(cf), Bytes.toBytes(cq), Bytes.toBytes("value")); + p.addColumn(cf.getBytes(), cq.getBytes(), "value".getBytes()); ht.put(p); } @@ -191,12 +191,12 @@ public class TestFuzzyRowFilterEndToEnd { for(int i=0; i < rows.length; i++){ Put p = new Put(Bytes.toBytesBinary(rows[i])); - p.addColumn(Bytes.toBytes(cf), Bytes.toBytes(cq), Bytes.toBytes("value")); + p.addColumn(cf.getBytes(), cq.getBytes(), "value".getBytes()); ht.put(p); } Put p = new Put(Bytes.toBytesBinary(badRow)); - p.addColumn(Bytes.toBytes(cf), Bytes.toBytes(cq), Bytes.toBytes("value")); + p.addColumn(cf.getBytes(), cq.getBytes(), "value".getBytes()); ht.put(p); TEST_UTIL.flush(); @@ -248,7 +248,7 @@ public class TestFuzzyRowFilterEndToEnd { Put p = new Put(rk); p.setDurability(Durability.SKIP_WAL); - p.addColumn(Bytes.toBytes(cf), cq, Bytes.toBytes(c)); + p.addColumn(cf.getBytes(), cq, Bytes.toBytes(c)); ht.put(p); } } @@ -333,9 +333,9 @@ public class TestFuzzyRowFilterEndToEnd { String cf = "f"; Scan scan = new Scan(); - scan.addFamily(Bytes.toBytes(cf)); + scan.addFamily(cf.getBytes()); scan.setFilter(filter); - List regions = TEST_UTIL.getHBaseCluster().getRegions(TableName.valueOf(table)); + List regions = TEST_UTIL.getHBaseCluster().getRegions(table.getBytes()); HRegion first = regions.get(0); first.getScanner(scan); RegionScanner scanner = first.getScanner(scan); @@ -385,7 +385,7 @@ public class TestFuzzyRowFilterEndToEnd { Put p = new Put(rk); p.setDurability(Durability.SKIP_WAL); - p.addColumn(Bytes.toBytes(cf), cq, Bytes.toBytes(c)); + p.addColumn(cf.getBytes(), cq, Bytes.toBytes(c)); ht.put(p); LOG.info("Inserting: rk: " + Bytes.toStringBinary(rk) + " cq: " + Bytes.toStringBinary(cq)); @@ -435,7 +435,7 @@ public class TestFuzzyRowFilterEndToEnd { throws IOException { String cf = "f"; Scan scan = new Scan(); - scan.addFamily(Bytes.toBytes(cf)); + scan.addFamily(cf.getBytes()); FilterList filterList = new FilterList(Operator.MUST_PASS_ALL, filter1, filter2); scan.setFilter(filterList); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java index f53195471b9..e3b78ede593 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/filter/TestScanRowPrefix.java @@ -31,7 +31,6 @@ import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.client.Table; import org.apache.hadoop.hbase.testclassification.FilterTests; import org.apache.hadoop.hbase.testclassification.MediumTests; -import org.apache.hadoop.hbase.util.Bytes; import org.junit.Assert; import org.junit.ClassRule; import org.junit.Rule; @@ -85,7 +84,7 @@ public class TestScanRowPrefix extends FilterTestingCluster { for (byte[] rowId: rowIds) { Put p = new Put(rowId); // Use the rowId as the column qualifier - p.addColumn(Bytes.toBytes("F"), rowId, Bytes.toBytes("Dummy value")); + p.addColumn("F".getBytes(), rowId, "Dummy value".getBytes()); table.put(p); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestEncodedSeekers.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestEncodedSeekers.java index 12f2121872f..350a316a979 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestEncodedSeekers.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/encoding/TestEncodedSeekers.java @@ -149,7 +149,7 @@ public class TestEncodedSeekers { private void doPuts(HRegion region) throws IOException{ LoadTestKVGenerator dataGenerator = new LoadTestKVGenerator(MIN_VALUE_SIZE, MAX_VALUE_SIZE); for (int i = 0; i < NUM_ROWS; ++i) { - byte[] key = Bytes.toBytes(LoadTestKVGenerator.md5PrefixedKey(i)); + byte[] key = LoadTestKVGenerator.md5PrefixedKey(i).getBytes(); for (int j = 0; j < NUM_COLS_PER_ROW; ++j) { Put put = new Put(key); put.setDurability(Durability.ASYNC_WAL); @@ -177,7 +177,7 @@ public class TestEncodedSeekers { private void doGets(Region region) throws IOException{ for (int i = 0; i < NUM_ROWS; ++i) { - final byte[] rowKey = Bytes.toBytes(LoadTestKVGenerator.md5PrefixedKey(i)); + final byte[] rowKey = LoadTestKVGenerator.md5PrefixedKey(i).getBytes(); for (int j = 0; j < NUM_COLS_PER_ROW; ++j) { final String qualStr = String.valueOf(j); if (VERBOSE) { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/RandomKeyValueUtil.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/RandomKeyValueUtil.java index 337a32ff513..b6ff37eca2a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/RandomKeyValueUtil.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/RandomKeyValueUtil.java @@ -18,7 +18,6 @@ package org.apache.hadoop.hbase.io.hfile; import org.apache.hadoop.hbase.KeyValue; -import org.apache.hadoop.hbase.util.Bytes; import java.util.Random; @@ -50,7 +49,7 @@ public class RandomKeyValueUtil { public static KeyValue randomKeyValue(Random rand) { return new KeyValue(randomRowOrQualifier(rand), - Bytes.toBytes(COLUMN_FAMILY_NAME), randomRowOrQualifier(rand), + COLUMN_FAMILY_NAME.getBytes(), randomRowOrQualifier(rand), randomValue(rand)); } @@ -61,7 +60,7 @@ public class RandomKeyValueUtil { - MIN_ROW_OR_QUALIFIER_LENGTH + 1); for (int i = 0; i < fieldLen; ++i) field.append(randomReadableChar(rand)); - return Bytes.toBytes(field.toString()); + return field.toString().getBytes(); } public static byte[] randomValue(Random rand) { @@ -70,7 +69,7 @@ public class RandomKeyValueUtil { v.append((char) (32 + rand.nextInt(95))); } - byte[] valueBytes = Bytes.toBytes(v.toString()); + byte[] valueBytes = v.toString().getBytes(); return valueBytes; } @@ -99,7 +98,7 @@ public class RandomKeyValueUtil { for (int j = 0; j < rand.nextInt(50); ++j) k.append(randomReadableChar(rand)); - byte[] keyBytes = Bytes.toBytes(k.toString()); + byte[] keyBytes = k.toString().getBytes(); return keyBytes; } @@ -118,7 +117,7 @@ public class RandomKeyValueUtil { for (int j = 0; j < suffixLength; ++j) k.append(randomReadableChar(rand)); - byte[] keyBytes = Bytes.toBytes(k.toString()); + byte[] keyBytes = k.toString().getBytes(); return keyBytes; } @@ -128,7 +127,7 @@ public class RandomKeyValueUtil { v.append((char) (32 + rand.nextInt(95))); } - byte[] valueBytes = Bytes.toBytes(v.toString()); + byte[] valueBytes = v.toString().getBytes(); return valueBytes; } } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFile.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFile.java index bc608be58e5..2a613deadaf 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFile.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFile.java @@ -289,7 +289,7 @@ public class TestHFile { } private byte[] getSomeKey(int rowId) { - KeyValue kv = new KeyValue(Bytes.toBytes(String.format(localFormatter, Integer.valueOf(rowId))), + KeyValue kv = new KeyValue(String.format(localFormatter, Integer.valueOf(rowId)).getBytes(), Bytes.toBytes("family"), Bytes.toBytes("qual"), HConstants.LATEST_TIMESTAMP, Type.Put); return kv.getKey(); } @@ -377,7 +377,7 @@ public class TestHFile { @Override public void write(DataOutput out) throws IOException { - out.write(Bytes.toBytes("something to test" + val)); + out.write(("something to test" + val).getBytes()); } @Override @@ -394,7 +394,7 @@ public class TestHFile { for (int i = 0; i < n; i++) { ByteBuff actual = reader.getMetaBlock("HFileMeta" + i, false).getBufferWithoutHeader(); ByteBuffer expected = - ByteBuffer.wrap(Bytes.toBytes("something to test" + i)); + ByteBuffer.wrap(("something to test" + i).getBytes()); assertEquals( "failed to match metadata", Bytes.toStringBinary(expected), @@ -451,8 +451,7 @@ public class TestHFile { .withOutputStream(fout) .withFileContext(meta) .create(); - KeyValue kv = new KeyValue(Bytes.toBytes("foo"), Bytes.toBytes("f1"), null, - Bytes.toBytes("value")); + KeyValue kv = new KeyValue("foo".getBytes(), "f1".getBytes(), null, "value".getBytes()); writer.append(kv); writer.close(); fout.close(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java index 937b6414575..78f85842df2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileBlockIndex.java @@ -276,7 +276,7 @@ public class TestHFileBlockIndex { new HFileBlockIndex.BlockIndexWriter(hbw, null, null); for (int i = 0; i < NUM_DATA_BLOCKS; ++i) { - hbw.startWriting(BlockType.DATA).write(Bytes.toBytes(String.valueOf(rand.nextInt(1000)))); + hbw.startWriting(BlockType.DATA).write(String.valueOf(rand.nextInt(1000)).getBytes()); long blockOffset = outputStream.getPos(); hbw.writeHeaderAndData(outputStream); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileEncryption.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileEncryption.java index d77af6db23f..5eeecefa487 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileEncryption.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileEncryption.java @@ -179,8 +179,7 @@ public class TestHFileEncryption { .withFileContext(fileContext) .create(); try { - KeyValue kv = new KeyValue(Bytes.toBytes("foo"), Bytes.toBytes("f1"), null, - Bytes.toBytes("value")); + KeyValue kv = new KeyValue("foo".getBytes(), "f1".getBytes(), null, "value".getBytes()); writer.append(kv); } finally { writer.close(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileSeek.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileSeek.java index d43c0dac560..4dc9c6889b1 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileSeek.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestHFileSeek.java @@ -36,7 +36,6 @@ import org.apache.hadoop.hbase.io.hfile.HFile.Reader; import org.apache.hadoop.hbase.io.hfile.HFile.Writer; import org.apache.hadoop.hbase.testclassification.IOTests; import org.apache.hadoop.hbase.testclassification.MediumTests; -import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.BytesWritable; import org.junit.ClassRule; import org.junit.experimental.categories.Category; @@ -67,8 +66,8 @@ public class TestHFileSeek extends TestCase { public static final HBaseClassTestRule CLASS_RULE = HBaseClassTestRule.forClass(TestHFileSeek.class); - private static final byte[] CF = Bytes.toBytes("f1"); - private static final byte[] QUAL = Bytes.toBytes("q1"); + private static final byte[] CF = "f1".getBytes(); + private static final byte[] QUAL = "q1".getBytes(); private static final boolean USE_PREAD = true; private MyOptions options; private Configuration conf; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/balancer/TestBaseLoadBalancer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/balancer/TestBaseLoadBalancer.java index 041167984e7..8a0365f50a4 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/balancer/TestBaseLoadBalancer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/balancer/TestBaseLoadBalancer.java @@ -52,7 +52,6 @@ import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster; import org.apache.hadoop.hbase.master.balancer.BaseLoadBalancer.Cluster.MoveRegionAction; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.MediumTests; -import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.net.DNSToSwitchMapping; import org.junit.BeforeClass; import org.junit.ClassRule; @@ -260,8 +259,8 @@ public class TestBaseLoadBalancer extends BalancerTestBase { Mockito.when(services.getServerManager()).thenReturn(sm); balancer.setMasterServices(services); RegionInfo hri1 = RegionInfoBuilder.newBuilder(TableName.valueOf(name.getMethodName())) - .setStartKey(Bytes.toBytes("key1")) - .setEndKey(Bytes.toBytes("key2")) + .setStartKey("key1".getBytes()) + .setEndKey("key2".getBytes()) .setSplit(false) .setRegionId(100) .build(); @@ -285,8 +284,8 @@ public class TestBaseLoadBalancer extends BalancerTestBase { List list2 = new ArrayList<>(); // create a region (region1) RegionInfo hri1 = RegionInfoBuilder.newBuilder(TableName.valueOf(name.getMethodName())) - .setStartKey(Bytes.toBytes("key1")) - .setEndKey(Bytes.toBytes("key2")) + .setStartKey("key1".getBytes()) + .setEndKey("key2".getBytes()) .setSplit(false) .setRegionId(100) .build(); @@ -294,8 +293,8 @@ public class TestBaseLoadBalancer extends BalancerTestBase { RegionInfo hri2 = RegionReplicaUtil.getRegionInfoForReplica(hri1, 1); // create a second region (region2) RegionInfo hri3 = RegionInfoBuilder.newBuilder(TableName.valueOf(name.getMethodName())) - .setStartKey(Bytes.toBytes("key2")) - .setEndKey(Bytes.toBytes("key3")) + .setStartKey("key2".getBytes()) + .setEndKey("key3".getBytes()) .setSplit(false) .setRegionId(101) .build(); @@ -359,8 +358,8 @@ public class TestBaseLoadBalancer extends BalancerTestBase { List list2 = new ArrayList<>(); // create a region (region1) RegionInfo hri1 = RegionInfoBuilder.newBuilder(TableName.valueOf(name.getMethodName())) - .setStartKey(Bytes.toBytes("key1")) - .setEndKey(Bytes.toBytes("key2")) + .setStartKey("key1".getBytes()) + .setEndKey("key2".getBytes()) .setSplit(false) .setRegionId(100) .build(); @@ -368,8 +367,8 @@ public class TestBaseLoadBalancer extends BalancerTestBase { RegionInfo hri2 = RegionReplicaUtil.getRegionInfoForReplica(hri1, 1); // create a second region (region2) RegionInfo hri3 = RegionInfoBuilder.newBuilder(TableName.valueOf(name.getMethodName())) - .setStartKey(Bytes.toBytes("key2")) - .setEndKey(Bytes.toBytes("key3")) + .setStartKey("key2".getBytes()) + .setEndKey("key3".getBytes()) .setSplit(false) .setRegionId(101) .build(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/locking/TestLockManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/locking/TestLockManager.java index d04ef79d5c7..9f964c6f46c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/locking/TestLockManager.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/locking/TestLockManager.java @@ -37,7 +37,6 @@ import org.apache.hadoop.hbase.procedure2.ProcedureExecutor; import org.apache.hadoop.hbase.procedure2.ProcedureTestingUtility; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.MediumTests; -import org.apache.hadoop.hbase.util.Bytes; import org.junit.After; import org.junit.AfterClass; import org.junit.BeforeClass; @@ -81,8 +80,7 @@ public class TestLockManager { UTIL.startMiniCluster(1); masterServices = UTIL.getMiniHBaseCluster().getMaster(); UTIL.getAdmin().createNamespace(NamespaceDescriptor.create(namespace).build()); - UTIL.createTable(tableName, new byte[][]{Bytes.toBytes("fam")}, - new byte[][] {Bytes.toBytes("1")}); + UTIL.createTable(tableName, new byte[][]{"fam".getBytes()}, new byte[][] {"1".getBytes()}); List regions = UTIL.getAdmin().getTableRegions(tableName); assert regions.size() > 0; tableRegions = new HRegionInfo[regions.size()]; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/MasterProcedureTestingUtility.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/MasterProcedureTestingUtility.java index b4d16c6544a..98c39781432 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/MasterProcedureTestingUtility.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/procedure/MasterProcedureTestingUtility.java @@ -289,7 +289,7 @@ public class MasterProcedureTestingUtility { TableDescriptor htd = master.getTableDescriptors().get(tableName); assertTrue(htd != null); - assertTrue(htd.hasColumnFamily(Bytes.toBytes(family))); + assertTrue(htd.hasColumnFamily(family.getBytes())); } public static void validateColumnFamilyDeletion(final HMaster master, final TableName tableName, @@ -297,7 +297,7 @@ public class MasterProcedureTestingUtility { // verify htd TableDescriptor htd = master.getTableDescriptors().get(tableName); assertTrue(htd != null); - assertFalse(htd.hasColumnFamily(Bytes.toBytes(family))); + assertFalse(htd.hasColumnFamily(family.getBytes())); // verify fs final FileSystem fs = master.getMasterFileSystem().getFileSystem(); @@ -314,7 +314,7 @@ public class MasterProcedureTestingUtility { TableDescriptor htd = master.getTableDescriptors().get(tableName); assertTrue(htd != null); - ColumnFamilyDescriptor hcfd = htd.getColumnFamily(Bytes.toBytes(family)); + ColumnFamilyDescriptor hcfd = htd.getColumnFamily(family.getBytes()); assertEquals(0, ColumnFamilyDescriptor.COMPARATOR.compare(hcfd, columnDescriptor)); } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/SimpleRSProcedureManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/SimpleRSProcedureManager.java index 80f931c0bc8..f5a858ab93a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/SimpleRSProcedureManager.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/SimpleRSProcedureManager.java @@ -33,7 +33,6 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.DaemonThreadFactory; import org.apache.hadoop.hbase.regionserver.RegionServerServices; -import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.zookeeper.ZKWatcher; import org.apache.hadoop.hbase.errorhandling.ForeignException; import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher; @@ -250,7 +249,7 @@ public class SimpleRSProcedureManager extends RegionServerProcedureManager { @Override public byte[] insideBarrier() throws ForeignException { execute(); - return Bytes.toBytes(SimpleMasterProcedureManager.SIMPLE_DATA); + return SimpleMasterProcedureManager.SIMPLE_DATA.getBytes(); } /** diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/TestProcedureManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/TestProcedureManager.java index 1b86b50cf18..63273de4bdb 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/TestProcedureManager.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/TestProcedureManager.java @@ -27,7 +27,6 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.MediumTests; -import org.apache.hadoop.hbase.util.Bytes; import org.junit.AfterClass; import org.junit.BeforeClass; import org.junit.ClassRule; @@ -69,6 +68,6 @@ public class TestProcedureManager { byte[] result = admin.execProcedureWithRet(SimpleMasterProcedureManager.SIMPLE_SIGNATURE, "mytest", new HashMap<>()); assertArrayEquals("Incorrect return data from execProcedure", - Bytes.toBytes(SimpleMasterProcedureManager.SIMPLE_DATA), result); + SimpleMasterProcedureManager.SIMPLE_DATA.getBytes(), result); } } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/TestZKProcedureControllers.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/TestZKProcedureControllers.java index 6b3de051594..143f30db487 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/TestZKProcedureControllers.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/procedure/TestZKProcedureControllers.java @@ -32,7 +32,6 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.errorhandling.ForeignExceptionDispatcher; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.MediumTests; -import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.hbase.zookeeper.ZKUtil; import org.apache.hadoop.hbase.zookeeper.ZKWatcher; @@ -68,7 +67,7 @@ public class TestZKProcedureControllers { private static final String CONTROLLER_NODE_NAME = "controller"; private static final VerificationMode once = Mockito.times(1); - private final byte[] memberData = Bytes.toBytes("data from member"); + private final byte[] memberData = new String("data from member").getBytes(); @BeforeClass public static void setupTest() throws Exception { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompoundBloomFilter.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompoundBloomFilter.java index 6da7abc5e6b..3d22b73fb51 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompoundBloomFilter.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompoundBloomFilter.java @@ -357,8 +357,8 @@ public class TestCompoundBloomFilter { @Test public void testCreateKey() { - byte[] row = Bytes.toBytes("myRow"); - byte[] qualifier = Bytes.toBytes("myQualifier"); + byte[] row = "myRow".getBytes(); + byte[] qualifier = "myQualifier".getBytes(); // Mimic what Storefile.createBloomKeyValue() does byte[] rowKey = KeyValueUtil.createFirstOnRow(row, 0, row.length, new byte[0], 0, 0, row, 0, 0).getKey(); byte[] rowColKey = KeyValueUtil.createFirstOnRow(row, 0, row.length, diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMultiLogThreshold.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMultiLogThreshold.java index 5ce32de3a71..6434ff1b5ff 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMultiLogThreshold.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMultiLogThreshold.java @@ -106,7 +106,7 @@ public class TestMultiLogThreshold { RegionAction.Builder rab = RegionAction.newBuilder(); rab.setRegion(RequestConverter.buildRegionSpecifier( HBaseProtos.RegionSpecifier.RegionSpecifierType.REGION_NAME, - Bytes.toBytes("someStuff" + i))); + new String("someStuff" + i).getBytes())); for (int j = 0; j < numAs; j++) { Action.Builder ab = Action.newBuilder(); rab.addAction(ab.build()); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasWithModifyTable.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasWithModifyTable.java index c2552ea9c9d..2a604b40981 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasWithModifyTable.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionReplicasWithModifyTable.java @@ -49,6 +49,7 @@ public class TestRegionReplicasWithModifyTable { private static final int NB_SERVERS = 3; private static Table table; + private static final byte[] row = "TestRegionReplicasWithModifyTable".getBytes(); private static final HBaseTestingUtility HTU = new HBaseTestingUtility(); private static final byte[] f = HConstants.CATALOG_FAMILY; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerReadRequestMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerReadRequestMetrics.java index 116f2de053a..5ad8e21fcab 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerReadRequestMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerReadRequestMetrics.java @@ -82,17 +82,17 @@ public class TestRegionServerReadRequestMetrics { LoggerFactory.getLogger(TestRegionServerReadRequestMetrics.class); private static final HBaseTestingUtility TEST_UTIL = new HBaseTestingUtility(); private static final TableName TABLE_NAME = TableName.valueOf("test"); - private static final byte[] CF1 = Bytes.toBytes("c1"); - private static final byte[] CF2 = Bytes.toBytes("c2"); + private static final byte[] CF1 = "c1".getBytes(); + private static final byte[] CF2 = "c2".getBytes(); - private static final byte[] ROW1 = Bytes.toBytes("a"); - private static final byte[] ROW2 = Bytes.toBytes("b"); - private static final byte[] ROW3 = Bytes.toBytes("c"); - private static final byte[] COL1 = Bytes.toBytes("q1"); - private static final byte[] COL2 = Bytes.toBytes("q2"); - private static final byte[] COL3 = Bytes.toBytes("q3"); - private static final byte[] VAL1 = Bytes.toBytes("v1"); - private static final byte[] VAL2 = Bytes.toBytes("v2"); + private static final byte[] ROW1 = "a".getBytes(); + private static final byte[] ROW2 = "b".getBytes(); + private static final byte[] ROW3 = "c".getBytes(); + private static final byte[] COL1 = "q1".getBytes(); + private static final byte[] COL2 = "q2".getBytes(); + private static final byte[] COL3 = "q3".getBytes(); + private static final byte[] VAL1 = "v1".getBytes(); + private static final byte[] VAL2 = "v2".getBytes(); private static final byte[] VAL3 = Bytes.toBytes(0L); private static final int MAX_TRY = 20; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java index b91bf11dae3..f24515d6314 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestSplitTransactionOnCluster.java @@ -455,12 +455,12 @@ public class TestSplitTransactionOnCluster { try { for (int i = 0; i <= 5; i++) { String row = "row" + i; - Put p = new Put(Bytes.toBytes(row)); + Put p = new Put(row.getBytes()); String val = "Val" + i; - p.addColumn(Bytes.toBytes("col"), Bytes.toBytes("ql"), Bytes.toBytes(val)); + p.addColumn("col".getBytes(), "ql".getBytes(), val.getBytes()); table.put(p); admin.flush(userTableName); - Delete d = new Delete(Bytes.toBytes(row)); + Delete d = new Delete(row.getBytes()); // Do a normal delete table.delete(d); admin.flush(userTableName); @@ -471,17 +471,17 @@ public class TestSplitTransactionOnCluster { .getRegionsOfTable(userTableName); assertEquals(1, regionsOfTable.size()); RegionInfo hRegionInfo = regionsOfTable.get(0); - Put p = new Put(Bytes.toBytes("row6")); - p.addColumn(Bytes.toBytes("col"), Bytes.toBytes("ql"), Bytes.toBytes("val")); + Put p = new Put("row6".getBytes()); + p.addColumn("col".getBytes(), "ql".getBytes(), "val".getBytes()); table.put(p); - p = new Put(Bytes.toBytes("row7")); - p.addColumn(Bytes.toBytes("col"), Bytes.toBytes("ql"), Bytes.toBytes("val")); + p = new Put("row7".getBytes()); + p.addColumn("col".getBytes(), "ql".getBytes(), "val".getBytes()); table.put(p); - p = new Put(Bytes.toBytes("row8")); - p.addColumn(Bytes.toBytes("col"), Bytes.toBytes("ql"), Bytes.toBytes("val")); + p = new Put("row8".getBytes()); + p.addColumn("col".getBytes(), "ql".getBytes(), "val".getBytes()); table.put(p); admin.flush(userTableName); - admin.splitRegionAsync(hRegionInfo.getRegionName(), Bytes.toBytes("row7")); + admin.splitRegionAsync(hRegionInfo.getRegionName(), "row7".getBytes()); regionsOfTable = cluster.getMaster() .getAssignmentManager().getRegionStates() .getRegionsOfTable(userTableName); @@ -630,7 +630,7 @@ public class TestSplitTransactionOnCluster { tableName); assertEquals("The specified table should be present.", true, tableExists); // exists works on stale and we see the put after the flush - byte[] b1 = Bytes.toBytes("row1"); + byte[] b1 = "row1".getBytes(); Get g = new Get(b1); g.setConsistency(Consistency.STRONG); // The following GET will make a trip to the meta to get the new location of the 1st daughter diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALMonotonicallyIncreasingSeqId.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALMonotonicallyIncreasingSeqId.java index 2333b277736..599260be9d8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALMonotonicallyIncreasingSeqId.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestWALMonotonicallyIncreasingSeqId.java @@ -143,7 +143,7 @@ public class TestWALMonotonicallyIncreasingSeqId { for (int i = 0; i < 100; i++) { byte[] row = Bytes.toBytes("putRow" + i); Put put = new Put(row); - put.addColumn(Bytes.toBytes("cf"), Bytes.toBytes(0), new byte[0]); + put.addColumn("cf".getBytes(), Bytes.toBytes(0), Bytes.toBytes("")); latch.await(); region.batchMutate(new Mutation[] { put }); Thread.sleep(10); @@ -168,7 +168,7 @@ public class TestWALMonotonicallyIncreasingSeqId { for (int i = 0; i < 100; i++) { byte[] row = Bytes.toBytes("incrementRow" + i); Increment inc = new Increment(row); - inc.addColumn(Bytes.toBytes("cf"), Bytes.toBytes(0), 1); + inc.addColumn("cf".getBytes(), Bytes.toBytes(0), 1); // inc.setDurability(Durability.ASYNC_WAL); region.increment(inc); latch.countDown(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestStoreHotnessProtector.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestStoreHotnessProtector.java index 67ac1c35216..6d4193474b2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestStoreHotnessProtector.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/throttle/TestStoreHotnessProtector.java @@ -39,7 +39,6 @@ import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.regionserver.Store; import org.apache.hadoop.hbase.testclassification.SmallTests; -import org.apache.hadoop.hbase.util.Bytes; import org.junit.Assert; import org.junit.ClassRule; import org.junit.Test; @@ -67,7 +66,7 @@ public class TestStoreHotnessProtector { Store mockStore1 = mock(Store.class); RegionInfo mockRegionInfo = mock(RegionInfo.class); - byte[] family = Bytes.toBytes("testF1"); + byte[] family = "testF1".getBytes(); when(mockRegion.getStore(family)).thenReturn(mockStore1); when(mockRegion.getRegionInfo()).thenReturn(mockRegionInfo); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestRegionReplicaReplicationEndpointNoMaster.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestRegionReplicaReplicationEndpointNoMaster.java index 1e4a49e815e..ab67d94bee9 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestRegionReplicaReplicationEndpointNoMaster.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/replication/regionserver/TestRegionReplicaReplicationEndpointNoMaster.java @@ -57,7 +57,6 @@ import org.apache.hadoop.hbase.replication.ReplicationEndpoint.ReplicateContext; import org.apache.hadoop.hbase.replication.regionserver.RegionReplicaReplicationEndpoint.RegionReplicaReplayCallable; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.ReplicationTests; -import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.ServerRegionReplicaUtil; import org.apache.hadoop.hbase.wal.WAL.Entry; import org.apache.hadoop.hbase.wal.WALEdit; @@ -91,7 +90,7 @@ public class TestRegionReplicaReplicationEndpointNoMaster { private static TableName tableName = TableName.valueOf( TestRegionReplicaReplicationEndpointNoMaster.class.getSimpleName()); private static Table table; - private static final byte[] row = Bytes.toBytes("TestRegionReplicaReplicator"); + private static final byte[] row = "TestRegionReplicaReplicator".getBytes(); private static HRegionServer rs0; private static HRegionServer rs1; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java index 67c43eeba21..110afcd1375 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/security/access/TestWithDisabledAuthorization.java @@ -876,7 +876,7 @@ public class TestWithDisabledAuthorization extends SecureTestUtil { public Object run() throws Exception { ACCESS_CONTROLLER.preCheckAndPut(ObserverContextImpl.createAndPrepare(RCP_ENV), TEST_ROW, TEST_FAMILY, TEST_Q1, CompareOperator.EQUAL, - new BinaryComparator(Bytes.toBytes("foo")), new Put(TEST_ROW), true); + new BinaryComparator("foo".getBytes()), new Put(TEST_ROW), true); return null; } }, SUPERUSER, USER_ADMIN, USER_RW, USER_RO, USER_OWNER, USER_CREATE, USER_QUAL, USER_NONE); @@ -887,7 +887,7 @@ public class TestWithDisabledAuthorization extends SecureTestUtil { public Object run() throws Exception { ACCESS_CONTROLLER.preCheckAndDelete(ObserverContextImpl.createAndPrepare(RCP_ENV), TEST_ROW, TEST_FAMILY, TEST_Q1, CompareOperator.EQUAL, - new BinaryComparator(Bytes.toBytes("foo")), new Delete(TEST_ROW), true); + new BinaryComparator("foo".getBytes()), new Delete(TEST_ROW), true); return null; } }, SUPERUSER, USER_ADMIN, USER_RW, USER_RO, USER_OWNER, USER_CREATE, USER_QUAL, USER_NONE); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/tool/TestLoadIncrementalHFiles.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/tool/TestLoadIncrementalHFiles.java index 7c04edc5b96..129823efb8a 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/tool/TestLoadIncrementalHFiles.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/tool/TestLoadIncrementalHFiles.java @@ -648,45 +648,45 @@ public class TestLoadIncrementalHFiles { first = "a"; last = "e"; - addStartEndKeysForTest(map, Bytes.toBytes(first), Bytes.toBytes(last)); + addStartEndKeysForTest(map, first.getBytes(), last.getBytes()); first = "r"; last = "s"; - addStartEndKeysForTest(map, Bytes.toBytes(first), Bytes.toBytes(last)); + addStartEndKeysForTest(map, first.getBytes(), last.getBytes()); first = "o"; last = "p"; - addStartEndKeysForTest(map, Bytes.toBytes(first), Bytes.toBytes(last)); + addStartEndKeysForTest(map, first.getBytes(), last.getBytes()); first = "g"; last = "k"; - addStartEndKeysForTest(map, Bytes.toBytes(first), Bytes.toBytes(last)); + addStartEndKeysForTest(map, first.getBytes(), last.getBytes()); first = "v"; last = "x"; - addStartEndKeysForTest(map, Bytes.toBytes(first), Bytes.toBytes(last)); + addStartEndKeysForTest(map, first.getBytes(), last.getBytes()); first = "c"; last = "i"; - addStartEndKeysForTest(map, Bytes.toBytes(first), Bytes.toBytes(last)); + addStartEndKeysForTest(map, first.getBytes(), last.getBytes()); first = "m"; last = "q"; - addStartEndKeysForTest(map, Bytes.toBytes(first), Bytes.toBytes(last)); + addStartEndKeysForTest(map, first.getBytes(), last.getBytes()); first = "s"; last = "t"; - addStartEndKeysForTest(map, Bytes.toBytes(first), Bytes.toBytes(last)); + addStartEndKeysForTest(map, first.getBytes(), last.getBytes()); first = "u"; last = "w"; - addStartEndKeysForTest(map, Bytes.toBytes(first), Bytes.toBytes(last)); + addStartEndKeysForTest(map, first.getBytes(), last.getBytes()); byte[][] keysArray = LoadIncrementalHFiles.inferBoundaries(map); byte[][] compare = new byte[3][]; - compare[0] = Bytes.toBytes("m"); - compare[1] = Bytes.toBytes("r"); - compare[2] = Bytes.toBytes("u"); + compare[0] = "m".getBytes(); + compare[1] = "r".getBytes(); + compare[2] = "u".getBytes(); assertEquals(3, keysArray.length); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBloomFilterChunk.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBloomFilterChunk.java index 57484d92399..a59d3475bd8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBloomFilterChunk.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestBloomFilterChunk.java @@ -58,7 +58,7 @@ public class TestBloomFilterChunk extends TestCase { (int) bf2.byteSize, bf2.hash, bf2.hashCount)); byte [] bkey = {1,2,3,4}; - byte [] bval = Bytes.toBytes("this is a much larger byte array"); + byte [] bval = "this is a much larger byte array".getBytes(); bf1.add(bkey, 0, bkey.length); bf1.add(bval, 1, bval.length-1); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionMover.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionMover.java index 0d768e4075b..670248d16ef 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionMover.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionMover.java @@ -85,7 +85,7 @@ public class TestRegionMover { .setColumnFamily(ColumnFamilyDescriptorBuilder.of("fam1")).build(); String startKey = "a"; String endKey = "z"; - admin.createTable(tableDesc, Bytes.toBytes(startKey), Bytes.toBytes(endKey), 9); + admin.createTable(tableDesc, startKey.getBytes(), endKey.getBytes(), 9); } @Test diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitter.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitter.java index 8fdcc5381cc..ace22388f5f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitter.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitter.java @@ -82,220 +82,218 @@ public class TestRegionSplitter { /** * Test creating a pre-split table using the HexStringSplit algorithm. */ - @Test - public void testCreatePresplitTableHex() throws Exception { - final List expectedBounds = new ArrayList<>(17); - expectedBounds.add(ArrayUtils.EMPTY_BYTE_ARRAY); - expectedBounds.add(Bytes.toBytes("10000000")); - expectedBounds.add(Bytes.toBytes("20000000")); - expectedBounds.add(Bytes.toBytes("30000000")); - expectedBounds.add(Bytes.toBytes("40000000")); - expectedBounds.add(Bytes.toBytes("50000000")); - expectedBounds.add(Bytes.toBytes("60000000")); - expectedBounds.add(Bytes.toBytes("70000000")); - expectedBounds.add(Bytes.toBytes("80000000")); - expectedBounds.add(Bytes.toBytes("90000000")); - expectedBounds.add(Bytes.toBytes("a0000000")); - expectedBounds.add(Bytes.toBytes("b0000000")); - expectedBounds.add(Bytes.toBytes("c0000000")); - expectedBounds.add(Bytes.toBytes("d0000000")); - expectedBounds.add(Bytes.toBytes("e0000000")); - expectedBounds.add(Bytes.toBytes("f0000000")); - expectedBounds.add(ArrayUtils.EMPTY_BYTE_ARRAY); + @Test + public void testCreatePresplitTableHex() throws Exception { + final List expectedBounds = new ArrayList<>(17); + expectedBounds.add(ArrayUtils.EMPTY_BYTE_ARRAY); + expectedBounds.add("10000000".getBytes()); + expectedBounds.add("20000000".getBytes()); + expectedBounds.add("30000000".getBytes()); + expectedBounds.add("40000000".getBytes()); + expectedBounds.add("50000000".getBytes()); + expectedBounds.add("60000000".getBytes()); + expectedBounds.add("70000000".getBytes()); + expectedBounds.add("80000000".getBytes()); + expectedBounds.add("90000000".getBytes()); + expectedBounds.add("a0000000".getBytes()); + expectedBounds.add("b0000000".getBytes()); + expectedBounds.add("c0000000".getBytes()); + expectedBounds.add("d0000000".getBytes()); + expectedBounds.add("e0000000".getBytes()); + expectedBounds.add("f0000000".getBytes()); + expectedBounds.add(ArrayUtils.EMPTY_BYTE_ARRAY); // Do table creation/pre-splitting and verification of region boundaries preSplitTableAndVerify(expectedBounds, HexStringSplit.class.getSimpleName(), TableName.valueOf(name.getMethodName())); - } + } - /** - * Test creating a pre-split table using the UniformSplit algorithm. - */ - @Test - public void testCreatePresplitTableUniform() throws Exception { - List expectedBounds = new ArrayList<>(17); - expectedBounds.add(ArrayUtils.EMPTY_BYTE_ARRAY); - expectedBounds.add(new byte[] { 0x10, 0, 0, 0, 0, 0, 0, 0}); - expectedBounds.add(new byte[] { 0x20, 0, 0, 0, 0, 0, 0, 0}); - expectedBounds.add(new byte[] { 0x30, 0, 0, 0, 0, 0, 0, 0}); - expectedBounds.add(new byte[] { 0x40, 0, 0, 0, 0, 0, 0, 0}); - expectedBounds.add(new byte[] { 0x50, 0, 0, 0, 0, 0, 0, 0 }); - expectedBounds.add(new byte[] { 0x60, 0, 0, 0, 0, 0, 0, 0 }); - expectedBounds.add(new byte[] { 0x70, 0, 0, 0, 0, 0, 0, 0 }); - expectedBounds.add(new byte[] { (byte) 0x80, 0, 0, 0, 0, 0, 0, 0 }); - expectedBounds.add(new byte[] { (byte) 0x90, 0, 0, 0, 0, 0, 0, 0 }); - expectedBounds.add(new byte[] {(byte)0xa0, 0, 0, 0, 0, 0, 0, 0}); - expectedBounds.add(new byte[] { (byte) 0xb0, 0, 0, 0, 0, 0, 0, 0 }); - expectedBounds.add(new byte[] { (byte) 0xc0, 0, 0, 0, 0, 0, 0, 0 }); - expectedBounds.add(new byte[] { (byte) 0xd0, 0, 0, 0, 0, 0, 0, 0 }); - expectedBounds.add(new byte[] {(byte)0xe0, 0, 0, 0, 0, 0, 0, 0}); - expectedBounds.add(new byte[] { (byte) 0xf0, 0, 0, 0, 0, 0, 0, 0 }); - expectedBounds.add(ArrayUtils.EMPTY_BYTE_ARRAY); + /** + * Test creating a pre-split table using the UniformSplit algorithm. + */ + @Test + public void testCreatePresplitTableUniform() throws Exception { + List expectedBounds = new ArrayList<>(17); + expectedBounds.add(ArrayUtils.EMPTY_BYTE_ARRAY); + expectedBounds.add(new byte[] { 0x10, 0, 0, 0, 0, 0, 0, 0}); + expectedBounds.add(new byte[] { 0x20, 0, 0, 0, 0, 0, 0, 0}); + expectedBounds.add(new byte[] { 0x30, 0, 0, 0, 0, 0, 0, 0}); + expectedBounds.add(new byte[] { 0x40, 0, 0, 0, 0, 0, 0, 0}); + expectedBounds.add(new byte[] { 0x50, 0, 0, 0, 0, 0, 0, 0 }); + expectedBounds.add(new byte[] { 0x60, 0, 0, 0, 0, 0, 0, 0 }); + expectedBounds.add(new byte[] { 0x70, 0, 0, 0, 0, 0, 0, 0 }); + expectedBounds.add(new byte[] { (byte) 0x80, 0, 0, 0, 0, 0, 0, 0 }); + expectedBounds.add(new byte[] { (byte) 0x90, 0, 0, 0, 0, 0, 0, 0 }); + expectedBounds.add(new byte[] {(byte)0xa0, 0, 0, 0, 0, 0, 0, 0}); + expectedBounds.add(new byte[] { (byte) 0xb0, 0, 0, 0, 0, 0, 0, 0 }); + expectedBounds.add(new byte[] { (byte) 0xc0, 0, 0, 0, 0, 0, 0, 0 }); + expectedBounds.add(new byte[] { (byte) 0xd0, 0, 0, 0, 0, 0, 0, 0 }); + expectedBounds.add(new byte[] {(byte)0xe0, 0, 0, 0, 0, 0, 0, 0}); + expectedBounds.add(new byte[] { (byte) 0xf0, 0, 0, 0, 0, 0, 0, 0 }); + expectedBounds.add(ArrayUtils.EMPTY_BYTE_ARRAY); - // Do table creation/pre-splitting and verification of region boundaries - preSplitTableAndVerify(expectedBounds, UniformSplit.class.getSimpleName(), - TableName.valueOf(name.getMethodName())); - } + // Do table creation/pre-splitting and verification of region boundaries + preSplitTableAndVerify(expectedBounds, UniformSplit.class.getSimpleName(), + TableName.valueOf(name.getMethodName())); + } - /** - * Unit tests for the HexStringSplit algorithm. Makes sure it divides up the - * space of keys in the way that we expect. - */ - @Test - public void unitTestHexStringSplit() { - HexStringSplit splitter = new HexStringSplit(); - // Check splitting while starting from scratch + /** + * Unit tests for the HexStringSplit algorithm. Makes sure it divides up the + * space of keys in the way that we expect. + */ + @Test + public void unitTestHexStringSplit() { + HexStringSplit splitter = new HexStringSplit(); + // Check splitting while starting from scratch - byte[][] twoRegionsSplits = splitter.split(2); - assertEquals(1, twoRegionsSplits.length); - assertArrayEquals(Bytes.toBytes("80000000"), twoRegionsSplits[0]); + byte[][] twoRegionsSplits = splitter.split(2); + assertEquals(1, twoRegionsSplits.length); + assertArrayEquals("80000000".getBytes(), twoRegionsSplits[0]); - byte[][] threeRegionsSplits = splitter.split(3); - assertEquals(2, threeRegionsSplits.length); - byte[] expectedSplit0 = Bytes.toBytes("55555555"); - assertArrayEquals(expectedSplit0, threeRegionsSplits[0]); - byte[] expectedSplit1 = Bytes.toBytes("aaaaaaaa"); - assertArrayEquals(expectedSplit1, threeRegionsSplits[1]); + byte[][] threeRegionsSplits = splitter.split(3); + assertEquals(2, threeRegionsSplits.length); + byte[] expectedSplit0 = "55555555".getBytes(); + assertArrayEquals(expectedSplit0, threeRegionsSplits[0]); + byte[] expectedSplit1 = "aaaaaaaa".getBytes(); + assertArrayEquals(expectedSplit1, threeRegionsSplits[1]); - // Check splitting existing regions that have start and end points - byte[] splitPoint = splitter.split(Bytes.toBytes("10000000"), Bytes.toBytes("30000000")); - assertArrayEquals(Bytes.toBytes("20000000"), splitPoint); + // Check splitting existing regions that have start and end points + byte[] splitPoint = splitter.split("10000000".getBytes(), "30000000".getBytes()); + assertArrayEquals("20000000".getBytes(), splitPoint); - byte[] lastRow = Bytes.toBytes("ffffffff"); - assertArrayEquals(lastRow, splitter.lastRow()); - byte[] firstRow = Bytes.toBytes("00000000"); - assertArrayEquals(firstRow, splitter.firstRow()); + byte[] lastRow = "ffffffff".getBytes(); + assertArrayEquals(lastRow, splitter.lastRow()); + byte[] firstRow = "00000000".getBytes(); + assertArrayEquals(firstRow, splitter.firstRow()); - // Halfway between 00... and 20... should be 10... - splitPoint = splitter.split(firstRow, Bytes.toBytes("20000000")); - assertArrayEquals(Bytes.toBytes("10000000"), splitPoint); + // Halfway between 00... and 20... should be 10... + splitPoint = splitter.split(firstRow, "20000000".getBytes()); + assertArrayEquals("10000000".getBytes(), splitPoint); - // Halfway between df... and ff... should be ef.... - splitPoint = splitter.split(Bytes.toBytes("dfffffff"), lastRow); - assertArrayEquals(Bytes.toBytes("efffffff"), splitPoint); + // Halfway between df... and ff... should be ef.... + splitPoint = splitter.split("dfffffff".getBytes(), lastRow); + assertArrayEquals("efffffff".getBytes(), splitPoint); - // Check splitting region with multiple mappers per region - byte[][] splits = splitter.split(Bytes.toBytes("00000000"), Bytes.toBytes("30000000"), - 3, false); - assertEquals(2, splits.length); - assertArrayEquals(Bytes.toBytes("10000000"), splits[0]); - assertArrayEquals(Bytes.toBytes("20000000"), splits[1]); + // Check splitting region with multiple mappers per region + byte[][] splits = splitter.split("00000000".getBytes(), "30000000".getBytes(), 3, false); + assertEquals(2, splits.length); + assertArrayEquals("10000000".getBytes(), splits[0]); + assertArrayEquals("20000000".getBytes(), splits[1]); - splits = splitter.split(Bytes.toBytes("00000000"), Bytes.toBytes("20000000"), 2, true); - assertEquals(3, splits.length); - assertArrayEquals(Bytes.toBytes("10000000"), splits[1]); - } + splits = splitter.split("00000000".getBytes(), "20000000".getBytes(), 2, true); + assertEquals(3, splits.length); + assertArrayEquals("10000000".getBytes(), splits[1]); + } - /** - * Unit tests for the DecimalStringSplit algorithm. Makes sure it divides up the - * space of keys in the way that we expect. - */ - @Test - public void unitTestDecimalStringSplit() { - DecimalStringSplit splitter = new DecimalStringSplit(); - // Check splitting while starting from scratch + /** + * Unit tests for the DecimalStringSplit algorithm. Makes sure it divides up the + * space of keys in the way that we expect. + */ + @Test + public void unitTestDecimalStringSplit() { + DecimalStringSplit splitter = new DecimalStringSplit(); + // Check splitting while starting from scratch - byte[][] twoRegionsSplits = splitter.split(2); - assertEquals(1, twoRegionsSplits.length); - assertArrayEquals(Bytes.toBytes("50000000"), twoRegionsSplits[0]); + byte[][] twoRegionsSplits = splitter.split(2); + assertEquals(1, twoRegionsSplits.length); + assertArrayEquals("50000000".getBytes(), twoRegionsSplits[0]); - byte[][] threeRegionsSplits = splitter.split(3); - assertEquals(2, threeRegionsSplits.length); - byte[] expectedSplit0 = Bytes.toBytes("33333333"); - assertArrayEquals(expectedSplit0, threeRegionsSplits[0]); - byte[] expectedSplit1 = Bytes.toBytes("66666666"); - assertArrayEquals(expectedSplit1, threeRegionsSplits[1]); + byte[][] threeRegionsSplits = splitter.split(3); + assertEquals(2, threeRegionsSplits.length); + byte[] expectedSplit0 = "33333333".getBytes(); + assertArrayEquals(expectedSplit0, threeRegionsSplits[0]); + byte[] expectedSplit1 = "66666666".getBytes(); + assertArrayEquals(expectedSplit1, threeRegionsSplits[1]); - // Check splitting existing regions that have start and end points - byte[] splitPoint = splitter.split(Bytes.toBytes("10000000"), Bytes.toBytes("30000000")); - assertArrayEquals(Bytes.toBytes("20000000"), splitPoint); + // Check splitting existing regions that have start and end points + byte[] splitPoint = splitter.split("10000000".getBytes(), "30000000".getBytes()); + assertArrayEquals("20000000".getBytes(), splitPoint); - byte[] lastRow = Bytes.toBytes("99999999"); - assertArrayEquals(lastRow, splitter.lastRow()); - byte[] firstRow = Bytes.toBytes("00000000"); - assertArrayEquals(firstRow, splitter.firstRow()); + byte[] lastRow = "99999999".getBytes(); + assertArrayEquals(lastRow, splitter.lastRow()); + byte[] firstRow = "00000000".getBytes(); + assertArrayEquals(firstRow, splitter.firstRow()); - // Halfway between 00... and 20... should be 10... - splitPoint = splitter.split(firstRow, Bytes.toBytes("20000000")); - assertArrayEquals(Bytes.toBytes("10000000"), splitPoint); + // Halfway between 00... and 20... should be 10... + splitPoint = splitter.split(firstRow, "20000000".getBytes()); + assertArrayEquals("10000000".getBytes(), splitPoint); - // Halfway between 00... and 19... should be 09... - splitPoint = splitter.split(firstRow, Bytes.toBytes("19999999")); - assertArrayEquals(Bytes.toBytes("09999999"), splitPoint); + // Halfway between 00... and 19... should be 09... + splitPoint = splitter.split(firstRow, "19999999".getBytes()); + assertArrayEquals("09999999".getBytes(), splitPoint); - // Halfway between 79... and 99... should be 89.... - splitPoint = splitter.split(Bytes.toBytes("79999999"), lastRow); - assertArrayEquals(Bytes.toBytes("89999999"), splitPoint); + // Halfway between 79... and 99... should be 89.... + splitPoint = splitter.split("79999999".getBytes(), lastRow); + assertArrayEquals("89999999".getBytes(), splitPoint); - // Check splitting region with multiple mappers per region - byte[][] splits = splitter.split(Bytes.toBytes("00000000"), Bytes.toBytes("30000000"), - 3, false); - assertEquals(2, splits.length); - assertArrayEquals(Bytes.toBytes("10000000"), splits[0]); - assertArrayEquals(Bytes.toBytes("20000000"), splits[1]); + // Check splitting region with multiple mappers per region + byte[][] splits = splitter.split("00000000".getBytes(), "30000000".getBytes(), 3, false); + assertEquals(2, splits.length); + assertArrayEquals("10000000".getBytes(), splits[0]); + assertArrayEquals("20000000".getBytes(), splits[1]); - splits = splitter.split(Bytes.toBytes("00000000"), Bytes.toBytes("20000000"), 2, true); - assertEquals(3, splits.length); - assertArrayEquals(Bytes.toBytes("10000000"), splits[1]); - } + splits = splitter.split("00000000".getBytes(), "20000000".getBytes(), 2, true); + assertEquals(3, splits.length); + assertArrayEquals("10000000".getBytes(), splits[1]); + } - /** - * Unit tests for the UniformSplit algorithm. Makes sure it divides up the space of - * keys in the way that we expect. - */ - @Test - public void unitTestUniformSplit() { - UniformSplit splitter = new UniformSplit(); + /** + * Unit tests for the UniformSplit algorithm. Makes sure it divides up the space of + * keys in the way that we expect. + */ + @Test + public void unitTestUniformSplit() { + UniformSplit splitter = new UniformSplit(); - // Check splitting while starting from scratch - try { - splitter.split(1); - throw new AssertionError("Splitting into <2 regions should have thrown exception"); - } catch (IllegalArgumentException e) { } + // Check splitting while starting from scratch + try { + splitter.split(1); + throw new AssertionError("Splitting into <2 regions should have thrown exception"); + } catch (IllegalArgumentException e) { } - byte[][] twoRegionsSplits = splitter.split(2); - assertEquals(1, twoRegionsSplits.length); - assertArrayEquals(twoRegionsSplits[0], new byte[] { (byte) 0x80, 0, 0, 0, 0, 0, 0, 0 }); + byte[][] twoRegionsSplits = splitter.split(2); + assertEquals(1, twoRegionsSplits.length); + assertArrayEquals(twoRegionsSplits[0], new byte[] { (byte) 0x80, 0, 0, 0, 0, 0, 0, 0 }); - byte[][] threeRegionsSplits = splitter.split(3); - assertEquals(2, threeRegionsSplits.length); - byte[] expectedSplit0 = new byte[] {0x55, 0x55, 0x55, 0x55, 0x55, 0x55, 0x55, 0x55}; - assertArrayEquals(expectedSplit0, threeRegionsSplits[0]); - byte[] expectedSplit1 = new byte[] {(byte)0xAA, (byte)0xAA, (byte)0xAA, (byte)0xAA, - (byte)0xAA, (byte)0xAA, (byte)0xAA, (byte)0xAA}; - assertArrayEquals(expectedSplit1, threeRegionsSplits[1]); + byte[][] threeRegionsSplits = splitter.split(3); + assertEquals(2, threeRegionsSplits.length); + byte[] expectedSplit0 = new byte[] {0x55, 0x55, 0x55, 0x55, 0x55, 0x55, 0x55, 0x55}; + assertArrayEquals(expectedSplit0, threeRegionsSplits[0]); + byte[] expectedSplit1 = new byte[] {(byte)0xAA, (byte)0xAA, (byte)0xAA, (byte)0xAA, + (byte)0xAA, (byte)0xAA, (byte)0xAA, (byte)0xAA}; + assertArrayEquals(expectedSplit1, threeRegionsSplits[1]); - // Check splitting existing regions that have start and end points - byte[] splitPoint = splitter.split(new byte[] {0x10}, new byte[] {0x30}); - assertArrayEquals(new byte[] { 0x20 }, splitPoint); + // Check splitting existing regions that have start and end points + byte[] splitPoint = splitter.split(new byte[] {0x10}, new byte[] {0x30}); + assertArrayEquals(new byte[] { 0x20 }, splitPoint); - byte[] lastRow = new byte[] {xFF, xFF, xFF, xFF, xFF, xFF, xFF, xFF}; - assertArrayEquals(lastRow, splitter.lastRow()); - byte[] firstRow = ArrayUtils.EMPTY_BYTE_ARRAY; - assertArrayEquals(firstRow, splitter.firstRow()); + byte[] lastRow = new byte[] {xFF, xFF, xFF, xFF, xFF, xFF, xFF, xFF}; + assertArrayEquals(lastRow, splitter.lastRow()); + byte[] firstRow = ArrayUtils.EMPTY_BYTE_ARRAY; + assertArrayEquals(firstRow, splitter.firstRow()); - splitPoint = splitter.split(firstRow, new byte[] {0x20}); - assertArrayEquals(splitPoint, new byte[] { 0x10 }); + splitPoint = splitter.split(firstRow, new byte[] {0x20}); + assertArrayEquals(splitPoint, new byte[] { 0x10 }); - splitPoint = splitter.split(new byte[] {(byte)0xdf, xFF, xFF, xFF, xFF, - xFF, xFF, xFF}, lastRow); - assertArrayEquals(splitPoint, new byte[] { (byte) 0xef, xFF, xFF, xFF, xFF, xFF, xFF, xFF}); + splitPoint = splitter.split(new byte[] {(byte)0xdf, xFF, xFF, xFF, xFF, + xFF, xFF, xFF}, lastRow); + assertArrayEquals(splitPoint, new byte[] { (byte) 0xef, xFF, xFF, xFF, xFF, xFF, xFF, xFF + }); - splitPoint = splitter.split(new byte[] {'a', 'a', 'a'}, new byte[] {'a', 'a', 'b'}); - assertArrayEquals(splitPoint, new byte[] { 'a', 'a', 'a', (byte) 0x80 }); + splitPoint = splitter.split(new byte[] {'a', 'a', 'a'}, new byte[] {'a', 'a', 'b'}); + assertArrayEquals(splitPoint, new byte[] { 'a', 'a', 'a', (byte) 0x80 }); - // Check splitting region with multiple mappers per region - byte[][] splits = splitter.split(new byte[] {'a', 'a', 'a'}, new byte[] {'a', 'a', 'd'}, - 3, false); - assertEquals(2, splits.length); - assertArrayEquals(splits[0], new byte[]{'a', 'a', 'b'}); - assertArrayEquals(splits[1], new byte[]{'a', 'a', 'c'}); + // Check splitting region with multiple mappers per region + byte[][] splits = splitter.split(new byte[] {'a', 'a', 'a'}, new byte[] {'a', 'a', 'd'}, 3, false); + assertEquals(2, splits.length); + assertArrayEquals(splits[0], new byte[]{'a', 'a', 'b'}); + assertArrayEquals(splits[1], new byte[]{'a', 'a', 'c'}); - splits = splitter.split(new byte[] {'a', 'a', 'a'}, new byte[] {'a', 'a', 'e'}, 2, true); - assertEquals(3, splits.length); - assertArrayEquals(splits[1], new byte[] { 'a', 'a', 'c'}); - } + splits = splitter.split(new byte[] {'a', 'a', 'a'}, new byte[] {'a', 'a', 'e'}, 2, true); + assertEquals(3, splits.length); + assertArrayEquals(splits[1], new byte[] { 'a', 'a', 'c'}); + } @Test public void testUserInput() { @@ -364,21 +362,21 @@ public class TestRegionSplitter { } } - /** - * Creates a pre-split table with expectedBounds.size()+1 regions, then - * verifies that the region boundaries are the same as the expected - * region boundaries in expectedBounds. - * @throws Various junit assertions - */ - private void preSplitTableAndVerify(List expectedBounds, - String splitClass, TableName tableName) throws Exception { - final int numRegions = expectedBounds.size()-1; - final Configuration conf = UTIL.getConfiguration(); - conf.setInt("split.count", numRegions); - SplitAlgorithm splitAlgo = RegionSplitter.newSplitAlgoInstance(conf, splitClass); - RegionSplitter.createPresplitTable(tableName, splitAlgo, new String[] { CF_NAME }, conf); - verifyBounds(expectedBounds, tableName); - } + /** + * Creates a pre-split table with expectedBounds.size()+1 regions, then + * verifies that the region boundaries are the same as the expected + * region boundaries in expectedBounds. + * @throws Various junit assertions + */ + private void preSplitTableAndVerify(List expectedBounds, + String splitClass, TableName tableName) throws Exception { + final int numRegions = expectedBounds.size()-1; + final Configuration conf = UTIL.getConfiguration(); + conf.setInt("split.count", numRegions); + SplitAlgorithm splitAlgo = RegionSplitter.newSplitAlgoInstance(conf, splitClass); + RegionSplitter.createPresplitTable(tableName, splitAlgo, new String[] { CF_NAME }, conf); + verifyBounds(expectedBounds, tableName); + } @Test public void noopRollingSplit() throws Exception { @@ -388,58 +386,60 @@ public class TestRegionSplitter { "UniformSplit", expectedBounds); } - private void rollingSplitAndVerify(TableName tableName, String splitClass, - List expectedBounds) throws Exception { - final Configuration conf = UTIL.getConfiguration(); + private void rollingSplitAndVerify(TableName tableName, String splitClass, + List expectedBounds) throws Exception { + final Configuration conf = UTIL.getConfiguration(); - // Set this larger than the number of splits so RegionSplitter won't block - conf.setInt("split.outstanding", 5); - SplitAlgorithm splitAlgo = RegionSplitter.newSplitAlgoInstance(conf, splitClass); - RegionSplitter.rollingSplit(tableName, splitAlgo, conf); - verifyBounds(expectedBounds, tableName); - } + // Set this larger than the number of splits so RegionSplitter won't block + conf.setInt("split.outstanding", 5); + SplitAlgorithm splitAlgo = RegionSplitter.newSplitAlgoInstance(conf, splitClass); + RegionSplitter.rollingSplit(tableName, splitAlgo, conf); + verifyBounds(expectedBounds, tableName); + } - private void verifyBounds(List expectedBounds, TableName tableName) - throws Exception { - // Get region boundaries from the cluster and verify their endpoints - final int numRegions = expectedBounds.size()-1; - try (Table table = UTIL.getConnection().getTable(tableName); - RegionLocator locator = UTIL.getConnection().getRegionLocator(tableName)) { - final List regionInfoMap = locator.getAllRegionLocations(); - assertEquals(numRegions, regionInfoMap.size()); - for (HRegionLocation entry : regionInfoMap) { - final HRegionInfo regionInfo = entry.getRegionInfo(); - byte[] regionStart = regionInfo.getStartKey(); - byte[] regionEnd = regionInfo.getEndKey(); + private void verifyBounds(List expectedBounds, TableName tableName) + throws Exception { + // Get region boundaries from the cluster and verify their endpoints + final int numRegions = expectedBounds.size()-1; + try (Table table = UTIL.getConnection().getTable(tableName); + RegionLocator locator = UTIL.getConnection().getRegionLocator(tableName)) { + final List regionInfoMap = locator.getAllRegionLocations(); + assertEquals(numRegions, regionInfoMap.size()); + for (HRegionLocation entry : regionInfoMap) { + final HRegionInfo regionInfo = entry.getRegionInfo(); + byte[] regionStart = regionInfo.getStartKey(); + byte[] regionEnd = regionInfo.getEndKey(); - // This region's start key should be one of the region boundaries - int startBoundaryIndex = indexOfBytes(expectedBounds, regionStart); - assertNotSame(-1, startBoundaryIndex); + // This region's start key should be one of the region boundaries + int startBoundaryIndex = indexOfBytes(expectedBounds, regionStart); + assertNotSame(-1, startBoundaryIndex); - // This region's end key should be the region boundary that comes - // after the starting boundary. - byte[] expectedRegionEnd = expectedBounds.get(startBoundaryIndex + 1); - assertEquals(0, Bytes.compareTo(regionEnd, expectedRegionEnd)); + // This region's end key should be the region boundary that comes + // after the starting boundary. + byte[] expectedRegionEnd = expectedBounds.get(startBoundaryIndex + 1); + assertEquals(0, Bytes.compareTo(regionEnd, expectedRegionEnd)); + } } } - } - /** - * List.indexOf() doesn't really work for a List<byte[]>, because byte[] - * doesn't override equals(). This method checks whether a list contains - * a given element by checking each element using the byte array comparator. - * @return the index of the first element that equals compareTo, or -1 if no elements are equal. - */ - static private int indexOfBytes(List list, byte[] compareTo) { - int listIndex = 0; - for(byte[] elem: list) { - if(Bytes.BYTES_COMPARATOR.compare(elem, compareTo) == 0) { - return listIndex; - } - listIndex++; + /** + * List.indexOf() doesn't really work for a List<byte[]>, because byte[] + * doesn't override equals(). This method checks whether a list contains + * a given element by checking each element using the byte array + * comparator. + * @return the index of the first element that equals compareTo, or -1 + * if no elements are equal. + */ + static private int indexOfBytes(List list, byte[] compareTo) { + int listIndex = 0; + for(byte[] elem: list) { + if(Bytes.BYTES_COMPARATOR.compare(elem, compareTo) == 0) { + return listIndex; + } + listIndex++; + } + return -1; } - return -1; - } } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java index c9cca855748..993cf7ffbd4 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZooKeeperACL.java @@ -22,10 +22,8 @@ import static org.junit.Assert.assertFalse; import static org.junit.Assert.assertTrue; import java.io.File; -import java.io.FileOutputStream; +import java.io.FileWriter; import java.io.IOException; -import java.io.OutputStreamWriter; -import java.nio.charset.StandardCharsets; import java.util.ArrayList; import java.util.List; import javax.security.auth.login.AppConfigurationEntry; @@ -67,19 +65,19 @@ public class TestZooKeeperACL { @BeforeClass public static void setUpBeforeClass() throws Exception { File saslConfFile = File.createTempFile("tmp", "jaas.conf"); - try (OutputStreamWriter fwriter = new OutputStreamWriter( - new FileOutputStream(saslConfFile), StandardCharsets.UTF_8)) { - fwriter.write( - "Server {\n" + - "org.apache.zookeeper.server.auth.DigestLoginModule required\n" + - "user_hbase=\"secret\";\n" + - "};\n" + - "Client {\n" + - "org.apache.zookeeper.server.auth.DigestLoginModule required\n" + - "username=\"hbase\"\n" + - "password=\"secret\";\n" + - "};" + "\n"); - } + FileWriter fwriter = new FileWriter(saslConfFile); + + fwriter.write("" + + "Server {\n" + + "org.apache.zookeeper.server.auth.DigestLoginModule required\n" + + "user_hbase=\"secret\";\n" + + "};\n" + + "Client {\n" + + "org.apache.zookeeper.server.auth.DigestLoginModule required\n" + + "username=\"hbase\"\n" + + "password=\"secret\";\n" + + "};" + "\n"); + fwriter.close(); System.setProperty("java.security.auth.login.config", saslConfFile.getAbsolutePath()); System.setProperty("zookeeper.authProvider.1", @@ -281,11 +279,10 @@ public class TestZooKeeperACL { assertEquals(testJaasConfig, secureZKAvailable); // Define Jaas configuration without ZooKeeper Jaas config File saslConfFile = File.createTempFile("tmp", "fakeJaas.conf"); - try (OutputStreamWriter fwriter = new OutputStreamWriter( - new FileOutputStream(saslConfFile), StandardCharsets.UTF_8)) { - fwriter.write(""); - } + FileWriter fwriter = new FileWriter(saslConfFile); + fwriter.write(""); + fwriter.close(); System.setProperty("java.security.auth.login.config", saslConfFile.getAbsolutePath()); diff --git a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java index b158bb7f348..730d0dd4cc2 100644 --- a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java +++ b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java @@ -34,7 +34,6 @@ import java.util.Random; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; -import org.apache.hadoop.hbase.util.Bytes; import org.apache.yetus.audience.InterfaceAudience; import org.apache.zookeeper.server.NIOServerCnxnFactory; import org.apache.zookeeper.server.ZooKeeperServer; @@ -55,7 +54,6 @@ public class MiniZooKeeperCluster { private static final int TICK_TIME = 2000; private static final int DEFAULT_CONNECTION_TIMEOUT = 30000; - private static final byte[] STATIC_BYTES = Bytes.toBytes("stat"); private int connectionTimeout; private boolean started; @@ -408,7 +406,7 @@ public class MiniZooKeeperCluster { Socket sock = new Socket("localhost", port); try { OutputStream outstream = sock.getOutputStream(); - outstream.write(STATIC_BYTES); + outstream.write("stat".getBytes()); outstream.flush(); } finally { sock.close(); @@ -438,7 +436,7 @@ public class MiniZooKeeperCluster { BufferedReader reader = null; try { OutputStream outstream = sock.getOutputStream(); - outstream.write(STATIC_BYTES); + outstream.write("stat".getBytes()); outstream.flush(); Reader isr = new InputStreamReader(sock.getInputStream()); diff --git a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestRecoverableZooKeeper.java b/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestRecoverableZooKeeper.java index 47e0d11c846..1e7db058463 100644 --- a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestRecoverableZooKeeper.java +++ b/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestRecoverableZooKeeper.java @@ -81,14 +81,14 @@ public class TestRecoverableZooKeeper { String ensemble = ZKConfig.getZKQuorumServersString(conf); RecoverableZooKeeper rzk = ZKUtil.connect(conf, ensemble, zkw); rzk.create(znode, new byte[0], Ids.OPEN_ACL_UNSAFE, CreateMode.PERSISTENT); - rzk.setData(znode, Bytes.toBytes("OPENING"), 0); + rzk.setData(znode, "OPENING".getBytes(), 0); Field zkField = RecoverableZooKeeper.class.getDeclaredField("zk"); zkField.setAccessible(true); int timeout = conf.getInt(HConstants.ZK_SESSION_TIMEOUT, HConstants.DEFAULT_ZK_SESSION_TIMEOUT); ZookeeperStub zkStub = new ZookeeperStub(ensemble, timeout, zkw); zkStub.setThrowExceptionInNumOperations(1); zkField.set(rzk, zkStub); - byte[] opened = Bytes.toBytes("OPENED"); + byte[] opened = "OPENED".getBytes(); rzk.setData(znode, opened, 1); byte[] data = rzk.getData(znode, false, new Stat()); assertTrue(Bytes.equals(opened, data)); diff --git a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKUtil.java b/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKUtil.java index f103a189d4f..1508441181c 100644 --- a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKUtil.java +++ b/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/zookeeper/TestZKUtil.java @@ -157,7 +157,7 @@ public class TestZKUtil { String quorumServers = ZKConfig.getZKQuorumServersString(c); int sessionTimeout = 5 * 1000; // 5 seconds ZooKeeper zk = new ZooKeeper(quorumServers, sessionTimeout, EmptyWatcher.instance); - zk.addAuthInfo("digest", Bytes.toBytes("hbase:rox")); + zk.addAuthInfo("digest", "hbase:rox".getBytes()); // Save the previous ACL Stat s = null; @@ -223,7 +223,7 @@ public class TestZKUtil { // Restore the ACL ZooKeeper zk3 = new ZooKeeper(quorumServers, sessionTimeout, EmptyWatcher.instance); - zk3.addAuthInfo("digest", Bytes.toBytes("hbase:rox")); + zk3.addAuthInfo("digest", "hbase:rox".getBytes()); try { zk3.setACL("/", oldACL, -1); } finally {