From dace419238e30cef8104adce17056f0392c94b50 Mon Sep 17 00:00:00 2001 From: Michael Stack Date: Thu, 1 Mar 2012 17:53:03 +0000 Subject: [PATCH] HBASE-4403 Adopt interface stability/audience classifications from Hadoop AND HBASE-5502 region_mover.rb fails to load regions back to original server for regions only containing empty tables git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1295710 13f79535-47bb-0310-9956-ffa450edef68 --- bin/region_mover.rb | 3 ++- src/main/java/org/apache/hadoop/hbase/Abortable.java | 3 +++ src/main/java/org/apache/hadoop/hbase/Chore.java | 2 ++ .../org/apache/hadoop/hbase/ClockOutOfSyncException.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/ClusterStatus.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/Coprocessor.java | 5 +++++ .../org/apache/hadoop/hbase/CoprocessorEnvironment.java | 4 ++++ .../java/org/apache/hadoop/hbase/DoNotRetryIOException.java | 5 +++++ .../org/apache/hadoop/hbase/DroppedSnapshotException.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/EmptyWatcher.java | 2 ++ .../java/org/apache/hadoop/hbase/HBaseConfiguration.java | 4 ++++ .../java/org/apache/hadoop/hbase/HColumnDescriptor.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/HConstants.java | 4 ++++ .../org/apache/hadoop/hbase/HDFSBlocksDistribution.java | 3 +++ src/main/java/org/apache/hadoop/hbase/HRegionInfo.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/HRegionLocation.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/HServerLoad.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java | 4 ++++ .../hadoop/hbase/InvalidFamilyOperationException.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/KeyValue.java | 4 ++++ .../java/org/apache/hadoop/hbase/LocalHBaseCluster.java | 4 ++++ .../java/org/apache/hadoop/hbase/MasterAddressTracker.java | 2 ++ .../org/apache/hadoop/hbase/MasterNotRunningException.java | 5 +++++ .../hadoop/hbase/NotAllMetaRegionsOnlineException.java | 4 ++++ .../org/apache/hadoop/hbase/NotServingRegionException.java | 4 ++++ .../java/org/apache/hadoop/hbase/PleaseHoldException.java | 6 +++++- src/main/java/org/apache/hadoop/hbase/RegionException.java | 5 +++++ .../org/apache/hadoop/hbase/RemoteExceptionHandler.java | 2 ++ src/main/java/org/apache/hadoop/hbase/Server.java | 2 ++ src/main/java/org/apache/hadoop/hbase/ServerName.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/Stoppable.java | 3 +++ src/main/java/org/apache/hadoop/hbase/TableDescriptors.java | 5 +++++ .../java/org/apache/hadoop/hbase/TableExistsException.java | 5 +++++ .../org/apache/hadoop/hbase/TableNotDisabledException.java | 4 ++++ .../org/apache/hadoop/hbase/TableNotEnabledException.java | 4 ++++ .../org/apache/hadoop/hbase/TableNotFoundException.java | 5 +++++ .../org/apache/hadoop/hbase/UnknownRegionException.java | 5 +++++ .../org/apache/hadoop/hbase/UnknownRowLockException.java | 5 +++++ .../org/apache/hadoop/hbase/UnknownScannerException.java | 5 +++++ .../java/org/apache/hadoop/hbase/VersionAnnotation.java | 3 +++ .../java/org/apache/hadoop/hbase/YouAreDeadException.java | 5 +++++ .../apache/hadoop/hbase/ZooKeeperConnectionException.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/avro/AvroServer.java | 2 ++ src/main/java/org/apache/hadoop/hbase/avro/AvroUtil.java | 2 ++ .../org/apache/hadoop/hbase/catalog/CatalogTracker.java | 2 ++ .../java/org/apache/hadoop/hbase/catalog/MetaEditor.java | 2 ++ .../hadoop/hbase/catalog/MetaMigrationRemovingHTD.java | 1 + .../java/org/apache/hadoop/hbase/catalog/MetaReader.java | 2 ++ .../org/apache/hadoop/hbase/catalog/RootLocationEditor.java | 2 ++ .../apache/hadoop/hbase/client/AbstractClientScanner.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/client/Action.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/Append.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/Attributes.java | 5 +++++ .../java/org/apache/hadoop/hbase/client/ClientScanner.java | 4 ++++ .../org/apache/hadoop/hbase/client/ConnectionUtils.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/Delete.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/Get.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/HBaseAdmin.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/HConnection.java | 4 ++++ .../org/apache/hadoop/hbase/client/HConnectionManager.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/HTable.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/HTableFactory.java | 4 ++++ .../org/apache/hadoop/hbase/client/HTableInterface.java | 4 ++++ .../apache/hadoop/hbase/client/HTableInterfaceFactory.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/HTablePool.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/HTableUtil.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/client/Increment.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/IsolationLevel.java | 5 +++++ .../java/org/apache/hadoop/hbase/client/MetaScanner.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/MultiAction.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/MultiPut.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/MultiResponse.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/Mutation.java | 4 ++++ .../hadoop/hbase/client/NoServerForRegionException.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/Operation.java | 4 ++++ .../apache/hadoop/hbase/client/OperationWithAttributes.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/Put.java | 4 ++++ .../apache/hadoop/hbase/client/RegionOfflineException.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/Result.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/ResultScanner.java | 5 +++++ .../hadoop/hbase/client/RetriesExhaustedException.java | 5 +++++ .../hbase/client/RetriesExhaustedWithDetailsException.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/Row.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/RowLock.java | 5 +++++ .../java/org/apache/hadoop/hbase/client/RowMutations.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/client/Scan.java | 4 ++++ .../org/apache/hadoop/hbase/client/ScannerCallable.java | 4 ++++ .../apache/hadoop/hbase/client/ScannerTimeoutException.java | 4 ++++ .../java/org/apache/hadoop/hbase/client/ServerCallable.java | 4 ++++ .../hadoop/hbase/client/UnmodifyableHColumnDescriptor.java | 4 ++++ .../apache/hadoop/hbase/client/UnmodifyableHRegionInfo.java | 4 ++++ .../hadoop/hbase/client/UnmodifyableHTableDescriptor.java | 4 ++++ .../hadoop/hbase/client/coprocessor/AggregationClient.java | 4 ++++ .../org/apache/hadoop/hbase/client/coprocessor/Batch.java | 4 ++++ .../org/apache/hadoop/hbase/client/coprocessor/Exec.java | 4 ++++ .../apache/hadoop/hbase/client/coprocessor/ExecResult.java | 4 ++++ .../hbase/client/coprocessor/LongColumnInterpreter.java | 4 ++++ .../org/apache/hadoop/hbase/client/metrics/ScanMetrics.java | 4 ++++ .../org/apache/hadoop/hbase/constraint/BaseConstraint.java | 2 ++ .../java/org/apache/hadoop/hbase/constraint/Constraint.java | 2 ++ .../apache/hadoop/hbase/constraint/ConstraintException.java | 2 ++ .../apache/hadoop/hbase/constraint/ConstraintProcessor.java | 2 ++ .../org/apache/hadoop/hbase/constraint/Constraints.java | 2 ++ .../hadoop/hbase/coprocessor/AggregateImplementation.java | 4 ++++ .../apache/hadoop/hbase/coprocessor/AggregateProtocol.java | 4 ++++ .../hadoop/hbase/coprocessor/BaseEndpointCoprocessor.java | 4 ++++ .../apache/hadoop/hbase/coprocessor/BaseMasterObserver.java | 4 ++++ .../apache/hadoop/hbase/coprocessor/BaseRegionObserver.java | 5 +++++ .../apache/hadoop/hbase/coprocessor/ColumnInterpreter.java | 4 ++++ .../hadoop/hbase/coprocessor/CoprocessorException.java | 4 ++++ .../apache/hadoop/hbase/coprocessor/CoprocessorHost.java | 4 ++++ .../hbase/coprocessor/MasterCoprocessorEnvironment.java | 4 ++++ .../org/apache/hadoop/hbase/coprocessor/MasterObserver.java | 4 ++++ .../hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java | 4 ++++ .../hadoop/hbase/coprocessor/MultiRowMutationProtocol.java | 4 ++++ .../apache/hadoop/hbase/coprocessor/ObserverContext.java | 4 ++++ .../hbase/coprocessor/RegionCoprocessorEnvironment.java | 4 ++++ .../org/apache/hadoop/hbase/coprocessor/RegionObserver.java | 4 ++++ .../hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java | 4 ++++ .../org/apache/hadoop/hbase/coprocessor/WALObserver.java | 4 ++++ .../java/org/apache/hadoop/hbase/executor/EventHandler.java | 2 ++ .../org/apache/hadoop/hbase/executor/ExecutorService.java | 2 ++ .../apache/hadoop/hbase/executor/RegionTransitionData.java | 2 ++ .../org/apache/hadoop/hbase/filter/BinaryComparator.java | 4 ++++ .../apache/hadoop/hbase/filter/BinaryPrefixComparator.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/BitComparator.java | 5 +++++ .../apache/hadoop/hbase/filter/ColumnCountGetFilter.java | 4 ++++ .../apache/hadoop/hbase/filter/ColumnPaginationFilter.java | 4 ++++ .../org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java | 4 ++++ .../org/apache/hadoop/hbase/filter/ColumnRangeFilter.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/CompareFilter.java | 4 ++++ .../apache/hadoop/hbase/filter/DependentColumnFilter.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/FamilyFilter.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/filter/Filter.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/FilterBase.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/FilterList.java | 4 ++++ .../org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java | 4 ++++ .../org/apache/hadoop/hbase/filter/InclusiveStopFilter.java | 4 ++++ .../hadoop/hbase/filter/IncompatibleFilterException.java | 5 +++++ .../hadoop/hbase/filter/InvalidRowFilterException.java | 5 +++++ .../java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java | 5 +++++ .../hadoop/hbase/filter/MultipleColumnPrefixFilter.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/NullComparator.java | 5 +++++ .../java/org/apache/hadoop/hbase/filter/PageFilter.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/ParseConstants.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/ParseFilter.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/PrefixFilter.java | 4 ++++ .../org/apache/hadoop/hbase/filter/QualifierFilter.java | 4 ++++ .../org/apache/hadoop/hbase/filter/RandomRowFilter.java | 4 ++++ .../apache/hadoop/hbase/filter/RegexStringComparator.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java | 4 ++++ .../hadoop/hbase/filter/SingleColumnValueExcludeFilter.java | 4 ++++ .../apache/hadoop/hbase/filter/SingleColumnValueFilter.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/SkipFilter.java | 4 ++++ .../org/apache/hadoop/hbase/filter/SubstringComparator.java | 4 ++++ .../org/apache/hadoop/hbase/filter/TimestampsFilter.java | 4 ++++ .../java/org/apache/hadoop/hbase/filter/ValueFilter.java | 4 ++++ .../org/apache/hadoop/hbase/filter/WhileMatchFilter.java | 4 ++++ .../hadoop/hbase/filter/WritableByteArrayComparable.java | 4 ++++ .../java/org/apache/hadoop/hbase/io/CodeToClassAndBack.java | 3 +++ .../org/apache/hadoop/hbase/io/DataOutputOutputStream.java | 2 -- .../java/org/apache/hadoop/hbase/io/DoubleOutputStream.java | 3 +++ .../org/apache/hadoop/hbase/io/HalfStoreFileReader.java | 2 ++ .../java/org/apache/hadoop/hbase/io/HbaseMapWritable.java | 2 ++ .../org/apache/hadoop/hbase/io/HbaseObjectWritable.java | 2 ++ src/main/java/org/apache/hadoop/hbase/io/HeapSize.java | 3 +++ .../org/apache/hadoop/hbase/io/ImmutableBytesWritable.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/io/Reference.java | 2 ++ src/main/java/org/apache/hadoop/hbase/io/TimeRange.java | 4 ++++ .../java/org/apache/hadoop/hbase/io/WritableWithSize.java | 3 +++ .../hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java | 2 ++ .../apache/hadoop/hbase/io/encoding/CompressionState.java | 2 ++ .../hadoop/hbase/io/encoding/CopyKeyDataBlockEncoder.java | 2 ++ .../apache/hadoop/hbase/io/encoding/DataBlockEncoder.java | 2 ++ .../apache/hadoop/hbase/io/encoding/DataBlockEncoding.java | 2 ++ .../hadoop/hbase/io/encoding/DiffKeyDeltaEncoder.java | 2 ++ .../apache/hadoop/hbase/io/encoding/EncodedDataBlock.java | 2 ++ .../hbase/io/encoding/EncoderBufferTooSmallException.java | 3 +++ .../hadoop/hbase/io/encoding/FastDiffDeltaEncoder.java | 2 ++ .../hadoop/hbase/io/encoding/PrefixKeyDeltaEncoder.java | 2 ++ .../apache/hadoop/hbase/io/hfile/AbstractHFileReader.java | 2 ++ .../apache/hadoop/hbase/io/hfile/AbstractHFileWriter.java | 2 ++ .../java/org/apache/hadoop/hbase/io/hfile/BlockCache.java | 2 ++ .../hbase/io/hfile/BlockCacheColumnFamilySummary.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java | 2 ++ .../java/org/apache/hadoop/hbase/io/hfile/BlockType.java | 2 ++ .../hadoop/hbase/io/hfile/BoundedRangeFileInputStream.java | 2 ++ .../java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java | 2 ++ .../java/org/apache/hadoop/hbase/io/hfile/CacheStats.java | 3 +++ .../java/org/apache/hadoop/hbase/io/hfile/Cacheable.java | 3 +++ .../apache/hadoop/hbase/io/hfile/CacheableDeserializer.java | 3 +++ .../java/org/apache/hadoop/hbase/io/hfile/CachedBlock.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/CachedBlockQueue.java | 2 ++ .../java/org/apache/hadoop/hbase/io/hfile/Compression.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/DoubleBlockCache.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java | 2 ++ src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java | 2 ++ .../java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java | 2 ++ .../apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java | 2 ++ .../hadoop/hbase/io/hfile/HFileDataBlockEncoderImpl.java | 2 ++ .../apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.java | 4 ++++ .../org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java | 2 ++ .../java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/HFileWriterV1.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java | 3 +++ .../apache/hadoop/hbase/io/hfile/InvalidHFileException.java | 3 +++ .../org/apache/hadoop/hbase/io/hfile/LruBlockCache.java | 2 ++ .../apache/hadoop/hbase/io/hfile/NoOpDataBlockEncoder.java | 2 ++ .../hadoop/hbase/io/hfile/ReusableStreamGzipCodec.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/SimpleBlockCache.java | 2 ++ .../apache/hadoop/hbase/io/hfile/slab/SingleSizeCache.java | 2 ++ .../java/org/apache/hadoop/hbase/io/hfile/slab/Slab.java | 2 ++ .../org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java | 2 ++ .../hadoop/hbase/io/hfile/slab/SlabItemActionWatcher.java | 2 ++ .../java/org/apache/hadoop/hbase/ipc/ConnectionHeader.java | 2 ++ .../org/apache/hadoop/hbase/ipc/CoprocessorProtocol.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java | 3 +++ .../java/org/apache/hadoop/hbase/ipc/ExecRPCInvoker.java | 2 ++ src/main/java/org/apache/hadoop/hbase/ipc/HBaseClient.java | 2 ++ src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPC.java | 2 ++ .../org/apache/hadoop/hbase/ipc/HBaseRPCErrorHandler.java | 3 +++ .../org/apache/hadoop/hbase/ipc/HBaseRPCStatistics.java | 2 ++ .../java/org/apache/hadoop/hbase/ipc/HBaseRpcMetrics.java | 2 ++ src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java | 2 ++ .../java/org/apache/hadoop/hbase/ipc/HMasterInterface.java | 4 ++++ .../org/apache/hadoop/hbase/ipc/HMasterRegionInterface.java | 2 ++ .../java/org/apache/hadoop/hbase/ipc/HRegionInterface.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/ipc/Invocation.java | 2 ++ .../java/org/apache/hadoop/hbase/ipc/ProtocolSignature.java | 2 ++ .../java/org/apache/hadoop/hbase/ipc/RequestContext.java | 2 ++ src/main/java/org/apache/hadoop/hbase/ipc/ResponseFlag.java | 3 +++ src/main/java/org/apache/hadoop/hbase/ipc/RpcEngine.java | 2 ++ src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java | 2 ++ .../hadoop/hbase/ipc/ServerNotRunningYetException.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/ipc/Status.java | 3 +++ .../java/org/apache/hadoop/hbase/ipc/VersionedProtocol.java | 5 +++++ .../java/org/apache/hadoop/hbase/ipc/WritableRpcEngine.java | 2 ++ .../java/org/apache/hadoop/hbase/mapreduce/CellCounter.java | 4 ++++ .../java/org/apache/hadoop/hbase/mapreduce/CopyTable.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/mapreduce/Driver.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/mapreduce/Export.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/GroupingTableMapper.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/HFileOutputFormat.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/HRegionPartitioner.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/IdentityTableMapper.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/IdentityTableReducer.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java | 4 ++++ .../java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/KeyValueSortReducer.java | 4 ++++ .../hadoop/hbase/mapreduce/LoadIncrementalHFiles.java | 4 ++++ .../hadoop/hbase/mapreduce/MultiTableOutputFormat.java | 4 ++++ .../org/apache/hadoop/hbase/mapreduce/PutSortReducer.java | 4 ++++ .../java/org/apache/hadoop/hbase/mapreduce/RowCounter.java | 4 ++++ .../hadoop/hbase/mapreduce/SimpleTotalOrderPartitioner.java | 4 ++++ .../org/apache/hadoop/hbase/mapreduce/TableInputFormat.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/TableInputFormatBase.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java | 4 ++++ .../java/org/apache/hadoop/hbase/mapreduce/TableMapper.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/TableOutputCommitter.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/TableOutputFormat.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/TableRecordReader.java | 4 ++++ .../hadoop/hbase/mapreduce/TableRecordReaderImpl.java | 4 ++++ .../org/apache/hadoop/hbase/mapreduce/TableReducer.java | 4 ++++ .../java/org/apache/hadoop/hbase/mapreduce/TableSplit.java | 4 ++++ .../apache/hadoop/hbase/mapreduce/TsvImporterMapper.java | 4 ++++ .../org/apache/hadoop/hbase/master/ActiveMasterManager.java | 2 ++ .../java/org/apache/hadoop/hbase/master/AssignCallable.java | 2 ++ .../org/apache/hadoop/hbase/master/AssignmentManager.java | 2 ++ .../java/org/apache/hadoop/hbase/master/BulkAssigner.java | 2 ++ .../java/org/apache/hadoop/hbase/master/BulkReOpen.java | 2 ++ .../java/org/apache/hadoop/hbase/master/CatalogJanitor.java | 2 ++ .../java/org/apache/hadoop/hbase/master/DeadServer.java | 2 ++ .../org/apache/hadoop/hbase/master/DefaultLoadBalancer.java | 2 ++ src/main/java/org/apache/hadoop/hbase/master/HMaster.java | 2 ++ .../org/apache/hadoop/hbase/master/HMasterCommandLine.java | 2 ++ .../java/org/apache/hadoop/hbase/master/LoadBalancer.java | 2 ++ .../org/apache/hadoop/hbase/master/LoadBalancerFactory.java | 2 ++ .../java/org/apache/hadoop/hbase/master/LogCleaner.java | 2 ++ .../org/apache/hadoop/hbase/master/LogCleanerDelegate.java | 2 ++ .../apache/hadoop/hbase/master/MasterCoprocessorHost.java | 2 ++ .../org/apache/hadoop/hbase/master/MasterDumpServlet.java | 2 ++ .../org/apache/hadoop/hbase/master/MasterFileSystem.java | 2 ++ .../java/org/apache/hadoop/hbase/master/MasterServices.java | 2 ++ .../org/apache/hadoop/hbase/master/MasterStatusServlet.java | 2 ++ .../java/org/apache/hadoop/hbase/master/RegionPlan.java | 2 ++ .../java/org/apache/hadoop/hbase/master/ServerAndLoad.java | 2 ++ .../java/org/apache/hadoop/hbase/master/ServerManager.java | 2 ++ .../org/apache/hadoop/hbase/master/SplitLogManager.java | 2 ++ .../apache/hadoop/hbase/master/TimeToLiveLogCleaner.java | 2 ++ .../org/apache/hadoop/hbase/master/UnAssignCallable.java | 2 ++ .../hadoop/hbase/master/handler/ClosedRegionHandler.java | 2 ++ .../hadoop/hbase/master/handler/CreateTableHandler.java | 2 ++ .../hadoop/hbase/master/handler/DeleteTableHandler.java | 2 ++ .../hadoop/hbase/master/handler/DisableTableHandler.java | 2 ++ .../hadoop/hbase/master/handler/EnableTableHandler.java | 2 ++ .../hbase/master/handler/MetaServerShutdownHandler.java | 2 ++ .../hadoop/hbase/master/handler/ModifyTableHandler.java | 2 ++ .../hadoop/hbase/master/handler/OpenedRegionHandler.java | 2 ++ .../hadoop/hbase/master/handler/ServerShutdownHandler.java | 2 ++ .../hadoop/hbase/master/handler/SplitRegionHandler.java | 2 ++ .../hadoop/hbase/master/handler/TableAddFamilyHandler.java | 2 ++ .../hbase/master/handler/TableDeleteFamilyHandler.java | 2 ++ .../hadoop/hbase/master/handler/TableEventHandler.java | 2 ++ .../hbase/master/handler/TableModifyFamilyHandler.java | 2 ++ .../hadoop/hbase/master/handler/TotesHRegionInfo.java | 2 ++ .../apache/hadoop/hbase/master/metrics/MasterMetrics.java | 2 ++ .../hadoop/hbase/master/metrics/MasterStatistics.java | 2 ++ .../java/org/apache/hadoop/hbase/metrics/HBaseInfo.java | 2 ++ .../org/apache/hadoop/hbase/metrics/MetricsMBeanBase.java | 2 ++ .../java/org/apache/hadoop/hbase/metrics/MetricsRate.java | 2 ++ .../java/org/apache/hadoop/hbase/metrics/MetricsString.java | 2 ++ .../hbase/metrics/PersistentMetricsTimeVaryingRate.java | 2 ++ .../hadoop/hbase/metrics/file/TimeStampingFileContext.java | 2 ++ .../org/apache/hadoop/hbase/migration/HRegionInfo090x.java | 2 ++ .../org/apache/hadoop/hbase/monitoring/LogMonitoring.java | 2 ++ .../hbase/monitoring/MemoryBoundedLogMessageBuffer.java | 3 +++ .../apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java | 2 ++ .../hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java | 2 ++ .../org/apache/hadoop/hbase/monitoring/MonitoredTask.java | 3 +++ .../apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java | 2 ++ .../apache/hadoop/hbase/monitoring/StateDumpServlet.java | 2 ++ .../org/apache/hadoop/hbase/monitoring/TaskMonitor.java | 2 ++ .../apache/hadoop/hbase/monitoring/ThreadMonitoring.java | 3 +++ .../hadoop/hbase/regionserver/ChangedReadersObserver.java | 3 +++ .../org/apache/hadoop/hbase/regionserver/ColumnCount.java | 3 +++ .../org/apache/hadoop/hbase/regionserver/ColumnTracker.java | 2 ++ .../hadoop/hbase/regionserver/CompactSplitThread.java | 2 ++ .../hadoop/hbase/regionserver/CompactionRequestor.java | 3 +++ .../hbase/regionserver/ConstantSizeRegionSplitPolicy.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/DebugPrint.java | 3 +++ .../org/apache/hadoop/hbase/regionserver/DeleteTracker.java | 3 +++ .../hadoop/hbase/regionserver/ExplicitColumnTracker.java | 2 ++ .../apache/hadoop/hbase/regionserver/FlushRequester.java | 3 +++ .../hbase/regionserver/GetClosestRowBeforeTracker.java | 2 ++ .../java/org/apache/hadoop/hbase/regionserver/HRegion.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/HRegionServer.java | 2 ++ .../hadoop/hbase/regionserver/HRegionServerCommandLine.java | 2 ++ .../hadoop/hbase/regionserver/HRegionThriftServer.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/InternalScan.java | 2 ++ .../apache/hadoop/hbase/regionserver/InternalScanner.java | 2 ++ .../hbase/regionserver/KeyPrefixRegionSplitPolicy.java | 4 +++- .../org/apache/hadoop/hbase/regionserver/KeyValueHeap.java | 2 ++ .../apache/hadoop/hbase/regionserver/KeyValueScanner.java | 2 ++ .../hadoop/hbase/regionserver/KeyValueSkipListSet.java | 2 ++ .../apache/hadoop/hbase/regionserver/LeaseException.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/LeaseListener.java | 3 +++ .../java/org/apache/hadoop/hbase/regionserver/Leases.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/LogRoller.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/LruHashMap.java | 2 ++ .../java/org/apache/hadoop/hbase/regionserver/MemStore.java | 2 ++ .../apache/hadoop/hbase/regionserver/MemStoreFlusher.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/MemStoreLAB.java | 2 ++ .../hbase/regionserver/MultiVersionConsistencyControl.java | 2 ++ .../hbase/regionserver/NoSuchColumnFamilyException.java | 2 ++ .../hadoop/hbase/regionserver/NonLazyKeyValueScanner.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/OnlineRegions.java | 2 ++ .../apache/hadoop/hbase/regionserver/OperationStatus.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/RSDumpServlet.java | 2 ++ .../apache/hadoop/hbase/regionserver/RSStatusServlet.java | 2 ++ .../regionserver/RegionAlreadyInTransitionException.java | 6 ++++++ .../hadoop/hbase/regionserver/RegionOpeningState.java | 3 +++ .../org/apache/hadoop/hbase/regionserver/RegionScanner.java | 2 ++ .../hadoop/hbase/regionserver/RegionServerAccounting.java | 3 +++ .../hbase/regionserver/RegionServerRunningException.java | 3 +++ .../hadoop/hbase/regionserver/RegionServerServices.java | 2 ++ .../hbase/regionserver/RegionServerStoppedException.java | 3 +++ .../apache/hadoop/hbase/regionserver/RegionSplitPolicy.java | 2 ++ .../hadoop/hbase/regionserver/ReplicationService.java | 2 ++ .../hadoop/hbase/regionserver/ReplicationSinkService.java | 2 ++ .../hadoop/hbase/regionserver/ReplicationSourceService.java | 2 ++ .../apache/hadoop/hbase/regionserver/ScanDeleteTracker.java | 2 ++ .../apache/hadoop/hbase/regionserver/ScanQueryMatcher.java | 2 ++ .../hbase/regionserver/ScanWildcardColumnTracker.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/ShutdownHook.java | 2 ++ .../apache/hadoop/hbase/regionserver/SplitLogWorker.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/SplitRequest.java | 2 ++ .../apache/hadoop/hbase/regionserver/SplitTransaction.java | 2 ++ .../java/org/apache/hadoop/hbase/regionserver/Store.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/StoreFile.java | 2 ++ .../apache/hadoop/hbase/regionserver/StoreFileScanner.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/StoreFlusher.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/StoreScanner.java | 2 ++ .../apache/hadoop/hbase/regionserver/TimeRangeTracker.java | 2 ++ .../hadoop/hbase/regionserver/WrongRegionException.java | 5 +++++ .../hbase/regionserver/compactions/CompactSelection.java | 2 ++ .../hbase/regionserver/compactions/CompactionProgress.java | 3 +++ .../hbase/regionserver/compactions/CompactionRequest.java | 2 ++ .../hadoop/hbase/regionserver/handler/CloseMetaHandler.java | 2 ++ .../hbase/regionserver/handler/CloseRegionHandler.java | 2 ++ .../hadoop/hbase/regionserver/handler/CloseRootHandler.java | 2 ++ .../hadoop/hbase/regionserver/handler/OpenMetaHandler.java | 2 ++ .../hbase/regionserver/handler/OpenRegionHandler.java | 2 ++ .../hadoop/hbase/regionserver/handler/OpenRootHandler.java | 2 ++ .../regionserver/metrics/RegionServerDynamicMetrics.java | 2 ++ .../regionserver/metrics/RegionServerDynamicStatistics.java | 2 ++ .../hbase/regionserver/metrics/RegionServerMetrics.java | 2 ++ .../hbase/regionserver/metrics/RegionServerStatistics.java | 2 ++ .../hadoop/hbase/regionserver/metrics/SchemaConfigured.java | 2 ++ .../hadoop/hbase/regionserver/metrics/SchemaMetrics.java | 2 ++ .../hbase/regionserver/wal/FailedLogCloseException.java | 3 +++ .../java/org/apache/hadoop/hbase/regionserver/wal/HLog.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/wal/HLogKey.java | 2 ++ .../hadoop/hbase/regionserver/wal/HLogPrettyPrinter.java | 4 ++++ .../apache/hadoop/hbase/regionserver/wal/HLogSplitter.java | 2 ++ .../regionserver/wal/OrphanHLogAfterSplitException.java | 3 +++ .../hbase/regionserver/wal/SequenceFileLogReader.java | 2 ++ .../hbase/regionserver/wal/SequenceFileLogWriter.java | 2 ++ .../hadoop/hbase/regionserver/wal/WALActionsListener.java | 3 +++ .../hadoop/hbase/regionserver/wal/WALCoprocessorHost.java | 2 ++ .../org/apache/hadoop/hbase/regionserver/wal/WALEdit.java | 2 ++ .../apache/hadoop/hbase/replication/ReplicationPeer.java | 2 ++ .../hadoop/hbase/replication/ReplicationZookeeper.java | 2 ++ .../hbase/replication/master/ReplicationLogCleaner.java | 2 ++ .../hadoop/hbase/replication/regionserver/Replication.java | 2 ++ .../hbase/replication/regionserver/ReplicationSink.java | 2 ++ .../replication/regionserver/ReplicationSinkMetrics.java | 2 ++ .../hbase/replication/regionserver/ReplicationSource.java | 2 ++ .../regionserver/ReplicationSourceInterface.java | 2 ++ .../replication/regionserver/ReplicationSourceManager.java | 2 ++ .../replication/regionserver/ReplicationSourceMetrics.java | 2 ++ .../replication/regionserver/ReplicationStatistics.java | 2 ++ src/main/java/org/apache/hadoop/hbase/rest/Constants.java | 5 +++++ .../java/org/apache/hadoop/hbase/rest/ExistsResource.java | 3 +++ src/main/java/org/apache/hadoop/hbase/rest/Main.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/MultiRowResource.java | 2 ++ .../apache/hadoop/hbase/rest/ProtobufMessageHandler.java | 3 +++ src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/RegionsResource.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/ResourceBase.java | 3 +++ .../java/org/apache/hadoop/hbase/rest/ResourceConfig.java | 3 +++ .../java/org/apache/hadoop/hbase/rest/ResultGenerator.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/RootResource.java | 2 ++ src/main/java/org/apache/hadoop/hbase/rest/RowResource.java | 2 ++ .../org/apache/hadoop/hbase/rest/RowResultGenerator.java | 2 ++ src/main/java/org/apache/hadoop/hbase/rest/RowSpec.java | 2 ++ .../apache/hadoop/hbase/rest/ScannerInstanceResource.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/ScannerResource.java | 2 ++ .../apache/hadoop/hbase/rest/ScannerResultGenerator.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/SchemaResource.java | 2 ++ .../hadoop/hbase/rest/StorageClusterStatusResource.java | 2 ++ .../hadoop/hbase/rest/StorageClusterVersionResource.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/TableResource.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/VersionResource.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/client/Client.java | 4 ++++ .../java/org/apache/hadoop/hbase/rest/client/Cluster.java | 5 +++++ .../org/apache/hadoop/hbase/rest/client/RemoteAdmin.java | 4 ++++ .../org/apache/hadoop/hbase/rest/client/RemoteHTable.java | 4 ++++ .../java/org/apache/hadoop/hbase/rest/client/Response.java | 4 ++++ .../apache/hadoop/hbase/rest/filter/GZIPRequestStream.java | 3 +++ .../apache/hadoop/hbase/rest/filter/GZIPRequestWrapper.java | 3 +++ .../apache/hadoop/hbase/rest/filter/GZIPResponseStream.java | 3 +++ .../hadoop/hbase/rest/filter/GZIPResponseWrapper.java | 3 +++ .../org/apache/hadoop/hbase/rest/filter/GzipFilter.java | 3 +++ .../org/apache/hadoop/hbase/rest/metrics/RESTMetrics.java | 2 ++ .../apache/hadoop/hbase/rest/metrics/RESTStatistics.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/model/CellModel.java | 2 ++ .../org/apache/hadoop/hbase/rest/model/CellSetModel.java | 2 ++ .../apache/hadoop/hbase/rest/model/ColumnSchemaModel.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/model/RowModel.java | 2 ++ .../org/apache/hadoop/hbase/rest/model/ScannerModel.java | 2 ++ .../hadoop/hbase/rest/model/StorageClusterStatusModel.java | 2 ++ .../hadoop/hbase/rest/model/StorageClusterVersionModel.java | 3 +++ .../org/apache/hadoop/hbase/rest/model/TableInfoModel.java | 2 ++ .../org/apache/hadoop/hbase/rest/model/TableListModel.java | 2 ++ .../java/org/apache/hadoop/hbase/rest/model/TableModel.java | 3 +++ .../apache/hadoop/hbase/rest/model/TableRegionModel.java | 2 ++ .../apache/hadoop/hbase/rest/model/TableSchemaModel.java | 2 ++ .../org/apache/hadoop/hbase/rest/model/VersionModel.java | 2 ++ .../hadoop/hbase/rest/provider/JAXBContextResolver.java | 2 ++ .../rest/provider/consumer/ProtobufMessageBodyConsumer.java | 2 ++ .../provider/producer/PlainTextMessageBodyProducer.java | 2 ++ .../rest/provider/producer/ProtobufMessageBodyProducer.java | 2 ++ .../java/org/apache/hadoop/hbase/security/KerberosInfo.java | 3 +++ .../java/org/apache/hadoop/hbase/security/TokenInfo.java | 3 +++ src/main/java/org/apache/hadoop/hbase/security/User.java | 2 ++ src/main/java/org/apache/hadoop/hbase/thrift/CallQueue.java | 2 ++ .../hadoop/hbase/thrift/HThreadedSelectorServerArgs.java | 2 ++ .../hadoop/hbase/thrift/HbaseHandlerMetricsProxy.java | 2 ++ .../hadoop/hbase/thrift/TBoundedThreadPoolServer.java | 2 ++ .../java/org/apache/hadoop/hbase/thrift/ThriftMetrics.java | 2 ++ .../java/org/apache/hadoop/hbase/thrift/ThriftServer.java | 2 ++ .../org/apache/hadoop/hbase/thrift/ThriftServerRunner.java | 2 ++ .../org/apache/hadoop/hbase/thrift/ThriftUtilities.java | 2 ++ .../hadoop/hbase/thrift2/ThriftHBaseServiceHandler.java | 2 ++ .../java/org/apache/hadoop/hbase/thrift2/ThriftServer.java | 2 ++ .../org/apache/hadoop/hbase/thrift2/ThriftUtilities.java | 2 ++ .../org/apache/hadoop/hbase/util/AbstractHBaseTool.java | 2 ++ src/main/java/org/apache/hadoop/hbase/util/Addressing.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/Base64.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java | 3 +++ .../java/org/apache/hadoop/hbase/util/BloomFilterBase.java | 2 ++ .../org/apache/hadoop/hbase/util/BloomFilterFactory.java | 2 ++ .../org/apache/hadoop/hbase/util/BloomFilterWriter.java | 2 ++ .../java/org/apache/hadoop/hbase/util/ByteBloomFilter.java | 2 ++ .../apache/hadoop/hbase/util/ByteBufferOutputStream.java | 5 +++++ .../java/org/apache/hadoop/hbase/util/ByteBufferUtils.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/Bytes.java | 4 ++++ .../apache/hadoop/hbase/util/CancelableProgressable.java | 3 +++ src/main/java/org/apache/hadoop/hbase/util/ClassSize.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/Classes.java | 5 +++++ .../apache/hadoop/hbase/util/CollectionBackedScanner.java | 2 ++ .../org/apache/hadoop/hbase/util/CompoundBloomFilter.java | 2 ++ .../apache/hadoop/hbase/util/CompoundBloomFilterBase.java | 2 ++ .../apache/hadoop/hbase/util/CompoundBloomFilterWriter.java | 2 ++ .../java/org/apache/hadoop/hbase/util/CompressionTest.java | 4 ++++ .../apache/hadoop/hbase/util/DefaultEnvironmentEdge.java | 5 +++++ .../org/apache/hadoop/hbase/util/DirectMemoryUtils.java | 5 +++++ .../java/org/apache/hadoop/hbase/util/EnvironmentEdge.java | 5 +++++ .../apache/hadoop/hbase/util/EnvironmentEdgeManager.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/FSMapRUtils.java | 3 +++ .../org/apache/hadoop/hbase/util/FSTableDescriptors.java | 2 ++ src/main/java/org/apache/hadoop/hbase/util/FSUtils.java | 4 ++++ .../hadoop/hbase/util/FileSystemVersionException.java | 5 +++++ .../java/org/apache/hadoop/hbase/util/HBaseConfTool.java | 2 ++ src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java | 4 ++++ .../java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/HMerge.java | 2 ++ src/main/java/org/apache/hadoop/hbase/util/HasThread.java | 3 +++ src/main/java/org/apache/hadoop/hbase/util/Hash.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/HashedBytes.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/IdLock.java | 3 +++ .../hadoop/hbase/util/IncrementingEnvironmentEdge.java | 3 +++ src/main/java/org/apache/hadoop/hbase/util/InfoServer.java | 2 ++ .../java/org/apache/hadoop/hbase/util/JVMClusterUtil.java | 2 ++ src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/JvmVersion.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/KeyRange.java | 3 +++ src/main/java/org/apache/hadoop/hbase/util/Keying.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/MD5Hash.java | 4 ++++ .../org/apache/hadoop/hbase/util/ManualEnvironmentEdge.java | 3 +++ src/main/java/org/apache/hadoop/hbase/util/Merge.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/MetaUtils.java | 2 ++ src/main/java/org/apache/hadoop/hbase/util/Methods.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/MurmurHash.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/Objects.java | 2 ++ src/main/java/org/apache/hadoop/hbase/util/Pair.java | 5 +++++ .../java/org/apache/hadoop/hbase/util/PairOfSameType.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/PoolMap.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/ProtoUtil.java | 3 +++ .../org/apache/hadoop/hbase/util/RegionSplitCalculator.java | 2 ++ .../java/org/apache/hadoop/hbase/util/RegionSplitter.java | 2 ++ .../java/org/apache/hadoop/hbase/util/RetryCounter.java | 2 ++ .../org/apache/hadoop/hbase/util/RetryCounterFactory.java | 3 +++ .../org/apache/hadoop/hbase/util/ServerCommandLine.java | 2 ++ src/main/java/org/apache/hadoop/hbase/util/Sleeper.java | 4 ++++ .../org/apache/hadoop/hbase/util/SoftValueSortedMap.java | 5 +++++ .../org/apache/hadoop/hbase/util/SortedCopyOnWriteSet.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/Strings.java | 5 +++++ src/main/java/org/apache/hadoop/hbase/util/Threads.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/VersionInfo.java | 4 ++++ src/main/java/org/apache/hadoop/hbase/util/Writables.java | 2 ++ .../apache/hadoop/hbase/util/hbck/OfflineMetaRepair.java | 4 ++++ .../java/org/apache/hadoop/hbase/zookeeper/ClusterId.java | 2 ++ .../apache/hadoop/hbase/zookeeper/ClusterStatusTracker.java | 2 ++ .../hadoop/hbase/zookeeper/DrainingServerTracker.java | 2 ++ .../java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java | 4 ++++ .../hadoop/hbase/zookeeper/MasterSchemaChangeTracker.java | 2 ++ .../org/apache/hadoop/hbase/zookeeper/MetaNodeTracker.java | 2 ++ .../apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java | 4 ++++ .../apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java | 4 ++++ .../apache/hadoop/hbase/zookeeper/RegionServerTracker.java | 2 ++ .../apache/hadoop/hbase/zookeeper/RootRegionTracker.java | 2 ++ .../apache/hadoop/hbase/zookeeper/SchemaChangeTracker.java | 2 ++ .../java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java | 4 ++++ .../java/org/apache/hadoop/hbase/zookeeper/ZKConfig.java | 4 ++++ .../org/apache/hadoop/hbase/zookeeper/ZKLeaderManager.java | 4 ++++ .../org/apache/hadoop/hbase/zookeeper/ZKServerTool.java | 4 ++++ .../java/org/apache/hadoop/hbase/zookeeper/ZKSplitLog.java | 2 ++ .../java/org/apache/hadoop/hbase/zookeeper/ZKTable.java | 2 ++ src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java | 4 ++++ .../apache/hadoop/hbase/zookeeper/ZooKeeperListener.java | 5 +++++ .../apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java | 4 ++++ .../org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java | 4 ++++ .../java/org/apache/hadoop/hbase/HBaseTestingUtility.java | 4 ++++ src/test/java/org/apache/hadoop/hbase/MiniHBaseCluster.java | 4 ++++ 579 files changed, 1726 insertions(+), 5 deletions(-) diff --git a/bin/region_mover.rb b/bin/region_mover.rb index 028f9a89cbe..94da5d90bd9 100644 --- a/bin/region_mover.rb +++ b/bin/region_mover.rb @@ -364,7 +364,8 @@ def loadRegions(options, hostname) for r in regions exists = false begin - exists = isSuccessfulScan(admin, r) + isSuccessfulScan(admin, r) + exists = true rescue org.apache.hadoop.hbase.NotServingRegionException => e $LOG.info("Failed scan of " + e.message) end diff --git a/src/main/java/org/apache/hadoop/hbase/Abortable.java b/src/main/java/org/apache/hadoop/hbase/Abortable.java index 03249c12171..403e6f21fd8 100644 --- a/src/main/java/org/apache/hadoop/hbase/Abortable.java +++ b/src/main/java/org/apache/hadoop/hbase/Abortable.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Interface to support the aborting of a given server or client. *

@@ -27,6 +29,7 @@ package org.apache.hadoop.hbase; *

* Implemented by the Master, RegionServer, and TableServers (client). */ +@InterfaceAudience.Private public interface Abortable { /** * Abort the server or client. diff --git a/src/main/java/org/apache/hadoop/hbase/Chore.java b/src/main/java/org/apache/hadoop/hbase/Chore.java index 38f476bbfb8..58a0407e0ce 100644 --- a/src/main/java/org/apache/hadoop/hbase/Chore.java +++ b/src/main/java/org/apache/hadoop/hbase/Chore.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.HasThread; import org.apache.hadoop.hbase.util.Sleeper; @@ -34,6 +35,7 @@ import org.apache.hadoop.hbase.util.Sleeper; *

Don't subclass Chore if the task relies on being woken up for something to * do, such as an entry being added to a queue, etc. */ +@InterfaceAudience.Private public abstract class Chore extends HasThread { private final Log LOG = LogFactory.getLog(this.getClass()); private final Sleeper sleeper; diff --git a/src/main/java/org/apache/hadoop/hbase/ClockOutOfSyncException.java b/src/main/java/org/apache/hadoop/hbase/ClockOutOfSyncException.java index 5c51e4bb8b5..788eeed1dea 100644 --- a/src/main/java/org/apache/hadoop/hbase/ClockOutOfSyncException.java +++ b/src/main/java/org/apache/hadoop/hbase/ClockOutOfSyncException.java @@ -21,11 +21,16 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * This exception is thrown by the master when a region server clock skew is * too high. */ @SuppressWarnings("serial") +@InterfaceAudience.Public +@InterfaceStability.Stable public class ClockOutOfSyncException extends IOException { public ClockOutOfSyncException(String message) { super(message); diff --git a/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java b/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java index 7618e30baf5..9df4c102e65 100644 --- a/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java +++ b/src/main/java/org/apache/hadoop/hbase/ClusterStatus.java @@ -31,6 +31,8 @@ import java.util.HashMap; import java.util.Map; import java.util.TreeMap; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.master.AssignmentManager.RegionState; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.VersionMismatchException; @@ -54,6 +56,8 @@ import org.apache.hadoop.io.VersionedWritable; *

  • The unique cluster ID
  • * */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ClusterStatus extends VersionedWritable { /** * Version for object serialization. Incremented for changes in serialized diff --git a/src/main/java/org/apache/hadoop/hbase/Coprocessor.java b/src/main/java/org/apache/hadoop/hbase/Coprocessor.java index c0cb4636715..364141d3374 100644 --- a/src/main/java/org/apache/hadoop/hbase/Coprocessor.java +++ b/src/main/java/org/apache/hadoop/hbase/Coprocessor.java @@ -18,9 +18,14 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Coprocess interface. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface Coprocessor { static final int VERSION = 1; diff --git a/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java b/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java index dbb5cc94631..c426cf7a603 100644 --- a/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java +++ b/src/main/java/org/apache/hadoop/hbase/CoprocessorEnvironment.java @@ -18,12 +18,16 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.client.HTableInterface; /** * Coprocessor environment state. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface CoprocessorEnvironment { /** @return the Coprocessor interface version */ diff --git a/src/main/java/org/apache/hadoop/hbase/DoNotRetryIOException.java b/src/main/java/org/apache/hadoop/hbase/DoNotRetryIOException.java index 98c5b9bb422..7bb74255f12 100644 --- a/src/main/java/org/apache/hadoop/hbase/DoNotRetryIOException.java +++ b/src/main/java/org/apache/hadoop/hbase/DoNotRetryIOException.java @@ -21,10 +21,15 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Subclass if exception is not meant to be retried: e.g. * {@link UnknownScannerException} */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class DoNotRetryIOException extends IOException { private static final long serialVersionUID = 1197446454511704139L; diff --git a/src/main/java/org/apache/hadoop/hbase/DroppedSnapshotException.java b/src/main/java/org/apache/hadoop/hbase/DroppedSnapshotException.java index 9b1d021c853..bf86e94124e 100644 --- a/src/main/java/org/apache/hadoop/hbase/DroppedSnapshotException.java +++ b/src/main/java/org/apache/hadoop/hbase/DroppedSnapshotException.java @@ -16,11 +16,16 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Thrown during flush if the possibility snapshot content was not properly * persisted into store files. Response should include replay of hlog content. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class DroppedSnapshotException extends IOException { private static final long serialVersionUID = -5463156580831677374L; diff --git a/src/main/java/org/apache/hadoop/hbase/EmptyWatcher.java b/src/main/java/org/apache/hadoop/hbase/EmptyWatcher.java index e0e0a288992..9881ec2223a 100644 --- a/src/main/java/org/apache/hadoop/hbase/EmptyWatcher.java +++ b/src/main/java/org/apache/hadoop/hbase/EmptyWatcher.java @@ -19,12 +19,14 @@ */ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.zookeeper.Watcher; import org.apache.zookeeper.WatchedEvent; /** * An empty ZooKeeper watcher */ +@InterfaceAudience.Private public class EmptyWatcher implements Watcher { public static EmptyWatcher instance = new EmptyWatcher(); private EmptyWatcher() {} diff --git a/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java b/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java index 04e99b16116..862d1345cbd 100644 --- a/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java +++ b/src/main/java/org/apache/hadoop/hbase/HBaseConfiguration.java @@ -23,12 +23,16 @@ import java.util.Map.Entry; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.util.VersionInfo; /** * Adds HBase configuration files to a Configuration */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class HBaseConfiguration extends Configuration { private static final Log LOG = LogFactory.getLog(HBaseConfiguration.class); diff --git a/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java b/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java index 324ec46f34d..e598554d3ed 100644 --- a/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java +++ b/src/main/java/org/apache/hadoop/hbase/HColumnDescriptor.java @@ -26,6 +26,8 @@ import java.util.Collections; import java.util.HashMap; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; import org.apache.hadoop.hbase.io.hfile.Compression; @@ -45,6 +47,8 @@ import org.apache.hadoop.io.WritableComparable; * column and recreating it. If there is data stored in the column, it will be * deleted when the column is deleted. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HColumnDescriptor implements WritableComparable { // For future backward compatibility diff --git a/src/main/java/org/apache/hadoop/hbase/HConstants.java b/src/main/java/org/apache/hadoop/hbase/HConstants.java index 17cb0e3d9fd..e60ce043f81 100644 --- a/src/main/java/org/apache/hadoop/hbase/HConstants.java +++ b/src/main/java/org/apache/hadoop/hbase/HConstants.java @@ -25,12 +25,16 @@ import java.util.List; import java.util.UUID; import java.util.regex.Pattern; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.ipc.HRegionInterface; import org.apache.hadoop.hbase.util.Bytes; /** * HConstants holds a bunch of HBase-related constants */ +@InterfaceAudience.Public +@InterfaceStability.Stable public final class HConstants { /** * Status codes used for return values of bulk operations. diff --git a/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java b/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java index 29c59afb40a..1716d15b4e9 100644 --- a/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java +++ b/src/main/java/org/apache/hadoop/hbase/HDFSBlocksDistribution.java @@ -27,10 +27,13 @@ import java.util.NavigableSet; import java.util.TreeMap; import java.util.TreeSet; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Data structure to describe the distribution of HDFS blocks amount hosts */ +@InterfaceAudience.Private public class HDFSBlocksDistribution { private Map hostAndWeights = null; private long uniqueBlocksTotalWeight = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java b/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java index 74cb8217b35..98f79fc24bd 100644 --- a/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java +++ b/src/main/java/org/apache/hadoop/hbase/HRegionInfo.java @@ -27,6 +27,8 @@ import java.util.Arrays; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -44,6 +46,8 @@ import org.apache.hadoop.io.WritableComparable; * Contains HRegion id, start and end keys, a reference to this * HRegions' table descriptor, etc. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HRegionInfo extends VersionedWritable implements WritableComparable { // VERSION == 0 when HRegionInfo had an HTableDescriptor inside it. diff --git a/src/main/java/org/apache/hadoop/hbase/HRegionLocation.java b/src/main/java/org/apache/hadoop/hbase/HRegionLocation.java index 85fb91dc7d0..ef92556686f 100644 --- a/src/main/java/org/apache/hadoop/hbase/HRegionLocation.java +++ b/src/main/java/org/apache/hadoop/hbase/HRegionLocation.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Addressing; /** @@ -28,6 +30,8 @@ import org.apache.hadoop.hbase.util.Addressing; * instances are the same if they refer to the same 'location' (the same * hostname and port), though they may be carrying different regions. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HRegionLocation implements Comparable { private final HRegionInfo regionInfo; private final String hostname; diff --git a/src/main/java/org/apache/hadoop/hbase/HServerLoad.java b/src/main/java/org/apache/hadoop/hbase/HServerLoad.java index acded62dbad..5d703b9be0c 100644 --- a/src/main/java/org/apache/hadoop/hbase/HServerLoad.java +++ b/src/main/java/org/apache/hadoop/hbase/HServerLoad.java @@ -29,6 +29,8 @@ import java.util.Set; import java.util.TreeMap; import java.util.TreeSet; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Strings; import org.apache.hadoop.io.VersionedWritable; @@ -38,6 +40,8 @@ import org.apache.hadoop.io.WritableUtils; /** * This class is used exporting current state of load on a RegionServer. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HServerLoad extends VersionedWritable implements WritableComparable { private static final byte VERSION = 2; diff --git a/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java b/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java index f79739768e9..fc5e53e0b7d 100644 --- a/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java +++ b/src/main/java/org/apache/hadoop/hbase/HTableDescriptor.java @@ -31,6 +31,8 @@ import java.util.Set; import java.util.TreeMap; import java.util.regex.Matcher; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.io.hfile.Compression; @@ -45,6 +47,8 @@ import org.apache.hadoop.io.WritableComparable; * .META. , is the table is read only, the maximum size of the memstore, * when the region split should occur, coprocessors associated with it etc... */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HTableDescriptor implements WritableComparable { /** diff --git a/src/main/java/org/apache/hadoop/hbase/InvalidFamilyOperationException.java b/src/main/java/org/apache/hadoop/hbase/InvalidFamilyOperationException.java index bb2b6669d87..fa590fced82 100644 --- a/src/main/java/org/apache/hadoop/hbase/InvalidFamilyOperationException.java +++ b/src/main/java/org/apache/hadoop/hbase/InvalidFamilyOperationException.java @@ -21,10 +21,15 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Thrown if a request is table schema modification is requested but * made for an invalid family name. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class InvalidFamilyOperationException extends IOException { private static final long serialVersionUID = 1L << 22 - 1L; /** default constructor */ diff --git a/src/main/java/org/apache/hadoop/hbase/KeyValue.java b/src/main/java/org/apache/hadoop/hbase/KeyValue.java index b8fb08556b2..63112157246 100644 --- a/src/main/java/org/apache/hadoop/hbase/KeyValue.java +++ b/src/main/java/org/apache/hadoop/hbase/KeyValue.java @@ -30,6 +30,8 @@ import java.util.Map; import com.google.common.primitives.Longs; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.io.HeapSize; import org.apache.hadoop.hbase.io.hfile.HFile; import org.apache.hadoop.hbase.util.Bytes; @@ -63,6 +65,8 @@ import org.apache.hadoop.io.Writable; * be < Integer.MAX_SIZE. * The column does not contain the family/qualifier delimiter, {@link #COLUMN_FAMILY_DELIMITER} */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class KeyValue implements Writable, HeapSize { static final Log LOG = LogFactory.getLog(KeyValue.class); // TODO: Group Key-only comparators and operations into a Key class, just diff --git a/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java b/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java index 5fcc39c87a0..cc8a6897647 100644 --- a/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java +++ b/src/main/java/org/apache/hadoop/hbase/LocalHBaseCluster.java @@ -27,6 +27,8 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.regionserver.HRegionServer; @@ -55,6 +57,8 @@ import org.apache.hadoop.hbase.util.JVMClusterUtil; * instead of 60000. * */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class LocalHBaseCluster { static final Log LOG = LogFactory.getLog(LocalHBaseCluster.class); private final List masterThreads = diff --git a/src/main/java/org/apache/hadoop/hbase/MasterAddressTracker.java b/src/main/java/org/apache/hadoop/hbase/MasterAddressTracker.java index 2fdc6ba3c7c..a246abedc0c 100644 --- a/src/main/java/org/apache/hadoop/hbase/MasterAddressTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/MasterAddressTracker.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.zookeeper.ZooKeeperNodeTracker; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; @@ -36,6 +37,7 @@ import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; *

    * You can get the current master via {@link #getMasterAddress()} */ +@InterfaceAudience.Private public class MasterAddressTracker extends ZooKeeperNodeTracker { /** * Construct a master address listener with the specified diff --git a/src/main/java/org/apache/hadoop/hbase/MasterNotRunningException.java b/src/main/java/org/apache/hadoop/hbase/MasterNotRunningException.java index 6cf564c7bba..c32316f59a2 100644 --- a/src/main/java/org/apache/hadoop/hbase/MasterNotRunningException.java +++ b/src/main/java/org/apache/hadoop/hbase/MasterNotRunningException.java @@ -21,9 +21,14 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Thrown if the master is not running */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class MasterNotRunningException extends IOException { private static final long serialVersionUID = 1L << 23 - 1L; /** default constructor */ diff --git a/src/main/java/org/apache/hadoop/hbase/NotAllMetaRegionsOnlineException.java b/src/main/java/org/apache/hadoop/hbase/NotAllMetaRegionsOnlineException.java index 2c275e3586a..2b9812d3449 100644 --- a/src/main/java/org/apache/hadoop/hbase/NotAllMetaRegionsOnlineException.java +++ b/src/main/java/org/apache/hadoop/hbase/NotAllMetaRegionsOnlineException.java @@ -20,11 +20,15 @@ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.DoNotRetryIOException; /** * Thrown when an operation requires the root and all meta regions to be online */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class NotAllMetaRegionsOnlineException extends DoNotRetryIOException { private static final long serialVersionUID = 6439786157874827523L; /** diff --git a/src/main/java/org/apache/hadoop/hbase/NotServingRegionException.java b/src/main/java/org/apache/hadoop/hbase/NotServingRegionException.java index 32da8cb7ff0..d29713cff31 100644 --- a/src/main/java/org/apache/hadoop/hbase/NotServingRegionException.java +++ b/src/main/java/org/apache/hadoop/hbase/NotServingRegionException.java @@ -21,12 +21,16 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; /** * Thrown by a region server if it is sent a request for a region it is not * serving. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class NotServingRegionException extends IOException { private static final long serialVersionUID = 1L << 17 - 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/PleaseHoldException.java b/src/main/java/org/apache/hadoop/hbase/PleaseHoldException.java index 309dca41343..6866ee48048 100644 --- a/src/main/java/org/apache/hadoop/hbase/PleaseHoldException.java +++ b/src/main/java/org/apache/hadoop/hbase/PleaseHoldException.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; /** * This exception is thrown by the master when a region server was shut down and @@ -28,8 +30,10 @@ import java.io.IOException; * operations */ @SuppressWarnings("serial") +@InterfaceAudience.Public +@InterfaceStability.Stable public class PleaseHoldException extends IOException { public PleaseHoldException(String message) { super(message); } -} \ No newline at end of file +} diff --git a/src/main/java/org/apache/hadoop/hbase/RegionException.java b/src/main/java/org/apache/hadoop/hbase/RegionException.java index 63063a5e95e..2f3debf4e1e 100644 --- a/src/main/java/org/apache/hadoop/hbase/RegionException.java +++ b/src/main/java/org/apache/hadoop/hbase/RegionException.java @@ -20,10 +20,15 @@ package org.apache.hadoop.hbase; import java.io.IOException; + +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; /** * Thrown when something happens related to region handling. * Subclasses have to be more specific. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class RegionException extends IOException { private static final long serialVersionUID = 1473510258071111371L; diff --git a/src/main/java/org/apache/hadoop/hbase/RemoteExceptionHandler.java b/src/main/java/org/apache/hadoop/hbase/RemoteExceptionHandler.java index 485c254cd40..af68677be5b 100644 --- a/src/main/java/org/apache/hadoop/hbase/RemoteExceptionHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/RemoteExceptionHandler.java @@ -23,12 +23,14 @@ import java.io.IOException; import java.lang.reflect.Constructor; import java.lang.reflect.InvocationTargetException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.ipc.RemoteException; /** * An immutable class which contains a static method for handling * org.apache.hadoop.ipc.RemoteException exceptions. */ +@InterfaceAudience.Private public class RemoteExceptionHandler { /* Not instantiable */ private RemoteExceptionHandler() {super();} diff --git a/src/main/java/org/apache/hadoop/hbase/Server.java b/src/main/java/org/apache/hadoop/hbase/Server.java index de19e2c2a91..9235c7dfe31 100644 --- a/src/main/java/org/apache/hadoop/hbase/Server.java +++ b/src/main/java/org/apache/hadoop/hbase/Server.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.catalog.CatalogTracker; import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; @@ -27,6 +28,7 @@ import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; * Defines the set of shared functions implemented by HBase servers (Masters * and RegionServers). */ +@InterfaceAudience.Private public interface Server extends Abortable, Stoppable { /** * Gets the configuration object for this server. diff --git a/src/main/java/org/apache/hadoop/hbase/ServerName.java b/src/main/java/org/apache/hadoop/hbase/ServerName.java index d6d869a6926..9046dbd8941 100644 --- a/src/main/java/org/apache/hadoop/hbase/ServerName.java +++ b/src/main/java/org/apache/hadoop/hbase/ServerName.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase; import java.util.Collection; import java.util.regex.Pattern; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Addressing; import org.apache.hadoop.hbase.util.Bytes; @@ -43,6 +45,8 @@ import org.apache.hadoop.hbase.util.Bytes; * *

    Immutable. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ServerName implements Comparable { /** * Version for this class. diff --git a/src/main/java/org/apache/hadoop/hbase/Stoppable.java b/src/main/java/org/apache/hadoop/hbase/Stoppable.java index 74d4f4a689f..391fa28ef80 100644 --- a/src/main/java/org/apache/hadoop/hbase/Stoppable.java +++ b/src/main/java/org/apache/hadoop/hbase/Stoppable.java @@ -19,9 +19,12 @@ */ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Implementers are Stoppable. */ +@InterfaceAudience.Private public interface Stoppable { /** * Stop this service. diff --git a/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java b/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java index 145589504b5..8b8162ae017 100644 --- a/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java +++ b/src/main/java/org/apache/hadoop/hbase/TableDescriptors.java @@ -21,10 +21,15 @@ import java.io.FileNotFoundException; import java.io.IOException; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Get, remove and modify table descriptors. * Used by servers to host descriptors. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface TableDescriptors { /** * @param tablename diff --git a/src/main/java/org/apache/hadoop/hbase/TableExistsException.java b/src/main/java/org/apache/hadoop/hbase/TableExistsException.java index 5fde21920c4..c21fd214535 100644 --- a/src/main/java/org/apache/hadoop/hbase/TableExistsException.java +++ b/src/main/java/org/apache/hadoop/hbase/TableExistsException.java @@ -17,9 +17,14 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Thrown when a table exists but should not */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableExistsException extends IOException { private static final long serialVersionUID = 1L << 7 - 1L; /** default constructor */ diff --git a/src/main/java/org/apache/hadoop/hbase/TableNotDisabledException.java b/src/main/java/org/apache/hadoop/hbase/TableNotDisabledException.java index 42878001a7e..95d185da90c 100644 --- a/src/main/java/org/apache/hadoop/hbase/TableNotDisabledException.java +++ b/src/main/java/org/apache/hadoop/hbase/TableNotDisabledException.java @@ -21,11 +21,15 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; /** * Thrown if a table should be offline but is not */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableNotDisabledException extends IOException { private static final long serialVersionUID = 1L << 19 - 1L; /** default constructor */ diff --git a/src/main/java/org/apache/hadoop/hbase/TableNotEnabledException.java b/src/main/java/org/apache/hadoop/hbase/TableNotEnabledException.java index d0392d52f6e..a3163d9955b 100644 --- a/src/main/java/org/apache/hadoop/hbase/TableNotEnabledException.java +++ b/src/main/java/org/apache/hadoop/hbase/TableNotEnabledException.java @@ -21,11 +21,15 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; /** * Thrown if a table should be enabled but is not */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableNotEnabledException extends IOException { private static final long serialVersionUID = 262144L; /** default constructor */ diff --git a/src/main/java/org/apache/hadoop/hbase/TableNotFoundException.java b/src/main/java/org/apache/hadoop/hbase/TableNotFoundException.java index dc6da43e470..f9ff640cadf 100644 --- a/src/main/java/org/apache/hadoop/hbase/TableNotFoundException.java +++ b/src/main/java/org/apache/hadoop/hbase/TableNotFoundException.java @@ -19,7 +19,12 @@ */ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** Thrown when a table can not be located */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableNotFoundException extends RegionException { private static final long serialVersionUID = 993179627856392526L; diff --git a/src/main/java/org/apache/hadoop/hbase/UnknownRegionException.java b/src/main/java/org/apache/hadoop/hbase/UnknownRegionException.java index e87f42a6013..1be340e6c70 100644 --- a/src/main/java/org/apache/hadoop/hbase/UnknownRegionException.java +++ b/src/main/java/org/apache/hadoop/hbase/UnknownRegionException.java @@ -21,9 +21,14 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Thrown when we are asked to operate on a region we know nothing about. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class UnknownRegionException extends IOException { private static final long serialVersionUID = 1968858760475205392L; diff --git a/src/main/java/org/apache/hadoop/hbase/UnknownRowLockException.java b/src/main/java/org/apache/hadoop/hbase/UnknownRowLockException.java index 8ca50a9d807..05194f58462 100644 --- a/src/main/java/org/apache/hadoop/hbase/UnknownRowLockException.java +++ b/src/main/java/org/apache/hadoop/hbase/UnknownRowLockException.java @@ -19,10 +19,15 @@ */ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Thrown if a region server is passed an unknown row lock id */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class UnknownRowLockException extends DoNotRetryIOException { private static final long serialVersionUID = 993179627856392526L; diff --git a/src/main/java/org/apache/hadoop/hbase/UnknownScannerException.java b/src/main/java/org/apache/hadoop/hbase/UnknownScannerException.java index 13f2f6c3953..656ebdf725b 100644 --- a/src/main/java/org/apache/hadoop/hbase/UnknownScannerException.java +++ b/src/main/java/org/apache/hadoop/hbase/UnknownScannerException.java @@ -19,6 +19,9 @@ */ package org.apache.hadoop.hbase; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Thrown if a region server is passed an unknown scanner id. @@ -26,6 +29,8 @@ package org.apache.hadoop.hbase; * scanner lease on the serverside has expired OR the serverside is closing * down and has cancelled all leases. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class UnknownScannerException extends DoNotRetryIOException { private static final long serialVersionUID = 993179627856392526L; diff --git a/src/main/java/org/apache/hadoop/hbase/VersionAnnotation.java b/src/main/java/org/apache/hadoop/hbase/VersionAnnotation.java index ecea5800295..c9efee1a816 100644 --- a/src/main/java/org/apache/hadoop/hbase/VersionAnnotation.java +++ b/src/main/java/org/apache/hadoop/hbase/VersionAnnotation.java @@ -19,12 +19,15 @@ package org.apache.hadoop.hbase; import java.lang.annotation.*; +import org.apache.hadoop.classification.InterfaceAudience; + /** * A package attribute that captures the version of hbase that was compiled. * Copied down from hadoop. All is same except name of interface. */ @Retention(RetentionPolicy.RUNTIME) @Target(ElementType.PACKAGE) +@InterfaceAudience.Private public @interface VersionAnnotation { /** diff --git a/src/main/java/org/apache/hadoop/hbase/YouAreDeadException.java b/src/main/java/org/apache/hadoop/hbase/YouAreDeadException.java index fcd2ccdd2ed..c7003412e77 100644 --- a/src/main/java/org/apache/hadoop/hbase/YouAreDeadException.java +++ b/src/main/java/org/apache/hadoop/hbase/YouAreDeadException.java @@ -21,12 +21,17 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * This exception is thrown by the master when a region server reports and is * already being processed as dead. This can happen when a region server loses * its session but didn't figure it yet. */ @SuppressWarnings("serial") +@InterfaceAudience.Public +@InterfaceStability.Stable public class YouAreDeadException extends IOException { public YouAreDeadException(String message) { super(message); diff --git a/src/main/java/org/apache/hadoop/hbase/ZooKeeperConnectionException.java b/src/main/java/org/apache/hadoop/hbase/ZooKeeperConnectionException.java index ad48b25a822..779bce0bea6 100644 --- a/src/main/java/org/apache/hadoop/hbase/ZooKeeperConnectionException.java +++ b/src/main/java/org/apache/hadoop/hbase/ZooKeeperConnectionException.java @@ -21,9 +21,14 @@ package org.apache.hadoop.hbase; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Thrown if the client can't connect to zookeeper */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ZooKeeperConnectionException extends IOException { private static final long serialVersionUID = 1L << 23 - 1L; /** default constructor */ diff --git a/src/main/java/org/apache/hadoop/hbase/avro/AvroServer.java b/src/main/java/org/apache/hadoop/hbase/avro/AvroServer.java index f45bbbd8f90..9ba09b51fae 100644 --- a/src/main/java/org/apache/hadoop/hbase/avro/AvroServer.java +++ b/src/main/java/org/apache/hadoop/hbase/avro/AvroServer.java @@ -30,6 +30,7 @@ import org.apache.avro.ipc.specific.SpecificResponder; import org.apache.avro.util.Utf8; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HTableDescriptor; @@ -59,6 +60,7 @@ import org.apache.hadoop.hbase.util.Bytes; /** * Start an Avro server */ +@InterfaceAudience.Private public class AvroServer { /** diff --git a/src/main/java/org/apache/hadoop/hbase/avro/AvroUtil.java b/src/main/java/org/apache/hadoop/hbase/avro/AvroUtil.java index abd2ae6ace7..f3a7e9a34b6 100644 --- a/src/main/java/org/apache/hadoop/hbase/avro/AvroUtil.java +++ b/src/main/java/org/apache/hadoop/hbase/avro/AvroUtil.java @@ -27,6 +27,7 @@ import org.apache.avro.Schema; import org.apache.avro.generic.GenericArray; import org.apache.avro.generic.GenericData; import org.apache.avro.util.Utf8; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.ClusterStatus; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HServerAddress; @@ -59,6 +60,7 @@ import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.io.hfile.Compression; import org.apache.hadoop.hbase.util.Bytes; +@InterfaceAudience.Private public class AvroUtil { // diff --git a/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java b/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java index 3ba3d016813..beaff9791b6 100644 --- a/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/catalog/CatalogTracker.java @@ -28,6 +28,7 @@ import java.util.concurrent.atomic.AtomicBoolean; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.HRegionInfo; @@ -57,6 +58,7 @@ import org.apache.hadoop.ipc.RemoteException; *

    Call {@link #start()} to start up operation. Call {@link #stop()}} to * interrupt waits and close up shop. */ +@InterfaceAudience.Private public class CatalogTracker { // TODO: This class needs a rethink. The original intent was that it would be // the one-stop-shop for root and meta locations and that it would get this diff --git a/src/main/java/org/apache/hadoop/hbase/catalog/MetaEditor.java b/src/main/java/org/apache/hadoop/hbase/catalog/MetaEditor.java index 19fee5c4acd..8873512047c 100644 --- a/src/main/java/org/apache/hadoop/hbase/catalog/MetaEditor.java +++ b/src/main/java/org/apache/hadoop/hbase/catalog/MetaEditor.java @@ -24,6 +24,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.NotAllMetaRegionsOnlineException; @@ -40,6 +41,7 @@ import org.apache.hadoop.hbase.util.Writables; * TODO: Put MetaReader and MetaEditor together; doesn't make sense having * them distinct. */ +@InterfaceAudience.Private public class MetaEditor { // TODO: Strip CatalogTracker from this class. Its all over and in the end // its only used to get its Configuration so we can get associated diff --git a/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java b/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java index 01aa5156193..db1a1f6fad1 100644 --- a/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java +++ b/src/main/java/org/apache/hadoop/hbase/catalog/MetaMigrationRemovingHTD.java @@ -24,6 +24,7 @@ import java.util.Set; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; diff --git a/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java b/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java index 77a121b33a4..0129ee94b5a 100644 --- a/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java +++ b/src/main/java/org/apache/hadoop/hbase/catalog/MetaReader.java @@ -27,6 +27,7 @@ import java.util.TreeMap; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; @@ -46,6 +47,7 @@ import org.apache.hadoop.ipc.RemoteException; /** * Reads region and assignment information from .META.. */ +@InterfaceAudience.Private public class MetaReader { // TODO: Strip CatalogTracker from this class. Its all over and in the end // its only used to get its Configuration so we can get associated diff --git a/src/main/java/org/apache/hadoop/hbase/catalog/RootLocationEditor.java b/src/main/java/org/apache/hadoop/hbase/catalog/RootLocationEditor.java index 1cbf1b6561b..c90864a44b9 100644 --- a/src/main/java/org/apache/hadoop/hbase/catalog/RootLocationEditor.java +++ b/src/main/java/org/apache/hadoop/hbase/catalog/RootLocationEditor.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.catalog; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.zookeeper.ZKUtil; @@ -30,6 +31,7 @@ import org.apache.zookeeper.KeeperException; /** * Makes changes to the location of -ROOT- in ZooKeeper. */ +@InterfaceAudience.Private public class RootLocationEditor { private static final Log LOG = LogFactory.getLog(RootLocationEditor.class); diff --git a/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java b/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java index 047304723b4..2bb0687affb 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/client/AbstractClientScanner.java @@ -20,9 +20,14 @@ package org.apache.hadoop.hbase.client; import java.io.IOException; import java.util.Iterator; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Helper class for custom client scanners. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class AbstractClientScanner implements ResultScanner { @Override diff --git a/src/main/java/org/apache/hadoop/hbase/client/Action.java b/src/main/java/org/apache/hadoop/hbase/client/Action.java index 40b0f2ef54a..bda5c155230 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Action.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Action.java @@ -23,6 +23,8 @@ import java.io.DataInput; import java.io.DataOutput; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.io.HbaseObjectWritable; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.Writable; @@ -32,6 +34,8 @@ import org.apache.hadoop.io.Writable; * {@link HTable::batch} to associate the action with it's region and maintain * the index from the original request. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Action implements Writable, Comparable { private Row action; diff --git a/src/main/java/org/apache/hadoop/hbase/client/Append.java b/src/main/java/org/apache/hadoop/hbase/client/Append.java index a47480c7c52..7c352c8d96f 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Append.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Append.java @@ -25,6 +25,8 @@ import java.util.Arrays; import java.util.List; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.Writable; @@ -41,6 +43,8 @@ import org.apache.hadoop.io.Writable; * row to append to. At least one column to append must be specified using the * {@link #add(byte[], byte[], byte[])} method. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Append extends Mutation implements Row { // TODO: refactor to derive from Put? private static final String RETURN_RESULTS = "_rr_"; diff --git a/src/main/java/org/apache/hadoop/hbase/client/Attributes.java b/src/main/java/org/apache/hadoop/hbase/client/Attributes.java index 5999e3a1b16..e7e6a4da869 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Attributes.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Attributes.java @@ -22,6 +22,11 @@ package org.apache.hadoop.hbase.client; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + +@InterfaceAudience.Public +@InterfaceStability.Stable public interface Attributes { /** * Sets an attribute. diff --git a/src/main/java/org/apache/hadoop/hbase/client/ClientScanner.java b/src/main/java/org/apache/hadoop/hbase/client/ClientScanner.java index 103e39b733f..97293aa4aa1 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/ClientScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/client/ClientScanner.java @@ -23,6 +23,8 @@ import java.util.LinkedList; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HConstants; @@ -40,6 +42,8 @@ import org.apache.hadoop.io.DataOutputBuffer; * If there are multiple regions in a table, this scanner will iterate * through them all. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ClientScanner extends AbstractClientScanner { private final Log LOG = LogFactory.getLog(this.getClass()); private Scan scan; diff --git a/src/main/java/org/apache/hadoop/hbase/client/ConnectionUtils.java b/src/main/java/org/apache/hadoop/hbase/client/ConnectionUtils.java index f5d774e1b20..1278545365d 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/ConnectionUtils.java +++ b/src/main/java/org/apache/hadoop/hbase/client/ConnectionUtils.java @@ -17,6 +17,8 @@ */ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; @@ -24,6 +26,8 @@ import org.apache.hadoop.hbase.HConstants; * Utility used by client connections such as {@link HConnection} and * {@link ServerCallable} */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ConnectionUtils { /** * Calculate pause time. diff --git a/src/main/java/org/apache/hadoop/hbase/client/Delete.java b/src/main/java/org/apache/hadoop/hbase/client/Delete.java index 7cf1b894a00..66ef14e146b 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Delete.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Delete.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; @@ -64,6 +66,8 @@ import java.util.Map; * deleteFamily -- then you need to use the method overrides that take a * timestamp. The constructor timestamp is not referenced. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Delete extends Mutation implements Writable, Row, Comparable { private static final byte DELETE_VERSION = (byte)3; diff --git a/src/main/java/org/apache/hadoop/hbase/client/Get.java b/src/main/java/org/apache/hadoop/hbase/client/Get.java index 93c9e8961e5..f4d3e93733a 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Get.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Get.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.filter.Filter; @@ -63,6 +65,8 @@ import java.util.TreeSet; *

    * To add a filter, execute {@link #setFilter(Filter) setFilter}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Get extends OperationWithAttributes implements Writable, Row, Comparable { private static final byte GET_VERSION = (byte)2; diff --git a/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java b/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java index 73eca295cf5..1420556747b 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java +++ b/src/main/java/org/apache/hadoop/hbase/client/HBaseAdmin.java @@ -32,6 +32,8 @@ import java.util.regex.Pattern; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.ClusterStatus; @@ -72,6 +74,8 @@ import org.apache.hadoop.util.StringUtils; *

    Currently HBaseAdmin instances are not expected to be long-lived. For * example, an HBaseAdmin instance will not ride over a Master restart. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HBaseAdmin implements Abortable, Closeable { private final Log LOG = LogFactory.getLog(this.getClass().getName()); // private final HConnection connection; diff --git a/src/main/java/org/apache/hadoop/hbase/client/HConnection.java b/src/main/java/org/apache/hadoop/hbase/client/HConnection.java index c7b730b062d..9704633226c 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/HConnection.java +++ b/src/main/java/org/apache/hadoop/hbase/client/HConnection.java @@ -25,6 +25,8 @@ import java.util.List; import java.util.Map; import java.util.concurrent.ExecutorService; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.HRegionInfo; @@ -57,6 +59,8 @@ import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; * * @see HConnectionManager */ +@InterfaceAudience.Public +@InterfaceStability.Stable public interface HConnection extends Abortable, Closeable { /** * @return Configuration instance being used by this HConnection instance. diff --git a/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java b/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java index 34b6ba8c9da..206292f13ba 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java +++ b/src/main/java/org/apache/hadoop/hbase/client/HConnectionManager.java @@ -47,6 +47,8 @@ import java.util.concurrent.atomic.AtomicInteger; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -131,6 +133,8 @@ import org.apache.zookeeper.KeeperException; * cleanup to the client. */ @SuppressWarnings("serial") +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HConnectionManager { // An LRU Map of HConnectionKey -> HConnection (TableServer). All // access must be synchronized. This map is not private because tests diff --git a/src/main/java/org/apache/hadoop/hbase/client/HTable.java b/src/main/java/org/apache/hadoop/hbase/client/HTable.java index 29b8004b943..b30c9a2ae8c 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/HTable.java +++ b/src/main/java/org/apache/hadoop/hbase/client/HTable.java @@ -39,6 +39,8 @@ import java.util.concurrent.atomic.AtomicInteger; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HConstants; @@ -99,6 +101,8 @@ import org.apache.hadoop.hbase.util.Writables; * @see HConnection * @see HConnectionManager */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class HTable implements HTableInterface { private static final Log LOG = LogFactory.getLog(HTable.class); private HConnection connection; diff --git a/src/main/java/org/apache/hadoop/hbase/client/HTableFactory.java b/src/main/java/org/apache/hadoop/hbase/client/HTableFactory.java index 90f6cb9450f..0cb6cc837f4 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/HTableFactory.java +++ b/src/main/java/org/apache/hadoop/hbase/client/HTableFactory.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import java.io.IOException; @@ -28,6 +30,8 @@ import java.io.IOException; * * @since 0.21.0 */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class HTableFactory implements HTableInterfaceFactory { @Override public HTableInterface createHTableInterface(Configuration config, diff --git a/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java b/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java index 9b3c68b439d..04150adf7af 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java +++ b/src/main/java/org/apache/hadoop/hbase/client/HTableInterface.java @@ -23,6 +23,8 @@ import java.io.Closeable; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.KeyValue; @@ -36,6 +38,8 @@ import java.util.Map; * * @since 0.21.0 */ +@InterfaceAudience.Public +@InterfaceStability.Stable public interface HTableInterface extends Closeable { /** diff --git a/src/main/java/org/apache/hadoop/hbase/client/HTableInterfaceFactory.java b/src/main/java/org/apache/hadoop/hbase/client/HTableInterfaceFactory.java index ab7efcc5529..5c9c061573a 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/HTableInterfaceFactory.java +++ b/src/main/java/org/apache/hadoop/hbase/client/HTableInterfaceFactory.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.client; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; @@ -29,6 +31,8 @@ import org.apache.hadoop.conf.Configuration; * * @since 0.21.0 */ +@InterfaceAudience.Public +@InterfaceStability.Stable public interface HTableInterfaceFactory { /** diff --git a/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java b/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java index a5c198fc2c6..47381f4a26f 100755 --- a/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java +++ b/src/main/java/org/apache/hadoop/hbase/client/HTablePool.java @@ -25,6 +25,8 @@ import java.util.Collection; import java.util.List; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HTableDescriptor; @@ -56,6 +58,8 @@ import org.apache.hadoop.hbase.util.PoolMap.PoolType; * Pool will manage its own connections to the cluster. See * {@link HConnectionManager}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class HTablePool implements Closeable { private final PoolMap tables; private final int maxSize; diff --git a/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java b/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java index bc0872af804..e917612bcdc 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java +++ b/src/main/java/org/apache/hadoop/hbase/client/HTableUtil.java @@ -25,6 +25,9 @@ import java.util.ArrayList; import java.util.HashMap; import java.util.List; import java.util.Map; + +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HRegionLocation; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Put; @@ -35,6 +38,8 @@ import org.apache.hadoop.hbase.client.Row; * * */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class HTableUtil { private static final int INITIAL_LIST_SIZE = 250; diff --git a/src/main/java/org/apache/hadoop/hbase/client/Increment.java b/src/main/java/org/apache/hadoop/hbase/client/Increment.java index ef0b65458fd..6ed314ff44f 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Increment.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Increment.java @@ -27,6 +27,8 @@ import java.util.NavigableMap; import java.util.Set; import java.util.TreeMap; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.io.TimeRange; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.Writable; @@ -43,6 +45,8 @@ import org.apache.hadoop.io.Writable; * to increment. At least one column to increment must be specified using the * {@link #addColumn(byte[], byte[], long)} method. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Increment implements Row { private static final byte INCREMENT_VERSION = (byte)2; diff --git a/src/main/java/org/apache/hadoop/hbase/client/IsolationLevel.java b/src/main/java/org/apache/hadoop/hbase/client/IsolationLevel.java index 2631b464961..fb910d84580 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/IsolationLevel.java +++ b/src/main/java/org/apache/hadoop/hbase/client/IsolationLevel.java @@ -20,6 +20,9 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Specify Isolation levels in Scan operations. *

    @@ -29,6 +32,8 @@ package org.apache.hadoop.hbase.client; * should return data that is being modified by transactions that might * not have been committed yet. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public enum IsolationLevel { READ_COMMITTED(1), diff --git a/src/main/java/org/apache/hadoop/hbase/client/MetaScanner.java b/src/main/java/org/apache/hadoop/hbase/client/MetaScanner.java index 2e4aee5c412..f404999110b 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/MetaScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/client/MetaScanner.java @@ -28,6 +28,8 @@ import java.util.TreeMap; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HRegionInfo; @@ -45,6 +47,8 @@ import org.apache.hadoop.hbase.util.Writables; * Although public visibility, this is not a public-facing API and may evolve in * minor releases. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class MetaScanner { private static final Log LOG = LogFactory.getLog(MetaScanner.class); /** diff --git a/src/main/java/org/apache/hadoop/hbase/client/MultiAction.java b/src/main/java/org/apache/hadoop/hbase/client/MultiAction.java index 6a864c88be9..e59f3abe67c 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/MultiAction.java +++ b/src/main/java/org/apache/hadoop/hbase/client/MultiAction.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.client; import org.apache.hadoop.io.Writable; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.io.HbaseObjectWritable; import org.apache.hadoop.hbase.util.Bytes; @@ -36,6 +38,8 @@ import java.util.TreeMap; * Container for Actions (i.e. Get, Delete, or Put), which are grouped by * regionName. Intended to be used with HConnectionManager.processBatch() */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public final class MultiAction implements Writable { // map of regions to lists of puts/gets/deletes for that region. diff --git a/src/main/java/org/apache/hadoop/hbase/client/MultiPut.java b/src/main/java/org/apache/hadoop/hbase/client/MultiPut.java index 9235e2dabad..5ef2ddb876d 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/MultiPut.java +++ b/src/main/java/org/apache/hadoop/hbase/client/MultiPut.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HServerAddress; import org.apache.hadoop.hbase.util.Bytes; @@ -42,6 +44,8 @@ import java.util.TreeSet; * @deprecated Use MultiAction instead * Data type class for putting multiple regions worth of puts in one RPC. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class MultiPut extends Operation implements Writable { public HServerAddress address; // client code ONLY diff --git a/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java b/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java index 290e4c71f8d..8887329e7d3 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java +++ b/src/main/java/org/apache/hadoop/hbase/client/MultiResponse.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.client; import org.apache.hadoop.io.Writable; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.io.HbaseObjectWritable; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Pair; @@ -42,6 +44,8 @@ import java.util.TreeMap; /** * A container for Result objects, grouped by regionName. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class MultiResponse implements Writable { // map of regionName to list of (Results paired to the original index for that diff --git a/src/main/java/org/apache/hadoop/hbase/client/Mutation.java b/src/main/java/org/apache/hadoop/hbase/client/Mutation.java index c246470ed38..0958a691f9f 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Mutation.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Mutation.java @@ -27,10 +27,14 @@ import java.util.Map; import java.util.TreeMap; import java.util.UUID; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class Mutation extends OperationWithAttributes { // Attribute used in Mutations to indicate the originating cluster. private static final String CLUSTER_ID_ATTR = "_c.id_"; diff --git a/src/main/java/org/apache/hadoop/hbase/client/NoServerForRegionException.java b/src/main/java/org/apache/hadoop/hbase/client/NoServerForRegionException.java index 4f339144367..c9fedfb4f0d 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/NoServerForRegionException.java +++ b/src/main/java/org/apache/hadoop/hbase/client/NoServerForRegionException.java @@ -19,11 +19,15 @@ */ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.RegionException; /** * Thrown when no region server can be found for a region */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class NoServerForRegionException extends RegionException { private static final long serialVersionUID = 1L << 11 - 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/client/Operation.java b/src/main/java/org/apache/hadoop/hbase/client/Operation.java index dedd2e26ed4..716c772d498 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Operation.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Operation.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.client; import java.io.IOException; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.codehaus.jackson.map.ObjectMapper; /** @@ -29,6 +31,8 @@ import org.codehaus.jackson.map.ObjectMapper; * (e.g. Put, Get, Delete, Scan, Next, etc.) * Contains methods for exposure to logging and debugging tools. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class Operation { // TODO make this configurable private static final int DEFAULT_MAX_COLS = 5; diff --git a/src/main/java/org/apache/hadoop/hbase/client/OperationWithAttributes.java b/src/main/java/org/apache/hadoop/hbase/client/OperationWithAttributes.java index 698b31a5bfd..92abc287922 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/OperationWithAttributes.java +++ b/src/main/java/org/apache/hadoop/hbase/client/OperationWithAttributes.java @@ -27,10 +27,14 @@ import java.util.Collections; import java.util.HashMap; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.ClassSize; import org.apache.hadoop.io.WritableUtils; +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class OperationWithAttributes extends Operation implements Attributes { // a opaque blob of attributes private Map attributes; diff --git a/src/main/java/org/apache/hadoop/hbase/client/Put.java b/src/main/java/org/apache/hadoop/hbase/client/Put.java index c09b339d598..9ff267fedcb 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Put.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Put.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.io.HeapSize; @@ -43,6 +45,8 @@ import java.util.TreeMap; * for each column to be inserted, execute {@link #add(byte[], byte[], byte[]) add} or * {@link #add(byte[], byte[], long, byte[]) add} if setting the timestamp. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Put extends Mutation implements HeapSize, Writable, Row, Comparable { private static final byte PUT_VERSION = (byte)2; diff --git a/src/main/java/org/apache/hadoop/hbase/client/RegionOfflineException.java b/src/main/java/org/apache/hadoop/hbase/client/RegionOfflineException.java index d223860054d..0e47cade16d 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/RegionOfflineException.java +++ b/src/main/java/org/apache/hadoop/hbase/client/RegionOfflineException.java @@ -19,9 +19,13 @@ */ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.RegionException; /** Thrown when a table can not be located */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class RegionOfflineException extends RegionException { private static final long serialVersionUID = 466008402L; /** default constructor */ diff --git a/src/main/java/org/apache/hadoop/hbase/client/Result.java b/src/main/java/org/apache/hadoop/hbase/client/Result.java index 57f5ecb599e..df0b3ef7967 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Result.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Result.java @@ -31,6 +31,8 @@ import java.util.Map; import java.util.NavigableMap; import java.util.TreeMap; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue.SplitKeyValue; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; @@ -65,6 +67,8 @@ import org.apache.hadoop.io.Writable; * through {@link KeyValue#getRow()}, {@link KeyValue#getFamily()}, {@link KeyValue#getQualifier()}, * {@link KeyValue#getTimestamp()}, and {@link KeyValue#getValue()}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Result implements Writable, WritableWithSize { private static final byte RESULT_VERSION = (byte)1; diff --git a/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java b/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java index 684301893e4..064b0ad536a 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/client/ResultScanner.java @@ -22,10 +22,15 @@ package org.apache.hadoop.hbase.client; import java.io.Closeable; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Interface for client-side scanning. * Go to {@link HTable} to obtain instances. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public interface ResultScanner extends Closeable, Iterable { /** diff --git a/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedException.java b/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedException.java index b9042f623f6..4b06cf6bd75 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedException.java +++ b/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedException.java @@ -19,10 +19,15 @@ import java.io.IOException; import java.util.Date; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Exception thrown by HTable methods when an attempt to do something (like * commit changes) fails after a bunch of retries. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class RetriesExhaustedException extends IOException { private static final long serialVersionUID = 1876775844L; diff --git a/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java b/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java index 4a98c5e17d6..1fde5ac513b 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java +++ b/src/main/java/org/apache/hadoop/hbase/client/RetriesExhaustedWithDetailsException.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HServerAddress; import org.apache.hadoop.hbase.regionserver.NoSuchColumnFamilyException; @@ -42,6 +44,8 @@ import java.util.Set; * {@link #getCause(int)}, {@link #getRow(int)} and {@link #getAddress(int)}. */ @SuppressWarnings("serial") +@InterfaceAudience.Public +@InterfaceStability.Stable public class RetriesExhaustedWithDetailsException extends RetriesExhaustedException { List exceptions; diff --git a/src/main/java/org/apache/hadoop/hbase/client/Row.java b/src/main/java/org/apache/hadoop/hbase/client/Row.java index cd332bd57f4..64d032cae84 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Row.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Row.java @@ -19,11 +19,15 @@ */ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.io.WritableComparable; /** * Has a row. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public interface Row extends WritableComparable { /** * @return The row. diff --git a/src/main/java/org/apache/hadoop/hbase/client/RowLock.java b/src/main/java/org/apache/hadoop/hbase/client/RowLock.java index 56b0787b24d..24edc8b51e8 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/RowLock.java +++ b/src/main/java/org/apache/hadoop/hbase/client/RowLock.java @@ -19,9 +19,14 @@ */ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Holds row name and lock id. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class RowLock { private byte [] row = null; private long lockId = -1L; diff --git a/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java b/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java index 228c7981e14..ad277ac7047 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java +++ b/src/main/java/org/apache/hadoop/hbase/client/RowMutations.java @@ -25,6 +25,8 @@ import java.util.Arrays; import java.util.Collections; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.io.HbaseObjectWritable; import org.apache.hadoop.hbase.util.Bytes; @@ -36,6 +38,8 @@ import org.apache.hadoop.hbase.util.Bytes; * The mutations are performed in the order in which they * were added. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class RowMutations implements Row { private List mutations = new ArrayList(); private byte [] row; diff --git a/src/main/java/org/apache/hadoop/hbase/client/Scan.java b/src/main/java/org/apache/hadoop/hbase/client/Scan.java index cc4c6e8d610..66f84bf4b56 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/Scan.java +++ b/src/main/java/org/apache/hadoop/hbase/client/Scan.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.filter.Filter; @@ -80,6 +82,8 @@ import java.util.TreeSet; * Expert: To explicitly disable server-side block caching for this scan, * execute {@link #setCacheBlocks(boolean)}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Scan extends OperationWithAttributes implements Writable { private static final String RAW_ATTR = "_raw_"; private static final String ISOLATION_LEVEL = "_isolationlevel_"; diff --git a/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java b/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java index 4a4679c1435..9903df39451 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java +++ b/src/main/java/org/apache/hadoop/hbase/client/ScannerCallable.java @@ -23,6 +23,8 @@ import java.net.UnknownHostException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.client.metrics.ScanMetrics; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HRegionInfo; @@ -36,6 +38,8 @@ import org.apache.hadoop.net.DNS; * Retries scanner operations such as create, next, etc. * Used by {@link ResultScanner}s made by {@link HTable}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ScannerCallable extends ServerCallable { private static final Log LOG = LogFactory.getLog(ScannerCallable.class); private long scannerId = -1L; diff --git a/src/main/java/org/apache/hadoop/hbase/client/ScannerTimeoutException.java b/src/main/java/org/apache/hadoop/hbase/client/ScannerTimeoutException.java index 5a10b0ed116..f72cc84b9cd 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/ScannerTimeoutException.java +++ b/src/main/java/org/apache/hadoop/hbase/client/ScannerTimeoutException.java @@ -20,11 +20,15 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.DoNotRetryIOException; /** * Thrown when a scanner has timed out. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ScannerTimeoutException extends DoNotRetryIOException { private static final long serialVersionUID = 8788838690290688313L; diff --git a/src/main/java/org/apache/hadoop/hbase/client/ServerCallable.java b/src/main/java/org/apache/hadoop/hbase/client/ServerCallable.java index fcf43e755a2..ddcf9ad6dcd 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/ServerCallable.java +++ b/src/main/java/org/apache/hadoop/hbase/client/ServerCallable.java @@ -28,6 +28,8 @@ import java.util.ArrayList; import java.util.List; import java.util.concurrent.Callable; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HConstants; @@ -48,6 +50,8 @@ import org.apache.hadoop.ipc.RemoteException; * @see HConnection#getRegionServerWithoutRetries(ServerCallable) * @param the class that the ServerCallable handles */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class ServerCallable implements Callable { protected final HConnection connection; protected final byte [] tableName; diff --git a/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHColumnDescriptor.java b/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHColumnDescriptor.java index 301ea12b52b..485e811cc40 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHColumnDescriptor.java +++ b/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHColumnDescriptor.java @@ -20,12 +20,16 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.io.hfile.Compression; /** * Immutable HColumnDescriptor */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class UnmodifyableHColumnDescriptor extends HColumnDescriptor { /** diff --git a/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.java b/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.java index 412f770ac03..9b9e904b8fe 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.java +++ b/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHRegionInfo.java @@ -20,8 +20,12 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HRegionInfo; +@InterfaceAudience.Public +@InterfaceStability.Evolving class UnmodifyableHRegionInfo extends HRegionInfo { /* * Creates an unmodifyable copy of an HRegionInfo diff --git a/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java b/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java index 27d1faa4c5f..49e814b732c 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java +++ b/src/main/java/org/apache/hadoop/hbase/client/UnmodifyableHTableDescriptor.java @@ -20,12 +20,16 @@ package org.apache.hadoop.hbase.client; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; /** * Read-only table descriptor. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class UnmodifyableHTableDescriptor extends HTableDescriptor { /** Default constructor */ public UnmodifyableHTableDescriptor() { diff --git a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java index 865ffb1f444..0314cf8d976 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java +++ b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/AggregationClient.java @@ -31,6 +31,8 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; @@ -62,6 +64,8 @@ import org.apache.hadoop.hbase.util.Pair; * parameter type. For average and std, it returns a double value. For row * count, it returns a long value. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class AggregationClient { private static final Log log = LogFactory.getLog(AggregationClient.class); diff --git a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java index d430a389338..aad9ad947c9 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java +++ b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java @@ -23,6 +23,8 @@ package org.apache.hadoop.hbase.client.coprocessor; import org.apache.commons.lang.reflect.MethodUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.ipc.CoprocessorProtocol; import java.io.IOException; @@ -36,6 +38,8 @@ import java.lang.reflect.Proxy; * A collection of interfaces and utilities used for interacting with custom RPC * interfaces exposed by Coprocessors. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class Batch { private static Log LOG = LogFactory.getLog(Batch.class); diff --git a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Exec.java b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Exec.java index 75535b5efc4..4d437757345 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Exec.java +++ b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Exec.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.client.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Row; @@ -51,6 +53,8 @@ import java.lang.reflect.Method; * @see org.apache.hadoop.hbase.client.HTable#coprocessorExec(Class, byte[], byte[], org.apache.hadoop.hbase.client.coprocessor.Batch.Call) * @see org.apache.hadoop.hbase.client.HTable#coprocessorExec(Class, byte[], byte[], org.apache.hadoop.hbase.client.coprocessor.Batch.Call, org.apache.hadoop.hbase.client.coprocessor.Batch.Callback) */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class Exec extends Invocation implements Row { /** Row key used as a reference for any region lookups */ private byte[] referenceRow; diff --git a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/ExecResult.java b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/ExecResult.java index 065b3f22463..3046994be07 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/ExecResult.java +++ b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/ExecResult.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.client.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.io.HbaseObjectWritable; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Classes; @@ -46,6 +48,8 @@ import java.io.Serializable; * @see org.apache.hadoop.hbase.client.HTable#coprocessorExec(Class, byte[], byte[], org.apache.hadoop.hbase.client.coprocessor.Batch.Call) * @see org.apache.hadoop.hbase.client.HTable#coprocessorExec(Class, byte[], byte[], org.apache.hadoop.hbase.client.coprocessor.Batch.Call, org.apache.hadoop.hbase.client.coprocessor.Batch.Callback) */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ExecResult implements Writable { private byte[] regionName; private Object value; diff --git a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java index c37b5fde8de..485a1a89484 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java +++ b/src/main/java/org/apache/hadoop/hbase/client/coprocessor/LongColumnInterpreter.java @@ -23,6 +23,8 @@ import java.io.DataInput; import java.io.DataOutput; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.coprocessor.ColumnInterpreter; import org.apache.hadoop.hbase.util.Bytes; @@ -35,6 +37,8 @@ import org.apache.hadoop.hbase.util.Bytes; * TestAggregateProtocol methods for its sample usage. * Its methods handle null arguments gracefully. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class LongColumnInterpreter implements ColumnInterpreter { public Long getValue(byte[] colFamily, byte[] colQualifier, KeyValue kv) diff --git a/src/main/java/org/apache/hadoop/hbase/client/metrics/ScanMetrics.java b/src/main/java/org/apache/hadoop/hbase/client/metrics/ScanMetrics.java index c5faafb4578..57f2af239bf 100644 --- a/src/main/java/org/apache/hadoop/hbase/client/metrics/ScanMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/client/metrics/ScanMetrics.java @@ -26,6 +26,8 @@ import java.util.Collection; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.io.Writable; import org.apache.hadoop.metrics.util.MetricsBase; @@ -45,6 +47,8 @@ import org.apache.hadoop.metrics.util.MetricsTimeVaryingLong; * However, there is no need for this. So they are defined under scan operation * for now. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ScanMetrics implements Writable { private static final byte SCANMETRICS_VERSION = (byte)1; diff --git a/src/main/java/org/apache/hadoop/hbase/constraint/BaseConstraint.java b/src/main/java/org/apache/hadoop/hbase/constraint/BaseConstraint.java index 974738ccac6..1fd6753b9ff 100644 --- a/src/main/java/org/apache/hadoop/hbase/constraint/BaseConstraint.java +++ b/src/main/java/org/apache/hadoop/hbase/constraint/BaseConstraint.java @@ -17,12 +17,14 @@ */ package org.apache.hadoop.hbase.constraint; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configured; /** * Base class to use when actually implementing a {@link Constraint}. It takes * care of getting and setting of configuration for the constraint. */ +@InterfaceAudience.Private public abstract class BaseConstraint extends Configured implements Constraint { } diff --git a/src/main/java/org/apache/hadoop/hbase/constraint/Constraint.java b/src/main/java/org/apache/hadoop/hbase/constraint/Constraint.java index 5a55d1a4d66..4253296ae44 100644 --- a/src/main/java/org/apache/hadoop/hbase/constraint/Constraint.java +++ b/src/main/java/org/apache/hadoop/hbase/constraint/Constraint.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hbase.constraint; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.client.Put; @@ -61,6 +62,7 @@ import org.apache.hadoop.hbase.client.Put; * @see BaseConstraint * @see Constraints */ +@InterfaceAudience.Private public interface Constraint extends Configurable { /** diff --git a/src/main/java/org/apache/hadoop/hbase/constraint/ConstraintException.java b/src/main/java/org/apache/hadoop/hbase/constraint/ConstraintException.java index 1831421e3dd..754b1c53896 100644 --- a/src/main/java/org/apache/hadoop/hbase/constraint/ConstraintException.java +++ b/src/main/java/org/apache/hadoop/hbase/constraint/ConstraintException.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hbase.constraint; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.client.Put; @@ -27,6 +28,7 @@ import org.apache.hadoop.hbase.client.Put; * should fail every time for the same {@link Put} (it should be * idempotent). */ +@InterfaceAudience.Private public class ConstraintException extends DoNotRetryIOException { private static final long serialVersionUID = 1197446454511704140L; diff --git a/src/main/java/org/apache/hadoop/hbase/constraint/ConstraintProcessor.java b/src/main/java/org/apache/hadoop/hbase/constraint/ConstraintProcessor.java index 522369c0926..ee26a5e95a9 100644 --- a/src/main/java/org/apache/hadoop/hbase/constraint/ConstraintProcessor.java +++ b/src/main/java/org/apache/hadoop/hbase/constraint/ConstraintProcessor.java @@ -23,6 +23,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.CoprocessorEnvironment; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.client.Put; @@ -37,6 +38,7 @@ import org.apache.hadoop.hbase.regionserver.wal.WALEdit; * This is an ease of use mechanism - all the functionality here could be * implemented on any given system by a coprocessor. */ +@InterfaceAudience.Private public class ConstraintProcessor extends BaseRegionObserver { private static final Log LOG = LogFactory.getLog(ConstraintProcessor.class); diff --git a/src/main/java/org/apache/hadoop/hbase/constraint/Constraints.java b/src/main/java/org/apache/hadoop/hbase/constraint/Constraints.java index c78d894f6b8..d207953f0a7 100644 --- a/src/main/java/org/apache/hadoop/hbase/constraint/Constraints.java +++ b/src/main/java/org/apache/hadoop/hbase/constraint/Constraints.java @@ -31,6 +31,7 @@ import java.util.regex.Pattern; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; @@ -45,6 +46,7 @@ import org.apache.hadoop.hbase.util.Pair; * NOTE: this class is NOT thread safe. Concurrent setting/enabling/disabling of * constraints can cause constraints to be run at incorrect times or not at all. */ +@InterfaceAudience.Private public final class Constraints { private static final int DEFAULT_PRIORITY = -1; diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java index 579aa5505e1..6d23f81852f 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java @@ -26,6 +26,8 @@ import java.util.NavigableSet; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter; @@ -37,6 +39,8 @@ import org.apache.hadoop.hbase.util.Pair; * A concrete AggregateProtocol implementation. Its system level coprocessor * that computes the aggregate function at a region level. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class AggregateImplementation extends BaseEndpointCoprocessor implements AggregateProtocol { protected static Log log = LogFactory.getLog(AggregateImplementation.class); diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java index e654c0cb991..ef1b8a6ff8c 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateProtocol.java @@ -23,6 +23,8 @@ package org.apache.hadoop.hbase.coprocessor; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.client.coprocessor.AggregationClient; import org.apache.hadoop.hbase.ipc.CoprocessorProtocol; @@ -38,6 +40,8 @@ import org.apache.hadoop.hbase.util.Pair; * be picked. Refer to {@link AggregationClient} for some general conditions on * input parameters. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface AggregateProtocol extends CoprocessorProtocol { public static final long VERSION = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEndpointCoprocessor.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEndpointCoprocessor.java index 3a787fdf55d..28396df6b4c 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEndpointCoprocessor.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseEndpointCoprocessor.java @@ -18,6 +18,8 @@ package org.apache.hadoop.hbase.coprocessor; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Coprocessor; import org.apache.hadoop.hbase.CoprocessorEnvironment; import org.apache.hadoop.hbase.ipc.CoprocessorProtocol; @@ -34,6 +36,8 @@ import org.apache.hadoop.hbase.ipc.VersionedProtocol; * and Coprocessor to develop an Endpoint. But you won't be able to access * the region related resource, i.e., CoprocessorEnvironment. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class BaseEndpointCoprocessor implements Coprocessor, CoprocessorProtocol, VersionedProtocol { /** diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java index 428feb14b9e..3851a00cb3b 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseMasterObserver.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.CoprocessorEnvironment; @@ -29,6 +31,8 @@ import org.apache.hadoop.hbase.UnknownRegionException; import java.io.IOException; +@InterfaceAudience.Public +@InterfaceStability.Evolving public class BaseMasterObserver implements MasterObserver { @Override public void preCreateTable(ObserverContext ctx, diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java index 7a7b8962168..197eb7119d6 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRegionObserver.java @@ -20,6 +20,9 @@ import java.util.List; import java.util.Map; import com.google.common.collect.ImmutableList; + +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.CoprocessorEnvironment; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.KeyValue; @@ -47,6 +50,8 @@ import java.io.IOException; * By extending it, you can create your own region observer without * overriding all abstract methods of RegionObserver. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class BaseRegionObserver implements RegionObserver { @Override public void start(CoprocessorEnvironment e) throws IOException { } diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java index 149fe213464..ebdca93c5d7 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/ColumnInterpreter.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.coprocessor; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.coprocessor.LongColumnInterpreter; import org.apache.hadoop.io.Writable; @@ -45,6 +47,8 @@ import org.apache.hadoop.io.Writable; * @param Cell value data type * @param Promoted data type */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface ColumnInterpreter extends Writable { /** diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorException.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorException.java index d4344d0bc6f..fd27862cdd2 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorException.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorException.java @@ -19,11 +19,15 @@ */ package org.apache.hadoop.hbase.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.DoNotRetryIOException; /** * Thrown if a coprocessor encounters any exception. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class CoprocessorException extends DoNotRetryIOException { private static final long serialVersionUID = 4357922136679804887L; diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java index b6412ca67b7..18c13c4964f 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.coprocessor; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -50,6 +52,8 @@ import java.util.*; * @param the specific environment extension that a concrete implementation * provides */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class CoprocessorHost { public static final String REGION_COPROCESSOR_CONF_KEY = "hbase.coprocessor.region.classes"; diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java index 75f06536d0f..140962e7986 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterCoprocessorEnvironment.java @@ -20,9 +20,13 @@ package org.apache.hadoop.hbase.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.CoprocessorEnvironment; import org.apache.hadoop.hbase.master.MasterServices; +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface MasterCoprocessorEnvironment extends CoprocessorEnvironment { /** @return reference to the HMaster services */ MasterServices getMasterServices(); diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java index a36eb1e6bc4..d076e5002cc 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/MasterObserver.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.*; import java.io.IOException; @@ -28,6 +30,8 @@ import java.io.IOException; * Defines coprocessor hooks for interacting with operations on the * {@link org.apache.hadoop.hbase.master.HMaster} process. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface MasterObserver extends Coprocessor { /** diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java index c640eaea390..607764c41ac 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java @@ -22,6 +22,8 @@ import java.util.List; import java.util.SortedSet; import java.util.TreeSet; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.client.Mutation; @@ -34,6 +36,8 @@ import org.apache.hadoop.hbase.util.Bytes; * {@link HRegion#mutateRowsWithLocks(java.util.Collection, java.util.Collection)} * and Coprocessor endpoints. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class MultiRowMutationEndpoint extends BaseEndpointCoprocessor implements MultiRowMutationProtocol { diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationProtocol.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationProtocol.java index e8eea9f9058..29135821fe0 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationProtocol.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationProtocol.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.coprocessor; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.client.Mutation; import org.apache.hadoop.hbase.ipc.CoprocessorProtocol; import org.apache.hadoop.hbase.regionserver.HRegion; @@ -45,6 +47,8 @@ import org.apache.hadoop.hbase.regionserver.HRegion; * mrOp.mutateRows(mutations); * */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface MultiRowMutationProtocol extends CoprocessorProtocol { public void mutateRows(List mutations) throws IOException; } diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/ObserverContext.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/ObserverContext.java index d5cf6aa0c19..3008aa0a829 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/ObserverContext.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/ObserverContext.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.CoprocessorEnvironment; /** @@ -32,6 +34,8 @@ import org.apache.hadoop.hbase.CoprocessorEnvironment; * @param The {@link CoprocessorEnvironment} subclass applicable to the * revelant Observer interface. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ObserverContext { private E env; private boolean bypass; diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java index 9395a3828a6..eebd02b166d 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionCoprocessorEnvironment.java @@ -20,10 +20,14 @@ package org.apache.hadoop.hbase.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.CoprocessorEnvironment; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.RegionServerServices; +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface RegionCoprocessorEnvironment extends CoprocessorEnvironment { /** @return the region associated with this coprocessor */ public HRegion getRegion(); diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java index c0a418427f4..30c61ca898a 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/RegionObserver.java @@ -19,6 +19,8 @@ package org.apache.hadoop.hbase.coprocessor; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Coprocessor; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.KeyValue; @@ -45,6 +47,8 @@ import com.google.common.collect.ImmutableList; * Coprocessors implement this interface to observe and mediate client actions * on the region. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface RegionObserver extends Coprocessor { /** diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java index 2790abddd46..cee6b2c7e6d 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/WALCoprocessorEnvironment.java @@ -20,9 +20,13 @@ package org.apache.hadoop.hbase.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.CoprocessorEnvironment; import org.apache.hadoop.hbase.regionserver.wal.HLog; +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface WALCoprocessorEnvironment extends CoprocessorEnvironment { /** @return reference to the region server services */ public HLog getWAL(); diff --git a/src/main/java/org/apache/hadoop/hbase/coprocessor/WALObserver.java b/src/main/java/org/apache/hadoop/hbase/coprocessor/WALObserver.java index c90189d3e1e..d3926df5f07 100644 --- a/src/main/java/org/apache/hadoop/hbase/coprocessor/WALObserver.java +++ b/src/main/java/org/apache/hadoop/hbase/coprocessor/WALObserver.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.coprocessor; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Coprocessor; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.regionserver.wal.HLogKey; @@ -37,6 +39,8 @@ import java.io.IOException; * Defines coprocessor hooks for interacting with operations on the * {@link org.apache.hadoop.hbase.regionserver.wal.HLog}. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface WALObserver extends Coprocessor { /** diff --git a/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java b/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java index bdd3cf0d523..62298587bd5 100644 --- a/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/executor/EventHandler.java @@ -24,6 +24,7 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Server; @@ -53,6 +54,7 @@ import org.apache.hadoop.hbase.Server; * {@link #setListener(EventHandlerListener)}. * @see ExecutorService */ +@InterfaceAudience.Private public abstract class EventHandler implements Runnable, Comparable { private static final Log LOG = LogFactory.getLog(EventHandler.class); diff --git a/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java b/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java index 7fb4266ac6e..06b9228bf87 100644 --- a/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java +++ b/src/main/java/org/apache/hadoop/hbase/executor/ExecutorService.java @@ -38,6 +38,7 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.executor.EventHandler.EventHandlerListener; import org.apache.hadoop.hbase.executor.EventHandler.EventType; import org.apache.hadoop.hbase.monitoring.ThreadMonitoring; @@ -61,6 +62,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; * with {@link #registerListener(EventHandler.EventType, EventHandler.EventHandlerListener)}. Be sure * to deregister your listener when done via {@link #unregisterListener(EventHandler.EventType)}. */ +@InterfaceAudience.Private public class ExecutorService { private static final Log LOG = LogFactory.getLog(ExecutorService.class); diff --git a/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java b/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java index 2f5f0924f46..35d7b7044db 100644 --- a/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java +++ b/src/main/java/org/apache/hadoop/hbase/executor/RegionTransitionData.java @@ -23,6 +23,7 @@ import java.io.DataInput; import java.io.DataOutput; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.executor.EventHandler.EventType; import org.apache.hadoop.hbase.util.Bytes; @@ -32,6 +33,7 @@ import org.apache.hadoop.io.Writable; /** * Data serialized into ZooKeeper for region transitions. */ +@InterfaceAudience.Private public class RegionTransitionData implements Writable { /** * Type of transition event (offline, opening, opened, closing, closed). diff --git a/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java b/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java index 1e56948c3c3..654c28273ad 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/BinaryComparator.java @@ -20,12 +20,16 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; /** * A binary comparator which lexicographically compares against the specified * byte array using {@link org.apache.hadoop.hbase.util.Bytes#compareTo(byte[], byte[])}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class BinaryComparator extends WritableByteArrayComparable { /** Nullary constructor for Writable, do not use */ diff --git a/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java b/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java index 030341a307a..de86b7fc884 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/BinaryPrefixComparator.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; /** @@ -27,6 +29,8 @@ import org.apache.hadoop.hbase.util.Bytes; * up to the length of this byte array. For the rest it is similar to * {@link BinaryComparator}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class BinaryPrefixComparator extends WritableByteArrayComparable { /** Nullary constructor for Writable, do not use */ diff --git a/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java b/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java index bc29a42a69f..c74d056899f 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/BitComparator.java @@ -24,10 +24,15 @@ import java.io.DataInput; import java.io.DataOutput; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * A bit comparator which performs the specified bitwise operation on each of the bytes * with the specified byte array. Then returns whether the result is non-zero. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class BitComparator extends WritableByteArrayComparable { /** Nullary constructor for Writable, do not use */ diff --git a/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java index 26f11aa029f..52bc35a0d63 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/ColumnCountGetFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import java.io.DataInput; @@ -35,6 +37,8 @@ import com.google.common.base.Preconditions; * its quota of columns, {@link #filterAllRemaining()} returns true. This * makes this filter unsuitable as a Scan filter. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ColumnCountGetFilter extends FilterBase { private int limit = 0; private int count = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java index 85b0af77673..9143c13094b 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/ColumnPaginationFilter.java @@ -24,6 +24,8 @@ import java.io.DataOutput; import java.io.IOException; import java.util.ArrayList; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import com.google.common.base.Preconditions; @@ -32,6 +34,8 @@ import com.google.common.base.Preconditions; * This filter can be used for row-based indexing, where references to other tables are stored across many columns, * in order to efficient lookups and paginated results for end users. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ColumnPaginationFilter extends FilterBase { private int limit = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java index 931c7ad8d16..f0eba021964 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/ColumnPrefixFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; @@ -35,6 +37,8 @@ import com.google.common.base.Preconditions; * a particular prefix. For example, if prefix is 'an', it will pass keys with * columns like 'and', 'anti' but not keys with columns like 'ball', 'act'. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ColumnPrefixFilter extends FilterBase { protected byte [] prefix = null; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java index df8786c7b61..941cc7d9223 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/ColumnRangeFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; @@ -42,6 +44,8 @@ import com.google.common.base.Preconditions; * minColumnInclusive and maxColumnInclusive specify if the ranges are inclusive * or not. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ColumnRangeFilter extends FilterBase { protected byte[] minColumn = null; protected boolean minColumnInclusive = true; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java index 37090602966..3207ae4bd40 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/CompareFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.io.HbaseObjectWritable; import java.io.DataInput; @@ -43,6 +45,8 @@ import com.google.common.base.Preconditions; *

    * Multiple filters can be combined using {@link FilterList}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class CompareFilter extends FilterBase { /** Comparison operators. */ diff --git a/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java index 785bd9e9f6b..46c20e66988 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/DependentColumnFilter.java @@ -28,6 +28,8 @@ import java.util.List; import java.util.Set; import java.util.ArrayList; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; @@ -40,6 +42,8 @@ import com.google.common.base.Preconditions; * Not compatible with Scan.setBatch as operations need * full rows for correct filtering */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class DependentColumnFilter extends CompareFilter { protected byte[] columnFamily; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java index 63ec44a6a27..ff133be145f 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/FamilyFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import java.util.ArrayList; @@ -37,6 +39,8 @@ import java.util.ArrayList; * If an already known column family is looked for, use {@link org.apache.hadoop.hbase.client.Get#addFamily(byte[])} * directly rather than a filter. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class FamilyFilter extends CompareFilter { /** * Writable constructor, do not use. diff --git a/src/main/java/org/apache/hadoop/hbase/filter/Filter.java b/src/main/java/org/apache/hadoop/hbase/filter/Filter.java index 02ea5f5c97c..fa7a6ffa017 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/Filter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/Filter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.io.Writable; @@ -48,6 +50,8 @@ import java.util.List; * * @see FilterBase */ +@InterfaceAudience.Public +@InterfaceStability.Stable public interface Filter extends Writable { /** * Reset the state of the filter between rows. diff --git a/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java b/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java index 0d1b12339c1..adb674081f6 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/FilterBase.java @@ -19,6 +19,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import java.util.List; @@ -32,6 +34,8 @@ import java.util.ArrayList; * If you could instantiate FilterBase, it would end up being a "null" filter - * that is one that never filters anything. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class FilterBase implements Filter { /** diff --git a/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java b/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java index 216d0dbdc1b..e69d85fca35 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/FilterList.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; @@ -41,6 +43,8 @@ import java.util.List; * Defaults to {@link Operator#MUST_PASS_ALL}. *

    TODO: Fix creation of Configuration on serialization and deserialization. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class FilterList implements Filter { /** set operator */ public static enum Operator { diff --git a/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java index 7a068b4d981..f67bd8b8091 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/FirstKeyOnlyFilter.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import java.io.DataOutput; @@ -34,6 +36,8 @@ import com.google.common.base.Preconditions; *

    * This filter can be used to more efficiently perform row count operations. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class FirstKeyOnlyFilter extends FilterBase { private boolean foundKV = false; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java index 091800c8ebc..98d17ba9b51 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/InclusiveStopFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; @@ -37,6 +39,8 @@ import com.google.common.base.Preconditions; * * Use this filter to include the stop row, eg: [A,Z]. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class InclusiveStopFilter extends FilterBase { private byte [] stopRowKey; private boolean done = false; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/IncompatibleFilterException.java b/src/main/java/org/apache/hadoop/hbase/filter/IncompatibleFilterException.java index 75edf192cfb..0336f7ba513 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/IncompatibleFilterException.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/IncompatibleFilterException.java @@ -19,9 +19,14 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Used to indicate a filter incompatibility */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class IncompatibleFilterException extends RuntimeException { private static final long serialVersionUID = 3236763276623198231L; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/InvalidRowFilterException.java b/src/main/java/org/apache/hadoop/hbase/filter/InvalidRowFilterException.java index 14b8e31cfd5..1613468d858 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/InvalidRowFilterException.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/InvalidRowFilterException.java @@ -19,9 +19,14 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Used to indicate an invalid RowFilter. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class InvalidRowFilterException extends RuntimeException { private static final long serialVersionUID = 2667894046345657865L; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java index b2eb3a5fffb..54c031b5c82 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/KeyOnlyFilter.java @@ -22,6 +22,9 @@ package org.apache.hadoop.hbase.filter; import java.io.DataInput; import java.io.DataOutput; import java.io.IOException; + +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; @@ -36,6 +39,8 @@ import com.google.common.base.Preconditions; * This filter can be used to grab all of the keys without having to also grab * the values. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class KeyOnlyFilter extends FilterBase { boolean lenAsVal; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java index 233b295c97a..323894fd0fd 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/MultipleColumnPrefixFilter.java @@ -17,6 +17,8 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; @@ -33,6 +35,8 @@ import java.util.ArrayList; * a particular prefix. For example, if prefix is 'an', it will pass keys will * columns like 'and', 'anti' but not keys with columns like 'ball', 'act'. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class MultipleColumnPrefixFilter extends FilterBase { protected byte [] hint = null; protected TreeSet sortedPrefixes = createTreeSet(); diff --git a/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java b/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java index 45eb4778037..4af59655cb6 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/NullComparator.java @@ -20,10 +20,15 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * A binary comparator which lexicographically compares against the specified * byte array using {@link org.apache.hadoop.hbase.util.Bytes#compareTo(byte[], byte[])}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class NullComparator extends WritableByteArrayComparable { /** Nullary constructor for Writable, do not use */ diff --git a/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java index 9d66c757ceb..f663717585d 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/PageFilter.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import java.io.DataInput; @@ -39,6 +41,8 @@ import com.google.common.base.Preconditions; * individual HRegions by making sure that the page size is never exceeded * locally. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class PageFilter extends FilterBase { private long pageSize = Long.MAX_VALUE; private int rowsAccepted = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/ParseConstants.java b/src/main/java/org/apache/hadoop/hbase/filter/ParseConstants.java index 373d7a68fbd..1acbdaba0ad 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/ParseConstants.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/ParseConstants.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.ipc.HRegionInterface; import org.apache.hadoop.hbase.util.Bytes; import java.nio.ByteBuffer; @@ -29,6 +31,8 @@ import org.apache.hadoop.hbase.filter.*; * ParseConstants holds a bunch of constants related to parsing Filter Strings * Used by {@link ParseFilter} */ +@InterfaceAudience.Public +@InterfaceStability.Stable public final class ParseConstants { /** diff --git a/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java index 409d41ef83a..976e0434390 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/ParseFilter.java @@ -27,6 +27,8 @@ import java.util.*; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.util.Bytes; @@ -40,6 +42,8 @@ import org.apache.hadoop.hbase.util.Bytes; * This class addresses the HBASE-4168 JIRA. More documentaton on this * Filter Language can be found at: https://issues.apache.org/jira/browse/HBASE-4176 */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ParseFilter { private static final Log LOG = LogFactory.getLog(ParseFilter.class); diff --git a/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java index 9de199b1c0a..2b4506efe1c 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/PrefixFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; @@ -34,6 +36,8 @@ import com.google.common.base.Preconditions; /** * Pass results that have same row prefix. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class PrefixFilter extends FilterBase { protected byte [] prefix = null; protected boolean passedPrefix = false; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java index cd69277611c..5be86c35f4f 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/QualifierFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Get; @@ -38,6 +40,8 @@ import java.util.ArrayList; * If an already known column qualifier is looked for, use {@link Get#addColumn} * directly rather than a filter. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class QualifierFilter extends CompareFilter { /** diff --git a/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java index c23ac9b2fcf..94c3f7e0011 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/RandomRowFilter.java @@ -25,12 +25,16 @@ import java.io.DataOutput; import java.io.IOException; import java.util.Random; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; /** * A filter that includes rows based on a chance. * */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class RandomRowFilter extends FilterBase { protected static final Random random = new Random(); diff --git a/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java b/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java index 210de0d3b86..afd6ceb1530 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/RegexStringComparator.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.util.Bytes; @@ -54,6 +56,8 @@ import java.util.regex.Pattern; * "{3}[\\d]{1,3})?)(\\/[0-9]+)?")); * */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class RegexStringComparator extends WritableByteArrayComparable { private static final Log LOG = LogFactory.getLog(RegexStringComparator.class); diff --git a/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java index 3b00efd2e7e..e12667810f6 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/RowFilter.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.filter; import java.util.ArrayList; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Scan; @@ -37,6 +39,8 @@ import org.apache.hadoop.hbase.client.Scan; * If an already known row range needs to be scanned, use {@link Scan} start * and stop rows directly rather than a filter. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class RowFilter extends CompareFilter { private boolean filterOutRow = false; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java index 7c7607f13da..201b6ff736c 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueExcludeFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; @@ -31,6 +33,8 @@ import java.util.ArrayList; * {@link SingleColumnValueFilter}, if the tested column value is not actually * needed as input (besides for the filtering itself). */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class SingleColumnValueExcludeFilter extends SingleColumnValueFilter { /** diff --git a/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java index 7e4b2aedfaa..0a24ada7fb7 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/SingleColumnValueFilter.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.filter; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; @@ -64,6 +66,8 @@ import com.google.common.base.Preconditions; *

    * To filter based on the value of all scanned columns, use {@link ValueFilter}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class SingleColumnValueFilter extends FilterBase { static final Log LOG = LogFactory.getLog(SingleColumnValueFilter.class); diff --git a/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java index 8be40eece55..c89506105a8 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/SkipFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import java.io.DataInput; @@ -45,6 +47,8 @@ import java.util.List; * Without this filter, the other non-zero valued columns in the row would still * be emitted. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class SkipFilter extends FilterBase { private boolean filterRow = false; private Filter filter; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java b/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java index 90bc718a3de..d64e3212d05 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/SubstringComparator.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; import java.io.DataInput; @@ -40,6 +42,8 @@ import java.io.IOException; * new SubstringComparator("substr")); * */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class SubstringComparator extends WritableByteArrayComparable { private String substr; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java index 551cbab38b6..84eee0df9f2 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/TimestampsFilter.java @@ -25,6 +25,8 @@ import java.util.List; import java.util.TreeSet; import java.util.ArrayList; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import com.google.common.base.Preconditions; @@ -37,6 +39,8 @@ import com.google.common.base.Preconditions; * {@link org.apache.hadoop.hbase.client.Scan#setTimeRange(long, long)}, {@link org.apache.hadoop.hbase.client.Get#setTimeStamp(long)}, * or {@link org.apache.hadoop.hbase.client.Scan#setTimeStamp(long)}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TimestampsFilter extends FilterBase { TreeSet timestamps; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java index 24521296c6e..2f63e23ef03 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/ValueFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import java.util.ArrayList; @@ -37,6 +39,8 @@ import java.util.ArrayList; * To test the value of a single qualifier when scanning multiple qualifiers, * use {@link SingleColumnValueFilter}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ValueFilter extends CompareFilter { /** diff --git a/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java b/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java index b9fa92787ff..eb3f96b001d 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/WhileMatchFilter.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import java.io.DataInput; @@ -35,6 +37,8 @@ import java.util.List; * {@link org.apache.hadoop.hbase.filter.Filter#filterAllRemaining()} methods * returns true. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class WhileMatchFilter extends FilterBase { private boolean filterAllRemaining = false; private Filter filter; diff --git a/src/main/java/org/apache/hadoop/hbase/filter/WritableByteArrayComparable.java b/src/main/java/org/apache/hadoop/hbase/filter/WritableByteArrayComparable.java index c7731e3fde7..851158bc506 100644 --- a/src/main/java/org/apache/hadoop/hbase/filter/WritableByteArrayComparable.java +++ b/src/main/java/org/apache/hadoop/hbase/filter/WritableByteArrayComparable.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.filter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.Writable; @@ -27,6 +29,8 @@ import java.io.DataOutput; import java.io.IOException; /** Base class, combines Comparable and Writable. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class WritableByteArrayComparable implements Writable, Comparable { byte[] value; diff --git a/src/main/java/org/apache/hadoop/hbase/io/CodeToClassAndBack.java b/src/main/java/org/apache/hadoop/hbase/io/CodeToClassAndBack.java index 61a1c5e5b87..a2f26141159 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/CodeToClassAndBack.java +++ b/src/main/java/org/apache/hadoop/hbase/io/CodeToClassAndBack.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.io; import java.util.*; +import org.apache.hadoop.classification.InterfaceAudience; + /** * A Static Interface. * Instead of having this code in the the HbaseMapWritable code, where it @@ -29,6 +31,7 @@ import java.util.*; * it is put here in this static interface where the static final Maps are * loaded one time. Only byte[] and Cell are supported at this time. */ +@InterfaceAudience.Private public interface CodeToClassAndBack { /** * Static map that contains mapping from code to class diff --git a/src/main/java/org/apache/hadoop/hbase/io/DataOutputOutputStream.java b/src/main/java/org/apache/hadoop/hbase/io/DataOutputOutputStream.java index 10700c1d1e7..82123a868ba 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/DataOutputOutputStream.java +++ b/src/main/java/org/apache/hadoop/hbase/io/DataOutputOutputStream.java @@ -22,13 +22,11 @@ import java.io.IOException; import java.io.OutputStream; import org.apache.hadoop.classification.InterfaceAudience; -import org.apache.hadoop.classification.InterfaceStability; /** * OutputStream implementation that wraps a DataOutput. */ @InterfaceAudience.Private -@InterfaceStability.Unstable class DataOutputOutputStream extends OutputStream { private final DataOutput out; diff --git a/src/main/java/org/apache/hadoop/hbase/io/DoubleOutputStream.java b/src/main/java/org/apache/hadoop/hbase/io/DoubleOutputStream.java index cf33cd32e64..9062ed9e0ac 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/DoubleOutputStream.java +++ b/src/main/java/org/apache/hadoop/hbase/io/DoubleOutputStream.java @@ -22,12 +22,15 @@ package org.apache.hadoop.hbase.io; import java.io.IOException; import java.io.OutputStream; +import org.apache.hadoop.classification.InterfaceAudience; + /** * An output stream that writes to two streams on each operation. Does not * attempt to handle exceptions gracefully. If any operation other than * {@link #close()} fails on the first stream, it is not called on the second * stream. */ +@InterfaceAudience.Private public class DoubleOutputStream extends OutputStream { private OutputStream out1; private OutputStream out2; diff --git a/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java b/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java index c2ceb1e2361..3f302208490 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java +++ b/src/main/java/org/apache/hadoop/hbase/io/HalfStoreFileReader.java @@ -24,6 +24,7 @@ import java.nio.ByteBuffer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.KeyValue; @@ -46,6 +47,7 @@ import org.apache.hadoop.hbase.util.Bytes; * *

    This file is not splitable. Calls to {@link #midkey()} return null. */ +@InterfaceAudience.Private public class HalfStoreFileReader extends StoreFile.Reader { final Log LOG = LogFactory.getLog(HalfStoreFileReader.class); final boolean top; diff --git a/src/main/java/org/apache/hadoop/hbase/io/HbaseMapWritable.java b/src/main/java/org/apache/hadoop/hbase/io/HbaseMapWritable.java index 45eb49512f6..deb9b82eb14 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/HbaseMapWritable.java +++ b/src/main/java/org/apache/hadoop/hbase/io/HbaseMapWritable.java @@ -30,6 +30,7 @@ import java.util.SortedMap; import java.util.TreeMap; import java.util.concurrent.atomic.AtomicReference; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.util.Bytes; @@ -45,6 +46,7 @@ import org.apache.hadoop.util.ReflectionUtils; * @param key TODO: Parameter K is never used, could be removed. * @param value Expects a Writable or byte []. */ +@InterfaceAudience.Private public class HbaseMapWritable implements SortedMap, Configurable, Writable, CodeToClassAndBack{ private AtomicReference conf = null; diff --git a/src/main/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java b/src/main/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java index e67a377f27b..cbfa48987c7 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java +++ b/src/main/java/org/apache/hadoop/hbase/io/HbaseObjectWritable.java @@ -38,6 +38,7 @@ import java.util.NavigableSet; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; @@ -115,6 +116,7 @@ import com.google.protobuf.Message; * name and reflection to instantiate class was costing in excess of the cell * handling). */ +@InterfaceAudience.Private public class HbaseObjectWritable implements Writable, WritableWithSize, Configurable { protected final static Log LOG = LogFactory.getLog(HbaseObjectWritable.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/HeapSize.java b/src/main/java/org/apache/hadoop/hbase/io/HeapSize.java index bd788469ab2..f6142b66959 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/HeapSize.java +++ b/src/main/java/org/apache/hadoop/hbase/io/HeapSize.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.io; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Implementations can be asked for an estimate of their size in bytes. *

    @@ -37,6 +39,7 @@ package org.apache.hadoop.hbase.io; * } * */ +@InterfaceAudience.Private public interface HeapSize { /** * @return Approximate 'exclusive deep size' of implementing object. Includes diff --git a/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java b/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java index 0cd521377b6..926f12d5fe0 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java +++ b/src/main/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java @@ -26,6 +26,8 @@ import java.io.DataOutput; import java.util.Arrays; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.io.BytesWritable; import org.apache.hadoop.io.WritableComparable; import org.apache.hadoop.io.WritableComparator; @@ -39,6 +41,8 @@ import org.apache.hadoop.io.WritableComparator; * the underlying byte [] is not copied, just referenced. The backing * buffer is accessed when we go to serialize. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ImmutableBytesWritable implements WritableComparable { private byte[] bytes; diff --git a/src/main/java/org/apache/hadoop/hbase/io/Reference.java b/src/main/java/org/apache/hadoop/hbase/io/Reference.java index 99ecb7e85fb..636005913e4 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/Reference.java +++ b/src/main/java/org/apache/hadoop/hbase/io/Reference.java @@ -23,6 +23,7 @@ import java.io.DataInput; import java.io.DataOutput; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; @@ -48,6 +49,7 @@ import org.apache.hadoop.io.Writable; * Note, a region is itself not splitable if it has instances of store file * references. References are cleaned up by compactions. */ +@InterfaceAudience.Private public class Reference implements Writable { private byte [] splitkey; private Range region; diff --git a/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java b/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java index 12a9b6814d3..d13539360df 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java +++ b/src/main/java/org/apache/hadoop/hbase/io/TimeRange.java @@ -26,6 +26,8 @@ import java.io.IOException; import org.apache.hadoop.io.Writable; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; /** @@ -36,6 +38,8 @@ import org.apache.hadoop.hbase.util.Bytes; *

    * Only used internally; should not be accessed directly by clients. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TimeRange implements Writable { private long minStamp = 0L; private long maxStamp = Long.MAX_VALUE; diff --git a/src/main/java/org/apache/hadoop/hbase/io/WritableWithSize.java b/src/main/java/org/apache/hadoop/hbase/io/WritableWithSize.java index f8aefa10125..bd94043b7b3 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/WritableWithSize.java +++ b/src/main/java/org/apache/hadoop/hbase/io/WritableWithSize.java @@ -20,9 +20,12 @@ package org.apache.hadoop.hbase.io; +import org.apache.hadoop.classification.InterfaceAudience; + /** * An optional interface to 'size' writables. */ +@InterfaceAudience.Private public interface WritableWithSize { /** * Provide a size hint to the caller. write() should ideally diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java index 1dedbbdc59f..ce77e96d963 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/BufferedDataBlockEncoder.java @@ -21,6 +21,7 @@ import java.io.DataOutputStream; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue.SamePrefixComparator; import org.apache.hadoop.hbase.util.ByteBufferUtils; @@ -31,6 +32,7 @@ import org.apache.hadoop.io.WritableUtils; /** * Base class for all data block encoders that use a buffer. */ +@InterfaceAudience.Private abstract class BufferedDataBlockEncoder implements DataBlockEncoder { private static int INITIAL_KEY_BUFFER_SIZE = 512; diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/CompressionState.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/CompressionState.java index 319f782f76a..2fceb520be6 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/CompressionState.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/CompressionState.java @@ -18,12 +18,14 @@ package org.apache.hadoop.hbase.io.encoding; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.ByteBufferUtils; /** * Stores the state of data block encoder at the beginning of new key. */ +@InterfaceAudience.Private class CompressionState { int keyLength; int valueLength; diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/CopyKeyDataBlockEncoder.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/CopyKeyDataBlockEncoder.java index 548985b6c6c..9cb7559a255 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/CopyKeyDataBlockEncoder.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/CopyKeyDataBlockEncoder.java @@ -21,6 +21,7 @@ import java.io.DataOutputStream; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.ByteBufferUtils; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.RawComparator; @@ -29,6 +30,7 @@ import org.apache.hadoop.io.RawComparator; * Just copy data, do not do any kind of compression. Use for comparison and * benchmarking. */ +@InterfaceAudience.Private public class CopyKeyDataBlockEncoder extends BufferedDataBlockEncoder { @Override public void compressKeyValues(DataOutputStream out, diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java index b2f2319fe13..b9ac3736aa7 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoder.java @@ -21,6 +21,7 @@ import java.io.DataOutputStream; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.io.RawComparator; @@ -34,6 +35,7 @@ import org.apache.hadoop.io.RawComparator; * * It is designed to work fast enough to be feasible as in memory compression. */ +@InterfaceAudience.Private public interface DataBlockEncoder { /** * Compress KeyValues and write them to output buffer. diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java index 963b5da1976..a2770451c6f 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/DataBlockEncoding.java @@ -23,6 +23,7 @@ import java.util.HashMap; import java.util.List; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.Bytes; /** @@ -31,6 +32,7 @@ import org.apache.hadoop.hbase.util.Bytes; * want to add a new algorithm/version, assign it a new id. Announce the new id * in the HBase mailing list to prevent collisions. */ +@InterfaceAudience.Private public enum DataBlockEncoding { /** Disable data block encoding. */ diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/DiffKeyDeltaEncoder.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/DiffKeyDeltaEncoder.java index 7e518186ac1..5180bd2dc9d 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/DiffKeyDeltaEncoder.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/DiffKeyDeltaEncoder.java @@ -21,6 +21,7 @@ import java.io.DataOutputStream; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.ByteBufferUtils; import org.apache.hadoop.hbase.util.Bytes; @@ -47,6 +48,7 @@ import org.apache.hadoop.io.RawComparator; * - 1 byte: type (only if FLAG_SAME_TYPE is not set in the flag) * - ... bytes: value */ +@InterfaceAudience.Private public class DiffKeyDeltaEncoder extends BufferedDataBlockEncoder { static final int FLAG_SAME_KEY_LENGTH = 1; static final int FLAG_SAME_VALUE_LENGTH = 1 << 1; diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java index 279a393315c..47c2d00b4c1 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/EncodedDataBlock.java @@ -25,6 +25,7 @@ import java.nio.ByteBuffer; import java.util.Iterator; import org.apache.commons.lang.NotImplementedException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.io.compress.Compressor; @@ -32,6 +33,7 @@ import org.apache.hadoop.io.compress.Compressor; * Encapsulates a data block compressed using a particular encoding algorithm. * Useful for testing and benchmarking. */ +@InterfaceAudience.Private public class EncodedDataBlock { private static final int BUFFER_SIZE = 4 * 1024; protected DataBlockEncoder dataBlockEncoder; diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/EncoderBufferTooSmallException.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/EncoderBufferTooSmallException.java index 55195c18914..8193c13b047 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/EncoderBufferTooSmallException.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/EncoderBufferTooSmallException.java @@ -16,9 +16,12 @@ */ package org.apache.hadoop.hbase.io.encoding; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Internal error which indicates a bug in a data block encoding algorithm. */ +@InterfaceAudience.Private public class EncoderBufferTooSmallException extends RuntimeException { private static final long serialVersionUID = 4767495176134878737L; diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/FastDiffDeltaEncoder.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/FastDiffDeltaEncoder.java index 0ca02c8d9a4..48448d378a4 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/FastDiffDeltaEncoder.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/FastDiffDeltaEncoder.java @@ -22,6 +22,7 @@ import java.io.IOException; import java.io.OutputStream; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.ByteBufferUtils; import org.apache.hadoop.hbase.util.Bytes; @@ -52,6 +53,7 @@ import org.apache.hadoop.io.RawComparator; * - ... bytes: value (only if FLAG_SAME_VALUE is not set in the flag) * */ +@InterfaceAudience.Private public class FastDiffDeltaEncoder extends BufferedDataBlockEncoder { final int MASK_TIMESTAMP_LENGTH = (1 << 0) | (1 << 1) | (1 << 2); final int SHIFT_TIMESTAMP_LENGTH = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/io/encoding/PrefixKeyDeltaEncoder.java b/src/main/java/org/apache/hadoop/hbase/io/encoding/PrefixKeyDeltaEncoder.java index ac63ead6314..98ea7c965ec 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/encoding/PrefixKeyDeltaEncoder.java +++ b/src/main/java/org/apache/hadoop/hbase/io/encoding/PrefixKeyDeltaEncoder.java @@ -21,6 +21,7 @@ import java.io.DataOutputStream; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.ByteBufferUtils; import org.apache.hadoop.hbase.util.Bytes; @@ -40,6 +41,7 @@ import org.apache.hadoop.io.RawComparator; * In a worst case compressed KeyValue will be three bytes longer than original. * */ +@InterfaceAudience.Private public class PrefixKeyDeltaEncoder extends BufferedDataBlockEncoder { private int addKV(int prevKeyOffset, DataOutputStream out, diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java index 3d206d7c107..1f756b00c40 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileReader.java @@ -23,6 +23,7 @@ import java.io.IOException; import java.nio.ByteBuffer; import java.util.concurrent.atomic.AtomicLong; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.KeyValue; @@ -34,6 +35,7 @@ import org.apache.hadoop.io.RawComparator; /** * Common functionality needed by all versions of {@link HFile} readers. */ +@InterfaceAudience.Private public abstract class AbstractHFileReader extends SchemaConfigured implements HFile.Reader { diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileWriter.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileWriter.java index 52f063f5ec7..9e7e624647b 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileWriter.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/AbstractHFileWriter.java @@ -24,6 +24,7 @@ import java.util.ArrayList; import java.util.Arrays; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; @@ -39,6 +40,7 @@ import org.apache.hadoop.io.Writable; /** * Common functionality needed by all versions of {@link HFile} writers. */ +@InterfaceAudience.Private public abstract class AbstractHFileWriter extends SchemaConfigured implements HFile.Writer { diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java index 6f0e33e7162..2ba3ab29c14 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCache.java @@ -22,12 +22,14 @@ package org.apache.hadoop.hbase.io.hfile; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; /** * Block cache interface. Anything that implements the {@link Cacheable} * interface can be put in the cache. */ +@InterfaceAudience.Private public interface BlockCache { /** * Add block to cache. diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheColumnFamilySummary.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheColumnFamilySummary.java index 34513f1825e..ffed712395f 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheColumnFamilySummary.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheColumnFamilySummary.java @@ -23,6 +23,7 @@ import java.io.DataInput; import java.io.DataOutput; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.Path; import org.apache.hadoop.io.Writable; @@ -34,6 +35,7 @@ import org.apache.hadoop.io.Writable; * the owning Table is included in the summarization. * */ +@InterfaceAudience.Private public class BlockCacheColumnFamilySummary implements Writable, Comparable { private String table = ""; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java index d1ab403e647..ec0ddd7ccaf 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockCacheKey.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hbase.io.hfile; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.HeapSize; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; import org.apache.hadoop.hbase.util.Bytes; @@ -25,6 +26,7 @@ import org.apache.hadoop.hbase.util.ClassSize; /** * Cache Key for use with implementations of {@link BlockCache} */ +@InterfaceAudience.Private public class BlockCacheKey implements HeapSize { private final String hfileName; private final long offset; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java index 5797694a8d0..e3c4fe48561 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java @@ -25,6 +25,7 @@ import java.io.IOException; import java.io.OutputStream; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.Bytes; /** @@ -32,6 +33,7 @@ import org.apache.hadoop.hbase.util.Bytes; * must not be relied upon. The values in the enum appear in the order they * appear in a version 2 {@link HFile}. */ +@InterfaceAudience.Private public enum BlockType { // Scanned block section diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/BoundedRangeFileInputStream.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/BoundedRangeFileInputStream.java index f7da04d539b..c32a5e7ae62 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/BoundedRangeFileInputStream.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/BoundedRangeFileInputStream.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.io.hfile; import java.io.IOException; import java.io.InputStream; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FSDataInputStream; /** @@ -29,6 +30,7 @@ import org.apache.hadoop.fs.FSDataInputStream; * would not interfere with each other. * Copied from hadoop-335 tfile. */ +@InterfaceAudience.Private class BoundedRangeFileInputStream extends InputStream { private FSDataInputStream in; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java index 3c2e7229bfb..e130f6ba562 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java @@ -22,6 +22,7 @@ import java.lang.management.MemoryUsage; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; @@ -33,6 +34,7 @@ import org.apache.hadoop.util.StringUtils; /** * Stores all of the cache objects and configuration for a single HFile. */ +@InterfaceAudience.Private public class CacheConfig { private static final Log LOG = LogFactory.getLog(CacheConfig.class.getName()); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheStats.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheStats.java index 439d431d308..ef030751d88 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheStats.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheStats.java @@ -21,9 +21,12 @@ package org.apache.hadoop.hbase.io.hfile; import java.util.concurrent.atomic.AtomicLong; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Class that implements cache metrics. */ +@InterfaceAudience.Private public class CacheStats { /** Sliding window statistics. The number of metric periods to include in diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java index 13d3b4e3b8b..2073e272b32 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/Cacheable.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.io.hfile; import java.nio.ByteBuffer; + +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.HeapSize; import org.apache.hadoop.hbase.regionserver.metrics.SchemaMetrics; @@ -33,6 +35,7 @@ import org.apache.hadoop.hbase.regionserver.metrics.SchemaMetrics; * getSerializedLength() of 0. * */ +@InterfaceAudience.Private public interface Cacheable extends HeapSize { /** * Returns the length of the ByteBuffer required to serialized the object. If the diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheableDeserializer.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheableDeserializer.java index 6210c6a1a51..aa3b855a00e 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheableDeserializer.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheableDeserializer.java @@ -20,10 +20,13 @@ package org.apache.hadoop.hbase.io.hfile; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Interface for a deserializer. Throws an IOException if the serialized data is * incomplete or wrong. * */ +@InterfaceAudience.Private public interface CacheableDeserializer { /** * Returns the deserialized object. diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/CachedBlock.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/CachedBlock.java index a66ab851220..7320463f5d9 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/CachedBlock.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/CachedBlock.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.io.hfile; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.HeapSize; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.ClassSize; @@ -31,6 +32,7 @@ import org.apache.hadoop.hbase.util.ClassSize; * either instantiating as in-memory or handling the transition from single * to multiple access. */ +@InterfaceAudience.Private public class CachedBlock implements HeapSize, Comparable { public final static long PER_BLOCK_OVERHEAD = ClassSize.align( diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/CachedBlockQueue.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/CachedBlockQueue.java index 1637fbfb9ca..6ede4b4bc01 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/CachedBlockQueue.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/CachedBlockQueue.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.io.hfile; import com.google.common.collect.MinMaxPriorityQueue; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.HeapSize; /** @@ -36,6 +37,7 @@ import org.apache.hadoop.hbase.io.HeapSize; *

    Object used in this queue must implement {@link HeapSize} as well as * {@link Comparable}. */ +@InterfaceAudience.Private public class CachedBlockQueue implements HeapSize { private MinMaxPriorityQueue queue; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/Compression.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/Compression.java index b64bbe95a36..7a32610a8f4 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/Compression.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/Compression.java @@ -25,6 +25,7 @@ import java.io.OutputStream; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.io.compress.CodecPool; @@ -41,6 +42,7 @@ import org.apache.hadoop.util.ReflectionUtils; * Compression related stuff. * Copied from hadoop-3315 tfile. */ +@InterfaceAudience.Private public final class Compression { static final Log LOG = LogFactory.getLog(Compression.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/DoubleBlockCache.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/DoubleBlockCache.java index 1b721df8913..e952ef9bda1 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/DoubleBlockCache.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/DoubleBlockCache.java @@ -24,6 +24,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.io.HeapSize; import org.apache.hadoop.hbase.io.hfile.slab.SlabCache; @@ -37,6 +38,7 @@ import org.apache.hadoop.util.StringUtils; * combined size and hits and misses of both caches. * **/ +@InterfaceAudience.Private public class DoubleBlockCache implements BlockCache, HeapSize { static final Log LOG = LogFactory.getLog(DoubleBlockCache.class.getName()); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java index a3a3b919faa..efd145dc8b2 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/FixedFileTrailer.java @@ -29,6 +29,7 @@ import java.nio.ByteBuffer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.RawComparator; @@ -44,6 +45,7 @@ import com.google.common.io.NullOutputStream; * trailer size is fixed within a given {@link HFile} format version only, but * we always store the version number as the last four-byte integer of the file. */ +@InterfaceAudience.Private public class FixedFileTrailer { private static final Log LOG = LogFactory.getLog(FixedFileTrailer.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java index e765e77e8ba..07cf830d526 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java @@ -31,6 +31,7 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; @@ -117,6 +118,7 @@ import com.google.common.base.Preconditions; * that points at its file say for the case where an index lives apart from * an HFile instance? */ +@InterfaceAudience.Private public class HFile { static final Log LOG = LogFactory.getLog(HFile.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java index ef4cb9de121..093af408537 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java @@ -30,6 +30,7 @@ import java.io.IOException; import java.io.InputStream; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; @@ -74,6 +75,7 @@ import com.google.common.base.Preconditions; * The version 2 block representation in the block cache is the same as above, * except that the data section is always uncompressed in the cache. */ +@InterfaceAudience.Private public class HFileBlock extends SchemaConfigured implements Cacheable { public static final boolean FILL_HEADER = true; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java index cae93265194..343e7054598 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlockIndex.java @@ -34,6 +34,7 @@ import java.util.concurrent.atomic.AtomicReference; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.hbase.KeyValue; @@ -57,6 +58,7 @@ import org.apache.hadoop.util.StringUtils; * to use the reader can be found in {@link HFileReaderV2} and * TestHFileBlockIndex. */ +@InterfaceAudience.Private public class HFileBlockIndex { private static final Log LOG = LogFactory.getLog(HFileBlockIndex.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java index 37b0b7c9ba3..3376e78095a 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoder.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.io.hfile; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; import org.apache.hadoop.hbase.regionserver.StoreFile; import org.apache.hadoop.hbase.util.Pair; @@ -28,6 +29,7 @@ import org.apache.hadoop.hbase.util.Pair; * not set or the given block is not a data block (encoded or not), methods * should just return the unmodified block. */ +@InterfaceAudience.Private public interface HFileDataBlockEncoder { /** * Converts a block from the on-disk format to the in-cache format. Called in diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoderImpl.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoderImpl.java index 1759cffbce3..bd4f61273d0 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoderImpl.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileDataBlockEncoderImpl.java @@ -21,6 +21,7 @@ import java.io.DataOutputStream; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoder; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; import org.apache.hadoop.hbase.io.hfile.HFile.FileInfo; @@ -34,6 +35,7 @@ import com.google.common.base.Preconditions; * Do different kinds of data block encoding according to column family * options. */ +@InterfaceAudience.Private public class HFileDataBlockEncoderImpl implements HFileDataBlockEncoder { private final DataBlockEncoding onDisk; private final DataBlockEncoding inCache; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.java index b68e2af66e1..762b3216184 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.java @@ -34,6 +34,8 @@ import org.apache.commons.cli.ParseException; import org.apache.commons.cli.PosixParser; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -53,6 +55,8 @@ import org.apache.hadoop.hbase.util.Writables; /** * Implements pretty-printing functionality for {@link HFile}s. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HFilePrettyPrinter { private static final Log LOG = LogFactory.getLog(HFilePrettyPrinter.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java index d8dac00c7d7..75cec7ef1e5 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java @@ -27,6 +27,7 @@ import java.nio.ByteBuffer; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.KeyValue; @@ -46,6 +47,7 @@ import com.google.common.base.Preconditions; * even in cache only, i.e. HFile v1 blocks are always brought into cache * unencoded. */ +@InterfaceAudience.Private public class HFileReaderV1 extends AbstractHFileReader { private static final Log LOG = LogFactory.getLog(HFileReaderV1.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java index 33203cb0183..4dcae08546e 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java @@ -27,6 +27,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.KeyValue; @@ -41,6 +42,7 @@ import org.apache.hadoop.io.WritableUtils; /** * {@link HFile} reader for version 2. */ +@InterfaceAudience.Private public class HFileReaderV2 extends AbstractHFileReader { private static final Log LOG = LogFactory.getLog(HFileReaderV2.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java index b06878f68a3..9ae4bda2c55 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileScanner.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.io.hfile; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; /** @@ -36,6 +37,7 @@ import org.apache.hadoop.hbase.KeyValue; * you position the Scanner using the seekTo variants and then getKey and * getValue. */ +@InterfaceAudience.Private public interface HFileScanner { /** * SeekTo or just before the passed key. Examine the return diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV1.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV1.java index 080a14c178d..0b5225541a1 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV1.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV1.java @@ -29,6 +29,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; @@ -49,6 +50,7 @@ import org.apache.hadoop.io.compress.Compressor; /** * Writes version 1 HFiles. Mainly used for testing backwards-compatibility. */ +@InterfaceAudience.Private public class HFileWriterV1 extends AbstractHFileWriter { /** Meta data block name for bloom filter parameters. */ diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java index ae7a1347cdd..2fe36ff7e9b 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java @@ -28,6 +28,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; @@ -45,6 +46,7 @@ import org.apache.hadoop.io.WritableUtils; /** * Writes HFile format version 2. */ +@InterfaceAudience.Private public class HFileWriterV2 extends AbstractHFileWriter { static final Log LOG = LogFactory.getLog(HFileWriterV2.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java index c3840362da1..7068224f002 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/InlineBlockWriter.java @@ -22,11 +22,14 @@ package org.apache.hadoop.hbase.io.hfile; import java.io.DataOutput; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + /** * A way to write "inline" blocks into an {@link HFile}. Inline blocks are * interspersed with data blocks. For example, Bloom filter chunks and * leaf-level blocks of a multi-level block index are stored as inline blocks. */ +@InterfaceAudience.Private public interface InlineBlockWriter { /** diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/InvalidHFileException.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/InvalidHFileException.java index 87c7e07d7aa..b68929bf628 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/InvalidHFileException.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/InvalidHFileException.java @@ -19,9 +19,12 @@ package org.apache.hadoop.hbase.io.hfile; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Thrown when an invalid HFile format is detected */ +@InterfaceAudience.Private public class InvalidHFileException extends IOException { private static final long serialVersionUID = 4660352028739861249L; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/LruBlockCache.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/LruBlockCache.java index e9a13fdc4e6..b52e5d35bc3 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/LruBlockCache.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/LruBlockCache.java @@ -39,6 +39,7 @@ import java.util.concurrent.locks.ReentrantLock; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -91,6 +92,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; * to free). It then uses the priority chunk sizes to evict fairly according * to the relative sizes and usage. */ +@InterfaceAudience.Private public class LruBlockCache implements BlockCache, HeapSize { static final Log LOG = LogFactory.getLog(LruBlockCache.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/NoOpDataBlockEncoder.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/NoOpDataBlockEncoder.java index bba8b538ba2..82541df5610 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/NoOpDataBlockEncoder.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/NoOpDataBlockEncoder.java @@ -18,6 +18,7 @@ package org.apache.hadoop.hbase.io.hfile; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; import org.apache.hadoop.hbase.regionserver.StoreFile; import org.apache.hadoop.hbase.util.Pair; @@ -25,6 +26,7 @@ import org.apache.hadoop.hbase.util.Pair; /** * Does not perform any kind of encoding/decoding. */ +@InterfaceAudience.Private public class NoOpDataBlockEncoder implements HFileDataBlockEncoder { public static final NoOpDataBlockEncoder INSTANCE = diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/ReusableStreamGzipCodec.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/ReusableStreamGzipCodec.java index 2b1d48b5a60..7f372c54f64 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/ReusableStreamGzipCodec.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/ReusableStreamGzipCodec.java @@ -24,6 +24,7 @@ import java.util.zip.GZIPOutputStream; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.io.compress.CompressionOutputStream; import org.apache.hadoop.io.compress.CompressorStream; import org.apache.hadoop.io.compress.GzipCodec; @@ -33,6 +34,7 @@ import org.apache.hadoop.io.compress.zlib.ZlibFactory; * Fixes an inefficiency in Hadoop's Gzip codec, allowing to reuse compression * streams. */ +@InterfaceAudience.Private public class ReusableStreamGzipCodec extends GzipCodec { private static final Log LOG = LogFactory.getLog(Compression.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/SimpleBlockCache.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/SimpleBlockCache.java index 1d234301d77..87634ba2116 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/SimpleBlockCache.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/SimpleBlockCache.java @@ -25,12 +25,14 @@ import java.util.HashMap; import java.util.List; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; /** * Simple one RFile soft reference cache. */ +@InterfaceAudience.Private public class SimpleBlockCache implements BlockCache { private static class Ref extends SoftReference { public BlockCacheKey blockId; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SingleSizeCache.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SingleSizeCache.java index 4d9f518e34b..17952298d2e 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SingleSizeCache.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SingleSizeCache.java @@ -26,6 +26,7 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.io.HeapSize; import org.apache.hadoop.hbase.io.hfile.BlockCache; @@ -52,6 +53,7 @@ import com.google.common.collect.MapMaker; * ConcurrentLinkedHashMap. * **/ +@InterfaceAudience.Private public class SingleSizeCache implements BlockCache, HeapSize { private final Slab backingStore; private final ConcurrentMap backingMap; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/Slab.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/Slab.java index ed32980e2b4..82e1fff77ce 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/Slab.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/Slab.java @@ -25,6 +25,7 @@ import java.util.concurrent.LinkedBlockingQueue; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.ClassSize; import org.apache.hadoop.hbase.util.DirectMemoryUtils; import com.google.common.base.Preconditions; @@ -35,6 +36,7 @@ import com.google.common.base.Preconditions; * requisite size, then puts them all in a buffer. **/ +@InterfaceAudience.Private class Slab implements org.apache.hadoop.hbase.io.HeapSize { static final Log LOG = LogFactory.getLog(Slab.class); diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java index 6d343617a4a..251ccd40665 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabCache.java @@ -32,6 +32,7 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.io.HeapSize; import org.apache.hadoop.hbase.io.hfile.BlockCache; @@ -52,6 +53,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; * correct SingleSizeCache. * **/ +@InterfaceAudience.Private public class SlabCache implements SlabItemActionWatcher, BlockCache, HeapSize { private final ConcurrentHashMap backingStore; diff --git a/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabItemActionWatcher.java b/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabItemActionWatcher.java index dfd727f8ad7..82e16ab5d62 100644 --- a/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabItemActionWatcher.java +++ b/src/main/java/org/apache/hadoop/hbase/io/hfile/slab/SlabItemActionWatcher.java @@ -20,11 +20,13 @@ package org.apache.hadoop.hbase.io.hfile.slab; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.hfile.BlockCacheKey; /** * Interface for objects that want to know when actions occur in a SingleSizeCache. * */ +@InterfaceAudience.Private interface SlabItemActionWatcher { /** diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionHeader.java b/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionHeader.java index 63283101630..60e0c8c8ce4 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionHeader.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/ConnectionHeader.java @@ -25,12 +25,14 @@ import java.io.IOException; import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Writable; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.security.User; /** * The IPC connection header sent by the client to the server * on connection establishment. */ +@InterfaceAudience.Private class ConnectionHeader implements Writable { protected String protocol; diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/CoprocessorProtocol.java b/src/main/java/org/apache/hadoop/hbase/ipc/CoprocessorProtocol.java index 8211f037fcb..a00bcba172c 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/CoprocessorProtocol.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/CoprocessorProtocol.java @@ -19,6 +19,9 @@ */ package org.apache.hadoop.hbase.ipc; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * All custom RPC protocols to be exported by Coprocessors must extend this interface. * @@ -34,6 +37,8 @@ package org.apache.hadoop.hbase.ipc; * *

    */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface CoprocessorProtocol extends VersionedProtocol { public static final long VERSION = 1L; } diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java b/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java index 04eb02d9cb8..d63793e1db3 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/Delayable.java @@ -21,9 +21,12 @@ package org.apache.hadoop.hbase.ipc; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + /** * A call whose response can be delayed by the server. */ +@InterfaceAudience.Private public interface Delayable { /** * Signal that the call response should be delayed, thus freeing the RPC diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/ExecRPCInvoker.java b/src/main/java/org/apache/hadoop/hbase/ipc/ExecRPCInvoker.java index b8b290c89ca..05ae7174ae9 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/ExecRPCInvoker.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/ExecRPCInvoker.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.ipc; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.client.*; import org.apache.hadoop.hbase.client.coprocessor.Exec; @@ -36,6 +37,7 @@ import java.lang.reflect.Method; * separate RPC call for each method invocation (using a * {@link org.apache.hadoop.hbase.client.ServerCallable} instance). */ +@InterfaceAudience.Private public class ExecRPCInvoker implements InvocationHandler { // LOG is NOT in hbase subpackage intentionally so that the default HBase // DEBUG log level does NOT emit RPC-level logging. diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseClient.java b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseClient.java index 48a68b06b89..fd99eece816 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseClient.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseClient.java @@ -43,6 +43,7 @@ import javax.net.SocketFactory; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.security.User; @@ -66,6 +67,7 @@ import org.apache.hadoop.util.ReflectionUtils; * * @see HBaseServer */ +@InterfaceAudience.Private public class HBaseClient { private static final Log LOG = diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPC.java b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPC.java index 7eb9e0309f5..3a3a79f5026 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPC.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPC.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.ipc; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HConstants; @@ -65,6 +66,7 @@ import java.util.Map; * All methods in the protocol should throw only IOException. No field data of * the protocol instance is transmitted. */ +@InterfaceAudience.Private public class HBaseRPC { // Leave this out in the hadoop ipc package but keep class name. Do this // so that we dont' get the logging of this class's invocations by doing our diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCErrorHandler.java b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCErrorHandler.java index ad790b55828..fdd1ede5aab 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCErrorHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCErrorHandler.java @@ -20,9 +20,12 @@ package org.apache.hadoop.hbase.ipc; +import org.apache.hadoop.classification.InterfaceAudience; + /** * An interface for calling out of RPC for error conditions. */ +@InterfaceAudience.Private public interface HBaseRPCErrorHandler { /** * Take actions on the event of an OutOfMemoryError. diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCStatistics.java b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCStatistics.java index c9b0257a262..2c7380c6e11 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCStatistics.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRPCStatistics.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.ipc; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.metrics.util.MBeanUtil; import org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase; import org.apache.hadoop.metrics.util.MetricsRegistry; @@ -30,6 +31,7 @@ import javax.management.ObjectName; * Exports HBase RPC statistics recorded in {@link HBaseRpcMetrics} as an MBean * for JMX monitoring. */ +@InterfaceAudience.Private public class HBaseRPCStatistics extends MetricsDynamicMBeanBase { private final ObjectName mbeanName; diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcMetrics.java b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcMetrics.java index 828a56e5dd3..326ff8015d3 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseRpcMetrics.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.ipc; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.ipc.VersionedProtocol; import org.apache.hadoop.metrics.MetricsContext; import org.apache.hadoop.metrics.MetricsRecord; @@ -43,6 +44,7 @@ import java.lang.reflect.Method; *

    {@link #rpcQueueTime}.inc(time) * */ +@InterfaceAudience.Private public class HBaseRpcMetrics implements Updater { public static final String NAME_DELIM = "$"; private final MetricsRegistry registry = new MetricsRegistry(); diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java index 5e004042645..0ba352a6879 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/HBaseServer.java @@ -57,6 +57,7 @@ import java.util.concurrent.LinkedBlockingQueue; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.io.HbaseObjectWritable; import org.apache.hadoop.hbase.io.WritableWithSize; @@ -83,6 +84,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; * * @see HBaseClient */ +@InterfaceAudience.Private public abstract class HBaseServer implements RpcServer { /** diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/HMasterInterface.java b/src/main/java/org/apache/hadoop/hbase/ipc/HMasterInterface.java index 2ee0c626949..3bcf8997a47 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/HMasterInterface.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/HMasterInterface.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.ipc; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.ClusterStatus; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; @@ -42,6 +44,8 @@ import org.apache.hadoop.hbase.util.Pair; @KerberosInfo( serverPrincipal = "hbase.master.kerberos.principal") @TokenInfo("HBASE_AUTH_TOKEN") +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface HMasterInterface extends VersionedProtocol { /** * This Interfaces' version. Version changes when the Interface changes. diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/HMasterRegionInterface.java b/src/main/java/org/apache/hadoop/hbase/ipc/HMasterRegionInterface.java index dfb913353ed..fd9783016ef 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/HMasterRegionInterface.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/HMasterRegionInterface.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.ipc; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HServerLoad; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.security.KerberosInfo; @@ -34,6 +35,7 @@ import org.apache.hadoop.hbase.ipc.VersionedProtocol; @KerberosInfo( serverPrincipal = "hbase.master.kerberos.principal", clientPrincipal = "hbase.regionserver.kerberos.principal") +@InterfaceAudience.Private public interface HMasterRegionInterface extends VersionedProtocol { /** * This Interfaces' version. Version changes when the Interface changes. diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java b/src/main/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java index a3e766957fc..757f98e0a1a 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/HRegionInterface.java @@ -23,6 +23,8 @@ import java.io.IOException; import java.net.ConnectException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HServerInfo; @@ -61,6 +63,8 @@ import org.apache.hadoop.hbase.ipc.VersionedProtocol; @KerberosInfo( serverPrincipal = "hbase.regionserver.kerberos.principal") @TokenInfo("HBASE_AUTH_TOKEN") +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface HRegionInterface extends VersionedProtocol, Stoppable, Abortable { /** * This Interfaces' version. Version changes when the Interface changes. diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/Invocation.java b/src/main/java/org/apache/hadoop/hbase/ipc/Invocation.java index 055787e560d..f1f06b0be61 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/Invocation.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/Invocation.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.ipc; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.io.HbaseObjectWritable; @@ -32,6 +33,7 @@ import java.lang.reflect.Field; import java.lang.reflect.Method; /** A method invocation, including the method name and its parameters.*/ +@InterfaceAudience.Private public class Invocation extends VersionedWritable implements Configurable { protected String methodName; @SuppressWarnings("rawtypes") diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/ProtocolSignature.java b/src/main/java/org/apache/hadoop/hbase/ipc/ProtocolSignature.java index f345cee159b..b3d1df581cd 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/ProtocolSignature.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/ProtocolSignature.java @@ -25,10 +25,12 @@ import java.lang.reflect.Method; import java.util.Arrays; import java.util.HashMap; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.io.Writable; import org.apache.hadoop.io.WritableFactories; import org.apache.hadoop.io.WritableFactory; +@InterfaceAudience.Private public class ProtocolSignature implements Writable { static { // register a ctor WritableFactories.setFactory diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/RequestContext.java b/src/main/java/org/apache/hadoop/hbase/ipc/RequestContext.java index 7de87c14db7..4353c5ad927 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/RequestContext.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/RequestContext.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.ipc; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.security.User; import java.net.InetAddress; @@ -30,6 +31,7 @@ import java.net.InetAddress; * called outside the context of a RPC request, all values will be * null. */ +@InterfaceAudience.Private public class RequestContext { private static ThreadLocal instance = new ThreadLocal() { diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/ResponseFlag.java b/src/main/java/org/apache/hadoop/hbase/ipc/ResponseFlag.java index 88f5115f1b4..7354379e6fb 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/ResponseFlag.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/ResponseFlag.java @@ -17,10 +17,13 @@ */ package org.apache.hadoop.hbase.ipc; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Utility for managing the flag byte passed in response to a * {@link HBaseServer.Call} */ +@InterfaceAudience.Private class ResponseFlag { private static final byte ERROR_BIT = 0x1; private static final byte LENGTH_BIT = 0x2; diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/RpcEngine.java b/src/main/java/org/apache/hadoop/hbase/ipc/RpcEngine.java index d48aeaef2c4..dd0a1bcd2e7 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/RpcEngine.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/RpcEngine.java @@ -26,9 +26,11 @@ import javax.net.SocketFactory; import org.apache.hadoop.hbase.ipc.VersionedProtocol; import org.apache.hadoop.hbase.security.User; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; /** An RPC implementation. */ +@InterfaceAudience.Private interface RpcEngine { /** Construct a client-side proxy object. */ diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java b/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java index 7b2df77c248..ce2cb4eee74 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/RpcServer.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.ipc; import com.google.common.base.Function; import org.apache.hadoop.io.Writable; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.ipc.VersionedProtocol; import org.apache.hadoop.hbase.monitoring.MonitoredRPCHandler; @@ -30,6 +31,7 @@ import java.net.InetSocketAddress; /** */ +@InterfaceAudience.Private public interface RpcServer { void setSocketSendBufSize(int size); diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.java b/src/main/java/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.java index 7dd9b19541f..8ae0e91937b 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/ServerNotRunningYetException.java @@ -22,6 +22,10 @@ package org.apache.hadoop.hbase.ipc; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + +@SuppressWarnings("serial") +@InterfaceAudience.Private public class ServerNotRunningYetException extends IOException { public ServerNotRunningYetException(String s) { super(s); diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/Status.java b/src/main/java/org/apache/hadoop/hbase/ipc/Status.java index c61282f33c2..5d2280248f9 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/Status.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/Status.java @@ -17,9 +17,12 @@ */ package org.apache.hadoop.hbase.ipc; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Status of a Hadoop IPC call. */ +@InterfaceAudience.Private enum Status { SUCCESS (0), ERROR (1), diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/VersionedProtocol.java b/src/main/java/org/apache/hadoop/hbase/ipc/VersionedProtocol.java index 9568b1be5f3..3667c6db5cc 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/VersionedProtocol.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/VersionedProtocol.java @@ -20,11 +20,16 @@ package org.apache.hadoop.hbase.ipc; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Superclass of all protocols that use Hadoop RPC. * Subclasses of this interface are also supposed to have * a static final long versionID field. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface VersionedProtocol { /** diff --git a/src/main/java/org/apache/hadoop/hbase/ipc/WritableRpcEngine.java b/src/main/java/org/apache/hadoop/hbase/ipc/WritableRpcEngine.java index ccc2809c5ec..13164572f0d 100644 --- a/src/main/java/org/apache/hadoop/hbase/ipc/WritableRpcEngine.java +++ b/src/main/java/org/apache/hadoop/hbase/ipc/WritableRpcEngine.java @@ -47,11 +47,13 @@ import org.apache.hadoop.ipc.RPC; import org.apache.hadoop.hbase.ipc.VersionedProtocol; import org.apache.hadoop.hbase.security.User; import org.apache.hadoop.security.authorize.ServiceAuthorizationManager; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.*; import org.codehaus.jackson.map.ObjectMapper; /** An RpcEngine implementation for Writable data. */ +@InterfaceAudience.Private class WritableRpcEngine implements RpcEngine { // LOG is NOT in hbase subpackage intentionally so that the default HBase // DEBUG log level does NOT emit RPC-level logging. diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCounter.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCounter.java index 46d8c7179bd..32d66fb77e2 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCounter.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/CellCounter.java @@ -24,6 +24,8 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; @@ -61,6 +63,8 @@ import org.apache.hadoop.io.Text; * second a regex based or prefix based row filter to restrict the * count operation to a limited subset of rows from the table. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class CellCounter { private static final Log LOG = LogFactory.getLog(CellCounter.class.getName()); diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/CopyTable.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/CopyTable.java index 798f22799df..1fcd6c6f5fd 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/CopyTable.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/CopyTable.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.mapreduce; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HConstants; @@ -36,6 +38,8 @@ import java.util.Map; * It is also configurable with a start and time as well as a specification * of the region server implementation if different from the local cluster. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class CopyTable { final static String NAME = "copytable"; diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/Driver.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/Driver.java index dda4241698a..605a35bf25e 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/Driver.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/Driver.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.mapreduce; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.mapreduce.replication.VerifyReplication; import org.apache.hadoop.util.ProgramDriver; @@ -26,6 +28,8 @@ import org.apache.hadoop.util.ProgramDriver; * Driver for hbase mapreduce jobs. Select which to run by passing * name of job to this main. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Driver { /** * @param args diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/Export.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/Export.java index 46ab6fd7384..015d376ce6f 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/Export.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/Export.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.mapreduce; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -45,6 +47,8 @@ import org.apache.commons.logging.LogFactory; * Writes content to sequence files up in HDFS. Use {@link Import} to read it * back in again. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Export { private static final Log LOG = LogFactory.getLog(Export.class); final static String NAME = "export"; diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/GroupingTableMapper.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/GroupingTableMapper.java index c38337b2ce9..beaec973b70 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/GroupingTableMapper.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/GroupingTableMapper.java @@ -23,6 +23,8 @@ import java.io.IOException; import java.io.UnsupportedEncodingException; import java.util.ArrayList; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.KeyValue; @@ -36,6 +38,8 @@ import org.apache.hadoop.mapreduce.Job; /** * Extract grouping columns from input record. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class GroupingTableMapper extends TableMapper implements Configurable { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat.java index a7419d09d27..c7266038852 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/HFileOutputFormat.java @@ -35,6 +35,8 @@ import java.util.UUID; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.filecache.DistributedCache; import org.apache.hadoop.fs.FileSystem; @@ -71,6 +73,8 @@ import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat; * all HFiles being written. * @see KeyValueSortReducer */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class HFileOutputFormat extends FileOutputFormat { static Log LOG = LogFactory.getLog(HFileOutputFormat.class); static final String COMPRESSION_CONF_KEY = "hbase.hfileoutputformat.families.compression"; diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java index e42d5000e19..f36a50744f0 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/HRegionPartitioner.java @@ -23,6 +23,8 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -44,6 +46,8 @@ import org.apache.hadoop.mapreduce.Partitioner; * @param The type of the key. * @param The type of the value. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class HRegionPartitioner extends Partitioner implements Configurable { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/IdentityTableMapper.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/IdentityTableMapper.java index fd5d8fe6f68..0baf9f1c02d 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/IdentityTableMapper.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/IdentityTableMapper.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.mapreduce; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; @@ -29,6 +31,8 @@ import org.apache.hadoop.mapreduce.Job; /** * Pass the given key and record as-is to the reduce phase. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class IdentityTableMapper extends TableMapper { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/IdentityTableReducer.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/IdentityTableReducer.java index 25f466e86c4..7d6e4accc73 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/IdentityTableReducer.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/IdentityTableReducer.java @@ -23,6 +23,8 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.io.Writable; import org.apache.hadoop.mapreduce.OutputFormat; @@ -48,6 +50,8 @@ import org.apache.hadoop.mapreduce.OutputFormat; * {@link org.apache.hadoop.hbase.client.Delete Delete} define the * row and columns implicitly. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class IdentityTableReducer extends TableReducer { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java index dac1b7560e8..1054e42e316 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/Import.java @@ -23,6 +23,8 @@ import java.io.IOException; import java.util.Map; import java.util.TreeMap; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -43,6 +45,8 @@ import org.apache.hadoop.util.GenericOptionsParser; /** * Import data written by {@link Export}. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Import { final static String NAME = "import"; final static String CF_RENAME_PROP = "HBASE_IMPORTER_RENAME_CFS"; diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java index aed7846253a..f6f8c690d80 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java @@ -24,6 +24,8 @@ import org.apache.hadoop.hbase.util.Base64; import java.io.IOException; import java.util.ArrayList; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -50,6 +52,8 @@ import com.google.common.collect.Lists; * * @see ImportTsv#usage(String) */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ImportTsv { final static String NAME = "importtsv"; diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/KeyValueSortReducer.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/KeyValueSortReducer.java index 1f1567ed864..dde33fb342b 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/KeyValueSortReducer.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/KeyValueSortReducer.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.mapreduce; import java.util.TreeSet; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.mapreduce.Reducer; @@ -32,6 +34,8 @@ import org.apache.hadoop.mapreduce.Reducer; * memory sorting. * @see HFileOutputFormat */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class KeyValueSortReducer extends Reducer { protected void reduce(ImmutableBytesWritable row, java.lang.Iterable kvs, org.apache.hadoop.mapreduce.Reducer.Context context) diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java index 1a724fc8576..e5bb7ea22e0 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/LoadIncrementalHFiles.java @@ -44,6 +44,8 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; import org.apache.hadoop.fs.FileStatus; @@ -86,6 +88,8 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; * Tool to load the output of HFileOutputFormat into an existing table. * @see #usage() */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class LoadIncrementalHFiles extends Configured implements Tool { private static Log LOG = LogFactory.getLog(LoadIncrementalHFiles.class); diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java index 81d27466cf0..ab62c2570e8 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/MultiTableOutputFormat.java @@ -25,6 +25,8 @@ import java.util.Map; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.client.Delete; @@ -55,6 +57,8 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext; * because it is easy to rerun a bulk import). *

    */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class MultiTableOutputFormat extends OutputFormat { /** Set this to {@link #WAL_OFF} to turn off write-ahead logging (HLog) */ public static final String WAL_PROPERTY = "hbase.mapreduce.multitableoutputformat.wal"; diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/PutSortReducer.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/PutSortReducer.java index e76df8c19e2..2cc207e772c 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/PutSortReducer.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/PutSortReducer.java @@ -23,6 +23,8 @@ import java.util.Iterator; import java.util.List; import java.util.TreeSet; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; @@ -37,6 +39,8 @@ import org.apache.hadoop.util.StringUtils; * @see HFileOutputFormat * @see KeyValueSortReducer */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class PutSortReducer extends Reducer { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/RowCounter.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/RowCounter.java index ebf96978e12..c0806f1b2e5 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/RowCounter.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/RowCounter.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.mapreduce; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.KeyValue; @@ -37,6 +39,8 @@ import org.apache.hadoop.util.GenericOptionsParser; * A job with a just a map phase to count rows. Map outputs table rows IF the * input row has columns that have content. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class RowCounter { /** Name of this 'program'. */ diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/SimpleTotalOrderPartitioner.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/SimpleTotalOrderPartitioner.java index a7524cb95aa..6da259f0144 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/SimpleTotalOrderPartitioner.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/SimpleTotalOrderPartitioner.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.mapreduce; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; @@ -43,6 +45,8 @@ import org.apache.hadoop.mapreduce.Partitioner; * @see #START * @see #END */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class SimpleTotalOrderPartitioner extends Partitioner implements Configurable { private final static Log LOG = LogFactory.getLog(SimpleTotalOrderPartitioner.class); diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormat.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormat.java index 27abad569ce..699144901dc 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormat.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormat.java @@ -23,6 +23,8 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.KeyValue; @@ -34,6 +36,8 @@ import org.apache.hadoop.util.StringUtils; /** * Convert HBase tabular data into a format that is consumable by Map/Reduce. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableInputFormat extends TableInputFormatBase implements Configurable { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java index b275e4e0cfe..f84c2ba7ba3 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableInputFormatBase.java @@ -29,6 +29,8 @@ import javax.naming.NamingException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HServerAddress; import org.apache.hadoop.hbase.client.HTable; @@ -72,6 +74,8 @@ import org.apache.hadoop.net.DNS; * } * */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class TableInputFormatBase extends InputFormat { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java index 8d96b1d4cef..b18732c2c6f 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapReduceUtil.java @@ -34,6 +34,8 @@ import java.util.Set; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -57,6 +59,8 @@ import org.apache.hadoop.util.StringUtils; * Utility for {@link TableMapper} and {@link TableReducer} */ @SuppressWarnings("unchecked") +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableMapReduceUtil { static Log LOG = LogFactory.getLog(TableMapReduceUtil.class); diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapper.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapper.java index bbceb63dd6c..0fc4bf878bd 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapper.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableMapper.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.mapreduce; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.mapreduce.Mapper; @@ -31,6 +33,8 @@ import org.apache.hadoop.mapreduce.Mapper; * @param The type of the value. * @see org.apache.hadoop.mapreduce.Mapper */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class TableMapper extends Mapper { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputCommitter.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputCommitter.java index 5289da7e0dc..231e0301155 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputCommitter.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputCommitter.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.mapreduce; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.mapreduce.JobContext; import org.apache.hadoop.mapreduce.OutputCommitter; import org.apache.hadoop.mapreduce.TaskAttemptContext; @@ -28,6 +30,8 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext; /** * Small committer class that does not do anything. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableOutputCommitter extends OutputCommitter { @Override diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java index 89c9603335a..cd9b5289e0a 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableOutputFormat.java @@ -23,6 +23,8 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -45,6 +47,8 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext; * * @param The type of the key. Ignored in this class. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableOutputFormat extends OutputFormat implements Configurable { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReader.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReader.java index a55f82a4973..69ca306f972 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReader.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReader.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.mapreduce; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.client.Scan; @@ -33,6 +35,8 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext; * Iterate over an HBase table data, return (ImmutableBytesWritable, Result) * pairs. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableRecordReader extends RecordReader { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java index a4068396312..1c8a393ac51 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableRecordReaderImpl.java @@ -22,6 +22,8 @@ import java.lang.reflect.Method; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; @@ -41,6 +43,8 @@ import org.apache.hadoop.util.StringUtils; * Iterate over an HBase table data, return (ImmutableBytesWritable, Result) * pairs. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableRecordReaderImpl { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableReducer.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableReducer.java index d087f854abd..537cf49570d 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableReducer.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableReducer.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.mapreduce; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.io.Writable; import org.apache.hadoop.mapreduce.Reducer; @@ -39,6 +41,8 @@ import org.apache.hadoop.mapreduce.Reducer; * @param The type of the output key. * @see org.apache.hadoop.mapreduce.Reducer */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class TableReducer extends Reducer { } \ No newline at end of file diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSplit.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSplit.java index cfc232f6615..27add26de57 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSplit.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TableSplit.java @@ -24,6 +24,8 @@ import java.io.DataOutput; import java.io.IOException; import java.util.Arrays; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.Writable; @@ -33,6 +35,8 @@ import org.apache.hadoop.mapreduce.InputSplit; * A table split corresponds to a key range (low, high). All references to row * below refer to the key of the row. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TableSplit extends InputSplit implements Writable, Comparable { diff --git a/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java b/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java index 406dc9cabdf..9eb3642064e 100644 --- a/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java +++ b/src/main/java/org/apache/hadoop/hbase/mapreduce/TsvImporterMapper.java @@ -25,6 +25,8 @@ import org.apache.hadoop.hbase.util.Base64; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.mapreduce.Mapper; import org.apache.hadoop.mapreduce.Counter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import java.io.IOException; @@ -32,6 +34,8 @@ import java.io.IOException; /** * Write table content out to files in hdfs. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class TsvImporterMapper extends Mapper { diff --git a/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java b/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java index fe4710b87e3..7416ae28f27 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java +++ b/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java @@ -23,6 +23,7 @@ import java.util.concurrent.atomic.AtomicBoolean; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.ServerName; @@ -46,6 +47,7 @@ import org.apache.zookeeper.KeeperException; * #blockUntilBecomingActiveMaster() is called to wait until becoming * the active master of the cluster. */ +@InterfaceAudience.Private class ActiveMasterManager extends ZooKeeperListener { private static final Log LOG = LogFactory.getLog(ActiveMasterManager.class); diff --git a/src/main/java/org/apache/hadoop/hbase/master/AssignCallable.java b/src/main/java/org/apache/hadoop/hbase/master/AssignCallable.java index b233d105bab..d97321d9e77 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/AssignCallable.java +++ b/src/main/java/org/apache/hadoop/hbase/master/AssignCallable.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.master; import java.util.concurrent.Callable; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; /** @@ -29,6 +30,7 @@ import org.apache.hadoop.hbase.HRegionInfo; * Implementing as future callable we are able to act on the timeout * asynchronously. */ +@InterfaceAudience.Private public class AssignCallable implements Callable { private AssignmentManager assignmentManager; diff --git a/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java b/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java index bb58026f9fd..b9a8f882d79 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java +++ b/src/main/java/org/apache/hadoop/hbase/master/AssignmentManager.java @@ -45,6 +45,7 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Chore; import org.apache.hadoop.hbase.HConstants; @@ -98,6 +99,7 @@ import org.apache.zookeeper.data.Stat; *

    * Handles existing regions in transition during master failover. */ +@InterfaceAudience.Private public class AssignmentManager extends ZooKeeperListener { private static final Log LOG = LogFactory.getLog(AssignmentManager.class); diff --git a/src/main/java/org/apache/hadoop/hbase/master/BulkAssigner.java b/src/main/java/org/apache/hadoop/hbase/master/BulkAssigner.java index 588cf358177..2d6d6200179 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/BulkAssigner.java +++ b/src/main/java/org/apache/hadoop/hbase/master/BulkAssigner.java @@ -23,6 +23,7 @@ import java.io.IOException; import java.lang.Thread.UncaughtExceptionHandler; import java.util.concurrent.Executors; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Server; import com.google.common.util.concurrent.ThreadFactoryBuilder; @@ -35,6 +36,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; * the {@link #getUncaughtExceptionHandler()} is to abort the hosting * Server. */ +@InterfaceAudience.Private public abstract class BulkAssigner { protected final Server server; diff --git a/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java b/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java index 37e22cc6190..b9000d8b713 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java +++ b/src/main/java/org/apache/hadoop/hbase/master/BulkReOpen.java @@ -27,6 +27,7 @@ import java.util.concurrent.ExecutorService; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.ServerName; @@ -34,6 +35,7 @@ import org.apache.hadoop.hbase.ServerName; /** * Performs bulk reopen of the list of regions provided to it. */ +@InterfaceAudience.Private public class BulkReOpen extends BulkAssigner { private final Map> rsToRegions; private final AssignmentManager assignmentManager; diff --git a/src/main/java/org/apache/hadoop/hbase/master/CatalogJanitor.java b/src/main/java/org/apache/hadoop/hbase/master/CatalogJanitor.java index 4ee75ee948e..79d5fdd050a 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/CatalogJanitor.java +++ b/src/main/java/org/apache/hadoop/hbase/master/CatalogJanitor.java @@ -28,6 +28,7 @@ import java.util.concurrent.atomic.AtomicInteger; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -55,6 +56,7 @@ import org.apache.hadoop.hbase.util.Writables; * A janitor for the catalog tables. Scans the .META. catalog * table on a period looking for unused regions to garbage collect. */ +@InterfaceAudience.Private class CatalogJanitor extends Chore { private static final Log LOG = LogFactory.getLog(CatalogJanitor.class.getName()); private final Server server; diff --git a/src/main/java/org/apache/hadoop/hbase/master/DeadServer.java b/src/main/java/org/apache/hadoop/hbase/master/DeadServer.java index 26e57144554..adf79c62967 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/DeadServer.java +++ b/src/main/java/org/apache/hadoop/hbase/master/DeadServer.java @@ -25,11 +25,13 @@ import java.util.Iterator; import java.util.Set; import org.apache.commons.lang.NotImplementedException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.ServerName; /** * Class to hold dead servers list and utility querying dead server list. */ +@InterfaceAudience.Private public class DeadServer implements Set { /** * Set of known dead servers. On znode expiration, servers are added here. diff --git a/src/main/java/org/apache/hadoop/hbase/master/DefaultLoadBalancer.java b/src/main/java/org/apache/hadoop/hbase/master/DefaultLoadBalancer.java index 9b132c0286a..1fb923ff500 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/DefaultLoadBalancer.java +++ b/src/main/java/org/apache/hadoop/hbase/master/DefaultLoadBalancer.java @@ -36,6 +36,7 @@ import java.util.TreeMap; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.hbase.ClusterStatus; @@ -68,6 +69,7 @@ import com.google.common.collect.Sets; * *

    This classes produces plans for the {@link AssignmentManager} to execute. */ +@InterfaceAudience.Private public class DefaultLoadBalancer implements LoadBalancer { private static final Log LOG = LogFactory.getLog(LoadBalancer.class); private static final Random RANDOM = new Random(System.currentTimeMillis()); diff --git a/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/src/main/java/org/apache/hadoop/hbase/master/HMaster.java index 3e4b4d048cf..e2bbbd09661 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/HMaster.java +++ b/src/main/java/org/apache/hadoop/hbase/master/HMaster.java @@ -40,6 +40,7 @@ import javax.management.ObjectName; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Chore; import org.apache.hadoop.hbase.ClusterStatus; @@ -126,6 +127,7 @@ import org.apache.zookeeper.Watcher; * @see HMasterRegionInterface * @see Watcher */ +@InterfaceAudience.Private public class HMaster extends HasThread implements HMasterInterface, HMasterRegionInterface, MasterServices, Server { diff --git a/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java b/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java index ae9faf9a348..f2151f0e1d9 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java +++ b/src/main/java/org/apache/hadoop/hbase/master/HMasterCommandLine.java @@ -29,6 +29,7 @@ import org.apache.commons.cli.Options; import org.apache.commons.cli.ParseException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.LocalHBaseCluster; @@ -41,6 +42,7 @@ import org.apache.hadoop.hbase.util.ServerCommandLine; import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster; import org.apache.zookeeper.KeeperException; +@InterfaceAudience.Private public class HMasterCommandLine extends ServerCommandLine { private static final Log LOG = LogFactory.getLog(HMasterCommandLine.class); diff --git a/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java b/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java index 7d2dd74a569..11de9dd1227 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java +++ b/src/main/java/org/apache/hadoop/hbase/master/LoadBalancer.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.master; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.hbase.ClusterStatus; import org.apache.hadoop.hbase.HRegionInfo; @@ -43,6 +44,7 @@ import java.util.Map; * *

    This classes produces plans for the {@link AssignmentManager} to execute. */ +@InterfaceAudience.Private public interface LoadBalancer extends Configurable { /** diff --git a/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java b/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java index 3a22e468dd4..de6f1fb6fcc 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java +++ b/src/main/java/org/apache/hadoop/hbase/master/LoadBalancerFactory.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.master; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.util.ReflectionUtils; @@ -27,6 +28,7 @@ import org.apache.hadoop.util.ReflectionUtils; /** * The class that creates a load balancer from a conf. */ +@InterfaceAudience.Private public class LoadBalancerFactory { /** diff --git a/src/main/java/org/apache/hadoop/hbase/master/LogCleaner.java b/src/main/java/org/apache/hadoop/hbase/master/LogCleaner.java index b5fc6650193..1be3cb692e4 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/LogCleaner.java +++ b/src/main/java/org/apache/hadoop/hbase/master/LogCleaner.java @@ -25,6 +25,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -41,6 +42,7 @@ import static org.apache.hadoop.hbase.HConstants.HBASE_MASTER_LOGCLEANER_PLUGINS * This Chore, everytime it runs, will clear the HLogs in the old logs folder * that are deletable for each log cleaner in the chain. */ +@InterfaceAudience.Private public class LogCleaner extends Chore { static final Log LOG = LogFactory.getLog(LogCleaner.class.getName()); diff --git a/src/main/java/org/apache/hadoop/hbase/master/LogCleanerDelegate.java b/src/main/java/org/apache/hadoop/hbase/master/LogCleanerDelegate.java index 3c8eb0a1c2c..ffb3ef57c0b 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/LogCleanerDelegate.java +++ b/src/main/java/org/apache/hadoop/hbase/master/LogCleanerDelegate.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.master; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configurable; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.Stoppable; @@ -38,6 +39,7 @@ import org.apache.hadoop.hbase.Stoppable; * Since LogCleanerDelegates are created in LogsCleaner by reflection. Classes * that implements this interface should provide a default constructor. */ +@InterfaceAudience.Private public interface LogCleanerDelegate extends Configurable, Stoppable { /** * Should the master delete the log or keep it? diff --git a/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java b/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java index 4beafb21509..cafaf9c6615 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java +++ b/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.master; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.*; import org.apache.hadoop.hbase.coprocessor.*; @@ -33,6 +34,7 @@ import java.io.IOException; * operations. {@link HMaster} interacts with the loaded coprocessors * through this class. */ +@InterfaceAudience.Private public class MasterCoprocessorHost extends CoprocessorHost { diff --git a/src/main/java/org/apache/hadoop/hbase/master/MasterDumpServlet.java b/src/main/java/org/apache/hadoop/hbase/master/MasterDumpServlet.java index 876eda4137c..be6383873f3 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/MasterDumpServlet.java +++ b/src/main/java/org/apache/hadoop/hbase/master/MasterDumpServlet.java @@ -29,6 +29,7 @@ import java.util.NavigableMap; import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletResponse; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HServerInfo; import org.apache.hadoop.hbase.HServerLoad; @@ -39,6 +40,7 @@ import org.apache.hadoop.hbase.monitoring.StateDumpServlet; import org.apache.hadoop.hbase.monitoring.TaskMonitor; import org.apache.hadoop.util.ReflectionUtils; +@InterfaceAudience.Private public class MasterDumpServlet extends StateDumpServlet { private static final long serialVersionUID = 1L; private static final String LINE = diff --git a/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java b/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java index 411023485fb..8a7da2e4217 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java +++ b/src/main/java/org/apache/hadoop/hbase/master/MasterFileSystem.java @@ -29,6 +29,7 @@ import java.util.concurrent.locks.ReentrantLock; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -56,6 +57,7 @@ import org.apache.hadoop.hbase.util.FSUtils; * the underlying file system, including splitting log files, checking file * system status, etc. */ +@InterfaceAudience.Private public class MasterFileSystem { private static final Log LOG = LogFactory.getLog(MasterFileSystem.class.getName()); // HBase configuration diff --git a/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java b/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java index c4b42ae5038..fb6c7e56e7b 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java +++ b/src/main/java/org/apache/hadoop/hbase/master/MasterServices.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.master; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.TableDescriptors; @@ -34,6 +35,7 @@ import org.apache.hadoop.hbase.zookeeper.RegionServerTracker; /** * Services Master supplies */ +@InterfaceAudience.Private public interface MasterServices extends Server { /** * @return Master's instance of the {@link AssignmentManager} diff --git a/src/main/java/org/apache/hadoop/hbase/master/MasterStatusServlet.java b/src/main/java/org/apache/hadoop/hbase/master/MasterStatusServlet.java index 862db2e6826..a0aaaf420cc 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/MasterStatusServlet.java +++ b/src/main/java/org/apache/hadoop/hbase/master/MasterStatusServlet.java @@ -31,6 +31,7 @@ import javax.servlet.http.HttpServletResponse; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.client.HBaseAdmin; @@ -41,6 +42,7 @@ import org.apache.hadoop.hbase.tmpl.master.MasterStatusTmpl; * The servlet responsible for rendering the index page of the * master. */ +@InterfaceAudience.Private public class MasterStatusServlet extends HttpServlet { private static final Log LOG = LogFactory.getLog(MasterStatusServlet.class); private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/master/RegionPlan.java b/src/main/java/org/apache/hadoop/hbase/master/RegionPlan.java index 3c1fe83697e..6fbf9ab65c6 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/RegionPlan.java +++ b/src/main/java/org/apache/hadoop/hbase/master/RegionPlan.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hbase.master; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.ServerName; @@ -30,6 +31,7 @@ import org.apache.hadoop.hbase.ServerName; * The comparable implementation of this class compares only the region * information and not the source/dest server info. */ +@InterfaceAudience.Private public class RegionPlan implements Comparable { private final HRegionInfo hri; private final ServerName source; diff --git a/src/main/java/org/apache/hadoop/hbase/master/ServerAndLoad.java b/src/main/java/org/apache/hadoop/hbase/master/ServerAndLoad.java index 17d0c764315..8c7dadf0010 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/ServerAndLoad.java +++ b/src/main/java/org/apache/hadoop/hbase/master/ServerAndLoad.java @@ -18,11 +18,13 @@ package org.apache.hadoop.hbase.master; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.ServerName; /** * Data structure that holds servername and 'load'. */ +@InterfaceAudience.Private class ServerAndLoad implements Comparable { private final ServerName sn; private final int load; diff --git a/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java index 02bebb652db..6fb68796a36 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java +++ b/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java @@ -31,6 +31,7 @@ import java.util.concurrent.ConcurrentHashMap; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.ClockOutOfSyncException; import org.apache.hadoop.hbase.HRegionInfo; @@ -63,6 +64,7 @@ import org.apache.hadoop.hbase.monitoring.MonitoredTask; * was started). This is used to differentiate a restarted instance of a given * server from the original instance. */ +@InterfaceAudience.Private public class ServerManager { private static final Log LOG = LogFactory.getLog(ServerManager.class); diff --git a/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java b/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java index 79862c50165..585bc1e5052 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java +++ b/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java @@ -32,6 +32,7 @@ import java.util.concurrent.ConcurrentMap; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -93,6 +94,7 @@ import static org.apache.hadoop.hbase.master.SplitLogManager.TerminationStatus.* * again. If a task is resubmitted then there is a risk that old "delete task" * can delete the re-submission. */ +@InterfaceAudience.Private public class SplitLogManager extends ZooKeeperListener { private static final Log LOG = LogFactory.getLog(SplitLogManager.class); diff --git a/src/main/java/org/apache/hadoop/hbase/master/TimeToLiveLogCleaner.java b/src/main/java/org/apache/hadoop/hbase/master/TimeToLiveLogCleaner.java index dde820700d2..f29b0cf6e47 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/TimeToLiveLogCleaner.java +++ b/src/main/java/org/apache/hadoop/hbase/master/TimeToLiveLogCleaner.java @@ -23,6 +23,7 @@ import java.io.IOException; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.Path; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -31,6 +32,7 @@ import org.apache.commons.logging.LogFactory; * Log cleaner that uses the timestamp of the hlog to determine if it should * be deleted. By default they are allowed to live for 10 minutes. */ +@InterfaceAudience.Private public class TimeToLiveLogCleaner implements LogCleanerDelegate { static final Log LOG = LogFactory.getLog(TimeToLiveLogCleaner.class.getName()); private Configuration conf; diff --git a/src/main/java/org/apache/hadoop/hbase/master/UnAssignCallable.java b/src/main/java/org/apache/hadoop/hbase/master/UnAssignCallable.java index 2cbe7e08cb5..9fe8d857d30 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/UnAssignCallable.java +++ b/src/main/java/org/apache/hadoop/hbase/master/UnAssignCallable.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.master; import java.util.concurrent.Callable; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; /** @@ -28,6 +29,7 @@ import org.apache.hadoop.hbase.HRegionInfo; * taken for unassignment of a region in transition. Implementing as future * callable we are able to act on the timeout asynchronously. */ +@InterfaceAudience.Private public class UnAssignCallable implements Callable { private AssignmentManager assignmentManager; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/ClosedRegionHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/ClosedRegionHandler.java index 88f207a1af4..992d3332d3b 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/ClosedRegionHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/ClosedRegionHandler.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.master.handler; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.executor.EventHandler; @@ -34,6 +35,7 @@ import org.apache.hadoop.hbase.master.AssignmentManager; *

    * Otherwise, assigns the region to another server. */ +@InterfaceAudience.Private public class ClosedRegionHandler extends EventHandler implements TotesHRegionInfo { private static final Log LOG = LogFactory.getLog(ClosedRegionHandler.class); private final AssignmentManager assignmentManager; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/CreateTableHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/CreateTableHandler.java index 4600991cb4e..bb57f8db4b2 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/CreateTableHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/CreateTableHandler.java @@ -26,6 +26,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; @@ -48,6 +49,7 @@ import org.apache.zookeeper.KeeperException; /** * Handler to create a table. */ +@InterfaceAudience.Private public class CreateTableHandler extends EventHandler { private static final Log LOG = LogFactory.getLog(CreateTableHandler.class); private MasterFileSystem fileSystemManager; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/DeleteTableHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/DeleteTableHandler.java index 17dd2f9c79e..7de80d32258 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/DeleteTableHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/DeleteTableHandler.java @@ -24,6 +24,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.catalog.MetaEditor; @@ -34,6 +35,7 @@ import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Threads; import org.apache.zookeeper.KeeperException; +@InterfaceAudience.Private public class DeleteTableHandler extends TableEventHandler { private static final Log LOG = LogFactory.getLog(DeleteTableHandler.class); diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/DisableTableHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/DisableTableHandler.java index 5af0690207e..d9251130e5c 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/DisableTableHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/DisableTableHandler.java @@ -25,6 +25,7 @@ import java.util.concurrent.ExecutorService; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.TableNotEnabledException; @@ -40,6 +41,7 @@ import org.apache.zookeeper.KeeperException; /** * Handler to run disable of a table. */ +@InterfaceAudience.Private public class DisableTableHandler extends EventHandler { private static final Log LOG = LogFactory.getLog(DisableTableHandler.class); private final byte [] tableName; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/EnableTableHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/EnableTableHandler.java index 00f8e72738c..498f1e20de6 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/EnableTableHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/EnableTableHandler.java @@ -25,6 +25,7 @@ import java.util.concurrent.ExecutorService; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.TableNotDisabledException; @@ -40,6 +41,7 @@ import org.apache.zookeeper.KeeperException; /** * Handler to run enable of a table. */ +@InterfaceAudience.Private public class EnableTableHandler extends EventHandler { private static final Log LOG = LogFactory.getLog(EnableTableHandler.class); private final byte [] tableName; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/MetaServerShutdownHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/MetaServerShutdownHandler.java index 99d77e4ddb6..37cedbdf509 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/MetaServerShutdownHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/MetaServerShutdownHandler.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.master.handler; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.master.DeadServer; @@ -28,6 +29,7 @@ import org.apache.hadoop.hbase.master.MasterServices; * Shutdown handler for the server hosting -ROOT-, * .META., or both. */ +@InterfaceAudience.Private public class MetaServerShutdownHandler extends ServerShutdownHandler { private final boolean carryingRoot; private final boolean carryingMeta; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java index c0918164863..4a735ec3181 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/ModifyTableHandler.java @@ -22,12 +22,14 @@ package org.apache.hadoop.hbase.master.handler; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.ipc.HMasterInterface; import org.apache.hadoop.hbase.master.MasterServices; +@InterfaceAudience.Private public class ModifyTableHandler extends TableEventHandler { private final HTableDescriptor htd; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/OpenedRegionHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/OpenedRegionHandler.java index f171a5a3b8d..2254cf46b0e 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/OpenedRegionHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/OpenedRegionHandler.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.master.handler; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.ServerName; @@ -34,6 +35,7 @@ import org.apache.zookeeper.KeeperException; /** * Handles OPENED region event on Master. */ +@InterfaceAudience.Private public class OpenedRegionHandler extends EventHandler implements TotesHRegionInfo { private static final Log LOG = LogFactory.getLog(OpenedRegionHandler.class); private final AssignmentManager assignmentManager; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/ServerShutdownHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/ServerShutdownHandler.java index 4307d89a404..4068b0dc562 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/ServerShutdownHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/ServerShutdownHandler.java @@ -26,6 +26,7 @@ import java.util.NavigableMap; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; @@ -48,6 +49,7 @@ import org.apache.zookeeper.KeeperException; * Server-to-handle must be already in the deadservers lists. See * {@link ServerManager#expireServer(ServerName)} */ +@InterfaceAudience.Private public class ServerShutdownHandler extends EventHandler { private static final Log LOG = LogFactory.getLog(ServerShutdownHandler.class); private final ServerName serverName; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/SplitRegionHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/SplitRegionHandler.java index 2d544dd155f..ee97cfcdec3 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/SplitRegionHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/SplitRegionHandler.java @@ -23,6 +23,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.ServerName; @@ -36,6 +37,7 @@ import org.apache.zookeeper.KeeperException.NoNodeException; /** * Handles SPLIT region event on Master. */ +@InterfaceAudience.Private public class SplitRegionHandler extends EventHandler implements TotesHRegionInfo { private static final Log LOG = LogFactory.getLog(SplitRegionHandler.class); private final AssignmentManager assignmentManager; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/TableAddFamilyHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/TableAddFamilyHandler.java index d9933673b65..4ec3d684a1f 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/TableAddFamilyHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/TableAddFamilyHandler.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.master.handler; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; @@ -33,6 +34,7 @@ import org.apache.hadoop.hbase.master.MasterServices; /** * Handles adding a new family to an existing table. */ +@InterfaceAudience.Private public class TableAddFamilyHandler extends TableEventHandler { private final HColumnDescriptor familyDesc; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/TableDeleteFamilyHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/TableDeleteFamilyHandler.java index 07f67dd5719..bfa624b80ff 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/TableDeleteFamilyHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/TableDeleteFamilyHandler.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.master.handler; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.Server; @@ -32,6 +33,7 @@ import org.apache.hadoop.hbase.util.Bytes; /** * Handles adding a new family to an existing table. */ +@InterfaceAudience.Private public class TableDeleteFamilyHandler extends TableEventHandler { private final byte [] familyName; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/TableEventHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/TableEventHandler.java index af5b96a412a..421f01d2742 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/TableEventHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/TableEventHandler.java @@ -29,6 +29,7 @@ import java.util.TreeMap; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.InvalidFamilyOperationException; @@ -60,6 +61,7 @@ import com.google.common.collect.Maps; * later down in an async invocation of {@link #process()} (which currently has * no means of reporting back issues once started). */ +@InterfaceAudience.Private public abstract class TableEventHandler extends EventHandler { private static final Log LOG = LogFactory.getLog(TableEventHandler.class); protected final MasterServices masterServices; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/TableModifyFamilyHandler.java b/src/main/java/org/apache/hadoop/hbase/master/handler/TableModifyFamilyHandler.java index c6e292e0bfc..b4f8cd4c647 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/TableModifyFamilyHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/TableModifyFamilyHandler.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.master.handler; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; @@ -34,6 +35,7 @@ import org.apache.hadoop.hbase.util.Bytes; /** * Handles adding a new family to an existing table. */ +@InterfaceAudience.Private public class TableModifyFamilyHandler extends TableEventHandler { private final HColumnDescriptor familyDesc; diff --git a/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java b/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java index d08f64982fc..50d72580b43 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java +++ b/src/main/java/org/apache/hadoop/hbase/master/handler/TotesHRegionInfo.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.master.handler; import java.beans.EventHandler; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; /** @@ -28,6 +29,7 @@ import org.apache.hadoop.hbase.HRegionInfo; * This is a marker interface that can be put on {@link EventHandler}s that * have an {@link HRegionInfo}. */ +@InterfaceAudience.Private public interface TotesHRegionInfo { /** * @return HRegionInfo instance. diff --git a/src/main/java/org/apache/hadoop/hbase/master/metrics/MasterMetrics.java b/src/main/java/org/apache/hadoop/hbase/master/metrics/MasterMetrics.java index 9e4cf735104..83abc524eba 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/metrics/MasterMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/master/metrics/MasterMetrics.java @@ -21,6 +21,7 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.metrics.HBaseInfo; import org.apache.hadoop.hbase.metrics.MetricsRate; import org.apache.hadoop.hbase.metrics.PersistentMetricsTimeVaryingRate; @@ -41,6 +42,7 @@ import org.apache.hadoop.metrics.util.MetricsRegistry; * This class has a number of metrics variables that are publicly accessible; * these variables (objects) have methods to update their values. */ +@InterfaceAudience.Private public class MasterMetrics implements Updater { private final Log LOG = LogFactory.getLog(this.getClass()); private final MetricsRecord metricsRecord; diff --git a/src/main/java/org/apache/hadoop/hbase/master/metrics/MasterStatistics.java b/src/main/java/org/apache/hadoop/hbase/master/metrics/MasterStatistics.java index d8853486b2e..ced1d7e2690 100644 --- a/src/main/java/org/apache/hadoop/hbase/master/metrics/MasterStatistics.java +++ b/src/main/java/org/apache/hadoop/hbase/master/metrics/MasterStatistics.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.master.metrics; import javax.management.ObjectName; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.metrics.MetricsMBeanBase; import org.apache.hadoop.metrics.util.MBeanUtil; import org.apache.hadoop.metrics.util.MetricsRegistry; @@ -27,6 +28,7 @@ import org.apache.hadoop.metrics.util.MetricsRegistry; * Exports the {@link MasterMetrics} statistics as an MBean * for JMX. */ +@InterfaceAudience.Private public class MasterStatistics extends MetricsMBeanBase { private final ObjectName mbeanName; diff --git a/src/main/java/org/apache/hadoop/hbase/metrics/HBaseInfo.java b/src/main/java/org/apache/hadoop/hbase/metrics/HBaseInfo.java index fb65a652d3d..c4cc47abb9b 100644 --- a/src/main/java/org/apache/hadoop/hbase/metrics/HBaseInfo.java +++ b/src/main/java/org/apache/hadoop/hbase/metrics/HBaseInfo.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.metrics; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.metrics.MetricsMBeanBase; import org.apache.hadoop.metrics.MetricsContext; import org.apache.hadoop.metrics.MetricsRecord; @@ -31,6 +32,7 @@ import javax.management.ObjectName; /** * Exports HBase system information as an MBean for JMX observation. */ +@InterfaceAudience.Private public class HBaseInfo { protected static class HBaseInfoMBean extends MetricsMBeanBase { private final ObjectName mbeanName; diff --git a/src/main/java/org/apache/hadoop/hbase/metrics/MetricsMBeanBase.java b/src/main/java/org/apache/hadoop/hbase/metrics/MetricsMBeanBase.java index 37fdfc281ec..d23f37897af 100644 --- a/src/main/java/org/apache/hadoop/hbase/metrics/MetricsMBeanBase.java +++ b/src/main/java/org/apache/hadoop/hbase/metrics/MetricsMBeanBase.java @@ -32,6 +32,7 @@ import javax.management.ReflectionException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.metrics.util.MetricsBase; import org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase; import org.apache.hadoop.metrics.util.MetricsRegistry; @@ -43,6 +44,7 @@ import org.apache.hadoop.metrics.util.MetricsRegistry; * org.apache.hadoop.metrics.util implementations. * */ +@InterfaceAudience.Private public class MetricsMBeanBase extends MetricsDynamicMBeanBase { private static final Log LOG = LogFactory.getLog("org.apache.hadoop.hbase.metrics"); diff --git a/src/main/java/org/apache/hadoop/hbase/metrics/MetricsRate.java b/src/main/java/org/apache/hadoop/hbase/metrics/MetricsRate.java index 9e00d15f80e..18543b38629 100644 --- a/src/main/java/org/apache/hadoop/hbase/metrics/MetricsRate.java +++ b/src/main/java/org/apache/hadoop/hbase/metrics/MetricsRate.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.metrics; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.metrics.MetricsRecord; import org.apache.hadoop.metrics.util.MetricsBase; import org.apache.hadoop.metrics.util.MetricsRegistry; @@ -28,6 +29,7 @@ import org.apache.hadoop.util.StringUtils; * Publishes a rate based on a counter - you increment the counter each * time an event occurs (eg: an RPC call) and this publishes a rate. */ +@InterfaceAudience.Private public class MetricsRate extends MetricsBase { private static final Log LOG = LogFactory.getLog("org.apache.hadoop.hbase.metrics"); diff --git a/src/main/java/org/apache/hadoop/hbase/metrics/MetricsString.java b/src/main/java/org/apache/hadoop/hbase/metrics/MetricsString.java index 2ee8066f59d..09c9b8aa3b0 100644 --- a/src/main/java/org/apache/hadoop/hbase/metrics/MetricsString.java +++ b/src/main/java/org/apache/hadoop/hbase/metrics/MetricsString.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.metrics; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.metrics.MetricsRecord; import org.apache.hadoop.metrics.util.MetricsBase; import org.apache.hadoop.metrics.util.MetricsRegistry; @@ -26,6 +27,7 @@ import org.apache.hadoop.metrics.util.MetricsRegistry; /** * Publishes a string to the metrics collector */ +@InterfaceAudience.Private public class MetricsString extends MetricsBase { private static final Log LOG = LogFactory.getLog("org.apache.hadoop.hbase.metrics"); diff --git a/src/main/java/org/apache/hadoop/hbase/metrics/PersistentMetricsTimeVaryingRate.java b/src/main/java/org/apache/hadoop/hbase/metrics/PersistentMetricsTimeVaryingRate.java index cf2fc287dd4..bfad5a5f94a 100644 --- a/src/main/java/org/apache/hadoop/hbase/metrics/PersistentMetricsTimeVaryingRate.java +++ b/src/main/java/org/apache/hadoop/hbase/metrics/PersistentMetricsTimeVaryingRate.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.metrics; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.metrics.MetricsRecord; import org.apache.hadoop.metrics.util.MetricsRegistry; import org.apache.hadoop.metrics.util.MetricsTimeVaryingRate; @@ -28,6 +29,7 @@ import org.apache.hadoop.util.StringUtils; * This class extends MetricsTimeVaryingRate to let the metrics * persist past a pushMetric() call */ +@InterfaceAudience.Private public class PersistentMetricsTimeVaryingRate extends MetricsTimeVaryingRate { protected static final Log LOG = LogFactory.getLog("org.apache.hadoop.hbase.metrics"); diff --git a/src/main/java/org/apache/hadoop/hbase/metrics/file/TimeStampingFileContext.java b/src/main/java/org/apache/hadoop/hbase/metrics/file/TimeStampingFileContext.java index 000e0d3d976..600ce554ffe 100644 --- a/src/main/java/org/apache/hadoop/hbase/metrics/file/TimeStampingFileContext.java +++ b/src/main/java/org/apache/hadoop/hbase/metrics/file/TimeStampingFileContext.java @@ -27,6 +27,7 @@ import java.io.PrintWriter; import java.text.SimpleDateFormat; import java.util.Date; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.metrics.ContextFactory; import org.apache.hadoop.metrics.file.FileContext; import org.apache.hadoop.metrics.spi.OutputRecord; @@ -34,6 +35,7 @@ import org.apache.hadoop.metrics.spi.OutputRecord; /** * Add timestamp to {@link org.apache.hadoop.metrics.file.FileContext#emitRecord(String, String, OutputRecord)}. */ +@InterfaceAudience.Private public class TimeStampingFileContext extends FileContext { // Copies bunch of FileContext here because writer and file are private in // superclass. diff --git a/src/main/java/org/apache/hadoop/hbase/migration/HRegionInfo090x.java b/src/main/java/org/apache/hadoop/hbase/migration/HRegionInfo090x.java index eeb18e8793b..c183ba5869f 100644 --- a/src/main/java/org/apache/hadoop/hbase/migration/HRegionInfo090x.java +++ b/src/main/java/org/apache/hadoop/hbase/migration/HRegionInfo090x.java @@ -27,6 +27,7 @@ import java.util.Arrays; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.KeyValue; @@ -42,6 +43,7 @@ import org.apache.hadoop.io.WritableComparable; * Contains HRegion id, start and end keys, a reference to this * HRegions' table descriptor, etc. */ +@InterfaceAudience.Private public class HRegionInfo090x extends VersionedWritable implements WritableComparable{ private static final byte VERSION = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/monitoring/LogMonitoring.java b/src/main/java/org/apache/hadoop/hbase/monitoring/LogMonitoring.java index d121ee1f4ef..343c6551357 100644 --- a/src/main/java/org/apache/hadoop/hbase/monitoring/LogMonitoring.java +++ b/src/main/java/org/apache/hadoop/hbase/monitoring/LogMonitoring.java @@ -29,6 +29,7 @@ import java.nio.channels.FileChannel; import java.util.Enumeration; import java.util.Set; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.io.IOUtils; import org.apache.log4j.Appender; import org.apache.log4j.FileAppender; @@ -40,6 +41,7 @@ import com.google.common.collect.Sets; * Utility functions for reading the log4j logs that are * being written by HBase. */ +@InterfaceAudience.Private public abstract class LogMonitoring { public static Set getActiveLogFiles() throws IOException { Set ret = Sets.newHashSet(); diff --git a/src/main/java/org/apache/hadoop/hbase/monitoring/MemoryBoundedLogMessageBuffer.java b/src/main/java/org/apache/hadoop/hbase/monitoring/MemoryBoundedLogMessageBuffer.java index e8b7416843b..331673829af 100644 --- a/src/main/java/org/apache/hadoop/hbase/monitoring/MemoryBoundedLogMessageBuffer.java +++ b/src/main/java/org/apache/hadoop/hbase/monitoring/MemoryBoundedLogMessageBuffer.java @@ -25,6 +25,8 @@ import java.util.Date; import java.util.LinkedList; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; + import com.google.common.base.Charsets; import com.google.common.base.Preconditions; import com.google.common.collect.Lists; @@ -35,6 +37,7 @@ import com.google.common.collect.Lists; * automatically be removed one by one when the specified heap * usage is exhausted. */ +@InterfaceAudience.Private public class MemoryBoundedLogMessageBuffer { private final long maxSizeBytes; private long usage = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java b/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java index d4f97144d87..7dd2686ed26 100644 --- a/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandler.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.monitoring; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.Writable; @@ -27,6 +28,7 @@ import org.apache.hadoop.io.Writable; * handling frequent, short duration tasks. String concatenations and object * allocations are avoided in methods that will be hit by every RPC call. */ +@InterfaceAudience.Private public interface MonitoredRPCHandler extends MonitoredTask { public abstract String getRPC(); public abstract String getRPC(boolean withParams); diff --git a/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java b/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java index d68468bc812..c815080858a 100644 --- a/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java +++ b/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredRPCHandlerImpl.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.monitoring; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.client.Operation; import org.apache.hadoop.hbase.io.WritableWithSize; import org.apache.hadoop.hbase.util.Bytes; @@ -36,6 +37,7 @@ import java.util.Map; * handling frequent, short duration tasks. String concatenations and object * allocations are avoided in methods that will be hit by every RPC call. */ +@InterfaceAudience.Private public class MonitoredRPCHandlerImpl extends MonitoredTaskImpl implements MonitoredRPCHandler { private String clientAddress; diff --git a/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java b/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java index c59d4153797..77db07af8e2 100644 --- a/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java +++ b/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTask.java @@ -22,6 +22,9 @@ package org.apache.hadoop.hbase.monitoring; import java.io.IOException; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public interface MonitoredTask extends Cloneable { enum State { RUNNING, diff --git a/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java b/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java index 394129c8150..5191bee3c15 100644 --- a/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java +++ b/src/main/java/org/apache/hadoop/hbase/monitoring/MonitoredTaskImpl.java @@ -19,12 +19,14 @@ */ package org.apache.hadoop.hbase.monitoring; +import org.apache.hadoop.classification.InterfaceAudience; import org.codehaus.jackson.map.ObjectMapper; import java.io.IOException; import java.util.HashMap; import java.util.Map; +@InterfaceAudience.Private class MonitoredTaskImpl implements MonitoredTask { private long startTime; private long statusTime; diff --git a/src/main/java/org/apache/hadoop/hbase/monitoring/StateDumpServlet.java b/src/main/java/org/apache/hadoop/hbase/monitoring/StateDumpServlet.java index 604f10d9a50..d017944b7fb 100644 --- a/src/main/java/org/apache/hadoop/hbase/monitoring/StateDumpServlet.java +++ b/src/main/java/org/apache/hadoop/hbase/monitoring/StateDumpServlet.java @@ -26,10 +26,12 @@ import java.util.Map; import javax.servlet.http.HttpServlet; import javax.servlet.http.HttpServletRequest; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.executor.ExecutorService; import org.apache.hadoop.hbase.executor.ExecutorService.ExecutorStatus; import org.apache.hadoop.hbase.util.VersionInfo; +@InterfaceAudience.Private public abstract class StateDumpServlet extends HttpServlet { static final long DEFAULT_TAIL_KB = 100; private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/monitoring/TaskMonitor.java b/src/main/java/org/apache/hadoop/hbase/monitoring/TaskMonitor.java index fc9c8301e47..bdb754a5795 100644 --- a/src/main/java/org/apache/hadoop/hbase/monitoring/TaskMonitor.java +++ b/src/main/java/org/apache/hadoop/hbase/monitoring/TaskMonitor.java @@ -30,6 +30,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import com.google.common.annotations.VisibleForTesting; import com.google.common.collect.Lists; @@ -39,6 +40,7 @@ import com.google.common.collect.Lists; * A Task here is anything which takes more than a few seconds * and the user might want to inquire about the status */ +@InterfaceAudience.Private public class TaskMonitor { private static final Log LOG = LogFactory.getLog(TaskMonitor.class); diff --git a/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java b/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java index a3fa7064ef5..a3ff4b21ccd 100644 --- a/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java +++ b/src/main/java/org/apache/hadoop/hbase/monitoring/ThreadMonitoring.java @@ -23,6 +23,9 @@ import java.lang.management.ManagementFactory; import java.lang.management.ThreadInfo; import java.lang.management.ThreadMXBean; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public abstract class ThreadMonitoring { private static final ThreadMXBean threadBean = diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java index 82894e2dbb7..ffa73063563 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ChangedReadersObserver.java @@ -22,10 +22,13 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + /** * If set of MapFile.Readers in Store change, implementors are notified. */ +@InterfaceAudience.Private public interface ChangedReadersObserver { /** * Notify observers. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnCount.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnCount.java index a617d687fa9..6cb3dfe73fd 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnCount.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnCount.java @@ -19,11 +19,14 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Simple wrapper for a byte buffer and a counter. Does not copy. *

    * NOT thread-safe because it is not used in a multi-threaded context, yet. */ +@InterfaceAudience.Private public class ColumnCount { private final byte [] bytes; private final int offset; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java index 2eeaab18b40..1aca5e9f519 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ColumnTracker.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.regionserver.ScanQueryMatcher.MatchCode; /** @@ -40,6 +41,7 @@ import org.apache.hadoop.hbase.regionserver.ScanQueryMatcher.MatchCode; *

    * This class is NOT thread-safe as queries are never multi-threaded */ +@InterfaceAudience.Private public interface ColumnTracker { /** * Keeps track of the number of versions for the columns asked for diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java b/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java index 066ae528348..2913c2bcb09 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/CompactSplitThread.java @@ -28,6 +28,7 @@ import java.util.concurrent.TimeUnit; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HTableDescriptor; @@ -38,6 +39,7 @@ import com.google.common.base.Preconditions; /** * Compact region on request and then run split if appropriate */ +@InterfaceAudience.Private public class CompactSplitThread implements CompactionRequestor { static final Log LOG = LogFactory.getLog(CompactSplitThread.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java b/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java index 3fe928663ad..1c9bb726ab3 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/CompactionRequestor.java @@ -19,6 +19,9 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public interface CompactionRequestor { /** * @param r Region to compact diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java index e0c27f099e4..9dcde7c2acc 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy.java @@ -17,6 +17,7 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; /** @@ -25,6 +26,7 @@ import org.apache.hadoop.hbase.HConstants; * size. *

    This is the default split policy.

    */ +@InterfaceAudience.Private public class ConstantSizeRegionSplitPolicy extends RegionSplitPolicy { private long desiredMaxFileSize; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/DebugPrint.java b/src/main/java/org/apache/hadoop/hbase/regionserver/DebugPrint.java index e1d69c7ce5d..2263a8a5242 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/DebugPrint.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/DebugPrint.java @@ -23,6 +23,9 @@ import java.io.FileWriter; import java.io.IOException; import java.util.concurrent.atomic.AtomicBoolean; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class DebugPrint { private static final AtomicBoolean enabled = new AtomicBoolean(false); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java b/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java index 29630845423..f4837e506db 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/DeleteTracker.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; + /** * This interface is used for the tracking and enforcement of Deletes * during the course of a Get or Scan operation. @@ -28,6 +30,7 @@ package org.apache.hadoop.hbase.regionserver; *
  • {@link #isDeleted} when checking if a Put KeyValue has been deleted *
  • {@link #update} when reaching the end of a StoreFile */ +@InterfaceAudience.Private public interface DeleteTracker { /** diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ExplicitColumnTracker.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ExplicitColumnTracker.java index 49375cf1cc1..cdd91233ae4 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ExplicitColumnTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ExplicitColumnTracker.java @@ -23,6 +23,7 @@ import java.util.ArrayList; import java.util.List; import java.util.NavigableSet; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.regionserver.ScanQueryMatcher.MatchCode; @@ -47,6 +48,7 @@ import org.apache.hadoop.hbase.util.Bytes; *

    * This class is NOT thread-safe as queries are never multi-threaded */ +@InterfaceAudience.Private public class ExplicitColumnTracker implements ColumnTracker { private final int maxVersions; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/FlushRequester.java b/src/main/java/org/apache/hadoop/hbase/regionserver/FlushRequester.java index b843c91f7c4..27672b24851 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/FlushRequester.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/FlushRequester.java @@ -20,9 +20,12 @@ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Request a flush. */ +@InterfaceAudience.Private public interface FlushRequester { /** * Tell the listener the cache needs to be flushed. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/GetClosestRowBeforeTracker.java b/src/main/java/org/apache/hadoop/hbase/regionserver/GetClosestRowBeforeTracker.java index 3a26bbb9f23..8174cf5e0cb 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/GetClosestRowBeforeTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/GetClosestRowBeforeTracker.java @@ -24,6 +24,7 @@ import java.util.NavigableSet; import java.util.TreeMap; import java.util.TreeSet; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue.KVComparator; @@ -35,6 +36,7 @@ import org.apache.hadoop.hbase.util.Bytes; * implement the {@link DeleteTracker} interface since state spans rows (There * is no update nor reset method). */ +@InterfaceAudience.Private class GetClosestRowBeforeTracker { private final KeyValue targetkey; // Any cell w/ a ts older than this is expired. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java index 16277280c94..76ff42253ba 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java @@ -59,6 +59,7 @@ import java.util.concurrent.locks.ReentrantReadWriteLock; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileStatus; @@ -166,6 +167,7 @@ import com.google.common.collect.MutableClassToInstanceMap; * regionName is a unique identifier for this HRegion. (startKey, endKey] * defines the keyspace for this HRegion. */ +@InterfaceAudience.Private public class HRegion implements HeapSize { // , Writable{ public static final Log LOG = LogFactory.getLog(HRegion.class); static final String MERGEDIR = "merges"; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java index 5309aaf771a..35205a520f1 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java @@ -55,6 +55,7 @@ import javax.management.ObjectName; import org.apache.commons.lang.mutable.MutableDouble; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -167,6 +168,7 @@ import com.google.common.collect.Lists; * HRegionServer makes a set of HRegions available to clients. It checks in with * the HMaster. There are many HRegionServers in a single HBase deployment. */ +@InterfaceAudience.Private public class HRegionServer implements HRegionInterface, HBaseRPCErrorHandler, Runnable, RegionServerServices { diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServerCommandLine.java b/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServerCommandLine.java index 71b9985e188..c0cdde5d325 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServerCommandLine.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServerCommandLine.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.regionserver; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.LocalHBaseCluster; @@ -31,6 +32,7 @@ import org.apache.hadoop.hbase.util.ServerCommandLine; * Class responsible for parsing the command line and starting the * RegionServer. */ +@InterfaceAudience.Private public class HRegionServerCommandLine extends ServerCommandLine { private static final Log LOG = LogFactory.getLog(HRegionServerCommandLine.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionThriftServer.java b/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionThriftServer.java index 1972f35cac8..68442fe2445 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionThriftServer.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionThriftServer.java @@ -26,6 +26,7 @@ import java.util.Map; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HRegionLocation; import org.apache.hadoop.hbase.KeyValue; @@ -48,6 +49,7 @@ import org.apache.hadoop.hbase.thrift.generated.TRowResult; *

    * This can be enabled with hbase.regionserver.export.thrift set to true. */ +@InterfaceAudience.Private public class HRegionThriftServer extends Thread { public static final Log LOG = LogFactory.getLog(HRegionThriftServer.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScan.java b/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScan.java index db2e02d80a8..4cc0e0179e3 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScan.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScan.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.client.Get; import org.apache.hadoop.hbase.client.Scan; @@ -33,6 +34,7 @@ import org.apache.hadoop.hbase.client.Scan; * {@link #checkOnlyMemStore()} or to only read from StoreFiles with * {@link #checkOnlyStoreFiles()}. */ +@InterfaceAudience.Private class InternalScan extends Scan { private boolean memOnly = false; private boolean filesOnly = false; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java b/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java index 0f5f36c0029..4e91743652a 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/InternalScanner.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import java.io.Closeable; @@ -39,6 +40,7 @@ import java.util.List; * all the family members, as they may have been last updated at different * times. */ +@InterfaceAudience.Private public interface InternalScanner extends Closeable { /** * Grab the next row's worth of values. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/KeyPrefixRegionSplitPolicy.java b/src/main/java/org/apache/hadoop/hbase/regionserver/KeyPrefixRegionSplitPolicy.java index bfd89542996..0cdcbd533c8 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/KeyPrefixRegionSplitPolicy.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/KeyPrefixRegionSplitPolicy.java @@ -21,6 +21,7 @@ import java.util.Arrays; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; /** * A custom RegionSplitPolicy implementing a SplitPolicy that groups @@ -29,6 +30,7 @@ import org.apache.commons.logging.LogFactory; * This ensures that a region is not split "inside" a prefix of a row key. * I.e. rows can be co-located in a regionb by their prefix. */ +@InterfaceAudience.Private public class KeyPrefixRegionSplitPolicy extends IncreasingToUpperBoundRegionSplitPolicy { private static final Log LOG = LogFactory .getLog(KeyPrefixRegionSplitPolicy.class); @@ -75,4 +77,4 @@ public class KeyPrefixRegionSplitPolicy extends IncreasingToUpperBoundRegionSpli return splitPoint; } } -} \ No newline at end of file +} diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueHeap.java b/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueHeap.java index 87883a0805b..0d47d147872 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueHeap.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueHeap.java @@ -25,6 +25,7 @@ import java.util.Comparator; import java.util.List; import java.util.PriorityQueue; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue.KVComparator; @@ -40,6 +41,7 @@ import org.apache.hadoop.hbase.KeyValue.KVComparator; * also implements InternalScanner. WARNING: As is, if you try to use this * as an InternalScanner at the Store level, you will get runtime exceptions. */ +@InterfaceAudience.Private public class KeyValueHeap extends NonLazyKeyValueScanner implements KeyValueScanner, InternalScanner { private PriorityQueue heap = null; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java b/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java index 6a7d5c68b96..52c4b4f4629 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueScanner.java @@ -22,12 +22,14 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; import java.util.SortedSet; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Scan; /** * Scanner that returns the next KeyValue. */ +@InterfaceAudience.Private public interface KeyValueScanner { /** * Look at the next KeyValue in this scanner, but do not iterate scanner. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueSkipListSet.java b/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueSkipListSet.java index 51df1ee2a03..61aae05ac99 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueSkipListSet.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/KeyValueSkipListSet.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import java.util.Collection; @@ -43,6 +44,7 @@ import java.util.concurrent.ConcurrentSkipListMap; * has same attributes as ConcurrentSkipListSet: e.g. tolerant of concurrent * get and set and won't throw ConcurrentModificationException when iterating. */ +@InterfaceAudience.Private class KeyValueSkipListSet implements NavigableSet { private final ConcurrentNavigableMap delegatee; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseException.java b/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseException.java index cafbb28496d..ff1152a246e 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseException.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseException.java @@ -19,11 +19,13 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.DoNotRetryIOException; /** * Reports a problem with a lease */ +@InterfaceAudience.Private public class LeaseException extends DoNotRetryIOException { private static final long serialVersionUID = 8179703995292418650L; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java b/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java index a8437364a9b..0faa48e925a 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/LeaseListener.java @@ -19,6 +19,8 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; + /** * LeaseListener is an interface meant to be implemented by users of the Leases @@ -28,6 +30,7 @@ package org.apache.hadoop.hbase.regionserver; * lease. Users of the Leases class can use a LeaseListener subclass to, for * example, clean up resources after a lease has expired. */ +@InterfaceAudience.Private public interface LeaseListener { /** When a lease expires, this method is called. */ public void leaseExpired(); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/Leases.java b/src/main/java/org/apache/hadoop/hbase/regionserver/Leases.java index c5185210d47..b520f3ff040 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/Leases.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/Leases.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.HasThread; import java.util.ConcurrentModificationException; @@ -51,6 +52,7 @@ import java.io.IOException; * can be interrupted when there is something to do, rather than the Chore * sleep time which is invariant. */ +@InterfaceAudience.Private public class Leases extends HasThread { private static final Log LOG = LogFactory.getLog(Leases.class.getName()); private final int leasePeriod; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/LogRoller.java b/src/main/java/org/apache/hadoop/hbase/regionserver/LogRoller.java index 74429a19bea..4e1b491a758 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/LogRoller.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/LogRoller.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.*; import org.apache.hadoop.hbase.regionserver.wal.FailedLogCloseException; @@ -41,6 +42,7 @@ import java.util.concurrent.locks.ReentrantLock; * can be interrupted when there is something to do, rather than the Chore * sleep time which is invariant. */ +@InterfaceAudience.Private class LogRoller extends HasThread implements WALActionsListener { static final Log LOG = LogFactory.getLog(LogRoller.class); private final ReentrantLock rollLock = new ReentrantLock(); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/LruHashMap.java b/src/main/java/org/apache/hadoop/hbase/regionserver/LruHashMap.java index 161ae189e8c..f87a6bf9c6d 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/LruHashMap.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/LruHashMap.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.HeapSize; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.ClassSize; @@ -48,6 +49,7 @@ import java.util.Set; *

    * This class contains internal synchronization and is thread-safe. */ +@InterfaceAudience.Private public class LruHashMap implements HeapSize, Map { diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/MemStore.java b/src/main/java/org/apache/hadoop/hbase/regionserver/MemStore.java index 03ed7f10079..0592f400c67 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/MemStore.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/MemStore.java @@ -34,6 +34,7 @@ import java.util.concurrent.locks.ReentrantReadWriteLock; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HConstants; @@ -55,6 +56,7 @@ import org.apache.hadoop.hbase.util.ClassSize; * TODO: With new KVSLS, need to make sure we update HeapSize with difference * in KV size. */ +@InterfaceAudience.Private public class MemStore implements HeapSize { private static final Log LOG = LogFactory.getLog(MemStore.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java b/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java index beae44331b6..e529ff348a5 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreFlusher.java @@ -37,6 +37,7 @@ import java.util.concurrent.locks.ReentrantLock; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.DroppedSnapshotException; import org.apache.hadoop.hbase.HConstants; @@ -56,6 +57,7 @@ import com.google.common.base.Preconditions; * * @see FlushRequester */ +@InterfaceAudience.Private class MemStoreFlusher extends HasThread implements FlushRequester { static final Log LOG = LogFactory.getLog(MemStoreFlusher.class); // These two data members go together. Any entry in the one must have diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreLAB.java b/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreLAB.java index cbb76e8be09..26948978544 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreLAB.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/MemStoreLAB.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.regionserver; import java.util.concurrent.atomic.AtomicInteger; import java.util.concurrent.atomic.AtomicReference; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import com.google.common.base.Preconditions; @@ -46,6 +47,7 @@ import com.google.common.base.Preconditions; * Bytes.toLong/Bytes.toInt calls in KeyValue, but some of those are cached * anyway */ +@InterfaceAudience.Private public class MemStoreLAB { private AtomicReference curChunk = new AtomicReference(); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/MultiVersionConsistencyControl.java b/src/main/java/org/apache/hadoop/hbase/regionserver/MultiVersionConsistencyControl.java index 6b28f03436f..235fb413c56 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/MultiVersionConsistencyControl.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/MultiVersionConsistencyControl.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver; import java.util.LinkedList; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.ClassSize; @@ -33,6 +34,7 @@ import org.apache.commons.logging.Log; * a mechanism for writers to obtain new write numbers, then "commit" * the new writes for readers to read (thus forming atomic transactions). */ +@InterfaceAudience.Private public class MultiVersionConsistencyControl { private volatile long memstoreRead = 0; private volatile long memstoreWrite = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/NoSuchColumnFamilyException.java b/src/main/java/org/apache/hadoop/hbase/regionserver/NoSuchColumnFamilyException.java index 4881fc0f1a2..6b207189201 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/NoSuchColumnFamilyException.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/NoSuchColumnFamilyException.java @@ -19,11 +19,13 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.DoNotRetryIOException; /** * Thrown if request for nonexistent column family. */ +@InterfaceAudience.Private public class NoSuchColumnFamilyException extends DoNotRetryIOException { private static final long serialVersionUID = -6569952730832331274L; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/NonLazyKeyValueScanner.java b/src/main/java/org/apache/hadoop/hbase/regionserver/NonLazyKeyValueScanner.java index 6534e2c754f..4c639bb2678 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/NonLazyKeyValueScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/NonLazyKeyValueScanner.java @@ -23,6 +23,7 @@ import java.io.IOException; import java.util.SortedSet; import org.apache.commons.lang.NotImplementedException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Scan; @@ -30,6 +31,7 @@ import org.apache.hadoop.hbase.client.Scan; * A "non-lazy" scanner which always does a real seek operation. Most scanners * are inherited from this class. */ +@InterfaceAudience.Private public abstract class NonLazyKeyValueScanner implements KeyValueScanner { @Override diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java b/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java index e587380af42..da9e5cfc6a1 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/OnlineRegions.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Server; import java.io.IOException; @@ -28,6 +29,7 @@ import java.util.List; * Interface to Map of online regions. In the Map, the key is the region's * encoded name and the value is an {@link HRegion} instance. */ +@InterfaceAudience.Private interface OnlineRegions extends Server { /** * Add to online regions. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/OperationStatus.java b/src/main/java/org/apache/hadoop/hbase/regionserver/OperationStatus.java index 1b94ab5540e..fb1771045d4 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/OperationStatus.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/OperationStatus.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants.OperationStatusCode; /** * @@ -28,6 +29,7 @@ import org.apache.hadoop.hbase.HConstants.OperationStatusCode; * the operation status in future. * */ +@InterfaceAudience.Private public class OperationStatus { /** Singleton for successful operations. */ diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RSDumpServlet.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RSDumpServlet.java index df33d82070d..6ea72d242a1 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RSDumpServlet.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RSDumpServlet.java @@ -27,12 +27,14 @@ import java.util.Date; import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletResponse; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.monitoring.LogMonitoring; import org.apache.hadoop.hbase.monitoring.StateDumpServlet; import org.apache.hadoop.hbase.monitoring.TaskMonitor; import org.apache.hadoop.util.ReflectionUtils; +@InterfaceAudience.Private public class RSDumpServlet extends StateDumpServlet { private static final long serialVersionUID = 1L; private static final String LINE = diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RSStatusServlet.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RSStatusServlet.java index 7521cd4c6e1..3d97c7650b1 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RSStatusServlet.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RSStatusServlet.java @@ -25,9 +25,11 @@ import javax.servlet.ServletException; import javax.servlet.http.HttpServlet; import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletResponse; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.tmpl.regionserver.RSStatusTmpl; +@InterfaceAudience.Private public class RSStatusServlet extends HttpServlet { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionAlreadyInTransitionException.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionAlreadyInTransitionException.java index 1c21825f8a0..870d81dd3a5 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionAlreadyInTransitionException.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionAlreadyInTransitionException.java @@ -21,10 +21,16 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * This exception is thrown when a region server is asked to open or close * a region but it's already processing it */ +@SuppressWarnings("serial") +@InterfaceAudience.Public +@InterfaceStability.Stable public class RegionAlreadyInTransitionException extends IOException { public RegionAlreadyInTransitionException(String s) { diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionOpeningState.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionOpeningState.java index c5bcb4c8cbb..7c6200aa3e3 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionOpeningState.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionOpeningState.java @@ -19,6 +19,9 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public enum RegionOpeningState { OPENED, diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java index b10aecafa7e..3bc40b8a7a2 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionScanner.java @@ -19,11 +19,13 @@ */ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; /** * RegionScanner describes iterators over rows in an HRegion. */ +@InterfaceAudience.Private public interface RegionScanner extends InternalScanner { /** * @return The RegionInfo for this scanner. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerAccounting.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerAccounting.java index 05c842ecb4e..c145e0aa149 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerAccounting.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerAccounting.java @@ -21,10 +21,13 @@ package org.apache.hadoop.hbase.regionserver; import java.util.concurrent.atomic.AtomicLong; +import org.apache.hadoop.classification.InterfaceAudience; + /** * RegionServerAccounting keeps record of some basic real time information about * the Region Server. Currently, it only keeps record the global memstore size. */ +@InterfaceAudience.Private public class RegionServerAccounting { private final AtomicLong atomicGlobalMemstoreSize = new AtomicLong(0); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerRunningException.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerRunningException.java index ed36ed79abc..8045c510d32 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerRunningException.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerRunningException.java @@ -21,10 +21,13 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Thrown if the region server log directory exists (which indicates another * region server is running at the same address) */ +@InterfaceAudience.Private public class RegionServerRunningException extends IOException { private static final long serialVersionUID = 1L << 31 - 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java index 984ef479da4..9b7e88b34e1 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerServices.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.catalog.CatalogTracker; import org.apache.hadoop.hbase.ipc.RpcServer; import org.apache.hadoop.hbase.regionserver.wal.HLog; @@ -30,6 +31,7 @@ import org.apache.zookeeper.KeeperException; /** * Services provided by {@link HRegionServer} */ +@InterfaceAudience.Private public interface RegionServerServices extends OnlineRegions { /** * @return True if this regionserver is stopping. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerStoppedException.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerStoppedException.java index 45acb177ba2..1dd5f8d3a55 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerStoppedException.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionServerStoppedException.java @@ -21,10 +21,13 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Thrown by the region server when it is in shutting down state. */ @SuppressWarnings("serial") +@InterfaceAudience.Private public class RegionServerStoppedException extends IOException { public RegionServerStoppedException(String s) { diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionSplitPolicy.java b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionSplitPolicy.java index 9957abdcee7..f299f856d85 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/RegionSplitPolicy.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/RegionSplitPolicy.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; import org.apache.hadoop.hbase.HConstants; @@ -32,6 +33,7 @@ import com.google.common.base.Preconditions; * A split policy determines when a region should be split. * {@see ConstantSizeRegionSplitPolicy} */ +@InterfaceAudience.Private public abstract class RegionSplitPolicy extends Configured { private static final Class DEFAULT_SPLIT_POLICY_CLASS = IncreasingToUpperBoundRegionSplitPolicy.class; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java index 83ffcc70d9a..da43c8333b0 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationService.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; import org.apache.hadoop.hbase.Server; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -31,6 +32,7 @@ import org.apache.hadoop.fs.Path; * One such application is a cross-datacenter * replication service that can keep two hbase clusters in sync. */ +@InterfaceAudience.Private public interface ReplicationService { /** diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java index 1d221925e51..f245fb1ac14 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSinkService.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.regionserver.wal.HLog; /** @@ -28,6 +29,7 @@ import org.apache.hadoop.hbase.regionserver.wal.HLog; * This service allows an application to hook into the * regionserver and behave as a replication sink. */ +@InterfaceAudience.Private public interface ReplicationSinkService extends ReplicationService { /** diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java index 9e16d022b9d..0250be682e3 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ReplicationSourceService.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.regionserver.wal.WALActionsListener; /** @@ -28,6 +29,7 @@ import org.apache.hadoop.hbase.regionserver.wal.WALActionsListener; * This service allows an application to hook into the * regionserver and watch for new transactions. */ +@InterfaceAudience.Private public interface ReplicationSourceService extends ReplicationService { /** diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ScanDeleteTracker.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ScanDeleteTracker.java index b828669b4bc..ae0471f8d8b 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ScanDeleteTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ScanDeleteTracker.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.regionserver.DeleteTracker.DeleteResult; import org.apache.hadoop.hbase.util.Bytes; @@ -39,6 +40,7 @@ import org.apache.hadoop.hbase.util.Bytes; *

    * This class is NOT thread-safe as queries are never multi-threaded */ +@InterfaceAudience.Private public class ScanDeleteTracker implements DeleteTracker { private long familyStamp = -1L; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java index 52459f0a57d..001b1788a76 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java @@ -23,6 +23,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; import java.util.NavigableSet; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Scan; @@ -38,6 +39,7 @@ import org.apache.hadoop.hbase.regionserver.StoreScanner.ScanType; /** * A query matcher that is specifically designed for the scan case. */ +@InterfaceAudience.Private public class ScanQueryMatcher { // Optimization so we can skip lots of compares when we decide to skip // to the next row. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ScanWildcardColumnTracker.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ScanWildcardColumnTracker.java index 1402455668e..23e77f02d87 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ScanWildcardColumnTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ScanWildcardColumnTracker.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.regionserver.ScanQueryMatcher.MatchCode; @@ -30,6 +31,7 @@ import org.apache.hadoop.hbase.util.Bytes; /** * Keeps track of the columns for a scan if they are not explicitly specified */ +@InterfaceAudience.Private public class ScanWildcardColumnTracker implements ColumnTracker { private byte [] columnBuffer = null; private int columnOffset = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/ShutdownHook.java b/src/main/java/org/apache/hadoop/hbase/regionserver/ShutdownHook.java index 4ccf68e7de5..e3b230e05ed 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/ShutdownHook.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/ShutdownHook.java @@ -26,6 +26,7 @@ import java.util.Map; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -36,6 +37,7 @@ import org.apache.hadoop.hbase.util.Threads; * Manage regionserver shutdown hooks. * @see #install(Configuration, FileSystem, Stoppable, Thread) */ +@InterfaceAudience.Private public class ShutdownHook { private static final Log LOG = LogFactory.getLog(ShutdownHook.class); private static final String CLIENT_FINALIZER_DATA_METHOD = "clientFinalizer"; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java b/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java index 0f119e53ad2..2a0284a6cf7 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/SplitLogWorker.java @@ -28,6 +28,7 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -65,6 +66,7 @@ import org.apache.zookeeper.data.Stat; * is stripped of its ownership. Here we rely on the idempotency of the log * splitting task for correctness */ +@InterfaceAudience.Private public class SplitLogWorker extends ZooKeeperListener implements Runnable { private static final Log LOG = LogFactory.getLog(SplitLogWorker.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/SplitRequest.java b/src/main/java/org/apache/hadoop/hbase/regionserver/SplitRequest.java index 6bfcc906795..42a93eba193 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/SplitRequest.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/SplitRequest.java @@ -23,6 +23,7 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.RemoteExceptionHandler; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.util.StringUtils; @@ -32,6 +33,7 @@ import com.google.common.base.Preconditions; /** * Handles processing region splits. Put in a queue, owned by HRegionServer. */ +@InterfaceAudience.Private class SplitRequest implements Runnable { static final Log LOG = LogFactory.getLog(SplitRequest.class); private final HRegion parent; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java b/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java index cfaffd800c4..3f42efa07b8 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/SplitTransaction.java @@ -33,6 +33,7 @@ import java.util.concurrent.TimeUnit; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -82,6 +83,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; *

    This class is not thread safe. Caller needs ensure split is run by * one thread only. */ +@InterfaceAudience.Private public class SplitTransaction { private static final Log LOG = LogFactory.getLog(SplitTransaction.class); private static final String SPLITDIR = "splits"; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java b/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java index 3f2e390563a..d884598fa0a 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/Store.java @@ -40,6 +40,7 @@ import java.util.concurrent.locks.ReentrantReadWriteLock; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -104,6 +105,7 @@ import com.google.common.collect.Lists; *

    Locking and transactions are handled at a higher level. This API should * not be called directly but by an HRegion manager. */ +@InterfaceAudience.Private public class Store extends SchemaConfigured implements HeapSize { static final Log LOG = LogFactory.getLog(Store.class); protected final MemStore memstore; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java b/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java index 4a4cc6763fb..12ebc0a554c 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java @@ -36,6 +36,7 @@ import java.util.regex.Pattern; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -86,6 +87,7 @@ import com.google.common.collect.Ordering; * The reason for this weird pattern where you use a different instance for the * writer and a reader is that we write once but read a lot more. */ +@InterfaceAudience.Private public class StoreFile extends SchemaConfigured { static final Log LOG = LogFactory.getLog(StoreFile.class.getName()); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileScanner.java b/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileScanner.java index 49f8e813c38..6cb6640b253 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileScanner.java @@ -29,6 +29,7 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.io.hfile.HFileScanner; @@ -38,6 +39,7 @@ import org.apache.hadoop.hbase.regionserver.StoreFile.Reader; * KeyValueScanner adaptor over the Reader. It also provides hooks into * bloom filter things. */ +@InterfaceAudience.Private class StoreFileScanner implements KeyValueScanner { static final Log LOG = LogFactory.getLog(Store.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFlusher.java b/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFlusher.java index d2eb697bc8a..c42c7d5cfc4 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFlusher.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFlusher.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.monitoring.MonitoredTask; /** @@ -28,6 +29,7 @@ import org.apache.hadoop.hbase.monitoring.MonitoredTask; * A store flusher carries the state required to prepare/flush/commit the * store's cache. */ +@InterfaceAudience.Private interface StoreFlusher { /** diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java b/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java index f492f00b238..6a4a440f0a7 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java @@ -27,6 +27,7 @@ import java.util.NavigableSet; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; @@ -40,6 +41,7 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; * Scanner scans both the memstore and the HStore. Coalesce KeyValue stream * into List for a single row. */ +@InterfaceAudience.Private class StoreScanner extends NonLazyKeyValueScanner implements KeyValueScanner, InternalScanner, ChangedReadersObserver { static final Log LOG = LogFactory.getLog(StoreScanner.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/TimeRangeTracker.java b/src/main/java/org/apache/hadoop/hbase/regionserver/TimeRangeTracker.java index ec028d1d1a8..e5e025f8de7 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/TimeRangeTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/TimeRangeTracker.java @@ -23,6 +23,7 @@ import java.io.DataInput; import java.io.DataOutput; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValue.Type; import org.apache.hadoop.hbase.io.TimeRange; @@ -36,6 +37,7 @@ import org.apache.hadoop.io.Writable; * When writing StoreFiles, this information is stored in meta blocks and used * at read time to match against the required TimeRange. */ +@InterfaceAudience.Private public class TimeRangeTracker implements Writable { long minimumTimestamp = -1; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/WrongRegionException.java b/src/main/java/org/apache/hadoop/hbase/regionserver/WrongRegionException.java index 52b9a6c05db..953e785ba51 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/WrongRegionException.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/WrongRegionException.java @@ -21,9 +21,14 @@ package org.apache.hadoop.hbase.regionserver; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Thrown when a request contains a key which is not part of this region */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class WrongRegionException extends IOException { private static final long serialVersionUID = 993179627856392526L; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactSelection.java b/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactSelection.java index 26338e17eef..804adc4762e 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactSelection.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactSelection.java @@ -26,9 +26,11 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.regionserver.StoreFile; +@InterfaceAudience.Private public class CompactSelection { private static final long serialVersionUID = 1L; static final Log LOG = LogFactory.getLog(CompactSelection.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionProgress.java b/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionProgress.java index 9bc66e1413f..553eee0a9cc 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionProgress.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionProgress.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.regionserver.compactions; +import org.apache.hadoop.classification.InterfaceAudience; + /** * This class holds information relevant for tracking the progress of a * compaction. @@ -29,6 +31,7 @@ package org.apache.hadoop.hbase.regionserver.compactions; * total amount scheduled to be compacted. * */ +@InterfaceAudience.Private public class CompactionProgress { /** the total compacting key values in currently running compaction */ diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionRequest.java b/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionRequest.java index 7544b7144fb..fedc6f003fe 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionRequest.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/compactions/CompactionRequest.java @@ -26,6 +26,7 @@ import java.util.concurrent.ThreadPoolExecutor; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.RemoteExceptionHandler; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.regionserver.HRegionServer; @@ -43,6 +44,7 @@ import com.google.common.collect.Collections2; /** * This class holds all details necessary to run a compaction. */ +@InterfaceAudience.Private public class CompactionRequest implements Comparable, Runnable { static final Log LOG = LogFactory.getLog(CompactionRequest.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseMetaHandler.java b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseMetaHandler.java index 83544e2f0e9..6b984b484bc 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseMetaHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseMetaHandler.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver.handler; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.regionserver.RegionServerServices; @@ -26,6 +27,7 @@ import org.apache.hadoop.hbase.regionserver.RegionServerServices; /** * Handles closing of the root region on a region server. */ +@InterfaceAudience.Private public class CloseMetaHandler extends CloseRegionHandler { // Called when master tells us shutdown a region via close rpc public CloseMetaHandler(final Server server, diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseRegionHandler.java b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseRegionHandler.java index 78eb4e61912..082bd3c7bb5 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseRegionHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseRegionHandler.java @@ -23,6 +23,7 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.executor.EventHandler; @@ -34,6 +35,7 @@ import org.apache.zookeeper.KeeperException; /** * Handles closing of a region on a region server. */ +@InterfaceAudience.Private public class CloseRegionHandler extends EventHandler { // NOTE on priorities shutting down. There are none for close. There are some // for open. I think that is right. On shutdown, we want the meta to close diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseRootHandler.java b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseRootHandler.java index 4cfeec6549d..e2f47820664 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseRootHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/CloseRootHandler.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver.handler; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.regionserver.RegionServerServices; @@ -26,6 +27,7 @@ import org.apache.hadoop.hbase.regionserver.RegionServerServices; /** * Handles closing of the root region on a region server. */ +@InterfaceAudience.Private public class CloseRootHandler extends CloseRegionHandler { // This is executed after receiving an CLOSE RPC from the master for root. public CloseRootHandler(final Server server, diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenMetaHandler.java b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenMetaHandler.java index 66e57069d15..43974a33dcd 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenMetaHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenMetaHandler.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver.handler; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.Server; @@ -29,6 +30,7 @@ import org.apache.hadoop.hbase.regionserver.RegionServerServices; *

    * This is executed after receiving an OPEN RPC from the master for meta. */ +@InterfaceAudience.Private public class OpenMetaHandler extends OpenRegionHandler { public OpenMetaHandler(final Server server, final RegionServerServices rsServices, HRegionInfo regionInfo, diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.java b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.java index 8ba723d30ba..9bfd45ebaf7 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenRegionHandler.java @@ -24,6 +24,7 @@ import java.util.concurrent.atomic.AtomicBoolean; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.Server; @@ -39,6 +40,7 @@ import org.apache.zookeeper.KeeperException; *

    * This is executed after receiving an OPEN RPC from the master or client. */ +@InterfaceAudience.Private public class OpenRegionHandler extends EventHandler { private static final Log LOG = LogFactory.getLog(OpenRegionHandler.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenRootHandler.java b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenRootHandler.java index 9a4f01a838f..3090c758ad2 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenRootHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/handler/OpenRootHandler.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver.handler; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.Server; @@ -29,6 +30,7 @@ import org.apache.hadoop.hbase.regionserver.RegionServerServices; *

    * This is executed after receiving an OPEN RPC from the master for root. */ +@InterfaceAudience.Private public class OpenRootHandler extends OpenRegionHandler { public OpenRootHandler(final Server server, final RegionServerServices rsServices, HRegionInfo regionInfo, diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerDynamicMetrics.java b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerDynamicMetrics.java index c34874c86c1..3b46dd807ca 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerDynamicMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerDynamicMetrics.java @@ -27,6 +27,7 @@ import java.util.concurrent.atomic.AtomicInteger; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.regionserver.HRegion; import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.metrics.MetricsContext; @@ -49,6 +50,7 @@ import org.apache.hadoop.metrics.util.MetricsTimeVaryingRate; * for example: rpcQueueTime.inc(time) * */ +@InterfaceAudience.Private public class RegionServerDynamicMetrics implements Updater { private MetricsRecord metricsRecord; private MetricsContext context; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerDynamicStatistics.java b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerDynamicStatistics.java index c250d817053..aea3a10c611 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerDynamicStatistics.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerDynamicStatistics.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.regionserver.metrics; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.metrics.util.MBeanUtil; import org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase; import org.apache.hadoop.metrics.util.MetricsRegistry; @@ -31,6 +32,7 @@ import javax.management.ObjectName; * {@link RegionServerDynamicMetrics} as an MBean * for JMX monitoring. */ +@InterfaceAudience.Private public class RegionServerDynamicStatistics extends MetricsDynamicMBeanBase { private final ObjectName mbeanName; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerMetrics.java b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerMetrics.java index e3fcfdf3062..ad9345998df 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerMetrics.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.regionserver.metrics; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.hfile.HFile; import org.apache.hadoop.hbase.metrics.HBaseInfo; import org.apache.hadoop.hbase.metrics.MetricsRate; @@ -53,6 +54,7 @@ import java.util.List; * This class has a number of metrics variables that are publicly accessible; * these variables (objects) have methods to update their values. */ +@InterfaceAudience.Private public class RegionServerMetrics implements Updater { @SuppressWarnings({"FieldCanBeLocal"}) private final Log LOG = LogFactory.getLog(this.getClass()); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerStatistics.java b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerStatistics.java index 04fe7b10e44..0145e76ecf1 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerStatistics.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/RegionServerStatistics.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.regionserver.metrics; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.metrics.MetricsMBeanBase; import org.apache.hadoop.metrics.util.MBeanUtil; import org.apache.hadoop.metrics.util.MetricsRegistry; @@ -29,6 +30,7 @@ import javax.management.ObjectName; * Exports metrics recorded by {@link RegionServerMetrics} as an MBean * for JMX monitoring. */ +@InterfaceAudience.Private public class RegionServerStatistics extends MetricsMBeanBase { private final ObjectName mbeanName; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/SchemaConfigured.java b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/SchemaConfigured.java index 4821b53b233..945d8024c93 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/SchemaConfigured.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/SchemaConfigured.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.regionserver.metrics; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.io.HeapSize; @@ -37,6 +38,7 @@ import org.apache.hadoop.hbase.util.ClassSize; * from another similar object. For example, an HFile reader configures HFile * blocks it reads with its own table/CF name. */ +@InterfaceAudience.Private public class SchemaConfigured implements HeapSize, SchemaAware { private static final Log LOG = LogFactory.getLog(SchemaConfigured.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/SchemaMetrics.java b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/SchemaMetrics.java index b70766c3835..5ad74062759 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/SchemaMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/metrics/SchemaMetrics.java @@ -35,6 +35,7 @@ import java.util.regex.Pattern; import org.apache.commons.lang.mutable.MutableDouble; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.io.hfile.BlockType.BlockCategory; @@ -86,6 +87,7 @@ import org.apache.hadoop.hbase.util.Pair; * * */ +@InterfaceAudience.Private public class SchemaMetrics { public interface SchemaAware { diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/FailedLogCloseException.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/FailedLogCloseException.java index 393b1d237dc..755bb8d0cc9 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/FailedLogCloseException.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/FailedLogCloseException.java @@ -21,10 +21,13 @@ package org.apache.hadoop.hbase.regionserver.wal; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Thrown when we fail close of the write-ahead-log file. * Package private. Only used inside this package. */ +@InterfaceAudience.Private public class FailedLogCloseException extends IOException { private static final long serialVersionUID = 1759152841462990925L; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java index c945a99f9dc..b5049b12ceb 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLog.java @@ -50,6 +50,7 @@ import java.util.regex.Pattern; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileStatus; @@ -110,6 +111,7 @@ import org.apache.hadoop.util.StringUtils; * org.apache.hadoop.fs.Path, org.apache.hadoop.conf.Configuration)}. * */ +@InterfaceAudience.Private public class HLog implements Syncable { static final Log LOG = LogFactory.getLog(HLog.class); public static final byte [] METAFAMILY = Bytes.toBytes("METAFAMILY"); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java index f067221a258..311ea1bcc00 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogKey.java @@ -27,6 +27,7 @@ import java.util.HashMap; import java.util.Map; import java.util.UUID; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.io.WritableComparable; @@ -42,6 +43,7 @@ import org.apache.hadoop.io.WritableUtils; *

    Some Transactional edits (START, COMMIT, ABORT) will not have an * associated row. */ +@InterfaceAudience.Private public class HLogKey implements WritableComparable { // should be < 0 (@see #readFields(DataInput)) private static final int VERSION = -1; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogPrettyPrinter.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogPrettyPrinter.java index d0e85f35342..0acc7cbaa66 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogPrettyPrinter.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogPrettyPrinter.java @@ -32,6 +32,8 @@ import org.apache.commons.cli.HelpFormatter; import org.apache.commons.cli.Options; import org.apache.commons.cli.ParseException; import org.apache.commons.cli.PosixParser; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -54,6 +56,8 @@ import org.codehaus.jackson.map.ObjectMapper; * It can also toggle output of values. * */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HLogPrettyPrinter { private boolean outputValues; private boolean outputJSON; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogSplitter.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogSplitter.java index 193f15a0a57..5c8fc5e933b 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogSplitter.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/HLogSplitter.java @@ -37,6 +37,7 @@ import java.util.concurrent.atomic.AtomicReference; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -67,6 +68,7 @@ import com.google.common.collect.Lists; * files that are no longer being written to, into new files, one per region for * region to replay on startup. Delete the old log files when finished. */ +@InterfaceAudience.Private public class HLogSplitter { private static final String LOG_SPLITTER_IMPL = "hbase.hlog.splitter.impl"; diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/OrphanHLogAfterSplitException.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/OrphanHLogAfterSplitException.java index 1c93def668e..daa0947ab30 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/OrphanHLogAfterSplitException.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/OrphanHLogAfterSplitException.java @@ -21,6 +21,9 @@ package org.apache.hadoop.hbase.regionserver.wal; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class OrphanHLogAfterSplitException extends IOException { /** diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogReader.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogReader.java index d9cd6dead27..ff63a5fc2f9 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogReader.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogReader.java @@ -30,12 +30,14 @@ import java.util.Arrays; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.io.SequenceFile; +@InterfaceAudience.Private public class SequenceFileLogReader implements HLog.Reader { private static final Log LOG = LogFactory.getLog(SequenceFileLogReader.class); diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogWriter.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogWriter.java index bd31ead1af5..01ebb5c07fd 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogWriter.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/SequenceFileLogWriter.java @@ -28,6 +28,7 @@ import java.lang.reflect.Method; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; @@ -42,6 +43,7 @@ import org.apache.hadoop.io.compress.DefaultCodec; * Implementation of {@link HLog.Writer} that delegates to * SequenceFile.Writer. */ +@InterfaceAudience.Private public class SequenceFileLogWriter implements HLog.Writer { private final Log LOG = LogFactory.getLog(this.getClass()); // The sequence file we delegate to. diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java index 3cdf1bc03e0..08e86271710 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALActionsListener.java @@ -20,6 +20,8 @@ package org.apache.hadoop.hbase.regionserver.wal; import java.io.IOException; + +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; @@ -28,6 +30,7 @@ import org.apache.hadoop.hbase.HTableDescriptor; * Get notification of {@link HLog}/WAL log events. The invocations are inline * so make sure your implementation is fast else you'll slow hbase. */ +@InterfaceAudience.Private public interface WALActionsListener { /** diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALCoprocessorHost.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALCoprocessorHost.java index b14e190aa7f..3db016dd2b9 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALCoprocessorHost.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALCoprocessorHost.java @@ -26,12 +26,14 @@ import java.io.IOException; import org.apache.hadoop.hbase.Coprocessor; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.coprocessor.*; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; /** * Implements the coprocessor environment and runtime support for coprocessors * loaded within a {@link HLog}. */ +@InterfaceAudience.Private public class WALCoprocessorHost extends CoprocessorHost { diff --git a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALEdit.java b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALEdit.java index e1117ef6ea2..d8f317cac6e 100644 --- a/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALEdit.java +++ b/src/main/java/org/apache/hadoop/hbase/regionserver/wal/WALEdit.java @@ -27,6 +27,7 @@ import java.util.List; import java.util.NavigableMap; import java.util.TreeMap; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.HeapSize; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.util.Bytes; @@ -67,6 +68,7 @@ import org.apache.hadoop.io.Writable; * is an old style KeyValue or the new style WALEdit. * */ +@InterfaceAudience.Private public class WALEdit implements Writable, HeapSize { private final int VERSION_2 = -1; diff --git a/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeer.java b/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeer.java index 6495207bf7e..fbf66afc26f 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeer.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeer.java @@ -26,6 +26,7 @@ import java.util.concurrent.atomic.AtomicBoolean; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.ServerName; @@ -36,6 +37,7 @@ import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher; * communicate with remote peers and is responsible for answering to expired * sessions and re-establishing the ZK connections. */ +@InterfaceAudience.Private public class ReplicationPeer implements Abortable { private static final Log LOG = LogFactory.getLog(ReplicationPeer.class); diff --git a/src/main/java/org/apache/hadoop/hbase/replication/ReplicationZookeeper.java b/src/main/java/org/apache/hadoop/hbase/replication/ReplicationZookeeper.java index 8c9cb9b6925..9ef176ac6e8 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/ReplicationZookeeper.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/ReplicationZookeeper.java @@ -33,6 +33,7 @@ import java.util.concurrent.atomic.AtomicBoolean; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.HConstants; @@ -74,6 +75,7 @@ import org.apache.zookeeper.KeeperException.SessionExpiredException; * ... * */ +@InterfaceAudience.Private public class ReplicationZookeeper { private static final Log LOG = LogFactory.getLog(ReplicationZookeeper.class); diff --git a/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java b/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java index 05d03104ff2..cafde09b16b 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.replication.master; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.Abortable; @@ -40,6 +41,7 @@ import java.util.Set; * Implementation of a log cleaner that checks if a log is still scheduled for * replication before deleting it when its TTL is over. */ +@InterfaceAudience.Private public class ReplicationLogCleaner implements LogCleanerDelegate, Abortable { private static final Log LOG = LogFactory.getLog(ReplicationLogCleaner.class); private Configuration conf; diff --git a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/Replication.java b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/Replication.java index 464de865b08..250ea86f0b2 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/Replication.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/Replication.java @@ -24,6 +24,7 @@ import java.util.NavigableMap; import java.util.TreeMap; import java.util.concurrent.atomic.AtomicBoolean; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -49,6 +50,7 @@ import static org.apache.hadoop.hbase.HConstants.REPLICATION_SCOPE_LOCAL; /** * Gateway to Replication. Used by {@link org.apache.hadoop.hbase.regionserver.HRegionServer}. */ +@InterfaceAudience.Private public class Replication implements WALActionsListener, ReplicationSourceService, ReplicationSinkService { private boolean replication; diff --git a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java index 9c3f387f364..fcedc18a3b4 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSink.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.replication.regionserver; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Delete; @@ -53,6 +54,7 @@ import java.util.TreeMap; * * TODO make this class more like ReplicationSource wrt log handling */ +@InterfaceAudience.Private public class ReplicationSink { private static final Log LOG = LogFactory.getLog(ReplicationSink.class); diff --git a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSinkMetrics.java b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSinkMetrics.java index ae14375c602..6a4fe2c55ea 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSinkMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSinkMetrics.java @@ -18,6 +18,7 @@ * limitations under the License. */ package org.apache.hadoop.hbase.replication.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.metrics.MetricsRate; import org.apache.hadoop.metrics.MetricsContext; import org.apache.hadoop.metrics.MetricsRecord; @@ -32,6 +33,7 @@ import org.apache.hadoop.metrics.util.MetricsRegistry; * This class is for maintaining the various replication statistics * for a sink and publishing them through the metrics interfaces. */ +@InterfaceAudience.Private public class ReplicationSinkMetrics implements Updater { private final MetricsRecord metricsRecord; private MetricsRegistry registry = new MetricsRegistry(); diff --git a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java index ee0cc1286ea..8950c9f42f1 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSource.java @@ -39,6 +39,7 @@ import java.util.concurrent.atomic.AtomicBoolean; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; @@ -72,6 +73,7 @@ import org.apache.zookeeper.KeeperException; *

    * */ +@InterfaceAudience.Private public class ReplicationSource extends Thread implements ReplicationSourceInterface { diff --git a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java index 62a4bee3518..2fe395e20fe 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceInterface.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.replication.regionserver; import java.io.IOException; import java.util.concurrent.atomic.AtomicBoolean; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -30,6 +31,7 @@ import org.apache.hadoop.hbase.Stoppable; /** * Interface that defines a replication source */ +@InterfaceAudience.Private public interface ReplicationSourceInterface { /** diff --git a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java index 152c5f667f9..732d0869341 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceManager.java @@ -38,6 +38,7 @@ import java.util.concurrent.atomic.AtomicBoolean; import com.google.common.util.concurrent.ThreadFactoryBuilder; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -58,6 +59,7 @@ import org.apache.zookeeper.KeeperException; * tries to grab a lock in order to transfer all the queues in a local * old source. */ +@InterfaceAudience.Private public class ReplicationSourceManager { private static final Log LOG = LogFactory.getLog(ReplicationSourceManager.class); diff --git a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceMetrics.java b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceMetrics.java index 85b42579eb0..543e15df423 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationSourceMetrics.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.replication.regionserver; import java.io.UnsupportedEncodingException; import java.net.URLEncoder; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.metrics.MetricsRate; import org.apache.hadoop.metrics.MetricsContext; import org.apache.hadoop.metrics.MetricsRecord; @@ -35,6 +36,7 @@ import org.apache.hadoop.metrics.util.MetricsRegistry; * This class is for maintaining the various replication statistics * for a source and publishing them through the metrics interfaces. */ +@InterfaceAudience.Private public class ReplicationSourceMetrics implements Updater { private final MetricsRecord metricsRecord; private MetricsRegistry registry = new MetricsRegistry(); diff --git a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationStatistics.java b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationStatistics.java index 54ca3df75ff..ceeff5f4897 100644 --- a/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationStatistics.java +++ b/src/main/java/org/apache/hadoop/hbase/replication/regionserver/ReplicationStatistics.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.replication.regionserver; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.metrics.MetricsMBeanBase; import org.apache.hadoop.metrics.util.MBeanUtil; import org.apache.hadoop.metrics.util.MetricsRegistry; @@ -29,6 +30,7 @@ import javax.management.ObjectName; * Exports metrics recorded by {@link ReplicationSourceMetrics} as an MBean * for JMX monitoring. */ +@InterfaceAudience.Private public class ReplicationStatistics extends MetricsMBeanBase { private final ObjectName mbeanName; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/Constants.java b/src/main/java/org/apache/hadoop/hbase/rest/Constants.java index 55ca1c6bbb5..d6f5f6bb373 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/Constants.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/Constants.java @@ -20,9 +20,14 @@ package org.apache.hadoop.hbase.rest; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Common constants for org.apache.hadoop.hbase.rest */ +@InterfaceAudience.Public +@InterfaceStability.Stable public interface Constants { public static final String VERSION_STRING = "0.0.2"; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/ExistsResource.java b/src/main/java/org/apache/hadoop/hbase/rest/ExistsResource.java index 435c82b54ff..a1cc21028d1 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/ExistsResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/ExistsResource.java @@ -31,6 +31,9 @@ import javax.ws.rs.core.Response; import javax.ws.rs.core.UriInfo; import javax.ws.rs.core.Response.ResponseBuilder; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class ExistsResource extends ResourceBase { static CacheControl cacheControl; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/Main.java b/src/main/java/org/apache/hadoop/hbase/rest/Main.java index 7e2681bb933..22c01503c89 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/Main.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/Main.java @@ -28,6 +28,7 @@ import org.apache.commons.cli.ParseException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.rest.filter.GzipFilter; @@ -56,6 +57,7 @@ import com.sun.jersey.spi.container.servlet.ServletContainer; *

  • -ro --readonly : server mode
  • * */ +@InterfaceAudience.Private public class Main implements Constants { private static void printUsageAndExit(Options options, int exitCode) { diff --git a/src/main/java/org/apache/hadoop/hbase/rest/MultiRowResource.java b/src/main/java/org/apache/hadoop/hbase/rest/MultiRowResource.java index 6e62150299b..475c03b3d84 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/MultiRowResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/MultiRowResource.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.rest; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.rest.ResourceBase; import org.apache.hadoop.hbase.rest.RowSpec; @@ -36,6 +37,7 @@ import javax.ws.rs.core.Response; import javax.ws.rs.core.UriInfo; import java.io.IOException; +@InterfaceAudience.Private public class MultiRowResource extends ResourceBase { public static final String ROW_KEYS_PARAM_NAME = "row"; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java b/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java index 405cacea84e..cb802d05966 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/ProtobufMessageHandler.java @@ -22,11 +22,14 @@ package org.apache.hadoop.hbase.rest; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Common interface for models capable of supporting protobuf marshalling * and unmarshalling. Hooks up to the ProtobufMessageBodyConsumer and * ProtobufMessageBodyProducer adapters. */ +@InterfaceAudience.Private public abstract interface ProtobufMessageHandler { /** * @return the protobuf represention of the model diff --git a/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java b/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java index 1b83f474eef..23e96a8e20e 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/RESTServlet.java @@ -22,6 +22,7 @@ package org.apache.hadoop.hbase.rest; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.client.HTablePool; import org.apache.hadoop.hbase.rest.metrics.RESTMetrics; @@ -29,6 +30,7 @@ import org.apache.hadoop.hbase.rest.metrics.RESTMetrics; /** * Singleton class encapsulating global REST servlet state and functions. */ +@InterfaceAudience.Private public class RESTServlet implements Constants { private static RESTServlet INSTANCE; private final Configuration conf; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/RegionsResource.java b/src/main/java/org/apache/hadoop/hbase/rest/RegionsResource.java index bf85bc10d98..47202b21174 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/RegionsResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/RegionsResource.java @@ -36,6 +36,7 @@ import javax.ws.rs.core.Response.ResponseBuilder; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HServerAddress; import org.apache.hadoop.hbase.TableNotFoundException; @@ -45,6 +46,7 @@ import org.apache.hadoop.hbase.client.HTablePool; import org.apache.hadoop.hbase.rest.model.TableInfoModel; import org.apache.hadoop.hbase.rest.model.TableRegionModel; +@InterfaceAudience.Private public class RegionsResource extends ResourceBase { private static final Log LOG = LogFactory.getLog(RegionsResource.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/ResourceBase.java b/src/main/java/org/apache/hadoop/hbase/rest/ResourceBase.java index 6167ccc594c..f89442583e3 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/ResourceBase.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/ResourceBase.java @@ -22,6 +22,9 @@ package org.apache.hadoop.hbase.rest; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class ResourceBase implements Constants { RESTServlet servlet; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/ResourceConfig.java b/src/main/java/org/apache/hadoop/hbase/rest/ResourceConfig.java index 19c99e787e9..07b31126b80 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/ResourceConfig.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/ResourceConfig.java @@ -20,8 +20,11 @@ package org.apache.hadoop.hbase.rest; +import org.apache.hadoop.classification.InterfaceAudience; + import com.sun.jersey.api.core.PackagesResourceConfig; +@InterfaceAudience.Private public class ResourceConfig extends PackagesResourceConfig { public ResourceConfig() { super("org.apache.hadoop.hbase.rest"); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/ResultGenerator.java b/src/main/java/org/apache/hadoop/hbase/rest/ResultGenerator.java index 4e7edf96049..392b7d2d651 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/ResultGenerator.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/ResultGenerator.java @@ -23,10 +23,12 @@ package org.apache.hadoop.hbase.rest; import java.io.IOException; import java.util.Iterator; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.filter.Filter; import org.apache.hadoop.hbase.rest.model.ScannerModel; +@InterfaceAudience.Private public abstract class ResultGenerator implements Iterator { public static ResultGenerator fromRowSpec(final String table, diff --git a/src/main/java/org/apache/hadoop/hbase/rest/RootResource.java b/src/main/java/org/apache/hadoop/hbase/rest/RootResource.java index 4cf37a81e18..29b3654b35f 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/RootResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/RootResource.java @@ -36,12 +36,14 @@ import javax.ws.rs.core.Response.ResponseBuilder; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.rest.model.TableListModel; import org.apache.hadoop.hbase.rest.model.TableModel; @Path("/") +@InterfaceAudience.Private public class RootResource extends ResourceBase { private static final Log LOG = LogFactory.getLog(RootResource.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/RowResource.java b/src/main/java/org/apache/hadoop/hbase/rest/RowResource.java index adf1ff621f9..6a634bc9294 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/RowResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/RowResource.java @@ -39,6 +39,7 @@ import javax.ws.rs.core.Response.ResponseBuilder; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Delete; @@ -50,6 +51,7 @@ import org.apache.hadoop.hbase.rest.model.CellSetModel; import org.apache.hadoop.hbase.rest.model.RowModel; import org.apache.hadoop.hbase.util.Bytes; +@InterfaceAudience.Private public class RowResource extends ResourceBase { private static final Log LOG = LogFactory.getLog(RowResource.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/RowResultGenerator.java b/src/main/java/org/apache/hadoop/hbase/rest/RowResultGenerator.java index 74f30b4c387..7bab3b7095f 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/RowResultGenerator.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/RowResultGenerator.java @@ -28,6 +28,7 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.util.StringUtils; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.DoNotRetryIOException; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.KeyValue; @@ -38,6 +39,7 @@ import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hbase.filter.Filter; import org.apache.hadoop.hbase.regionserver.NoSuchColumnFamilyException; +@InterfaceAudience.Private public class RowResultGenerator extends ResultGenerator { private static final Log LOG = LogFactory.getLog(RowResultGenerator.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/RowSpec.java b/src/main/java/org/apache/hadoop/hbase/rest/RowSpec.java index 93d0ed1adeb..15bec51ef17 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/RowSpec.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/RowSpec.java @@ -25,6 +25,7 @@ import java.net.URLDecoder; import java.util.Collection; import java.util.TreeSet; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.util.Bytes; @@ -35,6 +36,7 @@ import org.apache.hadoop.hbase.util.Bytes; *

    * */ +@InterfaceAudience.Private public class RowSpec { public static final long DEFAULT_START_TIMESTAMP = 0; public static final long DEFAULT_END_TIMESTAMP = Long.MAX_VALUE; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/ScannerInstanceResource.java b/src/main/java/org/apache/hadoop/hbase/rest/ScannerInstanceResource.java index 75f10659016..15675f15b9d 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/ScannerInstanceResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/ScannerInstanceResource.java @@ -36,6 +36,7 @@ import javax.ws.rs.core.UriInfo; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.rest.model.CellModel; import org.apache.hadoop.hbase.rest.model.CellSetModel; @@ -43,6 +44,7 @@ import org.apache.hadoop.hbase.rest.model.RowModel; import org.apache.hadoop.hbase.util.Base64; import org.apache.hadoop.hbase.util.Bytes; +@InterfaceAudience.Private public class ScannerInstanceResource extends ResourceBase { private static final Log LOG = LogFactory.getLog(ScannerInstanceResource.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/ScannerResource.java b/src/main/java/org/apache/hadoop/hbase/rest/ScannerResource.java index dad5b4377d6..a82adeeffc8 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/ScannerResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/ScannerResource.java @@ -40,10 +40,12 @@ import javax.ws.rs.core.UriInfo; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.TableNotFoundException; import org.apache.hadoop.hbase.filter.Filter; import org.apache.hadoop.hbase.rest.model.ScannerModel; +@InterfaceAudience.Private public class ScannerResource extends ResourceBase { private static final Log LOG = LogFactory.getLog(ScannerResource.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/ScannerResultGenerator.java b/src/main/java/org/apache/hadoop/hbase/rest/ScannerResultGenerator.java index d4f1dfc8d4f..d82cf31c6b3 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/ScannerResultGenerator.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/ScannerResultGenerator.java @@ -25,6 +25,7 @@ import java.util.Iterator; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.UnknownScannerException; @@ -37,6 +38,7 @@ import org.apache.hadoop.hbase.filter.Filter; import org.apache.hadoop.hbase.rest.model.ScannerModel; import org.apache.hadoop.util.StringUtils; +@InterfaceAudience.Private public class ScannerResultGenerator extends ResultGenerator { private static final Log LOG = diff --git a/src/main/java/org/apache/hadoop/hbase/rest/SchemaResource.java b/src/main/java/org/apache/hadoop/hbase/rest/SchemaResource.java index 5e9a5be5525..b03fb8a48f5 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/SchemaResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/SchemaResource.java @@ -41,6 +41,7 @@ import javax.xml.namespace.QName; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.TableExistsException; @@ -52,6 +53,7 @@ import org.apache.hadoop.hbase.rest.model.ColumnSchemaModel; import org.apache.hadoop.hbase.rest.model.TableSchemaModel; import org.apache.hadoop.hbase.util.Bytes; +@InterfaceAudience.Private public class SchemaResource extends ResourceBase { private static final Log LOG = LogFactory.getLog(SchemaResource.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/StorageClusterStatusResource.java b/src/main/java/org/apache/hadoop/hbase/rest/StorageClusterStatusResource.java index cddade0c76f..01a805573d9 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/StorageClusterStatusResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/StorageClusterStatusResource.java @@ -34,12 +34,14 @@ import javax.ws.rs.core.UriInfo; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.ClusterStatus; import org.apache.hadoop.hbase.HServerLoad; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.rest.model.StorageClusterStatusModel; +@InterfaceAudience.Private public class StorageClusterStatusResource extends ResourceBase { private static final Log LOG = LogFactory.getLog(StorageClusterStatusResource.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/StorageClusterVersionResource.java b/src/main/java/org/apache/hadoop/hbase/rest/StorageClusterVersionResource.java index 106c9dcb41a..c662cfe2dbe 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/StorageClusterVersionResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/StorageClusterVersionResource.java @@ -34,9 +34,11 @@ import javax.ws.rs.core.Response.ResponseBuilder; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.rest.model.StorageClusterVersionModel; +@InterfaceAudience.Private public class StorageClusterVersionResource extends ResourceBase { private static final Log LOG = LogFactory.getLog(StorageClusterVersionResource.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java b/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java index 07f5f269488..d314604214f 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/TableResource.java @@ -27,8 +27,10 @@ import javax.ws.rs.Path; import javax.ws.rs.PathParam; import javax.ws.rs.QueryParam; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.client.HBaseAdmin; +@InterfaceAudience.Private public class TableResource extends ResourceBase { String table; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/VersionResource.java b/src/main/java/org/apache/hadoop/hbase/rest/VersionResource.java index 3d0a9b3f0b9..9f268e2e61c 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/VersionResource.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/VersionResource.java @@ -35,6 +35,7 @@ import javax.ws.rs.core.Response.ResponseBuilder; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.model.VersionModel; /** @@ -44,6 +45,7 @@ import org.apache.hadoop.hbase.rest.model.VersionModel; *

    * /version (alias for /version/rest) */ +@InterfaceAudience.Private public class VersionResource extends ResourceBase { private static final Log LOG = LogFactory.getLog(VersionResource.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java b/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java index 4ecbfa845a5..72f63b2ce36 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/client/Client.java @@ -38,11 +38,15 @@ import org.apache.commons.httpclient.params.HttpClientParams; import org.apache.commons.httpclient.params.HttpConnectionManagerParams; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; /** * A wrapper around HttpClient which provides some useful function and * semantics for interacting with the REST gateway. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Client { public static final Header[] EMPTY_HEADER_ARRAY = new Header[0]; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java b/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java index 46724470c07..623a576440e 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/client/Cluster.java @@ -24,10 +24,15 @@ import java.util.ArrayList; import java.util.Collections; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * A list of 'host:port' addresses of HTTP servers operating as a single * entity, for example multiple redundant web service gateways. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Cluster { protected List nodes = Collections.synchronizedList(new ArrayList()); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteAdmin.java b/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteAdmin.java index f7c0394a32a..d46a158b715 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteAdmin.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteAdmin.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.rest.client; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HTableDescriptor; @@ -29,6 +31,8 @@ import org.apache.hadoop.hbase.rest.Constants; import org.apache.hadoop.hbase.rest.model.TableSchemaModel; import org.apache.hadoop.hbase.util.Bytes; +@InterfaceAudience.Public +@InterfaceStability.Stable public class RemoteAdmin { final Client client; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteHTable.java b/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteHTable.java index 8acc5266fdd..56e31e1d3a2 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteHTable.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/client/RemoteHTable.java @@ -37,6 +37,8 @@ import org.apache.hadoop.hbase.client.coprocessor.Batch; import org.apache.hadoop.hbase.ipc.CoprocessorProtocol; import org.apache.hadoop.util.StringUtils; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HConstants; @@ -65,6 +67,8 @@ import org.apache.hadoop.hbase.util.Bytes; /** * HTable interface to remote tables accessed via REST gateway */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class RemoteHTable implements HTableInterface { private static final Log LOG = LogFactory.getLog(RemoteHTable.class); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java b/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java index 421065b895a..3bd89ab3a58 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/client/Response.java @@ -21,10 +21,14 @@ package org.apache.hadoop.hbase.rest.client; import org.apache.commons.httpclient.Header; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; /** * The HTTP result code, response headers, and body of a HTTP response. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Response { private int code; private Header[] headers; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPRequestStream.java b/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPRequestStream.java index 0bd5f653a98..d905bc580fd 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPRequestStream.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPRequestStream.java @@ -26,6 +26,9 @@ import java.util.zip.GZIPInputStream; import javax.servlet.ServletInputStream; import javax.servlet.http.HttpServletRequest; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class GZIPRequestStream extends ServletInputStream { private GZIPInputStream in; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPRequestWrapper.java b/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPRequestWrapper.java index 764576c208d..f26ae0341f2 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPRequestWrapper.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPRequestWrapper.java @@ -28,6 +28,9 @@ import javax.servlet.ServletInputStream; import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletRequestWrapper; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class GZIPRequestWrapper extends HttpServletRequestWrapper { private ServletInputStream is; private BufferedReader reader; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPResponseStream.java b/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPResponseStream.java index d27b37beaa1..22009535e9e 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPResponseStream.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPResponseStream.java @@ -26,6 +26,9 @@ import java.util.zip.GZIPOutputStream; import javax.servlet.ServletOutputStream; import javax.servlet.http.HttpServletResponse; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class GZIPResponseStream extends ServletOutputStream { private HttpServletResponse response; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPResponseWrapper.java b/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPResponseWrapper.java index 08a976b32fd..0da3c16a43f 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPResponseWrapper.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/filter/GZIPResponseWrapper.java @@ -27,6 +27,9 @@ import javax.servlet.ServletOutputStream; import javax.servlet.http.HttpServletResponse; import javax.servlet.http.HttpServletResponseWrapper; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class GZIPResponseWrapper extends HttpServletResponseWrapper { private HttpServletResponse response; private ServletOutputStream os; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/filter/GzipFilter.java b/src/main/java/org/apache/hadoop/hbase/rest/filter/GzipFilter.java index 9c1c3f60c09..ba47befdff1 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/filter/GzipFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/filter/GzipFilter.java @@ -35,6 +35,9 @@ import javax.servlet.ServletResponse; import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletResponse; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class GzipFilter implements Filter { private Set mimeTypes = new HashSet(); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/metrics/RESTMetrics.java b/src/main/java/org/apache/hadoop/hbase/rest/metrics/RESTMetrics.java index 284bbc54811..f4b397cd52c 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/metrics/RESTMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/metrics/RESTMetrics.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.rest.metrics; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.metrics.MetricsRate; import org.apache.hadoop.metrics.MetricsContext; @@ -29,6 +30,7 @@ import org.apache.hadoop.metrics.Updater; import org.apache.hadoop.metrics.jvm.JvmMetrics; import org.apache.hadoop.metrics.util.MetricsRegistry; +@InterfaceAudience.Private public class RESTMetrics implements Updater { private final MetricsRecord metricsRecord; private final MetricsRegistry registry = new MetricsRegistry(); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/metrics/RESTStatistics.java b/src/main/java/org/apache/hadoop/hbase/rest/metrics/RESTStatistics.java index d29d50de8eb..03035a2e446 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/metrics/RESTStatistics.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/metrics/RESTStatistics.java @@ -22,11 +22,13 @@ package org.apache.hadoop.hbase.rest.metrics; import javax.management.ObjectName; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.metrics.MetricsMBeanBase; import org.apache.hadoop.metrics.util.MBeanUtil; import org.apache.hadoop.metrics.util.MetricsRegistry; +@InterfaceAudience.Private public class RESTStatistics extends MetricsMBeanBase { private final ObjectName mbeanName; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java index 3413d00a9b6..d2e5d76565a 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/CellModel.java @@ -27,6 +27,7 @@ import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlRootElement; import javax.xml.bind.annotation.XmlValue; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; @@ -54,6 +55,7 @@ import com.google.protobuf.ByteString; * */ @XmlRootElement(name="Cell") +@InterfaceAudience.Private public class CellModel implements ProtobufMessageHandler, Serializable { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java index 7e7073c4bda..7b9bc06c095 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/CellSetModel.java @@ -28,6 +28,7 @@ import java.util.List; import javax.xml.bind.annotation.XmlRootElement; import javax.xml.bind.annotation.XmlElement; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; import org.apache.hadoop.hbase.rest.protobuf.generated.CellMessage.Cell; @@ -69,6 +70,7 @@ import com.google.protobuf.ByteString; * */ @XmlRootElement(name="CellSet") +@InterfaceAudience.Private public class CellSetModel implements Serializable, ProtobufMessageHandler { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java index caf53683e95..d8e27640109 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/ColumnSchemaModel.java @@ -29,6 +29,7 @@ import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlRootElement; import javax.xml.namespace.QName; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; @@ -43,6 +44,7 @@ import org.apache.hadoop.hbase.HConstants; * */ @XmlRootElement(name="ColumnSchema") +@InterfaceAudience.Private public class ColumnSchemaModel implements Serializable { private static final long serialVersionUID = 1L; private static QName BLOCKCACHE = new QName(HColumnDescriptor.BLOCKCACHE); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java index a987695554e..e97df1bba53 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/RowModel.java @@ -29,6 +29,7 @@ import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlElement; import javax.xml.bind.annotation.XmlRootElement; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; /** @@ -47,6 +48,7 @@ import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; * */ @XmlRootElement(name="Row") +@InterfaceAudience.Private public class RowModel implements ProtobufMessageHandler, Serializable { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java index d76bdb18f53..284af9b29a5 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/ScannerModel.java @@ -33,6 +33,7 @@ import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlElement; import javax.xml.bind.annotation.XmlRootElement; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.filter.*; @@ -68,6 +69,7 @@ import com.sun.jersey.api.json.JSONUnmarshaller; * */ @XmlRootElement(name="Scanner") +@InterfaceAudience.Private public class ScannerModel implements ProtobufMessageHandler, Serializable { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java index f45e902a6de..2ce6133a4de 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterStatusModel.java @@ -30,6 +30,7 @@ import javax.xml.bind.annotation.XmlElement; import javax.xml.bind.annotation.XmlElementWrapper; import javax.xml.bind.annotation.XmlRootElement; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; import org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus; import org.apache.hadoop.hbase.util.Bytes; @@ -86,6 +87,7 @@ import com.google.protobuf.ByteString; * */ @XmlRootElement(name="ClusterStatus") +@InterfaceAudience.Private public class StorageClusterStatusModel implements Serializable, ProtobufMessageHandler { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java index 05634795e4f..c6fecd18dd3 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/StorageClusterVersionModel.java @@ -25,6 +25,8 @@ import java.io.Serializable; import javax.xml.bind.annotation.XmlRootElement; import javax.xml.bind.annotation.XmlValue; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Simple representation of the version of the storage cluster * @@ -35,6 +37,7 @@ import javax.xml.bind.annotation.XmlValue; * */ @XmlRootElement(name="ClusterVersion") +@InterfaceAudience.Private public class StorageClusterVersionModel implements Serializable { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java index ce6fb96fe59..f983d27edd6 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/TableInfoModel.java @@ -29,6 +29,7 @@ import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlElement; import javax.xml.bind.annotation.XmlRootElement; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; import org.apache.hadoop.hbase.rest.protobuf.generated.TableInfoMessage.TableInfo; @@ -48,6 +49,7 @@ import com.google.protobuf.ByteString; * */ @XmlRootElement(name="TableInfo") +@InterfaceAudience.Private public class TableInfoModel implements Serializable, ProtobufMessageHandler { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java index 1c276c25706..a94a897da6c 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/TableListModel.java @@ -28,6 +28,7 @@ import java.util.List; import javax.xml.bind.annotation.XmlElementRef; import javax.xml.bind.annotation.XmlRootElement; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; import org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableList; @@ -35,6 +36,7 @@ import org.apache.hadoop.hbase.rest.protobuf.generated.TableListMessage.TableLis * Simple representation of a list of table names. */ @XmlRootElement(name="TableList") +@InterfaceAudience.Private public class TableListModel implements Serializable, ProtobufMessageHandler { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java index e1d33cd1b36..30ea8935027 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/TableModel.java @@ -25,6 +25,8 @@ import java.io.Serializable; import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlRootElement; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Simple representation of a table name. * @@ -37,6 +39,7 @@ import javax.xml.bind.annotation.XmlRootElement; * */ @XmlRootElement(name="table") +@InterfaceAudience.Private public class TableModel implements Serializable { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java index 67e7a049ace..353559529b6 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/TableRegionModel.java @@ -25,6 +25,7 @@ import java.io.Serializable; import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlRootElement; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.util.Bytes; @@ -44,6 +45,7 @@ import org.apache.hadoop.hbase.util.Bytes; * */ @XmlRootElement(name="Region") +@InterfaceAudience.Private public class TableRegionModel implements Serializable { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java index fa6e3a6654b..8886ef84893 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/TableSchemaModel.java @@ -34,6 +34,7 @@ import javax.xml.bind.annotation.XmlElement; import javax.xml.bind.annotation.XmlRootElement; import javax.xml.namespace.QName; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HTableDescriptor; @@ -58,6 +59,7 @@ import org.apache.hadoop.hbase.util.Bytes; * */ @XmlRootElement(name="TableSchema") +@InterfaceAudience.Private public class TableSchemaModel implements Serializable, ProtobufMessageHandler { private static final long serialVersionUID = 1L; private static final QName IS_META = new QName(HTableDescriptor.IS_META); diff --git a/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java b/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java index e4b6b0fbf83..427b58d80f7 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/model/VersionModel.java @@ -27,6 +27,7 @@ import javax.servlet.ServletContext; import javax.xml.bind.annotation.XmlAttribute; import javax.xml.bind.annotation.XmlRootElement; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; import org.apache.hadoop.hbase.rest.RESTServlet; import org.apache.hadoop.hbase.rest.protobuf.generated.VersionMessage.Version; @@ -45,6 +46,7 @@ import com.sun.jersey.spi.container.servlet.ServletContainer; * */ @XmlRootElement(name="Version") +@InterfaceAudience.Private public class VersionModel implements Serializable, ProtobufMessageHandler { private static final long serialVersionUID = 1L; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java b/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java index 0c2ab3d4a8a..4c4e7316249 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/provider/JAXBContextResolver.java @@ -28,6 +28,7 @@ import javax.ws.rs.ext.ContextResolver; import javax.ws.rs.ext.Provider; import javax.xml.bind.JAXBContext; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.model.CellModel; import org.apache.hadoop.hbase.rest.model.CellSetModel; import org.apache.hadoop.hbase.rest.model.ColumnSchemaModel; @@ -52,6 +53,7 @@ import com.sun.jersey.api.json.JSONJAXBContext; * created. */ @Provider +@InterfaceAudience.Private public class JAXBContextResolver implements ContextResolver { private final JAXBContext context; diff --git a/src/main/java/org/apache/hadoop/hbase/rest/provider/consumer/ProtobufMessageBodyConsumer.java b/src/main/java/org/apache/hadoop/hbase/rest/provider/consumer/ProtobufMessageBodyConsumer.java index 6fe2dd08de9..e6e0de59850 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/provider/consumer/ProtobufMessageBodyConsumer.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/provider/consumer/ProtobufMessageBodyConsumer.java @@ -35,6 +35,7 @@ import javax.ws.rs.ext.Provider; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.Constants; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; @@ -44,6 +45,7 @@ import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; */ @Provider @Consumes(Constants.MIMETYPE_PROTOBUF) +@InterfaceAudience.Private public class ProtobufMessageBodyConsumer implements MessageBodyReader { private static final Log LOG = diff --git a/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/PlainTextMessageBodyProducer.java b/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/PlainTextMessageBodyProducer.java index 092c69575b4..f2613eaf997 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/PlainTextMessageBodyProducer.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/PlainTextMessageBodyProducer.java @@ -32,6 +32,7 @@ import javax.ws.rs.core.MultivaluedMap; import javax.ws.rs.ext.MessageBodyWriter; import javax.ws.rs.ext.Provider; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.Constants; /** @@ -42,6 +43,7 @@ import org.apache.hadoop.hbase.rest.Constants; */ @Provider @Produces(Constants.MIMETYPE_TEXT) +@InterfaceAudience.Private public class PlainTextMessageBodyProducer implements MessageBodyWriter { diff --git a/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/ProtobufMessageBodyProducer.java b/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/ProtobufMessageBodyProducer.java index a1b4b70831a..11e5ee8d6d2 100644 --- a/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/ProtobufMessageBodyProducer.java +++ b/src/main/java/org/apache/hadoop/hbase/rest/provider/producer/ProtobufMessageBodyProducer.java @@ -33,6 +33,7 @@ import javax.ws.rs.core.MultivaluedMap; import javax.ws.rs.ext.MessageBodyWriter; import javax.ws.rs.ext.Provider; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.rest.Constants; import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; @@ -44,6 +45,7 @@ import org.apache.hadoop.hbase.rest.ProtobufMessageHandler; */ @Provider @Produces(Constants.MIMETYPE_PROTOBUF) +@InterfaceAudience.Private public class ProtobufMessageBodyProducer implements MessageBodyWriter { diff --git a/src/main/java/org/apache/hadoop/hbase/security/KerberosInfo.java b/src/main/java/org/apache/hadoop/hbase/security/KerberosInfo.java index da211457211..5e5e6ee2b6b 100644 --- a/src/main/java/org/apache/hadoop/hbase/security/KerberosInfo.java +++ b/src/main/java/org/apache/hadoop/hbase/security/KerberosInfo.java @@ -23,12 +23,15 @@ import java.lang.annotation.Retention; import java.lang.annotation.RetentionPolicy; import java.lang.annotation.Target; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Indicates Kerberos related information to be used for authorizing connections * over a given RPC protocol interface. */ @Retention(RetentionPolicy.RUNTIME) @Target(ElementType.TYPE) +@InterfaceAudience.Private public @interface KerberosInfo { /** Key for getting server's Kerberos principal name from Configuration */ String serverPrincipal(); diff --git a/src/main/java/org/apache/hadoop/hbase/security/TokenInfo.java b/src/main/java/org/apache/hadoop/hbase/security/TokenInfo.java index 22a9a5f05d9..27a0af17990 100644 --- a/src/main/java/org/apache/hadoop/hbase/security/TokenInfo.java +++ b/src/main/java/org/apache/hadoop/hbase/security/TokenInfo.java @@ -23,12 +23,15 @@ import java.lang.annotation.Retention; import java.lang.annotation.RetentionPolicy; import java.lang.annotation.Target; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Indicates Token related information to be used in authorizing connections * over a given RPC protocol interface. */ @Retention(RetentionPolicy.RUNTIME) @Target(ElementType.TYPE) +@InterfaceAudience.Private public @interface TokenInfo { /** The type of Token.getKind() to be handled */ String value(); diff --git a/src/main/java/org/apache/hadoop/hbase/security/User.java b/src/main/java/org/apache/hadoop/hbase/security/User.java index eb516a4444a..2e2a61e0052 100644 --- a/src/main/java/org/apache/hadoop/hbase/security/User.java +++ b/src/main/java/org/apache/hadoop/hbase/security/User.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.security; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.util.Methods; @@ -47,6 +48,7 @@ import org.apache.commons.logging.Log; * HBase, but can be extended as needs change. *

    */ +@InterfaceAudience.Private public abstract class User { public static final String HBASE_SECURITY_CONF_KEY = "hbase.security.authentication"; diff --git a/src/main/java/org/apache/hadoop/hbase/thrift/CallQueue.java b/src/main/java/org/apache/hadoop/hbase/thrift/CallQueue.java index 5bd47a86ae4..1727c59113a 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift/CallQueue.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift/CallQueue.java @@ -27,11 +27,13 @@ import java.util.concurrent.TimeUnit; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; /** * A BlockingQueue reports waiting time in queue and queue length to * ThriftMetrics. */ +@InterfaceAudience.Private public class CallQueue implements BlockingQueue { private static Log LOG = LogFactory.getLog(CallQueue.class); diff --git a/src/main/java/org/apache/hadoop/hbase/thrift/HThreadedSelectorServerArgs.java b/src/main/java/org/apache/hadoop/hbase/thrift/HThreadedSelectorServerArgs.java index 494f5a53c93..f19ec5fae45 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift/HThreadedSelectorServerArgs.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift/HThreadedSelectorServerArgs.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.thrift; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.thrift.server.TThreadedSelectorServer; import org.apache.thrift.transport.TNonblockingServerTransport; @@ -28,6 +29,7 @@ import org.slf4j.LoggerFactory; /** * A TThreadedSelectorServer.Args that reads hadoop configuration */ +@InterfaceAudience.Private public class HThreadedSelectorServerArgs extends TThreadedSelectorServer.Args { private static final Logger LOG = diff --git a/src/main/java/org/apache/hadoop/hbase/thrift/HbaseHandlerMetricsProxy.java b/src/main/java/org/apache/hadoop/hbase/thrift/HbaseHandlerMetricsProxy.java index 22fe4319107..039def706de 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift/HbaseHandlerMetricsProxy.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift/HbaseHandlerMetricsProxy.java @@ -25,6 +25,7 @@ import java.lang.reflect.Proxy; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.thrift.generated.Hbase; @@ -33,6 +34,7 @@ import org.apache.hadoop.hbase.thrift.generated.Hbase; * Converts a Hbase.Iface using InvocationHandler so that it reports process * time of each call to ThriftMetrics. */ +@InterfaceAudience.Private public class HbaseHandlerMetricsProxy implements InvocationHandler { public static final Log LOG = LogFactory.getLog( diff --git a/src/main/java/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.java b/src/main/java/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.java index c7e104b5a87..5a85e52a96c 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift/TBoundedThreadPoolServer.java @@ -27,6 +27,7 @@ import java.util.concurrent.TimeUnit; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.thrift.CallQueue.Call; import org.apache.hadoop.hbase.util.Threads; @@ -45,6 +46,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; /** * A bounded thread pool server customized for HBase. */ +@InterfaceAudience.Private public class TBoundedThreadPoolServer extends TServer { private static final String QUEUE_FULL_MSG = diff --git a/src/main/java/org/apache/hadoop/hbase/thrift/ThriftMetrics.java b/src/main/java/org/apache/hadoop/hbase/thrift/ThriftMetrics.java index f82d0c001b0..d6eb5cd1b9c 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift/ThriftMetrics.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift/ThriftMetrics.java @@ -23,6 +23,7 @@ import java.lang.reflect.Method; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.thrift.generated.Hbase; import org.apache.hadoop.metrics.MetricsContext; @@ -40,6 +41,7 @@ import org.apache.hadoop.metrics.util.MetricsTimeVaryingRate; * This class is for maintaining the various statistics of thrift server * and publishing them through the metrics interfaces. */ +@InterfaceAudience.Private public class ThriftMetrics implements Updater { public final static Log LOG = LogFactory.getLog(ThriftMetrics.class); public final static String CONTEXT_NAME = "thriftserver"; diff --git a/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java b/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java index 5812a7a5157..bf988433588 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServer.java @@ -28,6 +28,7 @@ import org.apache.commons.cli.Options; import org.apache.commons.cli.PosixParser; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.thrift.ThriftServerRunner.ImplType; @@ -39,6 +40,7 @@ import org.apache.hadoop.util.Shell.ExitCodeException; * Hbase API specified in the Hbase.thrift IDL file. The server runs in an * independent process. */ +@InterfaceAudience.Private public class ThriftServer { private static final Log LOG = LogFactory.getLog(ThriftServer.class); diff --git a/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServerRunner.java b/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServerRunner.java index 78e1a72f12f..4644d5e7a3a 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServerRunner.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift/ThriftServerRunner.java @@ -42,6 +42,7 @@ import org.apache.commons.cli.Option; import org.apache.commons.cli.OptionGroup; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HColumnDescriptor; @@ -102,6 +103,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; * ThriftServerRunner - this class starts up a Thrift server which implements * the Hbase API specified in the Hbase.thrift IDL file. */ +@InterfaceAudience.Private public class ThriftServerRunner implements Runnable { private static final Log LOG = LogFactory.getLog(ThriftServerRunner.class); diff --git a/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java b/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java index d7fa95e5796..f1329462e6e 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift/ThriftUtilities.java @@ -23,6 +23,7 @@ import java.util.ArrayList; import java.util.List; import java.util.TreeMap; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Result; @@ -35,6 +36,7 @@ import org.apache.hadoop.hbase.thrift.generated.TCell; import org.apache.hadoop.hbase.thrift.generated.TRowResult; import org.apache.hadoop.hbase.util.Bytes; +@InterfaceAudience.Private public class ThriftUtilities { /** diff --git a/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.java b/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.java index 2a8863f271d..bc6adecb1b2 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftHBaseServiceHandler.java @@ -44,6 +44,7 @@ import java.util.concurrent.atomic.AtomicInteger; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.client.Delete; import org.apache.hadoop.hbase.client.HTableInterface; @@ -65,6 +66,7 @@ import org.apache.thrift.TException; * This class is a glue object that connects Thrift RPC calls to the HBase client API primarily defined in the * HTableInterface. */ +@InterfaceAudience.Private public class ThriftHBaseServiceHandler implements THBaseService.Iface { // TODO: Size of pool configuraple diff --git a/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftServer.java b/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftServer.java index 721b663aacf..05b05a2f09f 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftServer.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftServer.java @@ -38,6 +38,7 @@ import org.apache.commons.cli.ParseException; import org.apache.commons.cli.PosixParser; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.thrift.CallQueue; @@ -65,6 +66,7 @@ import com.google.common.util.concurrent.ThreadFactoryBuilder; * ThriftServer - this class starts up a Thrift server which implements the HBase API specified in the * HbaseClient.thrift IDL file. */ +@InterfaceAudience.Private public class ThriftServer { private static final Log log = LogFactory.getLog(ThriftServer.class); diff --git a/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java b/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java index b63ed74e933..cd5797dbe66 100644 --- a/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java +++ b/src/main/java/org/apache/hadoop/hbase/thrift2/ThriftUtilities.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.thrift2; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.*; @@ -28,6 +29,7 @@ import java.io.IOException; import java.nio.ByteBuffer; import java.util.*; +@InterfaceAudience.Private public class ThriftUtilities { private ThriftUtilities() { diff --git a/src/main/java/org/apache/hadoop/hbase/util/AbstractHBaseTool.java b/src/main/java/org/apache/hadoop/hbase/util/AbstractHBaseTool.java index 779cdc1fd54..ab530858d1f 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/AbstractHBaseTool.java +++ b/src/main/java/org/apache/hadoop/hbase/util/AbstractHBaseTool.java @@ -27,6 +27,7 @@ import org.apache.commons.cli.Options; import org.apache.commons.cli.ParseException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.util.Tool; @@ -36,6 +37,7 @@ import org.apache.hadoop.util.ToolRunner; * Common base class used for HBase command-line tools. Simplifies workflow and * command-line argument parsing. */ +@InterfaceAudience.Private public abstract class AbstractHBaseTool implements Tool { private static final int EXIT_SUCCESS = 0; diff --git a/src/main/java/org/apache/hadoop/hbase/util/Addressing.java b/src/main/java/org/apache/hadoop/hbase/util/Addressing.java index 05c51ea440c..c5151ab36fc 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Addressing.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Addressing.java @@ -21,9 +21,14 @@ package org.apache.hadoop.hbase.util; import java.net.InetSocketAddress; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Utility for network addresses, resolving and naming. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class Addressing { public static final String VALID_PORT_REGEX = "[\\d]+"; public static final String HOSTNAME_PORT_SEPARATOR = ":"; diff --git a/src/main/java/org/apache/hadoop/hbase/util/Base64.java b/src/main/java/org/apache/hadoop/hbase/util/Base64.java index 892f8083c96..a76ce33e376 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Base64.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Base64.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.util; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import java.io.BufferedInputStream; import java.io.BufferedOutputStream; @@ -116,6 +118,8 @@ import java.util.zip.GZIPOutputStream; *
    * version: 2.2.1 */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Base64 { /* ******** P U B L I C F I E L D S ******** */ diff --git a/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java b/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java index 7dd2f68e43d..67caf194b96 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/util/BloomFilter.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.util; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Defines the general behavior of a bloom filter. * @@ -39,6 +41,7 @@ import java.nio.ByteBuffer; * * @see BloomFilterWriter for the ability to add elements to a Bloom filter */ +@InterfaceAudience.Private public interface BloomFilter extends BloomFilterBase { /** diff --git a/src/main/java/org/apache/hadoop/hbase/util/BloomFilterBase.java b/src/main/java/org/apache/hadoop/hbase/util/BloomFilterBase.java index ab218b77f75..e166f567a73 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/BloomFilterBase.java +++ b/src/main/java/org/apache/hadoop/hbase/util/BloomFilterBase.java @@ -19,11 +19,13 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.io.RawComparator; /** * Common methods Bloom filter methods required at read and write time. */ +@InterfaceAudience.Private public interface BloomFilterBase { /** diff --git a/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java b/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java index 418bd16ae9c..20b4a01c2f1 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java +++ b/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java @@ -25,6 +25,7 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.io.hfile.CacheConfig; @@ -36,6 +37,7 @@ import org.apache.hadoop.hbase.regionserver.StoreFile.BloomType; * Handles Bloom filter initialization based on configuration and serialized * metadata in the reader and writer of {@link StoreFile}. */ +@InterfaceAudience.Private public final class BloomFilterFactory { private static final Log LOG = diff --git a/src/main/java/org/apache/hadoop/hbase/util/BloomFilterWriter.java b/src/main/java/org/apache/hadoop/hbase/util/BloomFilterWriter.java index 46691fbfb4a..1ede0d0f1e4 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/BloomFilterWriter.java +++ b/src/main/java/org/apache/hadoop/hbase/util/BloomFilterWriter.java @@ -20,12 +20,14 @@ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.io.Writable; /** * Specifies methods needed to add elements to a Bloom filter and serialize the * resulting Bloom filter as a sequence of bytes. */ +@InterfaceAudience.Private public interface BloomFilterWriter extends BloomFilterBase { /** Allocate memory for the bloom filter data. */ diff --git a/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java b/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java index e9c50332a67..0cb82d7a7c6 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/util/ByteBloomFilter.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.io.RawComparator; import org.apache.hadoop.io.Writable; @@ -76,6 +77,7 @@ import java.util.Random; * href="http://portal.acm.org/citation.cfm?id=362692&dl=ACM&coll=portal"> * Space/Time Trade-Offs in Hash Coding with Allowable Errors */ +@InterfaceAudience.Private public class ByteBloomFilter implements BloomFilter, BloomFilterWriter { /** Current file format version */ diff --git a/src/main/java/org/apache/hadoop/hbase/util/ByteBufferOutputStream.java b/src/main/java/org/apache/hadoop/hbase/util/ByteBufferOutputStream.java index 599c15b1dcf..67c6e49402a 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/ByteBufferOutputStream.java +++ b/src/main/java/org/apache/hadoop/hbase/util/ByteBufferOutputStream.java @@ -26,9 +26,14 @@ import java.nio.ByteBuffer; import java.nio.channels.Channels; import java.nio.channels.WritableByteChannel; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Not thread safe! */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ByteBufferOutputStream extends OutputStream { protected ByteBuffer buf; diff --git a/src/main/java/org/apache/hadoop/hbase/util/ByteBufferUtils.java b/src/main/java/org/apache/hadoop/hbase/util/ByteBufferUtils.java index 453e1451f5b..21481bf3b5b 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/ByteBufferUtils.java +++ b/src/main/java/org/apache/hadoop/hbase/util/ByteBufferUtils.java @@ -23,6 +23,8 @@ import java.io.InputStream; import java.io.OutputStream; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.io.encoding. EncoderBufferTooSmallException; import org.apache.hadoop.io.WritableUtils; @@ -31,6 +33,8 @@ import org.apache.hadoop.io.WritableUtils; * Utility functions for working with byte buffers, such as reading/writing * variable-length long numbers. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public final class ByteBufferUtils { // "Compressed integer" serialization helper constants. diff --git a/src/main/java/org/apache/hadoop/hbase/util/Bytes.java b/src/main/java/org/apache/hadoop/hbase/util/Bytes.java index ead9a3bf2ad..de8e40b0b3d 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Bytes.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Bytes.java @@ -36,6 +36,8 @@ import java.util.Iterator; import org.apache.commons.lang.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.io.RawComparator; @@ -51,6 +53,8 @@ import com.google.common.annotations.VisibleForTesting; * comparisons, hash code generation, manufacturing keys for HashMaps or * HashSets, etc. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Bytes { private static final Log LOG = LogFactory.getLog(Bytes.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java b/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java index 6d7c3890c6b..cae258abb37 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java +++ b/src/main/java/org/apache/hadoop/hbase/util/CancelableProgressable.java @@ -19,12 +19,15 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Similar interface as {@link org.apache.hadoop.util.Progressable} but returns * a boolean to support canceling the operation. *

    * Used for doing updating of OPENING znode during log replay on region open. */ +@InterfaceAudience.Private public interface CancelableProgressable { /** diff --git a/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java b/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java index e18473067d5..b3fa9285c4b 100755 --- a/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java +++ b/src/main/java/org/apache/hadoop/hbase/util/ClassSize.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.util; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import java.lang.reflect.Field; import java.lang.reflect.Modifier; @@ -32,6 +34,8 @@ import java.lang.reflect.Modifier; * * The core of this class is taken from the Derby project */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class ClassSize { static final Log LOG = LogFactory.getLog(ClassSize.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/Classes.java b/src/main/java/org/apache/hadoop/hbase/util/Classes.java index 2b353e737b3..76551d918e9 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Classes.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Classes.java @@ -20,9 +20,14 @@ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Utilities for class manipulation. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Classes { /** diff --git a/src/main/java/org/apache/hadoop/hbase/util/CollectionBackedScanner.java b/src/main/java/org/apache/hadoop/hbase/util/CollectionBackedScanner.java index 25cec2af143..3d85ba37e43 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/CollectionBackedScanner.java +++ b/src/main/java/org/apache/hadoop/hbase/util/CollectionBackedScanner.java @@ -25,6 +25,7 @@ import java.util.Iterator; import java.util.List; import java.util.SortedSet; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.regionserver.NonLazyKeyValueScanner; @@ -32,6 +33,7 @@ import org.apache.hadoop.hbase.regionserver.NonLazyKeyValueScanner; * Utility scanner that wraps a sortable collection and serves * as a KeyValueScanner. */ +@InterfaceAudience.Private public class CollectionBackedScanner extends NonLazyKeyValueScanner { final private Iterable data; final KeyValue.KVComparator comparator; diff --git a/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilter.java b/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilter.java index 09e20a46c79..04cf70df82c 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilter.java +++ b/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilter.java @@ -24,6 +24,7 @@ import java.io.DataInput; import java.io.IOException; import java.nio.ByteBuffer; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.hfile.BlockType; import org.apache.hadoop.hbase.io.hfile.FixedFileTrailer; import org.apache.hadoop.hbase.io.hfile.HFile; @@ -38,6 +39,7 @@ import org.apache.hadoop.io.RawComparator; * block stream, and loaded on demand at query time. This class only provides * reading capabilities. */ +@InterfaceAudience.Private public class CompoundBloomFilter extends CompoundBloomFilterBase implements BloomFilter { diff --git a/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterBase.java b/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterBase.java index a096087047c..1273e59a9c7 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterBase.java +++ b/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterBase.java @@ -20,9 +20,11 @@ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.io.RawComparator; +@InterfaceAudience.Private public class CompoundBloomFilterBase implements BloomFilterBase { /** diff --git a/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterWriter.java b/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterWriter.java index 8571d9c841b..6012cef5812 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterWriter.java +++ b/src/main/java/org/apache/hadoop/hbase/util/CompoundBloomFilterWriter.java @@ -28,6 +28,7 @@ import java.util.Queue; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.io.hfile.BlockType; import org.apache.hadoop.hbase.io.hfile.HFileBlockIndex; import org.apache.hadoop.hbase.io.hfile.InlineBlockWriter; @@ -39,6 +40,7 @@ import org.apache.hadoop.io.Writable; * section of an {@link org.apache.hadoop.hbase.io.hfile.HFile} to the * {@link CompoundBloomFilter} class. */ +@InterfaceAudience.Private public class CompoundBloomFilterWriter extends CompoundBloomFilterBase implements BloomFilterWriter, InlineBlockWriter { diff --git a/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java b/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java index f37d262ce43..db2c8b4092d 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java +++ b/src/main/java/org/apache/hadoop/hbase/util/CompressionTest.java @@ -23,6 +23,8 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -36,6 +38,8 @@ import org.apache.hadoop.io.compress.Compressor; * Compression validation test. Checks compression is working. Be sure to run * on every node in your cluster. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class CompressionTest { static final Log LOG = LogFactory.getLog(CompressionTest.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/DefaultEnvironmentEdge.java b/src/main/java/org/apache/hadoop/hbase/util/DefaultEnvironmentEdge.java index 66f9192ed78..181bedc6713 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/DefaultEnvironmentEdge.java +++ b/src/main/java/org/apache/hadoop/hbase/util/DefaultEnvironmentEdge.java @@ -19,9 +19,14 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Default implementation of an environment edge. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class DefaultEnvironmentEdge implements EnvironmentEdge { diff --git a/src/main/java/org/apache/hadoop/hbase/util/DirectMemoryUtils.java b/src/main/java/org/apache/hadoop/hbase/util/DirectMemoryUtils.java index f9081f20748..b58894ea86e 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/DirectMemoryUtils.java +++ b/src/main/java/org/apache/hadoop/hbase/util/DirectMemoryUtils.java @@ -27,8 +27,13 @@ import java.lang.reflect.Method; import java.nio.ByteBuffer; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + import com.google.common.base.Preconditions; +@InterfaceAudience.Public +@InterfaceStability.Evolving public class DirectMemoryUtils { /** * @return the setting of -XX:MaxDirectMemorySize as a long. Returns 0 if diff --git a/src/main/java/org/apache/hadoop/hbase/util/EnvironmentEdge.java b/src/main/java/org/apache/hadoop/hbase/util/EnvironmentEdge.java index 16e65d34dc7..6514083ffda 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/EnvironmentEdge.java +++ b/src/main/java/org/apache/hadoop/hbase/util/EnvironmentEdge.java @@ -19,12 +19,17 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Has some basic interaction with the environment. Alternate implementations * can be used where required (eg in tests). * * @see EnvironmentEdgeManager */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public interface EnvironmentEdge { /** diff --git a/src/main/java/org/apache/hadoop/hbase/util/EnvironmentEdgeManager.java b/src/main/java/org/apache/hadoop/hbase/util/EnvironmentEdgeManager.java index 9984b4bd2f5..311db22b5a8 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/EnvironmentEdgeManager.java +++ b/src/main/java/org/apache/hadoop/hbase/util/EnvironmentEdgeManager.java @@ -19,11 +19,16 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Manages a singleton instance of the environment edge. This class shall * implement static versions of the interface {@link EnvironmentEdge}, then * defer to the delegate on invocation. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class EnvironmentEdgeManager { private static volatile EnvironmentEdge delegate = new DefaultEnvironmentEdge(); diff --git a/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java b/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java index 3d7787985ce..c9f5833e5db 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java +++ b/src/main/java/org/apache/hadoop/hbase/util/FSHDFSUtils.java @@ -25,6 +25,8 @@ import java.lang.reflect.InvocationTargetException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataOutputStream; import org.apache.hadoop.fs.FileSystem; @@ -38,6 +40,8 @@ import org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException; /** * Implementation for hdfs */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class FSHDFSUtils extends FSUtils{ private static final Log LOG = LogFactory.getLog(FSHDFSUtils.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/FSMapRUtils.java b/src/main/java/org/apache/hadoop/hbase/util/FSMapRUtils.java index e70b0d476e5..f6db7be7fff 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/FSMapRUtils.java +++ b/src/main/java/org/apache/hadoop/hbase/util/FSMapRUtils.java @@ -19,6 +19,8 @@ package org.apache.hadoop.hbase.util; import java.io.IOException; + +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -29,6 +31,7 @@ import org.apache.commons.logging.LogFactory; /** * MapR implementation. */ +@InterfaceAudience.Private public class FSMapRUtils extends FSUtils { private static final Log LOG = LogFactory.getLog(FSMapRUtils.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/FSTableDescriptors.java b/src/main/java/org/apache/hadoop/hbase/util/FSTableDescriptors.java index a59acde01b4..62cf6acf350 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/FSTableDescriptors.java +++ b/src/main/java/org/apache/hadoop/hbase/util/FSTableDescriptors.java @@ -31,6 +31,7 @@ import java.util.regex.Pattern; import org.apache.commons.lang.NotImplementedException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; @@ -61,6 +62,7 @@ import org.apache.hadoop.hbase.TableExistsException; * will not be lots of concurrent clients making table schema edits. If so, * the below needs a bit of a reworking and perhaps some supporting api in hdfs. */ +@InterfaceAudience.Private public class FSTableDescriptors implements TableDescriptors { private static final Log LOG = LogFactory.getLog(FSTableDescriptors.class); private final FileSystem fs; diff --git a/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java b/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java index 7dfbc15a8a6..d2d7efe361d 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java +++ b/src/main/java/org/apache/hadoop/hbase/util/FSUtils.java @@ -32,6 +32,8 @@ import java.util.Map; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.BlockLocation; import org.apache.hadoop.fs.FSDataInputStream; @@ -54,6 +56,8 @@ import org.apache.hadoop.util.StringUtils; /** * Utility methods for interacting with the underlying file system. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class FSUtils { private static final Log LOG = LogFactory.getLog(FSUtils.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/FileSystemVersionException.java b/src/main/java/org/apache/hadoop/hbase/util/FileSystemVersionException.java index 5235121fe38..d2ef255996b 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/FileSystemVersionException.java +++ b/src/main/java/org/apache/hadoop/hbase/util/FileSystemVersionException.java @@ -22,7 +22,12 @@ package org.apache.hadoop.hbase.util; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** Thrown when the file system needs to be upgraded */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class FileSystemVersionException extends IOException { private static final long serialVersionUID = 1004053363L; diff --git a/src/main/java/org/apache/hadoop/hbase/util/HBaseConfTool.java b/src/main/java/org/apache/hadoop/hbase/util/HBaseConfTool.java index 225f92c7883..526850b15ce 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/HBaseConfTool.java +++ b/src/main/java/org/apache/hadoop/hbase/util/HBaseConfTool.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -27,6 +28,7 @@ import org.apache.hadoop.hbase.HBaseConfiguration; * Tool that prints out a configuration. * Pass the configuration key on the command-line. */ +@InterfaceAudience.Private public class HBaseConfTool { public static void main(String args[]) { if (args.length < 1) { diff --git a/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java b/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java index e55b906cba3..720841c30f1 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java +++ b/src/main/java/org/apache/hadoop/hbase/util/HBaseFsck.java @@ -38,6 +38,8 @@ import java.util.concurrent.atomic.AtomicInteger; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FileStatus; @@ -84,6 +86,8 @@ import com.google.common.collect.TreeMultimap; * Check consistency among the in-memory states of the master and the * region server(s) and the state of data in HDFS. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HBaseFsck { public static final long DEFAULT_TIME_LAG = 60000; // default value of 1 minute public static final long DEFAULT_SLEEP_BEFORE_RERUN = 10000; diff --git a/src/main/java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java b/src/main/java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java index 77f60a021e7..5916d9c347e 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java +++ b/src/main/java/org/apache/hadoop/hbase/util/HBaseFsckRepair.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.util; import java.io.IOException; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.NotServingRegionException; @@ -33,6 +35,8 @@ import org.apache.hadoop.hbase.client.HConnectionManager; import org.apache.hadoop.hbase.ipc.HRegionInterface; import org.apache.zookeeper.KeeperException; +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HBaseFsckRepair { /** diff --git a/src/main/java/org/apache/hadoop/hbase/util/HMerge.java b/src/main/java/org/apache/hadoop/hbase/util/HMerge.java index a685aa77c39..f7e5255228c 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/HMerge.java +++ b/src/main/java/org/apache/hadoop/hbase/util/HMerge.java @@ -27,6 +27,7 @@ import java.util.Random; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -54,6 +55,7 @@ import org.apache.hadoop.hbase.regionserver.wal.HLog; * A non-instantiable class that has a static method capable of compacting * a table by merging adjacent regions. */ +@InterfaceAudience.Private class HMerge { // TODO: Where is this class used? How does it relate to Merge in same package? static final Log LOG = LogFactory.getLog(HMerge.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/HasThread.java b/src/main/java/org/apache/hadoop/hbase/util/HasThread.java index 076604f4af1..dbf9164d81d 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/HasThread.java +++ b/src/main/java/org/apache/hadoop/hbase/util/HasThread.java @@ -19,6 +19,8 @@ package org.apache.hadoop.hbase.util; import java.lang.Thread.UncaughtExceptionHandler; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Abstract class which contains a Thread and delegates the common Thread * methods to that instance. @@ -27,6 +29,7 @@ import java.lang.Thread.UncaughtExceptionHandler; * something internal to the JDK uses Thread.currentThread() as a monitor * lock. This can produce deadlocks like HBASE-4367, HBASE-4101, etc. */ +@InterfaceAudience.Private public abstract class HasThread implements Runnable { private final Thread thread; diff --git a/src/main/java/org/apache/hadoop/hbase/util/Hash.java b/src/main/java/org/apache/hadoop/hbase/util/Hash.java index 8a3dcf5bc9a..9857dc30ba8 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Hash.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Hash.java @@ -20,11 +20,15 @@ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; /** * This class represents a common API for hashing functions. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class Hash { /** Constant to denote invalid hash type. */ public static final int INVALID_HASH = -1; diff --git a/src/main/java/org/apache/hadoop/hbase/util/HashedBytes.java b/src/main/java/org/apache/hadoop/hbase/util/HashedBytes.java index e6471cbbfd1..2fdf3dca5fc 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/HashedBytes.java +++ b/src/main/java/org/apache/hadoop/hbase/util/HashedBytes.java @@ -21,10 +21,15 @@ package org.apache.hadoop.hbase.util; import java.util.Arrays; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * This class encapsulates a byte array and overrides hashCode and equals so * that it's identity is based on the data rather than the array instance. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class HashedBytes { private final byte[] bytes; diff --git a/src/main/java/org/apache/hadoop/hbase/util/IdLock.java b/src/main/java/org/apache/hadoop/hbase/util/IdLock.java index e9202dde61c..1e11ca73b74 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/IdLock.java +++ b/src/main/java/org/apache/hadoop/hbase/util/IdLock.java @@ -24,6 +24,8 @@ import java.io.InterruptedIOException; import java.util.concurrent.ConcurrentHashMap; import java.util.concurrent.ConcurrentMap; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Allows multiple concurrent clients to lock on a numeric id with a minimal * memory overhead. The intended usage is as follows: @@ -36,6 +38,7 @@ import java.util.concurrent.ConcurrentMap; * idLock.releaseLockEntry(lockEntry); * } */ +@InterfaceAudience.Private public class IdLock { /** An entry returned to the client as a lock object */ diff --git a/src/main/java/org/apache/hadoop/hbase/util/IncrementingEnvironmentEdge.java b/src/main/java/org/apache/hadoop/hbase/util/IncrementingEnvironmentEdge.java index e105b779c69..d61fb0f0940 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/IncrementingEnvironmentEdge.java +++ b/src/main/java/org/apache/hadoop/hbase/util/IncrementingEnvironmentEdge.java @@ -19,9 +19,12 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; + /** * Uses an incrementing algorithm instead of the default. */ +@InterfaceAudience.Private public class IncrementingEnvironmentEdge implements EnvironmentEdge { private long timeIncrement = 1; diff --git a/src/main/java/org/apache/hadoop/hbase/util/InfoServer.java b/src/main/java/org/apache/hadoop/hbase/util/InfoServer.java index 552974502bf..c2cbd1b73c5 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/InfoServer.java +++ b/src/main/java/org/apache/hadoop/hbase/util/InfoServer.java @@ -26,6 +26,7 @@ import java.lang.reflect.Method; import java.net.URL; import java.util.Map; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.http.HttpServer; @@ -41,6 +42,7 @@ import org.mortbay.jetty.servlet.DefaultServlet; * "/static/" -> points to common static files (src/hbase-webapps/static) * "/" -> the jsp server code from (src/hbase-webapps/) */ +@InterfaceAudience.Private public class InfoServer extends HttpServer { private final Configuration config; diff --git a/src/main/java/org/apache/hadoop/hbase/util/JVMClusterUtil.java b/src/main/java/org/apache/hadoop/hbase/util/JVMClusterUtil.java index 6056f73fee4..c21377c066a 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/JVMClusterUtil.java +++ b/src/main/java/org/apache/hadoop/hbase/util/JVMClusterUtil.java @@ -25,6 +25,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.hbase.master.HMaster; @@ -34,6 +35,7 @@ import org.apache.hadoop.hbase.regionserver.ShutdownHook; /** * Utility used running a cluster all in the one JVM. */ +@InterfaceAudience.Private public class JVMClusterUtil { private static final Log LOG = LogFactory.getLog(JVMClusterUtil.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java b/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java index f4f61d0210f..39715832db7 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java +++ b/src/main/java/org/apache/hadoop/hbase/util/JenkinsHash.java @@ -25,6 +25,9 @@ import static java.lang.Integer.rotateLeft; import java.io.FileInputStream; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Produces 32-bit hash for hash table lookup. * @@ -40,6 +43,8 @@ import java.io.IOException; * @see Has update on the * Dr. Dobbs Article */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class JenkinsHash extends Hash { private static final int BYTE_MASK = 0xff; diff --git a/src/main/java/org/apache/hadoop/hbase/util/JvmVersion.java b/src/main/java/org/apache/hadoop/hbase/util/JvmVersion.java index b7eb7e508e0..57701166ccd 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/JvmVersion.java +++ b/src/main/java/org/apache/hadoop/hbase/util/JvmVersion.java @@ -22,11 +22,16 @@ package org.apache.hadoop.hbase.util; import java.util.HashSet; import java.util.Set; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Certain JVM versions are known to be unstable with HBase. This * class has a utility function to determine whether the current JVM * is known to be unstable. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public abstract class JvmVersion { private static Set BAD_JVM_VERSIONS = new HashSet(); static { diff --git a/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java b/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java index 958edef8d5a..e7efc4d5dcf 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java +++ b/src/main/java/org/apache/hadoop/hbase/util/KeyRange.java @@ -19,9 +19,12 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; + /** * A key range use in split coverage. */ +@InterfaceAudience.Private public interface KeyRange { abstract byte[] getStartKey(); diff --git a/src/main/java/org/apache/hadoop/hbase/util/Keying.java b/src/main/java/org/apache/hadoop/hbase/util/Keying.java index 2e3d02761a1..43451ee99d4 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Keying.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Keying.java @@ -23,6 +23,9 @@ import java.util.StringTokenizer; import java.util.regex.Matcher; import java.util.regex.Pattern; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Utility creating hbase friendly keys. * Use fabricating row names or column qualifiers. @@ -30,6 +33,8 @@ import java.util.regex.Pattern; * http or https. * @see Bytes#split(byte[], byte[], int) */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Keying { private static final String SCHEME = "r:"; private static final Pattern URI_RE_PARSER = diff --git a/src/main/java/org/apache/hadoop/hbase/util/MD5Hash.java b/src/main/java/org/apache/hadoop/hbase/util/MD5Hash.java index b2998c9b040..a03bd364b58 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/MD5Hash.java +++ b/src/main/java/org/apache/hadoop/hbase/util/MD5Hash.java @@ -26,11 +26,15 @@ import java.security.NoSuchAlgorithmException; import org.apache.commons.codec.binary.Hex; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; /** * Utility class for MD5 * MD5 hash produces a 128-bit digest. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class MD5Hash { private static final Log LOG = LogFactory.getLog(MD5Hash.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/ManualEnvironmentEdge.java b/src/main/java/org/apache/hadoop/hbase/util/ManualEnvironmentEdge.java index d698df17ea9..853a9cd5c86 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/ManualEnvironmentEdge.java +++ b/src/main/java/org/apache/hadoop/hbase/util/ManualEnvironmentEdge.java @@ -19,10 +19,13 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; + /** * An environment edge that uses a manually set value. This is useful for testing events that are supposed to * happen in the same millisecond. */ +@InterfaceAudience.Private public class ManualEnvironmentEdge implements EnvironmentEdge { // Sometimes 0 ts might have a special value, so lets start with 1 diff --git a/src/main/java/org/apache/hadoop/hbase/util/Merge.java b/src/main/java/org/apache/hadoop/hbase/util/Merge.java index 67d0fda5681..57361e530e1 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Merge.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Merge.java @@ -22,6 +22,8 @@ package org.apache.hadoop.hbase.util; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; import org.apache.hadoop.fs.FileSystem; @@ -50,6 +52,8 @@ import java.util.List; * Utility that can merge any two regions in the same table: adjacent, * overlapping or disjoint. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class Merge extends Configured implements Tool { static final Log LOG = LogFactory.getLog(Merge.class); private Path rootdir; diff --git a/src/main/java/org/apache/hadoop/hbase/util/MetaUtils.java b/src/main/java/org/apache/hadoop/hbase/util/MetaUtils.java index af8d734a9f5..737f70955d0 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/MetaUtils.java +++ b/src/main/java/org/apache/hadoop/hbase/util/MetaUtils.java @@ -29,6 +29,7 @@ import java.util.TreeMap; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -54,6 +55,7 @@ import org.apache.hadoop.hbase.regionserver.wal.HLog; * how you use this class. If used during migrations, be careful when using * this class to check whether migration is needed. */ +@InterfaceAudience.Private public class MetaUtils { private static final Log LOG = LogFactory.getLog(MetaUtils.class); private final Configuration conf; diff --git a/src/main/java/org/apache/hadoop/hbase/util/Methods.java b/src/main/java/org/apache/hadoop/hbase/util/Methods.java index a6f48352b73..8f0a6e3b753 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Methods.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Methods.java @@ -26,7 +26,11 @@ import java.lang.reflect.UndeclaredThrowableException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; +@InterfaceAudience.Public +@InterfaceStability.Stable public class Methods { private static Log LOG = LogFactory.getLog(Methods.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/MurmurHash.java b/src/main/java/org/apache/hadoop/hbase/util/MurmurHash.java index 085bf1e3429..dc498376f3a 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/MurmurHash.java +++ b/src/main/java/org/apache/hadoop/hbase/util/MurmurHash.java @@ -20,6 +20,9 @@ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * This is a very fast, non-cryptographic hash suitable for general hash-based * lookup. See http://murmurhash.googlepages.com/ for more details. @@ -27,6 +30,8 @@ package org.apache.hadoop.hbase.util; *

    The C version of MurmurHash 2.0 found at that site was ported * to Java by Andrzej Bialecki (ab at getopt org).

    */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class MurmurHash extends Hash { private static MurmurHash _instance = new MurmurHash(); diff --git a/src/main/java/org/apache/hadoop/hbase/util/Objects.java b/src/main/java/org/apache/hadoop/hbase/util/Objects.java index 088b54a3bb7..42e88a822f0 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Objects.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Objects.java @@ -26,6 +26,7 @@ import java.util.Comparator; import java.util.HashMap; import java.util.List; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.client.Action; import org.apache.hadoop.hbase.client.Delete; @@ -37,6 +38,7 @@ import org.apache.hadoop.hbase.client.Put; /** * Utility methods for interacting with object instances. */ +@InterfaceAudience.Private public class Objects { private static class QuantityMap extends HashMap { public void increment(String type, int count) { diff --git a/src/main/java/org/apache/hadoop/hbase/util/Pair.java b/src/main/java/org/apache/hadoop/hbase/util/Pair.java index 9b4a567c93d..c0c2f2953d7 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Pair.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Pair.java @@ -22,11 +22,16 @@ package org.apache.hadoop.hbase.util; import java.io.Serializable; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * A generic class for pairs. * @param * @param */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Pair implements Serializable { private static final long serialVersionUID = -3986244606585552569L; diff --git a/src/main/java/org/apache/hadoop/hbase/util/PairOfSameType.java b/src/main/java/org/apache/hadoop/hbase/util/PairOfSameType.java index ddee30d2328..a83f89d14ab 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/PairOfSameType.java +++ b/src/main/java/org/apache/hadoop/hbase/util/PairOfSameType.java @@ -23,12 +23,16 @@ package org.apache.hadoop.hbase.util; import java.util.Iterator; import org.apache.commons.lang.NotImplementedException; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; /** * A generic, immutable class for pairs of objects both of type T. * @param * @see Pair if Types differ. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class PairOfSameType implements Iterable { private final T first; private final T second; diff --git a/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java b/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java index 1956e6b8ae5..8e2a8567144 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java +++ b/src/main/java/org/apache/hadoop/hbase/util/PoolMap.java @@ -32,6 +32,9 @@ import java.util.concurrent.ConcurrentLinkedQueue; import java.util.concurrent.CopyOnWriteArrayList; import java.util.concurrent.atomic.AtomicInteger; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * * The PoolMap maps a key to a collection of values, the elements @@ -50,6 +53,8 @@ import java.util.concurrent.atomic.AtomicInteger; * @param * the type of the resource being pooled */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class PoolMap implements Map { private PoolType poolType; diff --git a/src/main/java/org/apache/hadoop/hbase/util/ProtoUtil.java b/src/main/java/org/apache/hadoop/hbase/util/ProtoUtil.java index 92129676256..bdbe7b7c022 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/ProtoUtil.java +++ b/src/main/java/org/apache/hadoop/hbase/util/ProtoUtil.java @@ -21,6 +21,9 @@ package org.apache.hadoop.hbase.util; import java.io.DataInput; import java.io.IOException; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public abstract class ProtoUtil { /** diff --git a/src/main/java/org/apache/hadoop/hbase/util/RegionSplitCalculator.java b/src/main/java/org/apache/hadoop/hbase/util/RegionSplitCalculator.java index 69a87ab07a3..17678dd1435 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/RegionSplitCalculator.java +++ b/src/main/java/org/apache/hadoop/hbase/util/RegionSplitCalculator.java @@ -26,6 +26,7 @@ import java.util.TreeSet; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.util.Bytes.ByteArrayComparator; import com.google.common.collect.ArrayListMultimap; @@ -48,6 +49,7 @@ import com.google.common.collect.TreeMultimap; * * @param */ +@InterfaceAudience.Private public class RegionSplitCalculator { final static Log LOG = LogFactory.getLog(RegionSplitCalculator.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java b/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java index 02718832b7f..f6d088f1fee 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java +++ b/src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java @@ -40,6 +40,7 @@ import org.apache.commons.lang.ArrayUtils; import org.apache.commons.lang.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FSDataInputStream; import org.apache.hadoop.fs.FSDataOutputStream; @@ -134,6 +135,7 @@ import com.google.common.collect.Sets; * grows too large, use this script to perform a network IO safe rolling split * of all regions. */ +@InterfaceAudience.Private public class RegionSplitter { static final Log LOG = LogFactory.getLog(RegionSplitter.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/RetryCounter.java b/src/main/java/org/apache/hadoop/hbase/util/RetryCounter.java index c7b62c9d792..0d1c5eae6b0 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/RetryCounter.java +++ b/src/main/java/org/apache/hadoop/hbase/util/RetryCounter.java @@ -23,7 +23,9 @@ import java.util.concurrent.TimeUnit; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +@InterfaceAudience.Private public class RetryCounter { private static final Log LOG = LogFactory.getLog(RetryCounter.class); private final int maxRetries; diff --git a/src/main/java/org/apache/hadoop/hbase/util/RetryCounterFactory.java b/src/main/java/org/apache/hadoop/hbase/util/RetryCounterFactory.java index 445234e29b5..2bb33535ecf 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/RetryCounterFactory.java +++ b/src/main/java/org/apache/hadoop/hbase/util/RetryCounterFactory.java @@ -21,6 +21,9 @@ package org.apache.hadoop.hbase.util; import java.util.concurrent.TimeUnit; +import org.apache.hadoop.classification.InterfaceAudience; + +@InterfaceAudience.Private public class RetryCounterFactory { private final int maxRetries; private final int retryIntervalMillis; diff --git a/src/main/java/org/apache/hadoop/hbase/util/ServerCommandLine.java b/src/main/java/org/apache/hadoop/hbase/util/ServerCommandLine.java index b2f37705675..fccdd215fd8 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/ServerCommandLine.java +++ b/src/main/java/org/apache/hadoop/hbase/util/ServerCommandLine.java @@ -24,6 +24,7 @@ import java.lang.management.ManagementFactory; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configured; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -33,6 +34,7 @@ import org.apache.hadoop.util.ToolRunner; /** * Base class for command lines that start up various HBase daemons. */ +@InterfaceAudience.Private public abstract class ServerCommandLine extends Configured implements Tool { private static final Log LOG = LogFactory.getLog(ServerCommandLine.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java b/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java index a8d7d683073..c8b657ac836 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Sleeper.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.util; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Stoppable; /** @@ -29,6 +31,8 @@ import org.apache.hadoop.hbase.Stoppable; * will return if the flag is set (rather than go back to sleep until its * sleep time is up). */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Sleeper { private final Log LOG = LogFactory.getLog(this.getClass().getName()); private final int period; diff --git a/src/main/java/org/apache/hadoop/hbase/util/SoftValueSortedMap.java b/src/main/java/org/apache/hadoop/hbase/util/SoftValueSortedMap.java index 7229e146501..07622999ade 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/SoftValueSortedMap.java +++ b/src/main/java/org/apache/hadoop/hbase/util/SoftValueSortedMap.java @@ -33,6 +33,9 @@ import java.util.Set; import java.util.SortedMap; import java.util.TreeMap; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * A SortedMap implementation that uses Soft Reference values * internally to make it play well with the GC when in a low-memory @@ -41,6 +44,8 @@ import java.util.TreeMap; * @param key class * @param value class */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class SoftValueSortedMap implements SortedMap { private final SortedMap> internalMap; private final ReferenceQueue rq = new ReferenceQueue(); diff --git a/src/main/java/org/apache/hadoop/hbase/util/SortedCopyOnWriteSet.java b/src/main/java/org/apache/hadoop/hbase/util/SortedCopyOnWriteSet.java index f04f306e812..d3d0372553e 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/SortedCopyOnWriteSet.java +++ b/src/main/java/org/apache/hadoop/hbase/util/SortedCopyOnWriteSet.java @@ -26,6 +26,9 @@ import java.util.Iterator; import java.util.SortedSet; import java.util.TreeSet; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Simple {@link java.util.SortedSet} implementation that uses an internal * {@link java.util.TreeSet} to provide ordering. All mutation operations @@ -43,6 +46,8 @@ import java.util.TreeSet; * @param A class implementing {@link java.lang.Comparable} or able to be * compared by a provided comparator. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class SortedCopyOnWriteSet implements SortedSet { private SortedSet internalSet; diff --git a/src/main/java/org/apache/hadoop/hbase/util/Strings.java b/src/main/java/org/apache/hadoop/hbase/util/Strings.java index 9cbca9e5424..b14d0372139 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Strings.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Strings.java @@ -19,9 +19,14 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Utility for Strings. */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Strings { public final static String DEFAULT_SEPARATOR = "="; public final static String DEFAULT_KEYVALUE_SEPARATOR = ", "; diff --git a/src/main/java/org/apache/hadoop/hbase/util/Threads.java b/src/main/java/org/apache/hadoop/hbase/util/Threads.java index ce880a598b5..5d23df98990 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Threads.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Threads.java @@ -28,11 +28,15 @@ import java.util.concurrent.TimeUnit; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.util.ReflectionUtils; /** * Thread Utility */ +@InterfaceAudience.Public +@InterfaceStability.Stable public class Threads { protected static final Log LOG = LogFactory.getLog(Threads.class); diff --git a/src/main/java/org/apache/hadoop/hbase/util/VersionInfo.java b/src/main/java/org/apache/hadoop/hbase/util/VersionInfo.java index b516e16b6a2..fb16ad80d62 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/VersionInfo.java +++ b/src/main/java/org/apache/hadoop/hbase/util/VersionInfo.java @@ -23,6 +23,8 @@ package org.apache.hadoop.hbase.util; import org.apache.commons.logging.LogFactory; import java.io.PrintWriter; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.VersionAnnotation; import org.apache.hadoop.hbase.master.HMaster; import org.apache.commons.logging.Log; @@ -31,6 +33,8 @@ import org.apache.commons.logging.Log; * This class finds the package info for hbase and the VersionAnnotation * information. Taken from hadoop. Only name of annotation is different. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class VersionInfo { private static final Log LOG = LogFactory.getLog(VersionInfo.class.getName()); private static Package myPackage; diff --git a/src/main/java/org/apache/hadoop/hbase/util/Writables.java b/src/main/java/org/apache/hadoop/hbase/util/Writables.java index 1719e842368..3d20723bc10 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/Writables.java +++ b/src/main/java/org/apache/hadoop/hbase/util/Writables.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.util; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.migration.HRegionInfo090x; import org.apache.hadoop.io.DataInputBuffer; @@ -35,6 +36,7 @@ import java.util.List; /** * Utility class with methods for manipulating Writable objects */ +@InterfaceAudience.Private public class Writables { /** * @param w writable diff --git a/src/main/java/org/apache/hadoop/hbase/util/hbck/OfflineMetaRepair.java b/src/main/java/org/apache/hadoop/hbase/util/hbck/OfflineMetaRepair.java index 29e8bb2f51e..d57bb6bce3d 100644 --- a/src/main/java/org/apache/hadoop/hbase/util/hbck/OfflineMetaRepair.java +++ b/src/main/java/org/apache/hadoop/hbase/util/hbck/OfflineMetaRepair.java @@ -21,6 +21,8 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HConstants; @@ -38,6 +40,8 @@ import org.apache.hadoop.io.MultipleIOException; * * hbase org.apache.hadoop.hbase.util.hbck.OfflineMetaRepair ... */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class OfflineMetaRepair { private static final Log LOG = LogFactory.getLog(HBaseFsck.class.getName()); HBaseFsck fsck; diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ClusterId.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ClusterId.java index 0b1b647be49..3fa83e61ab8 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ClusterId.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ClusterId.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.zookeeper; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.util.Bytes; import org.apache.zookeeper.KeeperException; @@ -30,6 +31,7 @@ import org.apache.zookeeper.KeeperException; * master on startup, and is subsequently available to all watchers (including * clients). */ +@InterfaceAudience.Private public class ClusterId { private ZooKeeperWatcher watcher; private Abortable abortable; diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ClusterStatusTracker.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ClusterStatusTracker.java index 7e1a9528650..61e7367ed5a 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ClusterStatusTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ClusterStatusTracker.java @@ -21,6 +21,7 @@ package org.apache.hadoop.hbase.zookeeper; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.ClusterStatus; import org.apache.hadoop.hbase.util.Bytes; @@ -33,6 +34,7 @@ import org.apache.zookeeper.KeeperException; * cluster attributes up in zookeeper. * */ +@InterfaceAudience.Private public class ClusterStatusTracker extends ZooKeeperNodeTracker { private static final Log LOG = LogFactory.getLog(ClusterStatusTracker.class); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java index e90edf9b293..fcc2e70c5ba 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/DrainingServerTracker.java @@ -24,6 +24,7 @@ import java.util.TreeSet; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.HServerAddress; @@ -44,6 +45,7 @@ import org.apache.zookeeper.KeeperException; * {@link ServerManager#addServerToDrainList(ServerName)} * */ +@InterfaceAudience.Private public class DrainingServerTracker extends ZooKeeperListener { private static final Log LOG = LogFactory.getLog(DrainingServerTracker.class); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java index d551c6fc74a..b37b125fc51 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/HQuorumPeer.java @@ -31,6 +31,8 @@ import java.util.List; import java.util.Properties; import java.util.Map.Entry; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.net.DNS; @@ -47,6 +49,8 @@ import org.apache.zookeeper.server.quorum.QuorumPeerMain; * control over the process. This class uses {@link ZKConfig} to parse the * zoo.cfg and inject variables from HBase's site.xml configuration in. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HQuorumPeer { /** diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/MasterSchemaChangeTracker.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/MasterSchemaChangeTracker.java index 39c940e45c0..6c723b47127 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/MasterSchemaChangeTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/MasterSchemaChangeTracker.java @@ -25,6 +25,7 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.master.MasterServices; @@ -35,6 +36,7 @@ import org.apache.hadoop.hbase.util.Writables; import org.apache.hadoop.io.Writable; import org.apache.zookeeper.KeeperException; +@InterfaceAudience.Private public class MasterSchemaChangeTracker extends ZooKeeperNodeTracker { public static final Log LOG = LogFactory.getLog(MasterSchemaChangeTracker.class); private final MasterServices masterServices; diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaNodeTracker.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaNodeTracker.java index 4da6f962728..f4ad6735c04 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaNodeTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/MetaNodeTracker.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.zookeeper; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.HRegionInfo; @@ -29,6 +30,7 @@ import org.apache.hadoop.hbase.HRegionInfo; * receive any notification for that assignment. You will receive a * notification after META has been successfully assigned to a new location. */ +@InterfaceAudience.Private public class MetaNodeTracker extends ZooKeeperNodeTracker { /** * Creates a meta node tracker. diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java index 786ae210218..ea14c4b7234 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/MiniZooKeeperCluster.java @@ -34,6 +34,8 @@ import java.util.Random; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.hbase.HConstants; @@ -46,6 +48,8 @@ import org.apache.zookeeper.server.persistence.FileTxnLog; * of redoing it, we should contribute updates to their code which let us more * easily access testing helper objects. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class MiniZooKeeperCluster { private static final Log LOG = LogFactory.getLog(MiniZooKeeperCluster.class); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java index a097ec00a97..0e6d9ff263d 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java @@ -26,6 +26,8 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.RetryCounter; import org.apache.hadoop.hbase.util.RetryCounterFactory; @@ -61,6 +63,8 @@ import org.apache.zookeeper.data.Stat; * succeeded an the znode it created is "x-352-109". * @see "http://wiki.apache.org/hadoop/ZooKeeper/ErrorHandling" */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class RecoverableZooKeeper { private static final Log LOG = LogFactory.getLog(RecoverableZooKeeper.class); // the actual ZooKeeper client instance diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionServerTracker.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionServerTracker.java index 225df66f2a9..d95ff1409ab 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionServerTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/RegionServerTracker.java @@ -27,6 +27,7 @@ import java.util.TreeSet; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.master.ServerManager; @@ -42,6 +43,7 @@ import org.apache.zookeeper.KeeperException; *

    If an RS node gets deleted, this automatically handles calling of * {@link ServerManager#expireServer(ServerName)} */ +@InterfaceAudience.Private public class RegionServerTracker extends ZooKeeperListener { private static final Log LOG = LogFactory.getLog(RegionServerTracker.class); private NavigableSet regionServers = new TreeSet(); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/RootRegionTracker.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/RootRegionTracker.java index 48a8b3d3bde..3e0975bd085 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/RootRegionTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/RootRegionTracker.java @@ -19,6 +19,7 @@ */ package org.apache.hadoop.hbase.zookeeper; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.catalog.RootLocationEditor; @@ -32,6 +33,7 @@ import org.apache.zookeeper.KeeperException; * out of RegionServerServices. * This class has a watcher on the root location and notices changes. */ +@InterfaceAudience.Private public class RootRegionTracker extends ZooKeeperNodeTracker { /** * Creates a root region location tracker. diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/SchemaChangeTracker.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/SchemaChangeTracker.java index 9233eee223b..48d4ff774d8 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/SchemaChangeTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/SchemaChangeTracker.java @@ -20,6 +20,7 @@ package org.apache.hadoop.hbase.zookeeper; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.monitoring.MonitoredTask; import org.apache.hadoop.hbase.monitoring.TaskMonitor; @@ -40,6 +41,7 @@ import java.util.List; * alter schema requests from master and updates the status once the schema change * is complete. */ +@InterfaceAudience.Private public class SchemaChangeTracker extends ZooKeeperNodeTracker { public static final Log LOG = LogFactory.getLog(SchemaChangeTracker.class); private RegionServerServices regionServer = null; diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java index f14b026a0bb..e94b672f063 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKAssign.java @@ -23,6 +23,8 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.executor.RegionTransitionData; @@ -88,6 +90,8 @@ import org.apache.zookeeper.data.Stat; * * */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ZKAssign { private static final Log LOG = LogFactory.getLog(ZKAssign.class); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKConfig.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKConfig.java index 6d955acfe18..cb1539e7b87 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKConfig.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKConfig.java @@ -30,6 +30,8 @@ import java.util.Map.Entry; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.util.StringUtils; @@ -37,6 +39,8 @@ import org.apache.hadoop.util.StringUtils; /** * Utility methods for reading, parsing, and building zookeeper configuration. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ZKConfig { private static final Log LOG = LogFactory.getLog(ZKConfig.class); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKLeaderManager.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKLeaderManager.java index f20179994cd..2a30b9c4678 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKLeaderManager.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKLeaderManager.java @@ -22,6 +22,8 @@ import java.util.concurrent.atomic.AtomicBoolean; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Stoppable; import org.apache.hadoop.hbase.util.Bytes; import org.apache.zookeeper.KeeperException; @@ -36,6 +38,8 @@ import org.apache.zookeeper.KeeperException; * be removed, and all waiting instances will be notified, with the race * to claim the leader znode beginning all over again. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ZKLeaderManager extends ZooKeeperListener { private static Log LOG = LogFactory.getLog(ZKLeaderManager.class); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKServerTool.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKServerTool.java index 500bd3c146e..df0b8091b5f 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKServerTool.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKServerTool.java @@ -23,6 +23,8 @@ package org.apache.hadoop.hbase.zookeeper; import java.util.Properties; import java.util.Map.Entry; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; @@ -30,6 +32,8 @@ import org.apache.hadoop.hbase.HBaseConfiguration; * Tool for reading ZooKeeper servers from HBase XML configuration and producing * a line-by-line list for use by bash scripts. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ZKServerTool { /** * Run the tool. diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKSplitLog.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKSplitLog.java index 02034dc8672..30d7fe91444 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKSplitLog.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKSplitLog.java @@ -31,6 +31,7 @@ import java.util.concurrent.atomic.AtomicLong; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import org.apache.hadoop.hbase.HConstants; @@ -42,6 +43,7 @@ import org.apache.hadoop.hbase.util.Bytes; * Common methods and attributes used by {@link SplitLogManager} and * {@link SplitLogWorker} */ +@InterfaceAudience.Private public class ZKSplitLog { private static final Log LOG = LogFactory.getLog(ZKSplitLog.class); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKTable.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKTable.java index 4f8aea2c328..093f2890031 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKTable.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKTable.java @@ -27,6 +27,7 @@ import java.util.Set; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hbase.master.AssignmentManager; import org.apache.hadoop.hbase.util.Bytes; import org.apache.zookeeper.KeeperException; @@ -42,6 +43,7 @@ import org.apache.zookeeper.KeeperException; *

    To save on trips to the zookeeper ensemble, internally we cache table * state. */ +@InterfaceAudience.Private public class ZKTable { // A znode will exist under the table directory if it is in any of the // following states: {@link TableState#ENABLING} , {@link TableState#DISABLING}, diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java index 719a17691ab..28ab5bec5a8 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZKUtil.java @@ -32,6 +32,8 @@ import java.util.Properties; import org.apache.commons.lang.StringUtils; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.EmptyWatcher; import org.apache.hadoop.hbase.HConstants; @@ -58,6 +60,8 @@ import org.apache.zookeeper.data.Stat; * zookeeper exception, so callers of these methods must handle appropriately. * If ZK is required for the operation, the server will need to be aborted. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ZKUtil { private static final Log LOG = LogFactory.getLog(ZKUtil.class); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperListener.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperListener.java index bfd1ce86a6f..c6157d5e2ac 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperListener.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperListener.java @@ -19,6 +19,9 @@ */ package org.apache.hadoop.hbase.zookeeper; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; + /** * Base class for internal listeners of ZooKeeper events. @@ -32,6 +35,8 @@ package org.apache.hadoop.hbase.zookeeper; * Note that the watcher will be blocked when invoking methods in listeners so * they must not be long-running. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class ZooKeeperListener { // Reference to the zk watcher which also contains configuration and constants diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java index c6e607ef133..7a348bc0006 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperNodeTracker.java @@ -21,6 +21,8 @@ package org.apache.hadoop.hbase.zookeeper; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.hbase.Abortable; import org.apache.zookeeper.KeeperException; @@ -33,6 +35,8 @@ import org.apache.zookeeper.KeeperException; *

    This is the base class used by trackers in both the Master and * RegionServers. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public abstract class ZooKeeperNodeTracker extends ZooKeeperListener { static final Log LOG = LogFactory.getLog(ZooKeeperNodeTracker.class); diff --git a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java index 45d7d6a88df..83d27e6c7fc 100644 --- a/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java +++ b/src/main/java/org/apache/hadoop/hbase/zookeeper/ZooKeeperWatcher.java @@ -29,6 +29,8 @@ import java.util.concurrent.CountDownLatch; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Abortable; import org.apache.hadoop.hbase.HConstants; @@ -51,6 +53,8 @@ import org.apache.zookeeper.data.ACL; *

    This class also holds and manages the connection to ZooKeeper. Code to * deal with connection related events and exceptions are handled here. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class ZooKeeperWatcher implements Watcher, Abortable { private static final Log LOG = LogFactory.getLog(ZooKeeperWatcher.class); diff --git a/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java b/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java index 498075d1bb6..38eb6a84599 100644 --- a/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java +++ b/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java @@ -44,6 +44,8 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.impl.Jdk14Logger; import org.apache.commons.logging.impl.Log4JLogger; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; @@ -94,6 +96,8 @@ import org.apache.zookeeper.ZooKeeper; * hbase-site.xml for logging and test-run configuration. It does not set * logging levels nor make changes to configuration parameters. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class HBaseTestingUtility { private static final Log LOG = LogFactory.getLog(HBaseTestingUtility.class); private Configuration conf; diff --git a/src/test/java/org/apache/hadoop/hbase/MiniHBaseCluster.java b/src/test/java/org/apache/hadoop/hbase/MiniHBaseCluster.java index b36d12ca283..6af9188ca0e 100644 --- a/src/test/java/org/apache/hadoop/hbase/MiniHBaseCluster.java +++ b/src/test/java/org/apache/hadoop/hbase/MiniHBaseCluster.java @@ -26,6 +26,8 @@ import java.util.List; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.classification.InterfaceAudience; +import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.hbase.client.HConnectionManager; @@ -44,6 +46,8 @@ import org.apache.hadoop.io.MapWritable; * if we are running on DistributedFilesystem, create a FileSystem instance * each and will close down their instance on the way out. */ +@InterfaceAudience.Public +@InterfaceStability.Evolving public class MiniHBaseCluster { static final Log LOG = LogFactory.getLog(MiniHBaseCluster.class.getName()); private Configuration conf;