HBASE-10746 Bump the version of HTrace to 3.0

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1580076 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Michael Stack 2014-03-21 22:46:03 +00:00
parent ce71b349c1
commit c4610ecab2
28 changed files with 52 additions and 52 deletions

View File

@ -123,7 +123,7 @@
<artifactId>zookeeper</artifactId> <artifactId>zookeeper</artifactId>
</dependency> </dependency>
<dependency> <dependency>
<groupId>org.cloudera.htrace</groupId> <groupId>org.htrace</groupId>
<artifactId>htrace-core</artifactId> <artifactId>htrace-core</artifactId>
</dependency> </dependency>
<dependency> <dependency>

View File

@ -48,7 +48,7 @@ import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.client.coprocessor.Batch; import org.apache.hadoop.hbase.client.coprocessor.Batch;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager; import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import com.google.common.annotations.VisibleForTesting; import com.google.common.annotations.VisibleForTesting;

View File

@ -69,9 +69,9 @@ import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.security.token.Token; import org.apache.hadoop.security.token.Token;
import org.apache.hadoop.security.token.TokenIdentifier; import org.apache.hadoop.security.token.TokenIdentifier;
import org.apache.hadoop.security.token.TokenSelector; import org.apache.hadoop.security.token.TokenSelector;
import org.cloudera.htrace.Span; import org.htrace.Span;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
import javax.net.SocketFactory; import javax.net.SocketFactory;
import javax.security.sasl.SaslException; import javax.security.sasl.SaslException;

View File

@ -45,8 +45,8 @@ import org.apache.zookeeper.data.ACL;
import org.apache.zookeeper.data.Stat; import org.apache.zookeeper.data.Stat;
import org.apache.zookeeper.proto.CreateRequest; import org.apache.zookeeper.proto.CreateRequest;
import org.apache.zookeeper.proto.SetDataRequest; import org.apache.zookeeper.proto.SetDataRequest;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
/** /**
* A zookeeper that can handle 'recoverable' errors. * A zookeeper that can handle 'recoverable' errors.

View File

@ -192,7 +192,7 @@
<artifactId>commons-lang</artifactId> <artifactId>commons-lang</artifactId>
</dependency> </dependency>
<dependency> <dependency>
<groupId>org.cloudera.htrace</groupId> <groupId>org.htrace</groupId>
<artifactId>htrace-core</artifactId> <artifactId>htrace-core</artifactId>
</dependency> </dependency>
</dependencies> </dependencies>

View File

@ -315,7 +315,7 @@ public class IntegrationTestImportTsv implements Configurable, Tool {
HBaseTestingUtility.class, // hbase-server-test HBaseTestingUtility.class, // hbase-server-test
HBaseCommonTestingUtility.class, // hbase-common-test HBaseCommonTestingUtility.class, // hbase-common-test
com.google.common.collect.ListMultimap.class, // Guava com.google.common.collect.ListMultimap.class, // Guava
org.cloudera.htrace.Trace.class); // HTrace org.htrace.Trace.class); // HTrace
} }
/** /**

View File

@ -62,10 +62,10 @@ import org.apache.hadoop.hbase.regionserver.NoSuchColumnFamilyException;
import org.apache.hadoop.hbase.security.AccessDeniedException; import org.apache.hadoop.hbase.security.AccessDeniedException;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.LoadTestTool; import org.apache.hadoop.hbase.util.LoadTestTool;
import org.cloudera.htrace.Span; import org.htrace.Span;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
import org.cloudera.htrace.impl.AlwaysSampler; import org.htrace.impl.AlwaysSampler;
import org.junit.AfterClass; import org.junit.AfterClass;
import org.junit.BeforeClass; import org.junit.BeforeClass;
import org.junit.Test; import org.junit.Test;

View File

@ -35,10 +35,10 @@ import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.util.AbstractHBaseTool; import org.apache.hadoop.hbase.util.AbstractHBaseTool;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.util.ToolRunner; import org.apache.hadoop.util.ToolRunner;
import org.cloudera.htrace.Sampler; import org.htrace.Sampler;
import org.cloudera.htrace.Span; import org.htrace.Span;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
import org.junit.Before; import org.junit.Before;
import org.junit.Test; import org.junit.Test;
import org.junit.experimental.categories.Category; import org.junit.experimental.categories.Category;

View File

@ -438,7 +438,7 @@
</dependency> </dependency>
<!-- tracing Dependencies --> <!-- tracing Dependencies -->
<dependency> <dependency>
<groupId>org.cloudera.htrace</groupId> <groupId>org.htrace</groupId>
<artifactId>htrace-core</artifactId> <artifactId>htrace-core</artifactId>
</dependency> </dependency>
<dependency> <dependency>

View File

@ -25,9 +25,9 @@ import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.hbase.Server; import org.apache.hadoop.hbase.Server;
import org.cloudera.htrace.Span; import org.htrace.Span;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
/** /**
* Abstract base class for all HBase event handlers. Subclasses should * Abstract base class for all HBase event handlers. Subclasses should

View File

@ -41,8 +41,8 @@ import org.apache.hadoop.hbase.util.ByteBufferUtils;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
import org.apache.hadoop.hbase.util.IdLock; import org.apache.hadoop.hbase.util.IdLock;
import org.apache.hadoop.io.WritableUtils; import org.apache.hadoop.io.WritableUtils;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
/** /**
* {@link HFile} reader for version 2. * {@link HFile} reader for version 2.

View File

@ -27,8 +27,8 @@ import org.apache.hadoop.hbase.security.UserProvider;
import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.hbase.util.Pair;
import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.security.UserGroupInformation;
import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.util.StringUtils;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
import com.google.protobuf.Message; import com.google.protobuf.Message;

View File

@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.security.User;
import com.google.protobuf.BlockingService; import com.google.protobuf.BlockingService;
import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Bytes;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import java.net.InetAddress; import java.net.InetAddress;

View File

@ -110,7 +110,7 @@ import org.apache.hadoop.security.token.SecretManager;
import org.apache.hadoop.security.token.SecretManager.InvalidToken; import org.apache.hadoop.security.token.SecretManager.InvalidToken;
import org.apache.hadoop.security.token.TokenIdentifier; import org.apache.hadoop.security.token.TokenIdentifier;
import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.util.StringUtils;
import org.cloudera.htrace.TraceInfo; import org.htrace.TraceInfo;
import org.codehaus.jackson.map.ObjectMapper; import org.codehaus.jackson.map.ObjectMapper;
import com.google.common.util.concurrent.ThreadFactoryBuilder; import com.google.common.util.concurrent.ThreadFactoryBuilder;

View File

@ -682,7 +682,7 @@ public class TableMapReduceUtil {
org.jboss.netty.channel.ChannelFactory.class, org.jboss.netty.channel.ChannelFactory.class,
com.google.protobuf.Message.class, com.google.protobuf.Message.class,
com.google.common.collect.Lists.class, com.google.common.collect.Lists.class,
org.cloudera.htrace.Trace.class); org.htrace.Trace.class);
} }
/** /**

View File

@ -44,7 +44,7 @@ import org.apache.hadoop.hbase.master.TableLockManager;
import org.apache.hadoop.hbase.master.RegionState.State; import org.apache.hadoop.hbase.master.RegionState.State;
import org.apache.hadoop.hbase.master.TableLockManager.TableLock; import org.apache.hadoop.hbase.master.TableLockManager.TableLock;
import org.apache.zookeeper.KeeperException; import org.apache.zookeeper.KeeperException;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
/** /**
* Handler to run disable of a table. * Handler to run disable of a table.

View File

@ -48,7 +48,7 @@ import org.apache.hadoop.hbase.master.TableLockManager;
import org.apache.hadoop.hbase.master.TableLockManager.TableLock; import org.apache.hadoop.hbase.master.TableLockManager.TableLock;
import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.hbase.util.Pair;
import org.apache.zookeeper.KeeperException; import org.apache.zookeeper.KeeperException;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
/** /**
* Handler to run enable of a table. * Handler to run enable of a table.

View File

@ -49,8 +49,8 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
import org.apache.hadoop.hbase.util.HasThread; import org.apache.hadoop.hbase.util.HasThread;
import org.apache.hadoop.hbase.util.Threads; import org.apache.hadoop.hbase.util.Threads;
import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.util.StringUtils;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
import org.apache.hadoop.hbase.util.Counter; import org.apache.hadoop.hbase.util.Counter;
import com.google.common.base.Preconditions; import com.google.common.base.Preconditions;

View File

@ -70,8 +70,8 @@ import org.apache.hadoop.hbase.util.FSUtils;
import org.apache.hadoop.hbase.util.HasThread; import org.apache.hadoop.hbase.util.HasThread;
import org.apache.hadoop.hbase.util.Threads; import org.apache.hadoop.hbase.util.Threads;
import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.util.StringUtils;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
import com.google.common.annotations.VisibleForTesting; import com.google.common.annotations.VisibleForTesting;
import com.lmax.disruptor.BlockingWaitStrategy; import com.lmax.disruptor.BlockingWaitStrategy;

View File

@ -19,7 +19,7 @@
package org.apache.hadoop.hbase.regionserver.wal; package org.apache.hadoop.hbase.regionserver.wal;
import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceAudience;
import org.cloudera.htrace.Span; import org.htrace.Span;
import com.lmax.disruptor.EventFactory; import com.lmax.disruptor.EventFactory;

View File

@ -19,7 +19,7 @@
package org.apache.hadoop.hbase.trace; package org.apache.hadoop.hbase.trace;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.cloudera.htrace.HTraceConfiguration; import org.htrace.HTraceConfiguration;
public class HBaseHTraceConfiguration extends HTraceConfiguration { public class HBaseHTraceConfiguration extends HTraceConfiguration {

View File

@ -25,8 +25,8 @@ import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory; import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.util.ReflectionUtils; import org.apache.hadoop.util.ReflectionUtils;
import org.cloudera.htrace.SpanReceiver; import org.htrace.SpanReceiver;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
/** /**
* This class provides functions for reading the names of SpanReceivers from * This class provides functions for reading the names of SpanReceivers from

View File

@ -28,12 +28,12 @@ import org.apache.hadoop.hbase.MediumTests;
import org.apache.hadoop.hbase.Waiter; import org.apache.hadoop.hbase.Waiter;
import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.HTable;
import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Put;
import org.cloudera.htrace.Sampler; import org.htrace.Sampler;
import org.cloudera.htrace.Span; import org.htrace.Span;
import org.cloudera.htrace.Trace; import org.htrace.Trace;
import org.cloudera.htrace.TraceScope; import org.htrace.TraceScope;
import org.cloudera.htrace.TraceTree; import org.htrace.TraceTree;
import org.cloudera.htrace.impl.POJOSpanReceiver; import org.htrace.impl.POJOSpanReceiver;
import org.junit.AfterClass; import org.junit.AfterClass;
import org.junit.BeforeClass; import org.junit.BeforeClass;
import org.junit.Test; import org.junit.Test;

View File

@ -204,7 +204,7 @@
</dependency> </dependency>
<!-- Test Dependencies --> <!-- Test Dependencies -->
<dependency> <dependency>
<groupId>org.cloudera.htrace</groupId> <groupId>org.htrace</groupId>
<artifactId>htrace-core</artifactId> <artifactId>htrace-core</artifactId>
</dependency> </dependency>
</dependencies> </dependencies>

View File

@ -16,8 +16,8 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
# #
HTrace = org.cloudera.htrace.Trace HTrace = org.htrace.Trace
java_import org.cloudera.htrace.Sampler java_import org.htrace.Sampler
java_import org.apache.hadoop.hbase.trace.SpanReceiverHost java_import org.apache.hadoop.hbase.trace.SpanReceiverHost
module Shell module Shell

View File

@ -189,7 +189,7 @@
<version>${project.version}</version> <version>${project.version}</version>
</dependency> </dependency>
<dependency> <dependency>
<groupId>org.cloudera.htrace</groupId> <groupId>org.htrace</groupId>
<artifactId>htrace-core</artifactId> <artifactId>htrace-core</artifactId>
</dependency> </dependency>
<dependency> <dependency>

View File

@ -919,7 +919,7 @@
<jersey.version>1.8</jersey.version> <jersey.version>1.8</jersey.version>
<jruby.version>1.6.8</jruby.version> <jruby.version>1.6.8</jruby.version>
<junit.version>4.11</junit.version> <junit.version>4.11</junit.version>
<htrace.version>2.04</htrace.version> <htrace.version>3.0.3</htrace.version>
<log4j.version>1.2.17</log4j.version> <log4j.version>1.2.17</log4j.version>
<mockito-all.version>1.9.0</mockito-all.version> <mockito-all.version>1.9.0</mockito-all.version>
<protobuf.version>2.5.0</protobuf.version> <protobuf.version>2.5.0</protobuf.version>
@ -1366,7 +1366,7 @@
<scope>test</scope> <scope>test</scope>
</dependency> </dependency>
<dependency> <dependency>
<groupId>org.cloudera.htrace</groupId> <groupId>org.htrace</groupId>
<artifactId>htrace-core</artifactId> <artifactId>htrace-core</artifactId>
<version>${htrace.version}</version> <version>${htrace.version}</version>
</dependency> </dependency>

View File

@ -70,7 +70,7 @@
<programlisting><![CDATA[ <programlisting><![CDATA[
<property> <property>
<name>hbase.trace.spanreceiver.classes</name> <name>hbase.trace.spanreceiver.classes</name>
<value>org.cloudera.htrace.impl.LocalFileSpanReceiver</value> <value>org.htrace.impl.LocalFileSpanReceiver</value>
</property> </property>
<property> <property>
<name>hbase.local-file-span-receiver.path</name> <name>hbase.local-file-span-receiver.path</name>
@ -95,7 +95,7 @@
<programlisting><![CDATA[ <programlisting><![CDATA[
<property> <property>
<name>hbase.trace.spanreceiver.classes</name> <name>hbase.trace.spanreceiver.classes</name>
<value>org.cloudera.htrace.impl.ZipkinSpanReceiver</value> <value>org.htrace.impl.ZipkinSpanReceiver</value>
</property> </property>
<property> <property>
<name>hbase.zipkin.collector-hostname</name> <name>hbase.zipkin.collector-hostname</name>