#! /usr/bin/env bash # #/** # * Licensed to the Apache Software Foundation (ASF) under one # * or more contributor license agreements. See the NOTICE file # * distributed with this work for additional information # * regarding copyright ownership. The ASF licenses this file # * to you under the Apache License, Version 2.0 (the # * "License"); you may not use this file except in compliance # * with the License. You may obtain a copy of the License at # * # * http://www.apache.org/licenses/LICENSE-2.0 # * # * Unless required by applicable law or agreed to in writing, software # * distributed under the License is distributed on an "AS IS" BASIS, # * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # * See the License for the specific language governing permissions and # * limitations under the License. # */ # # The hbase command script. Based on the hadoop command script putting # in hbase classes, libs and configurations ahead of hadoop's. # # TODO: Narrow the amount of duplicated code. # # Environment Variables: # # JAVA_HOME The java implementation to use. Overrides JAVA_HOME. # # HBASE_CLASSPATH Extra Java CLASSPATH entries. # # HBASE_CLASSPATH_PREFIX Extra Java CLASSPATH entries that should be # prefixed to the system classpath. # # HBASE_HEAPSIZE The maximum amount of heap to use. # Default is unset and uses the JVMs default setting # (usually 1/4th of the available memory). # # HBASE_LIBRARY_PATH HBase additions to JAVA_LIBRARY_PATH for adding # native libraries. # # HBASE_OPTS Extra Java runtime options. # # HBASE_CONF_DIR Alternate conf dir. Default is ${HBASE_HOME}/conf. # # HBASE_ROOT_LOGGER The root appender. Default is INFO,console # # JRUBY_HOME JRuby path: $JRUBY_HOME/lib/jruby.jar should exist. # Defaults to the jar packaged with HBase. # # JRUBY_OPTS Extra options (eg '--1.9') passed to hbase. # Empty by default. # # HBASE_SHELL_OPTS Extra options passed to the hbase shell. # Empty by default. # bin=`dirname "$0"` bin=`cd "$bin">/dev/null; pwd` # This will set HBASE_HOME, etc. . "$bin"/hbase-config.sh cygwin=false case "`uname`" in CYGWIN*) cygwin=true;; esac # Detect if we are in hbase sources dir in_dev_env=false if [ -d "${HBASE_HOME}/target" ]; then in_dev_env=true fi # Detect if we are in the omnibus tarball in_omnibus_tarball="false" if [ -f "${HBASE_HOME}/bin/hbase-daemons.sh" ]; then in_omnibus_tarball="true" fi read -d '' options_string << EOF Options: --config DIR Configuration direction to use. Default: ./conf --hosts HOSTS Override the list in 'regionservers' file --auth-as-server Authenticate to ZooKeeper using servers configuration --internal-classpath Skip attempting to use client facing jars (WARNING: unstable results between versions) EOF # if no args specified, show usage if [ $# = 0 ]; then echo "Usage: hbase [] []" echo "$options_string" echo "" echo "Commands:" echo "Some commands take arguments. Pass no args or -h for usage." echo " shell Run the HBase shell" echo " hbck Run the HBase 'fsck' tool. Defaults read-only hbck1." echo " Pass '-j /path/to/HBCK2.jar' to run hbase-2.x HBCK2." echo " snapshot Tool for managing snapshots" if [ "${in_omnibus_tarball}" = "true" ]; then echo " wal Write-ahead-log analyzer" echo " hfile Store file analyzer" echo " zkcli Run the ZooKeeper shell" echo " master Run an HBase HMaster node" echo " regionserver Run an HBase HRegionServer node" echo " zookeeper Run a ZooKeeper server" echo " rest Run an HBase REST server" echo " thrift Run the HBase Thrift server" echo " thrift2 Run the HBase Thrift2 server" echo " clean Run the HBase clean up script" fi echo " classpath Dump hbase CLASSPATH" echo " mapredcp Dump CLASSPATH entries required by mapreduce" echo " pe Run PerformanceEvaluation" echo " ltt Run LoadTestTool" echo " canary Run the Canary tool" echo " version Print the version" echo " regionsplitter Run RegionSplitter tool" echo " rowcounter Run RowCounter tool" echo " cellcounter Run CellCounter tool" echo " pre-upgrade Run Pre-Upgrade validator tool" echo " CLASSNAME Run the class named CLASSNAME" exit 1 fi # get arguments COMMAND=$1 shift JAVA=$JAVA_HOME/bin/java # override default settings for this command, if applicable if [ -f "$HBASE_HOME/conf/hbase-env-$COMMAND.sh" ]; then . "$HBASE_HOME/conf/hbase-env-$COMMAND.sh" fi add_size_suffix() { # add an 'm' suffix if the argument is missing one, otherwise use whats there local val="$1" local lastchar=${val: -1} if [[ "mMgG" == *$lastchar* ]]; then echo $val else echo ${val}m fi } if [[ -n "$HBASE_HEAPSIZE" ]]; then JAVA_HEAP_MAX="-Xmx$(add_size_suffix $HBASE_HEAPSIZE)" fi if [[ -n "$HBASE_OFFHEAPSIZE" ]]; then JAVA_OFFHEAP_MAX="-XX:MaxDirectMemorySize=$(add_size_suffix $HBASE_OFFHEAPSIZE)" fi # so that filenames w/ spaces are handled correctly in loops below ORIG_IFS=$IFS IFS= # CLASSPATH initially contains $HBASE_CONF_DIR CLASSPATH="${HBASE_CONF_DIR}" CLASSPATH=${CLASSPATH}:$JAVA_HOME/lib/tools.jar add_to_cp_if_exists() { if [ -d "$@" ]; then CLASSPATH=${CLASSPATH}:"$@" fi } # For releases, add hbase & webapps to CLASSPATH # Webapps must come first else it messes up Jetty if [ -d "$HBASE_HOME/hbase-webapps" ]; then add_to_cp_if_exists "${HBASE_HOME}" fi #add if we are in a dev environment if [ -d "$HBASE_HOME/hbase-server/target/hbase-webapps" ]; then if [ "$COMMAND" = "thrift" ] ; then add_to_cp_if_exists "${HBASE_HOME}/hbase-thrift/target" elif [ "$COMMAND" = "thrift2" ] ; then add_to_cp_if_exists "${HBASE_HOME}/hbase-thrift/target" elif [ "$COMMAND" = "rest" ] ; then add_to_cp_if_exists "${HBASE_HOME}/hbase-rest/target" else add_to_cp_if_exists "${HBASE_HOME}/hbase-server/target" fi fi #add the hbase jars for each module for f in $HBASE_HOME/hbase-jars/hbase*.jar; do if [[ $f = *sources.jar ]] then : # Skip sources.jar elif [ -f $f ] then CLASSPATH=${CLASSPATH}:$f; fi done #If avail, add Hadoop to the CLASSPATH and to the JAVA_LIBRARY_PATH # Allow this functionality to be disabled if [ "$HBASE_DISABLE_HADOOP_CLASSPATH_LOOKUP" != "true" ] ; then HADOOP_IN_PATH=$(PATH="${HADOOP_HOME:-${HADOOP_PREFIX}}/bin:$PATH" which hadoop 2>/dev/null) fi # Add libs to CLASSPATH declare shaded_jar if [ "${INTERNAL_CLASSPATH}" != "true" ]; then # find our shaded jars declare shaded_client declare shaded_client_byo_hadoop declare shaded_mapreduce for f in "${HBASE_HOME}"/lib/shaded-clients/hbase-shaded-client*.jar; do if [[ "${f}" =~ byo-hadoop ]]; then shaded_client_byo_hadoop="${f}" else shaded_client="${f}" fi done for f in "${HBASE_HOME}"/lib/shaded-clients/hbase-shaded-mapreduce*.jar; do shaded_mapreduce="${f}" done # If command can use our shaded client, use it declare -a commands_in_client_jar=("classpath" "version") for c in "${commands_in_client_jar[@]}"; do if [ "${COMMAND}" = "${c}" ]; then if [ -n "${HADOOP_IN_PATH}" ] && [ -f "${HADOOP_IN_PATH}" ]; then # If we didn't find a jar above, this will just be blank and the # check below will then default back to the internal classpath. shaded_jar="${shaded_client_byo_hadoop}" else # If we didn't find a jar above, this will just be blank and the # check below will then default back to the internal classpath. shaded_jar="${shaded_client}" fi break fi done # If command needs our shaded mapreduce, use it # N.B "mapredcp" is not included here because in the shaded case it skips our built classpath declare -a commands_in_mr_jar=("hbck" "snapshot" "canary" "regionsplitter" "pre-upgrade") for c in "${commands_in_mr_jar[@]}"; do if [ "${COMMAND}" = "${c}" ]; then # If we didn't find a jar above, this will just be blank and the # check below will then default back to the internal classpath. shaded_jar="${shaded_mapreduce}" break fi done # Some commands specifically only can use shaded mapreduce when we'll get a full hadoop classpath at runtime if [ -n "${HADOOP_IN_PATH}" ] && [ -f "${HADOOP_IN_PATH}" ]; then declare -a commands_in_mr_need_hadoop=("backup" "restore" "rowcounter" "cellcounter") for c in "${commands_in_mr_need_hadoop[@]}"; do if [ "${COMMAND}" = "${c}" ]; then # If we didn't find a jar above, this will just be blank and the # check below will then default back to the internal classpath. shaded_jar="${shaded_mapreduce}" break fi done fi fi if [ -n "${shaded_jar}" ] && [ -f "${shaded_jar}" ]; then CLASSPATH="${CLASSPATH}:${shaded_jar}" # fall through to grabbing all the lib jars and hope we're in the omnibus tarball # # N.B. shell specifically can't rely on the shaded artifacts because RSGroups is only # available as non-shaded # # N.B. pe and ltt can't easily rely on shaded artifacts because they live in hbase-mapreduce:test-jar # and need some other jars that haven't been relocated. Currently enumerating that list # is too hard to be worth it. # else for f in $HBASE_HOME/lib/*.jar; do CLASSPATH=${CLASSPATH}:$f; done # make it easier to check for shaded/not later on. shaded_jar="" fi for f in "${HBASE_HOME}"/lib/client-facing-thirdparty/*.jar; do if [[ ! "${f}" =~ ^.*/htrace-core-3.*\.jar$ ]] && \ [ "${f}" != "htrace-core.jar$" ] && \ [[ ! "${f}" =~ ^.*/slf4j-log4j.*$ ]]; then CLASSPATH="${CLASSPATH}:${f}" fi done # default log directory & file if [ "$HBASE_LOG_DIR" = "" ]; then HBASE_LOG_DIR="$HBASE_HOME/logs" fi if [ "$HBASE_LOGFILE" = "" ]; then HBASE_LOGFILE='hbase.log' fi function append_path() { if [ -z "$1" ]; then echo "$2" else echo "$1:$2" fi } JAVA_PLATFORM="" # if HBASE_LIBRARY_PATH is defined lets use it as first or second option if [ "$HBASE_LIBRARY_PATH" != "" ]; then JAVA_LIBRARY_PATH=$(append_path "$JAVA_LIBRARY_PATH" "$HBASE_LIBRARY_PATH") fi #If configured and available, add Hadoop to the CLASSPATH and to the JAVA_LIBRARY_PATH if [ -n "${HADOOP_IN_PATH}" ] && [ -f "${HADOOP_IN_PATH}" ]; then HADOOP_JAVA_LIBRARY_PATH=$(HADOOP_CLASSPATH="$CLASSPATH" "${HADOOP_IN_PATH}" \ org.apache.hadoop.hbase.util.GetJavaProperty java.library.path) if [ -n "$HADOOP_JAVA_LIBRARY_PATH" ]; then JAVA_LIBRARY_PATH=$(append_path "${JAVA_LIBRARY_PATH}" "$HADOOP_JAVA_LIBRARY_PATH") fi CLASSPATH=$(append_path "${CLASSPATH}" "$(${HADOOP_IN_PATH} classpath 2>/dev/null)") else # Otherwise, if we're providing Hadoop we should include htrace 3 if we were built with a version that needs it. for f in "${HBASE_HOME}"/lib/client-facing-thirdparty/htrace-core-3*.jar "${HBASE_HOME}"/lib/client-facing-thirdparty/htrace-core.jar; do if [ -f "${f}" ]; then CLASSPATH="${CLASSPATH}:${f}" break fi done # Some commands require special handling when using shaded jars. For these cases, we rely on hbase-shaded-mapreduce # instead of hbase-shaded-client* because we make use of some IA.Private classes that aren't in the latter. However, # we don't invoke them using the "hadoop jar" command so we need to ensure there are some Hadoop classes available # when we're not doing runtime hadoop classpath lookup. # # luckily the set of classes we need are those packaged in the shaded-client. for c in "${commands_in_mr_jar[@]}"; do if [ "${COMMAND}" = "${c}" ] && [ -n "${shaded_jar}" ]; then CLASSPATH="${CLASSPATH}:${shaded_client:?We couldn\'t find the shaded client jar even though we did find the shaded MR jar. for command ${COMMAND} we need both. please use --internal-classpath as a workaround.}" break fi done fi # Add user-specified CLASSPATH last if [ "$HBASE_CLASSPATH" != "" ]; then CLASSPATH=${CLASSPATH}:${HBASE_CLASSPATH} fi # Add user-specified CLASSPATH prefix first if [ "$HBASE_CLASSPATH_PREFIX" != "" ]; then CLASSPATH=${HBASE_CLASSPATH_PREFIX}:${CLASSPATH} fi # cygwin path translation if $cygwin; then CLASSPATH=`cygpath -p -w "$CLASSPATH"` HBASE_HOME=`cygpath -d "$HBASE_HOME"` HBASE_LOG_DIR=`cygpath -d "$HBASE_LOG_DIR"` fi if [ -d "${HBASE_HOME}/build/native" -o -d "${HBASE_HOME}/lib/native" ]; then if [ -z $JAVA_PLATFORM ]; then JAVA_PLATFORM=`CLASSPATH=${CLASSPATH} ${JAVA} org.apache.hadoop.util.PlatformName | sed -e "s/ /_/g"` fi if [ -d "$HBASE_HOME/build/native" ]; then JAVA_LIBRARY_PATH=$(append_path "$JAVA_LIBRARY_PATH" "${HBASE_HOME}/build/native/${JAVA_PLATFORM}/lib") fi if [ -d "${HBASE_HOME}/lib/native" ]; then JAVA_LIBRARY_PATH=$(append_path "$JAVA_LIBRARY_PATH" "${HBASE_HOME}/lib/native/${JAVA_PLATFORM}") fi fi # cygwin path translation if $cygwin; then JAVA_LIBRARY_PATH=`cygpath -p "$JAVA_LIBRARY_PATH"` fi # restore ordinary behaviour unset IFS #Set the right GC options based on the what we are running declare -a server_cmds=("master" "regionserver" "thrift" "thrift2" "rest" "avro" "zookeeper") for cmd in ${server_cmds[@]}; do if [[ $cmd == $COMMAND ]]; then server=true break fi done if [[ $server ]]; then HBASE_OPTS="$HBASE_OPTS $SERVER_GC_OPTS" else HBASE_OPTS="$HBASE_OPTS $CLIENT_GC_OPTS" fi if [ "$AUTH_AS_SERVER" == "true" ] || [ "$COMMAND" = "hbck" ]; then if [ -n "$HBASE_SERVER_JAAS_OPTS" ]; then HBASE_OPTS="$HBASE_OPTS $HBASE_SERVER_JAAS_OPTS" else HBASE_OPTS="$HBASE_OPTS $HBASE_REGIONSERVER_OPTS" fi fi # check if the command needs jline declare -a jline_cmds=("zkcli" "org.apache.hadoop.hbase.zookeeper.ZKMainServer") for cmd in "${jline_cmds[@]}"; do if [[ $cmd == "$COMMAND" ]]; then jline_needed=true break fi done # for jruby # (1) for the commands which need jruby (see jruby_cmds defined below) # A. when JRUBY_HOME is specified explicitly, eg. export JRUBY_HOME=/usr/local/share/jruby # CLASSPATH and HBASE_OPTS are updated according to JRUBY_HOME specified # B. when JRUBY_HOME is not specified explicitly # add jruby packaged with HBase to CLASSPATH # (2) for other commands, do nothing # check if the commmand needs jruby declare -a jruby_cmds=("shell" "org.jruby.Main") for cmd in "${jruby_cmds[@]}"; do if [[ $cmd == "$COMMAND" ]]; then jruby_needed=true break fi done add_maven_deps_to_classpath() { f="${HBASE_HOME}/hbase-build-configuration/target/$1" if [ ! -f "${f}" ]; then echo "As this is a development environment, we need ${f} to be generated from maven (command: mvn install -DskipTests)" exit 1 fi CLASSPATH=${CLASSPATH}:$(cat "${f}") } #Add the development env class path stuff if $in_dev_env; then add_maven_deps_to_classpath "cached_classpath.txt" if [[ $jline_needed ]]; then add_maven_deps_to_classpath "cached_classpath_jline.txt" elif [[ $jruby_needed ]]; then add_maven_deps_to_classpath "cached_classpath_jruby.txt" fi fi # the command needs jruby if [[ $jruby_needed ]]; then if [ "$JRUBY_HOME" != "" ]; then # JRUBY_HOME is specified explicitly, eg. export JRUBY_HOME=/usr/local/share/jruby # add jruby.jar into CLASSPATH CLASSPATH="$JRUBY_HOME/lib/jruby.jar:$CLASSPATH" # add jruby to HBASE_OPTS HBASE_OPTS="$HBASE_OPTS -Djruby.home=$JRUBY_HOME -Djruby.lib=$JRUBY_HOME/lib" else # JRUBY_HOME is not specified explicitly if ! $in_dev_env; then # not in dev environment # add jruby packaged with HBase to CLASSPATH JRUBY_PACKAGED_WITH_HBASE="$HBASE_HOME/lib/ruby/*.jar" for jruby_jar in $JRUBY_PACKAGED_WITH_HBASE; do CLASSPATH=$jruby_jar:$CLASSPATH; done fi fi fi # figure out which class to run if [ "$COMMAND" = "shell" ] ; then #find the hbase ruby sources if [ -d "$HBASE_HOME/lib/ruby" ]; then HBASE_OPTS="$HBASE_OPTS -Dhbase.ruby.sources=$HBASE_HOME/lib/ruby" else HBASE_OPTS="$HBASE_OPTS -Dhbase.ruby.sources=$HBASE_HOME/hbase-shell/src/main/ruby" fi HBASE_OPTS="$HBASE_OPTS $HBASE_SHELL_OPTS" CLASS="org.jruby.Main -X+O ${JRUBY_OPTS} ${HBASE_HOME}/bin/hirb.rb" elif [ "$COMMAND" = "hbck" ] ; then # Look for the -j /path/to/HBCK2.jar parameter. Else pass through to hbck. case "${1}" in -j) # Found -j parameter. Add arg to CLASSPATH and set CLASS to HBCK2. shift JAR="${1}" if [ ! -f "${JAR}" ]; then echo "${JAR} file not found!" echo "Usage: hbase [] hbck -jar /path/to/HBCK2.jar []" exit 1 fi CLASSPATH="${JAR}:${CLASSPATH}"; CLASS="org.apache.hbase.HBCK2" shift # past argument=value ;; *) CLASS='org.apache.hadoop.hbase.util.HBaseFsck' ;; esac elif [ "$COMMAND" = "wal" ] ; then CLASS='org.apache.hadoop.hbase.wal.WALPrettyPrinter' elif [ "$COMMAND" = "hfile" ] ; then CLASS='org.apache.hadoop.hbase.io.hfile.HFilePrettyPrinter' elif [ "$COMMAND" = "zkcli" ] ; then CLASS="org.apache.hadoop.hbase.zookeeper.ZKMainServer" for f in $HBASE_HOME/lib/zkcli/*.jar; do CLASSPATH="${CLASSPATH}:$f"; done elif [ "$COMMAND" = "upgrade" ] ; then echo "This command was used to upgrade to HBase 0.96, it was removed in HBase 2.0.0." echo "Please follow the documentation at http://hbase.apache.org/book.html#upgrading." exit 1 elif [ "$COMMAND" = "snapshot" ] ; then SUBCOMMAND=$1 shift if [ "$SUBCOMMAND" = "create" ] ; then CLASS="org.apache.hadoop.hbase.snapshot.CreateSnapshot" elif [ "$SUBCOMMAND" = "info" ] ; then CLASS="org.apache.hadoop.hbase.snapshot.SnapshotInfo" elif [ "$SUBCOMMAND" = "export" ] ; then CLASS="org.apache.hadoop.hbase.snapshot.ExportSnapshot" else echo "Usage: hbase [] snapshot []" echo "$options_string" echo "" echo "Subcommands:" echo " create Create a new snapshot of a table" echo " info Tool for dumping snapshot information" echo " export Export an existing snapshot" exit 1 fi elif [ "$COMMAND" = "master" ] ; then CLASS='org.apache.hadoop.hbase.master.HMaster' if [ "$1" != "stop" ] && [ "$1" != "clear" ] ; then HBASE_OPTS="$HBASE_OPTS $HBASE_MASTER_OPTS" fi elif [ "$COMMAND" = "regionserver" ] ; then CLASS='org.apache.hadoop.hbase.regionserver.HRegionServer' if [ "$1" != "stop" ] ; then HBASE_OPTS="$HBASE_OPTS $HBASE_REGIONSERVER_OPTS" fi elif [ "$COMMAND" = "thrift" ] ; then CLASS='org.apache.hadoop.hbase.thrift.ThriftServer' if [ "$1" != "stop" ] ; then HBASE_OPTS="$HBASE_OPTS $HBASE_THRIFT_OPTS" fi elif [ "$COMMAND" = "thrift2" ] ; then CLASS='org.apache.hadoop.hbase.thrift2.ThriftServer' if [ "$1" != "stop" ] ; then HBASE_OPTS="$HBASE_OPTS $HBASE_THRIFT_OPTS" fi elif [ "$COMMAND" = "rest" ] ; then CLASS='org.apache.hadoop.hbase.rest.RESTServer' if [ "$1" != "stop" ] ; then HBASE_OPTS="$HBASE_OPTS $HBASE_REST_OPTS" fi elif [ "$COMMAND" = "zookeeper" ] ; then CLASS='org.apache.hadoop.hbase.zookeeper.HQuorumPeer' if [ "$1" != "stop" ] ; then HBASE_OPTS="$HBASE_OPTS $HBASE_ZOOKEEPER_OPTS" fi elif [ "$COMMAND" = "clean" ] ; then case $1 in --cleanZk|--cleanHdfs|--cleanAll) matches="yes" ;; *) ;; esac if [ $# -ne 1 -o "$matches" = "" ]; then echo "Usage: hbase clean (--cleanZk|--cleanHdfs|--cleanAll)" echo "Options: " echo " --cleanZk cleans hbase related data from zookeeper." echo " --cleanHdfs cleans hbase related data from hdfs." echo " --cleanAll cleans hbase related data from both zookeeper and hdfs." exit 1; fi "$bin"/hbase-cleanup.sh --config ${HBASE_CONF_DIR} $@ exit $? elif [ "$COMMAND" = "mapredcp" ] ; then # If we didn't find a jar above, this will just be blank and the # check below will then default back to the internal classpath. shaded_jar="${shaded_mapreduce}" if [ "${INTERNAL_CLASSPATH}" != "true" ] && [ -f "${shaded_jar}" ]; then echo -n "${shaded_jar}" for f in "${HBASE_HOME}"/lib/client-facing-thirdparty/*.jar; do if [[ ! "${f}" =~ ^.*/htrace-core-3.*\.jar$ ]] && \ [ "${f}" != "htrace-core.jar$" ] && \ [[ ! "${f}" =~ ^.*/slf4j-log4j.*$ ]]; then echo -n ":${f}" fi done echo "" exit 0 fi CLASS='org.apache.hadoop.hbase.util.MapreduceDependencyClasspathTool' elif [ "$COMMAND" = "classpath" ] ; then echo "$CLASSPATH" exit 0 elif [ "$COMMAND" = "pe" ] ; then CLASS='org.apache.hadoop.hbase.PerformanceEvaluation' HBASE_OPTS="$HBASE_OPTS $HBASE_PE_OPTS" elif [ "$COMMAND" = "ltt" ] ; then CLASS='org.apache.hadoop.hbase.util.LoadTestTool' HBASE_OPTS="$HBASE_OPTS $HBASE_LTT_OPTS" elif [ "$COMMAND" = "canary" ] ; then CLASS='org.apache.hadoop.hbase.tool.Canary' HBASE_OPTS="$HBASE_OPTS $HBASE_CANARY_OPTS" elif [ "$COMMAND" = "version" ] ; then CLASS='org.apache.hadoop.hbase.util.VersionInfo' elif [ "$COMMAND" = "regionsplitter" ] ; then CLASS='org.apache.hadoop.hbase.util.RegionSplitter' elif [ "$COMMAND" = "rowcounter" ] ; then CLASS='org.apache.hadoop.hbase.mapreduce.RowCounter' elif [ "$COMMAND" = "cellcounter" ] ; then CLASS='org.apache.hadoop.hbase.mapreduce.CellCounter' elif [ "$COMMAND" = "pre-upgrade" ] ; then CLASS='org.apache.hadoop.hbase.tool.PreUpgradeValidator' else CLASS=$COMMAND fi # Have JVM dump heap if we run out of memory. Files will be 'launch directory' # and are named like the following: java_pid21612.hprof. Apparently it doesn't # 'cost' to have this flag enabled. Its a 1.6 flag only. See: # http://blogs.sun.com/alanb/entry/outofmemoryerror_looks_a_bit_better HBASE_OPTS="$HBASE_OPTS -Dhbase.log.dir=$HBASE_LOG_DIR" HBASE_OPTS="$HBASE_OPTS -Dhbase.log.file=$HBASE_LOGFILE" HBASE_OPTS="$HBASE_OPTS -Dhbase.home.dir=$HBASE_HOME" HBASE_OPTS="$HBASE_OPTS -Dhbase.id.str=$HBASE_IDENT_STRING" HBASE_OPTS="$HBASE_OPTS -Dhbase.root.logger=${HBASE_ROOT_LOGGER:-INFO,console}" if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then HBASE_OPTS="$HBASE_OPTS -Djava.library.path=$JAVA_LIBRARY_PATH" export LD_LIBRARY_PATH="$LD_LIBRARY_PATH:$JAVA_LIBRARY_PATH" fi # Enable security logging on the master and regionserver only if [ "$COMMAND" = "master" ] || [ "$COMMAND" = "regionserver" ]; then HBASE_OPTS="$HBASE_OPTS -Dhbase.security.logger=${HBASE_SECURITY_LOGGER:-INFO,RFAS}" else HBASE_OPTS="$HBASE_OPTS -Dhbase.security.logger=${HBASE_SECURITY_LOGGER:-INFO,NullAppender}" fi HEAP_SETTINGS="$JAVA_HEAP_MAX $JAVA_OFFHEAP_MAX" # by now if we're running a command it means we need logging for f in ${HBASE_HOME}/lib/client-facing-thirdparty/slf4j-log4j*.jar; do if [ -f "${f}" ]; then CLASSPATH="${CLASSPATH}:${f}" break fi done # Exec unless HBASE_NOEXEC is set. export CLASSPATH if [ "${DEBUG}" = "true" ]; then echo "classpath=${CLASSPATH}" >&2 HBASE_OPTS="${HBASE_OPTS} -Xdiag" fi if [ "${HBASE_NOEXEC}" != "" ]; then "$JAVA" -Dproc_$COMMAND -XX:OnOutOfMemoryError="kill -9 %p" $HEAP_SETTINGS $HBASE_OPTS $CLASS "$@" else exec "$JAVA" -Dproc_$COMMAND -XX:OnOutOfMemoryError="kill -9 %p" $HEAP_SETTINGS $HBASE_OPTS $CLASS "$@" fi