175 lines
5.6 KiB
Bash
Executable File
175 lines
5.6 KiB
Bash
Executable File
#!/usr/bin/env bash
|
|
#
|
|
#/**
|
|
# * Copyright 2007 The Apache Software Foundation
|
|
# *
|
|
# * Licensed to the Apache Software Foundation (ASF) under one
|
|
# * or more contributor license agreements. See the NOTICE file
|
|
# * distributed with this work for additional information
|
|
# * regarding copyright ownership. The ASF licenses this file
|
|
# * to you under the Apache License, Version 2.0 (the
|
|
# * "License"); you may not use this file except in compliance
|
|
# * with the License. You may obtain a copy of the License at
|
|
# *
|
|
# * http://www.apache.org/licenses/LICENSE-2.0
|
|
# *
|
|
# * Unless required by applicable law or agreed to in writing, software
|
|
# * distributed under the License is distributed on an "AS IS" BASIS,
|
|
# * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# * See the License for the specific language governing permissions and
|
|
# * limitations under the License.
|
|
# */
|
|
#
|
|
# Run a shell command on all regionserver hosts.
|
|
#
|
|
# Environment Variables
|
|
#
|
|
# HBASE_REGIONSERVERS File naming remote hosts.
|
|
# Default is ${HADOOP_CONF_DIR}/regionservers
|
|
# HADOOP_CONF_DIR Alternate conf dir. Default is ${HADOOP_HOME}/conf.
|
|
# HBASE_CONF_DIR Alternate hbase conf dir. Default is ${HBASE_HOME}/conf.
|
|
# HADOOP_SLAVE_SLEEP Seconds to sleep between spawning remote commands.
|
|
# HADOOP_SLAVE_TIMEOUT Seconds to wait for timing out a remote command.
|
|
# HADOOP_SSH_OPTS Options passed to ssh when running remote commands.
|
|
#
|
|
# Modelled after $HADOOP_HOME/bin/slaves.sh.
|
|
|
|
usage="Usage: $0 [--config <hbase-confdir>] [--rs-only] [--master-only] [--graceful] [--maxthreads xx]"
|
|
|
|
bin=`dirname "$0"`
|
|
bin=`cd "$bin">/dev/null; pwd`
|
|
|
|
. "$bin"/hbase-config.sh
|
|
|
|
# start hbase daemons
|
|
errCode=$?
|
|
if [ $errCode -ne 0 ]
|
|
then
|
|
exit $errCode
|
|
fi
|
|
|
|
function usage() {
|
|
echo $usage
|
|
exit 1
|
|
}
|
|
|
|
|
|
RR_RS=1
|
|
RR_MASTER=1
|
|
RR_GRACEFUL=0
|
|
RR_MAXTHREADS=1
|
|
|
|
while [ $# -gt 0 ]; do
|
|
case "$1" in
|
|
--rs-only|-r)
|
|
RR_RS=1
|
|
RR_MASTER=0
|
|
RR_GRACEFUL=0
|
|
shift
|
|
;;
|
|
--master-only)
|
|
RR_RS=0
|
|
RR_MASTER=1
|
|
RR_GRACEFUL=0
|
|
shift
|
|
;;
|
|
--graceful)
|
|
RR_RS=0
|
|
RR_MASTER=0
|
|
RR_GRACEFUL=1
|
|
shift
|
|
;;
|
|
--maxthreads)
|
|
shift
|
|
RR_MAXTHREADS=$1
|
|
shift
|
|
;;
|
|
*)
|
|
echo Bad argument: $x
|
|
usage
|
|
exit 1
|
|
;;
|
|
esac
|
|
done
|
|
|
|
# quick function to get a value from the HBase config file
|
|
# HBASE-6504 - only take the first line of the output in case verbose gc is on
|
|
distMode=`HBASE_CONF_DIR=${HBASE_CONF_DIR} $bin/hbase org.apache.hadoop.hbase.util.HBaseConfTool hbase.cluster.distributed | head -n 1`
|
|
if [ "$distMode" == 'false' ]; then
|
|
if [ $RR_RS -ne 1 ] || [ $RR_MASTER -ne 1 ]; then
|
|
echo Cant do selective rolling restart if not running distributed
|
|
exit 1
|
|
fi
|
|
"$bin"/hbase-daemon.sh restart master
|
|
else
|
|
zparent=`$bin/hbase org.apache.hadoop.hbase.util.HBaseConfTool zookeeper.znode.parent`
|
|
if [ "$zparent" == "null" ]; then zparent="/hbase"; fi
|
|
|
|
if [ $RR_MASTER -eq 1 ]; then
|
|
# stop all masters before re-start to avoid races for master znode
|
|
"$bin"/hbase-daemon.sh --config "${HBASE_CONF_DIR}" stop master
|
|
"$bin"/hbase-daemons.sh --config "${HBASE_CONF_DIR}" \
|
|
--hosts "${HBASE_BACKUP_MASTERS}" stop master-backup
|
|
|
|
# make sure the master znode has been deleted before continuing
|
|
zmaster=`$bin/hbase org.apache.hadoop.hbase.util.HBaseConfTool zookeeper.znode.master`
|
|
if [ "$zmaster" == "null" ]; then zmaster="master"; fi
|
|
zmaster=$zparent/$zmaster
|
|
echo -n "Waiting for Master ZNode ${zmaster} to expire"
|
|
while ! "$bin"/hbase zkcli stat $zmaster 2>&1 | grep "Node does not exist"; do
|
|
echo -n "."
|
|
sleep 1
|
|
done
|
|
echo #force a newline
|
|
|
|
# all masters are down, now restart
|
|
"$bin"/hbase-daemon.sh --config "${HBASE_CONF_DIR}" start master
|
|
"$bin"/hbase-daemons.sh --config "${HBASE_CONF_DIR}" \
|
|
--hosts "${HBASE_BACKUP_MASTERS}" start master-backup
|
|
|
|
echo "Wait a minute for master to come up join cluster"
|
|
sleep 60
|
|
|
|
# Master joing cluster will start in cleaning out regions in transition.
|
|
# Wait until the master has cleaned out regions in transition before
|
|
# giving it a bunch of work to do; master is vulnerable during startup
|
|
zunassigned=`$bin/hbase org.apache.hadoop.hbase.util.HBaseConfTool zookeeper.znode.unassigned`
|
|
if [ "$zunassigned" == "null" ]; then zunassigned="region-in-transition"; fi
|
|
zunassigned="$zparent/$zunassigned"
|
|
echo -n "Waiting for ${zunassigned} to empty"
|
|
while true ; do
|
|
unassigned=`$bin/hbase zkcli stat ${zunassigned} 2>&1 |grep -e 'numChildren = '|sed -e 's,numChildren = ,,'`
|
|
if test 0 -eq ${unassigned}
|
|
then
|
|
break
|
|
else
|
|
echo -n " ${unassigned}"
|
|
fi
|
|
sleep 1
|
|
done
|
|
fi
|
|
|
|
if [ $RR_RS -eq 1 ]; then
|
|
# unlike the masters, roll all regionservers one-at-a-time
|
|
export HBASE_SLAVE_PARALLEL=false
|
|
"$bin"/hbase-daemons.sh --config "${HBASE_CONF_DIR}" \
|
|
--hosts "${HBASE_REGIONSERVERS}" restart regionserver
|
|
fi
|
|
|
|
if [ $RR_GRACEFUL -eq 1 ]; then
|
|
# gracefully restart all online regionservers
|
|
zkrs=`$bin/hbase org.apache.hadoop.hbase.util.HBaseConfTool zookeeper.znode.rs`
|
|
if [ "$zkrs" == "null" ]; then zkrs="rs"; fi
|
|
zkrs="$zparent/$zkrs"
|
|
online_regionservers=`$bin/hbase zkcli ls $zkrs 2>&1 | tail -1 | sed "s/\[//" | sed "s/\]//"`
|
|
for rs in $online_regionservers
|
|
do
|
|
rs_parts=(${rs//,/ })
|
|
hostname=${rs_parts[0]}
|
|
echo "Gracefully restarting: $hostname"
|
|
"$bin"/graceful_stop.sh --config "${HBASE_CONF_DIR}" --restart --reload --debug --maxthreads "${RR_MAXTHREADS}" "$hostname"
|
|
sleep 1
|
|
done
|
|
fi
|
|
fi
|