HADOOP-12850. pull shell code out of hadoop-dist
Signed-off-by: Steve Loughran <stevel@apache.org>
This commit is contained in:
parent
c58a6d53c5
commit
1cb2f93451
|
@ -0,0 +1,140 @@
|
|||
#!/usr/bin/env bash
|
||||
# Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
# contributor license agreements. See the NOTICE file distributed with
|
||||
# this work for additional information regarding copyright ownership.
|
||||
# The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
# (the "License"); you may not use this file except in compliance with
|
||||
# the License. You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
|
||||
# project.version
|
||||
VERSION=$1
|
||||
|
||||
# project.build.directory
|
||||
BASEDIR=$2
|
||||
|
||||
function run()
|
||||
{
|
||||
declare res
|
||||
|
||||
echo "\$ ${*}"
|
||||
"${@}"
|
||||
res=$?
|
||||
if [[ ${res} != 0 ]]; then
|
||||
echo
|
||||
echo "Failed!"
|
||||
echo
|
||||
exit "${res}"
|
||||
fi
|
||||
}
|
||||
|
||||
function findfileindir()
|
||||
{
|
||||
declare file="$1"
|
||||
declare dir="${2:-./share}"
|
||||
declare count
|
||||
|
||||
count=$(find "${dir}" -iname "${file}" | wc -l)
|
||||
|
||||
#shellcheck disable=SC2086
|
||||
echo ${count}
|
||||
}
|
||||
|
||||
function copyifnotexists()
|
||||
{
|
||||
declare src="$1"
|
||||
declare dest="$2"
|
||||
|
||||
declare srcname
|
||||
declare destdir
|
||||
|
||||
declare child
|
||||
declare childpath
|
||||
|
||||
if [[ -f "${src}" ]]; then
|
||||
srcname=${src##*/}
|
||||
if [[ "${srcname}" != *.jar ||
|
||||
$(findfileindir "${srcname}") -eq "0" ]]; then
|
||||
destdir=$(dirname "${dest}")
|
||||
mkdir -p "${destdir}"
|
||||
cp -p "${src}" "${dest}"
|
||||
fi
|
||||
else
|
||||
for childpath in "${src}"/*; do
|
||||
child="${childpath##*/}"
|
||||
if [[ "${child}" == "doc" ||
|
||||
"${child}" == "webapps" ]]; then
|
||||
mkdir -p "${dest}/${child}"
|
||||
cp -r "${src}/${child}"/* "${dest}/${child}"
|
||||
continue;
|
||||
fi
|
||||
copyifnotexists "${src}/${child}" "${dest}/${child}"
|
||||
done
|
||||
fi
|
||||
}
|
||||
|
||||
#Copy all contents as is except the lib.
|
||||
#for libs check for existence in share directory, if not exist then only copy.
|
||||
function copy()
|
||||
{
|
||||
declare src="$1"
|
||||
declare dest="$2"
|
||||
|
||||
declare child
|
||||
declare childpath
|
||||
|
||||
if [[ -d "${src}" ]]; then
|
||||
for childpath in "${src}"/*; do
|
||||
child="${childpath##*/}"
|
||||
|
||||
if [[ "${child}" == "share" ]]; then
|
||||
copyifnotexists "${src}/${child}" "${dest}/${child}"
|
||||
else
|
||||
if [[ -d "${src}/${child}" ]]; then
|
||||
mkdir -p "${dest}/${child}"
|
||||
cp -pr "${src}/${child}"/* "${dest}/${child}"
|
||||
else
|
||||
cp -pr "${src}/${child}" "${dest}/${child}"
|
||||
fi
|
||||
fi
|
||||
done
|
||||
fi
|
||||
}
|
||||
|
||||
# shellcheck disable=SC2164
|
||||
ROOT=$(cd "${BASEDIR}"/../..;pwd)
|
||||
echo
|
||||
echo "Current directory $(pwd)"
|
||||
echo
|
||||
run rm -rf "hadoop-${VERSION}"
|
||||
run mkdir "hadoop-${VERSION}"
|
||||
run cd "hadoop-${VERSION}"
|
||||
run cp -p "${ROOT}/LICENSE.txt" .
|
||||
run cp -p "${ROOT}/NOTICE.txt" .
|
||||
run cp -p "${ROOT}/README.txt" .
|
||||
|
||||
# Copy hadoop-common first so that it have always have all dependencies.
|
||||
# Remaining projects will copy only libraries which are not present already in 'share' directory.
|
||||
run copy "${ROOT}/hadoop-common-project/hadoop-common/target/hadoop-common-${VERSION}" .
|
||||
run copy "${ROOT}/hadoop-common-project/hadoop-nfs/target/hadoop-nfs-${VERSION}" .
|
||||
run copy "${ROOT}/hadoop-hdfs-project/hadoop-hdfs/target/hadoop-hdfs-${VERSION}" .
|
||||
run copy "${ROOT}/hadoop-hdfs-project/hadoop-hdfs-nfs/target/hadoop-hdfs-nfs-${VERSION}" .
|
||||
run copy "${ROOT}/hadoop-yarn-project/target/hadoop-yarn-project-${VERSION}" .
|
||||
run copy "${ROOT}/hadoop-mapreduce-project/target/hadoop-mapreduce-${VERSION}" .
|
||||
run copy "${ROOT}/hadoop-tools/hadoop-tools-dist/target/hadoop-tools-dist-${VERSION}" .
|
||||
|
||||
#copy httpfs and kms as is
|
||||
run cp -pr "${ROOT}/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-${VERSION}"/* .
|
||||
run cp -pr "${ROOT}/hadoop-common-project/hadoop-kms/target/hadoop-kms-${VERSION}"/* .
|
||||
|
||||
echo
|
||||
echo "Hadoop dist layout available at: ${BASEDIR}/hadoop-${VERSION}"
|
||||
echo
|
|
@ -0,0 +1,44 @@
|
|||
#!/usr/bin/env bash
|
||||
# Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
# contributor license agreements. See the NOTICE file distributed with
|
||||
# this work for additional information regarding copyright ownership.
|
||||
# The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
# (the "License"); you may not use this file except in compliance with
|
||||
# the License. You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
|
||||
|
||||
# project.version
|
||||
VERSION=$1
|
||||
|
||||
# project.build.directory
|
||||
BASEDIR=$2
|
||||
|
||||
function run()
|
||||
{
|
||||
declare res
|
||||
|
||||
echo "\$ ${*}"
|
||||
"${@}"
|
||||
res=$?
|
||||
if [[ ${res} != 0 ]]; then
|
||||
echo
|
||||
echo "Failed!"
|
||||
echo
|
||||
exit "${res}"
|
||||
fi
|
||||
}
|
||||
|
||||
|
||||
run tar cf "hadoop-${VERSION}.tar" "hadoop-${VERSION}"
|
||||
run gzip -f "hadoop-${VERSION}.tar"
|
||||
echo
|
||||
echo "Hadoop dist tar available at: ${BASEDIR}/hadoop-${VERSION}.tar.gz"
|
||||
echo
|
|
@ -83,151 +83,39 @@
|
|||
<build>
|
||||
<plugins>
|
||||
<plugin>
|
||||
<groupId>org.apache.maven.plugins</groupId>
|
||||
<artifactId>maven-antrun-plugin</artifactId>
|
||||
<groupId>org.codehaus.mojo</groupId>
|
||||
<artifactId>exec-maven-plugin</artifactId>
|
||||
<executions>
|
||||
<execution>
|
||||
<id>dist</id>
|
||||
<phase>prepare-package</phase>
|
||||
<goals>
|
||||
<goal>run</goal>
|
||||
<goal>exec</goal>
|
||||
</goals>
|
||||
<configuration>
|
||||
<target>
|
||||
<echo file="${project.build.directory}/dist-layout-stitching.sh">
|
||||
run() {
|
||||
echo "\$ ${@}"
|
||||
"${@}"
|
||||
res=$?
|
||||
if [ $res != 0 ]; then
|
||||
echo
|
||||
echo "Failed!"
|
||||
echo
|
||||
exit $res
|
||||
fi
|
||||
}
|
||||
|
||||
findFileInDir(){
|
||||
local file="$1";
|
||||
local dir="${2:-./share}";
|
||||
local count=$(find "$dir" -iname "$file"|wc -l)
|
||||
echo "$count";
|
||||
}
|
||||
|
||||
copyIfNotExists(){
|
||||
local src="$1"
|
||||
local srcName=$(basename "$src")
|
||||
local dest="$2";
|
||||
if [ -f "$src" ]; then
|
||||
if [[ "$srcName" != *.jar ]] || [ $(findFileInDir "$srcName") -eq "0" ]; then
|
||||
local destDir=$(dirname "$dest")
|
||||
mkdir -p "$destDir"
|
||||
cp "$src" "$dest"
|
||||
fi
|
||||
else
|
||||
for childPath in "$src"/* ;
|
||||
do
|
||||
child=$(basename "$childPath");
|
||||
if [ "$child" == "doc" ] || [ "$child" == "webapps" ]; then
|
||||
mkdir -p "$dest"/"$child"
|
||||
cp -r "$src"/"$child"/* "$dest"/"$child"
|
||||
continue;
|
||||
fi
|
||||
copyIfNotExists "$src"/"$child" "$dest"/"$child"
|
||||
done
|
||||
fi
|
||||
}
|
||||
|
||||
#Copy all contents as is except the lib.
|
||||
#for libs check for existence in share directory, if not exist then only copy.
|
||||
copy(){
|
||||
local src="$1";
|
||||
local dest="$2";
|
||||
if [ -d "$src" ]; then
|
||||
for childPath in "$src"/* ;
|
||||
do
|
||||
child=$(basename "$childPath");
|
||||
if [ "$child" == "share" ]; then
|
||||
copyIfNotExists "$src"/"$child" "$dest"/"$child"
|
||||
else
|
||||
if [ -d "$src"/"$child" ]; then
|
||||
mkdir -p "$dest"/"$child"
|
||||
cp -r "$src"/"$child"/* "$dest"/"$child"
|
||||
else
|
||||
cp -r "$src"/"$child" "$dest"/"$child"
|
||||
fi
|
||||
fi
|
||||
done
|
||||
fi
|
||||
}
|
||||
|
||||
# Shellcheck SC2086
|
||||
ROOT=$(cd "${project.build.directory}"/../..;pwd)
|
||||
echo
|
||||
echo "Current directory $(pwd)"
|
||||
echo
|
||||
run rm -rf hadoop-${project.version}
|
||||
run mkdir hadoop-${project.version}
|
||||
run cd hadoop-${project.version}
|
||||
run cp "$ROOT"/LICENSE.txt .
|
||||
run cp "$ROOT"/NOTICE.txt .
|
||||
run cp "$ROOT"/README.txt .
|
||||
|
||||
# Copy hadoop-common first so that it have always have all dependencies.
|
||||
# Remaining projects will copy only libraries which are not present already in 'share' directory.
|
||||
run copy "$ROOT"/hadoop-common-project/hadoop-common/target/hadoop-common-${project.version} .
|
||||
run copy "$ROOT"/hadoop-common-project/hadoop-nfs/target/hadoop-nfs-${project.version} .
|
||||
run copy "$ROOT"/hadoop-hdfs-project/hadoop-hdfs/target/hadoop-hdfs-${project.version} .
|
||||
run copy "$ROOT"/hadoop-hdfs-project/hadoop-hdfs-nfs/target/hadoop-hdfs-nfs-${project.version} .
|
||||
run copy "$ROOT"/hadoop-yarn-project/target/hadoop-yarn-project-${project.version} .
|
||||
run copy "$ROOT"/hadoop-mapreduce-project/target/hadoop-mapreduce-${project.version} .
|
||||
run copy "$ROOT"/hadoop-tools/hadoop-tools-dist/target/hadoop-tools-dist-${project.version} .
|
||||
|
||||
#copy httpfs and kms as is
|
||||
run cp -r "$ROOT"/hadoop-hdfs-project/hadoop-hdfs-httpfs/target/hadoop-hdfs-httpfs-${project.version}/* .
|
||||
run cp -r "$ROOT"/hadoop-common-project/hadoop-kms/target/hadoop-kms-${project.version}/* .
|
||||
|
||||
echo
|
||||
echo "Hadoop dist layout available at: ${project.build.directory}/hadoop-${project.version}"
|
||||
echo
|
||||
</echo>
|
||||
<exec executable="${shell-executable}" dir="${project.build.directory}" failonerror="true">
|
||||
<arg line="./dist-layout-stitching.sh"/>
|
||||
</exec>
|
||||
</target>
|
||||
<executable>${basedir}/../dev-support/bin/dist-layout-stitching</executable>
|
||||
<workingDirectory>${project.build.directory}</workingDirectory>
|
||||
<requiresOnline>false</requiresOnline>
|
||||
<arguments>
|
||||
<argument>${project.version}</argument>
|
||||
<argument>${project.build.directory}</argument>
|
||||
</arguments>
|
||||
</configuration>
|
||||
</execution>
|
||||
<execution>
|
||||
<id>tar</id>
|
||||
<phase>package</phase>
|
||||
<goals>
|
||||
<goal>run</goal>
|
||||
<goal>exec</goal>
|
||||
</goals>
|
||||
<configuration>
|
||||
<target if="tar">
|
||||
<echo file="${project.build.directory}/dist-tar-stitching.sh">
|
||||
run() {
|
||||
echo "\$ ${@}"
|
||||
"${@}"
|
||||
res=$?
|
||||
if [ $res != 0 ]; then
|
||||
echo
|
||||
echo "Failed!"
|
||||
echo
|
||||
exit $res
|
||||
fi
|
||||
}
|
||||
|
||||
run tar cf hadoop-${project.version}.tar hadoop-${project.version}
|
||||
run gzip -f hadoop-${project.version}.tar
|
||||
echo
|
||||
echo "Hadoop dist tar available at: ${project.build.directory}/hadoop-${project.version}.tar.gz"
|
||||
echo
|
||||
</echo>
|
||||
<exec executable="${shell-executable}" dir="${project.build.directory}" failonerror="true">
|
||||
<arg line="./dist-tar-stitching.sh"/>
|
||||
</exec>
|
||||
</target>
|
||||
<executable>${basedir}/../dev-support/bin/dist-tar-stitching</executable>
|
||||
<workingDirectory>${project.build.directory}</workingDirectory>
|
||||
<requiresOnline>false</requiresOnline>
|
||||
<arguments>
|
||||
<argument>${project.version}</argument>
|
||||
<argument>${project.build.directory}</argument>
|
||||
</arguments>
|
||||
</configuration>
|
||||
</execution>
|
||||
</executions>
|
||||
|
|
Loading…
Reference in New Issue