diff --git a/src/main/docbkx/developer.xml b/src/main/docbkx/developer.xml index bd2626d37b9..a2e9d46bab6 100644 --- a/src/main/docbkx/developer.xml +++ b/src/main/docbkx/developer.xml @@ -249,7 +249,7 @@ mvn clean package -DskipTests
- Making Release Candidate + Making a Release Candidate I'll explain by running through the process. See later in this section for more detail on particular steps. The Hadoop How To Release wiki page informs much of the below and may have more detail on particular sections so it is worth review. @@ -260,8 +260,9 @@ mvn clean package -DskipTests $ mvn clean org.codehaus.mojo:versions-maven-plugin:1.3.1:set -DnewVersion=0.96.0 Checkin the CHANGES.txt and version changes. - Now, build the src tarball. This tarball is hadoop version independent. It is just the pure src code and is without the hadoop1 or hadoop2 taint. - $ MAVEN_OPTS="-Xmx2g" mvn clean install javadoc:aggregate site assembly:single -DskipTests -Dassembly.file=hbase-assembly/src/main/assembly/src.xml -Prelease + Now, build the src tarball. This tarball is hadoop version independent. It is just the pure src code and documentation without an hadoop1 or hadoop2 taint. + Add the -Prelease profile when building; it checks files for licenses and will fail the build if unlicensed files present. + $ MAVEN_OPTS="-Xmx2g" mvn clean install -DskipTests assembly:single -Dassembly.file=hbase-assembly/src/main/assembly/src.xml -Prelease Undo the tarball and make sure it looks good (A good test is seeing if you can build from the undone tarball). Save it off to a version directory, i.e a directory somewhere where you are collecting all of the tarballs you will publish as part of the release candidate. For example if we were building a @@ -269,18 +270,24 @@ mvn clean package -DskipTests we will publish this directory as our release candidate up on people.apache.org/~you. Now we are into the making of the hadoop1 and hadoop2 specific builds. Lets do hadoop1 first. - - First generate the hadoop1 poms. See the generate-hadoopX-poms.sh script usage for what it expects by way of arguments. + First generate the hadoop1 poms. See the generate-hadoopX-poms.sh script usage for what it expects by way of arguments. You will find it in the dev-support subdirectory. In the below, we generate hadoop1 poms with a version of 0.96.0-hadoop1 (the script will look for a version of 0.96.0 in the current pom.xml). $ ./dev-support/generate-hadoopX-poms.sh 0.96.0 0.96.0-hadoop1 The script will work silently if all goes well. It will drop a pom.xml.hadoop1 beside all pom.xmls in all modules. - Now build the hadoop1 tarball. Note we ask for the generation of javadoc and site and we reference the new pom.xml.hadoop1 explicitly. - We also add the -Prelease profile when building; it checks files for licenses and will fail the build if unlicensed files present. - $ MAVEN_OPTS="-Xmx3g" mvn -f pom.xml.hadoop1 clean install -DskipTests javadoc:aggregate site assembly:single -Prelease + + Now build the hadoop1 tarball. Note how we reference the new pom.xml.hadoop1 explicitly. + We also add the -Prelease profile when building; it checks files for licenses and will fail the build if unlicensed files present. + Do it in two steps. First install into the local repository and then generate documentation and assemble the tarball + (Otherwise build complains that hbase modules are not in maven repo when we try to do it all in the one go especially on fresh repo). + It seems that you need the install goal in both steps. + $ MAVEN_OPTS="-Xmx3g" mvn -f pom.xml.hadoop1 clean install -DskipTests -Prelease +$ MAVEN_OPTS="-Xmx3g" mvn -f pom.xml.hadoop1 install -DskipTests site assembly:single -Prelease Undo the generated tarball and check it out. Look at doc. and see if it runs, etc. Are the set of modules appropriate: e.g. do we have a hbase-hadoop2-compat in the hadoop1 tarball? If good, copy the tarball to your version directory. -Now deploy hadoop1 hbase to mvn. Do the mvn deploy and tgz for a particular version all together in the one go else if you flip between hadoop1 and hadoop2 builds, + +I'll tag the release at this point since its looking good. If we find an issue later, we can delete the tag and start over. Release needs to be tagged when we do next step. +Now deploy hadoop1 hbase to mvn. Do the mvn deploy and tgz for a particular version all together in the one go else if you flip between hadoop1 and hadoop2 builds, you might mal-publish poms and hbase-default.xml's (the version interpolations won't match). This time we use the apache-release profile instead of just release profile when doing mvn deploy; it will invoke the apache pom referenced by our poms. It will also sign your artifacts published to mvn as long as your settings.xml in your local .m2 @@ -292,8 +299,9 @@ The last command above copies all artifacts for hadoop1 up to mvn repo. If no < Lets do the hadoop2 artifacts (read above hadoop1 section closely before coming here because we don't repeat explaination in the below). # Generate the hadoop2 poms. $ ./dev-support/generate-hadoopX-poms.sh 0.96.0 0.96.0-hadoop2 -# Build the hadoop2 tarball. -$ MAVEN_OPTS="-Xmx3g" mvn -f pom.xml.hadoop2 clean install -DskipTests javadoc:aggregate site assembly:single -Prelease +# Install the hbase hadoop2 jars into local repo then build the doc and tarball +$ MAVEN_OPTS="-Xmx3g" mvn -f pom.xml.hadoop2 clean install -DskipTests -Prelease +$ MAVEN_OPTS="-Xmx3g" mvn -f pom.xml.hadoop2 install -DskipTests site assembly:single -Prelease # Undo the tgz and check it out. If good, copy the tarball to your 'version directory'. Now deploy to mvn. $ MAVEN_OPTS="-Xmx3g" mvn -f pom.xml.hadoop2 deploy -DskipTests -Papache-release -Pgpg