340 lines
12 KiB
Groovy
340 lines
12 KiB
Groovy
/*
|
|
* Licensed to Elasticsearch under one or more contributor
|
|
* license agreements. See the NOTICE file distributed with
|
|
* this work for additional information regarding copyright
|
|
* ownership. Elasticsearch licenses this file to you under
|
|
* the Apache License, Version 2.0 (the "License"); you may
|
|
* not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing,
|
|
* software distributed under the License is distributed on an
|
|
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
* KIND, either express or implied. See the License for the
|
|
* specific language governing permissions and limitations
|
|
* under the License.
|
|
*/
|
|
|
|
import org.apache.tools.ant.taskdefs.condition.Os
|
|
import org.elasticsearch.gradle.BuildPlugin
|
|
import org.elasticsearch.gradle.EmptyDirTask
|
|
import org.elasticsearch.gradle.LoggedExec
|
|
import org.elasticsearch.gradle.MavenFilteringHack
|
|
import org.elasticsearch.gradle.VersionProperties
|
|
import org.elasticsearch.gradle.plugin.PluginBuildPlugin
|
|
|
|
import java.nio.file.Files
|
|
import java.nio.file.Path
|
|
|
|
// need this so Zip/Tar tasks get basic defaults...
|
|
apply plugin: 'base'
|
|
|
|
// CopySpec does not make it easy to create an empty directory so we
|
|
// create the directory that we want, and then point CopySpec to its
|
|
// parent to copy to the root of the distribution
|
|
ext.logsDir = new File(buildDir, 'logs-hack/logs')
|
|
task createLogsDir(type: EmptyDirTask) {
|
|
dir "${logsDir}"
|
|
dirMode 0755
|
|
}
|
|
ext.pluginsDir= new File(buildDir, 'plugins-hack/plugins')
|
|
task createPluginsDir(type: EmptyDirTask) {
|
|
dir "${pluginsDir}"
|
|
dirMode 0755
|
|
}
|
|
|
|
CopySpec archiveFiles(CopySpec modulesFiles, String distributionType, boolean oss) {
|
|
return copySpec {
|
|
into("elasticsearch-${version}") {
|
|
into('lib') {
|
|
with libFiles(oss)
|
|
}
|
|
into('config') {
|
|
dirMode 0750
|
|
fileMode 0660
|
|
with configFiles(distributionType, oss)
|
|
}
|
|
into('bin') {
|
|
with binFiles(distributionType, oss)
|
|
}
|
|
into('') {
|
|
from {
|
|
dirMode 0755
|
|
logsDir.getParent()
|
|
}
|
|
}
|
|
into('') {
|
|
from {
|
|
dirMode 0755
|
|
pluginsDir.getParent()
|
|
}
|
|
}
|
|
from(rootProject.projectDir) {
|
|
include 'README.textile'
|
|
}
|
|
from(rootProject.file('licenses')) {
|
|
include oss ? 'APACHE-LICENSE-2.0.txt' : 'ELASTIC-LICENSE.txt'
|
|
rename { 'LICENSE.txt' }
|
|
}
|
|
|
|
with noticeFile
|
|
into('modules') {
|
|
with modulesFiles
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// common config across all zip/tar
|
|
tasks.withType(AbstractArchiveTask) {
|
|
dependsOn createLogsDir, createPluginsDir
|
|
String subdir = it.name.substring('build'.size()).replaceAll(/[A-Z]/) { '-' + it.toLowerCase() }.substring(1)
|
|
destinationDir = file("${subdir}/build/distributions")
|
|
baseName = "elasticsearch${subdir.contains('oss') ? '-oss' : ''}"
|
|
}
|
|
|
|
Closure commonZipConfig = {
|
|
dirMode 0755
|
|
fileMode 0644
|
|
}
|
|
|
|
task buildIntegTestZip(type: Zip) {
|
|
configure(commonZipConfig)
|
|
with archiveFiles(transportModulesFiles, 'zip', true)
|
|
}
|
|
|
|
task buildWindowsZip(type: Zip) {
|
|
configure(commonZipConfig)
|
|
archiveClassifier = 'windows-x86_64'
|
|
with archiveFiles(modulesFiles(false), 'zip', false)
|
|
}
|
|
|
|
task buildOssWindowsZip(type: Zip) {
|
|
configure(commonZipConfig)
|
|
archiveClassifier = 'windows-x86_64'
|
|
with archiveFiles(modulesFiles(true), 'zip', true)
|
|
}
|
|
|
|
Closure commonTarConfig = {
|
|
extension = 'tar.gz'
|
|
compression = Compression.GZIP
|
|
dirMode 0755
|
|
fileMode 0644
|
|
}
|
|
|
|
task buildDarwinTar(type: Tar) {
|
|
configure(commonTarConfig)
|
|
archiveClassifier = 'darwin-x86_64'
|
|
with archiveFiles(modulesFiles(false), 'tar', false)
|
|
}
|
|
|
|
task buildOssDarwinTar(type: Tar) {
|
|
configure(commonTarConfig)
|
|
archiveClassifier = 'darwin-x86_64'
|
|
with archiveFiles(modulesFiles(true), 'tar', true)
|
|
}
|
|
|
|
task buildLinuxTar(type: Tar) {
|
|
configure(commonTarConfig)
|
|
archiveClassifier = 'linux-x86_64'
|
|
with archiveFiles(modulesFiles(false), 'tar', false)
|
|
}
|
|
|
|
task buildOssLinuxTar(type: Tar) {
|
|
configure(commonTarConfig)
|
|
archiveClassifier = 'linux-x86_64'
|
|
with archiveFiles(modulesFiles(true), 'tar', true)
|
|
}
|
|
|
|
Closure tarExists = { it -> new File('/bin/tar').exists() || new File('/usr/bin/tar').exists() || new File('/usr/local/bin/tar').exists() }
|
|
Closure unzipExists = { it -> new File('/bin/unzip').exists() || new File('/usr/bin/unzip').exists() || new File('/usr/local/bin/unzip').exists() }
|
|
|
|
// This configures the default artifact for the distribution specific
|
|
// subprojects. We have subprojects for two reasons:
|
|
// 1. Gradle project substitutions can only bind to the default
|
|
// configuration of a project
|
|
// 2. The integ-test-zip and zip distributions have the exact same
|
|
// filename, so they must be placed in different directories.
|
|
subprojects {
|
|
apply plugin: 'distribution'
|
|
|
|
String buildTask = "build${it.name.replaceAll(/-[a-z]/) { it.substring(1).toUpperCase() }.capitalize()}"
|
|
ext.buildDist = parent.tasks.getByName(buildTask)
|
|
artifacts {
|
|
'default' buildDist
|
|
}
|
|
|
|
// sanity checks if archives can be extracted
|
|
File archiveExtractionDir
|
|
if (project.name.contains('tar')) {
|
|
archiveExtractionDir = new File(buildDir, 'tar-extracted')
|
|
} else {
|
|
assert project.name.contains('zip')
|
|
archiveExtractionDir = new File(buildDir, 'zip-extracted')
|
|
}
|
|
task checkExtraction(type: LoggedExec) {
|
|
dependsOn buildDist
|
|
doFirst {
|
|
project.delete(archiveExtractionDir)
|
|
archiveExtractionDir.mkdirs()
|
|
}
|
|
}
|
|
check.dependsOn checkExtraction
|
|
if (project.name.contains('tar')) {
|
|
checkExtraction {
|
|
onlyIf tarExists
|
|
commandLine 'tar', '-xvzf', "${-> buildDist.outputs.files.singleFile}", '-C', archiveExtractionDir
|
|
}
|
|
} else {
|
|
assert project.name.contains('zip')
|
|
checkExtraction {
|
|
onlyIf unzipExists
|
|
commandLine 'unzip', "${-> buildDist.outputs.files.singleFile}", '-d', archiveExtractionDir
|
|
}
|
|
}
|
|
|
|
Closure toolExists
|
|
if (project.name.contains('tar')) {
|
|
toolExists = tarExists
|
|
} else {
|
|
assert project.name.contains('zip')
|
|
toolExists = unzipExists
|
|
}
|
|
|
|
|
|
task checkLicense {
|
|
dependsOn buildDist, checkExtraction
|
|
onlyIf toolExists
|
|
doLast {
|
|
String licenseFilename = null
|
|
if (project.name.contains('oss-') || project.name == 'integ-test-zip') {
|
|
licenseFilename = "APACHE-LICENSE-2.0.txt"
|
|
} else {
|
|
licenseFilename = "ELASTIC-LICENSE.txt"
|
|
}
|
|
final List<String> licenseLines = Files.readAllLines(rootDir.toPath().resolve("licenses/" + licenseFilename))
|
|
final Path licensePath = archiveExtractionDir.toPath().resolve("elasticsearch-${VersionProperties.elasticsearch}/LICENSE.txt")
|
|
assertLinesInFile(licensePath, licenseLines)
|
|
}
|
|
}
|
|
check.dependsOn checkLicense
|
|
|
|
task checkNotice {
|
|
dependsOn buildDist, checkExtraction
|
|
onlyIf toolExists
|
|
doLast {
|
|
final List<String> noticeLines = Arrays.asList("Elasticsearch", "Copyright 2009-2018 Elasticsearch")
|
|
final Path noticePath = archiveExtractionDir.toPath().resolve("elasticsearch-${VersionProperties.elasticsearch}/NOTICE.txt")
|
|
assertLinesInFile(noticePath, noticeLines)
|
|
}
|
|
}
|
|
check.dependsOn checkNotice
|
|
|
|
if (project.name == 'zip' || project.name == 'tar') {
|
|
project.ext.licenseName = 'Elastic License'
|
|
project.ext.licenseUrl = ext.elasticLicenseUrl
|
|
task checkMlCppNotice {
|
|
dependsOn buildDist, checkExtraction
|
|
onlyIf toolExists
|
|
doLast {
|
|
// this is just a small sample from the C++ notices, the idea being that if we've added these lines we've probably added all the required lines
|
|
final List<String> expectedLines = Arrays.asList("Apache log4cxx", "Boost Software License - Version 1.0 - August 17th, 2003")
|
|
final Path noticePath = archiveExtractionDir.toPath().resolve("elasticsearch-${VersionProperties.elasticsearch}/modules/x-pack-ml/NOTICE.txt")
|
|
final List<String> actualLines = Files.readAllLines(noticePath)
|
|
for (final String expectedLine : expectedLines) {
|
|
if (actualLines.contains(expectedLine) == false) {
|
|
throw new GradleException("expected [${noticePath}] to contain [${expectedLine}] but it did not")
|
|
}
|
|
}
|
|
}
|
|
}
|
|
check.dependsOn checkMlCppNotice
|
|
}
|
|
}
|
|
|
|
/*****************************************************************************
|
|
* Rest test config *
|
|
*****************************************************************************/
|
|
configure(subprojects.findAll { it.name == 'integ-test-zip' }) {
|
|
apply plugin: 'elasticsearch.standalone-rest-test'
|
|
apply plugin: 'elasticsearch.rest-test'
|
|
|
|
integTest {
|
|
includePackaged = true
|
|
}
|
|
|
|
integTestCluster {
|
|
dependsOn assemble
|
|
distribution = project.name
|
|
}
|
|
integTestRunner {
|
|
if (Os.isFamily(Os.FAMILY_WINDOWS) && System.getProperty('tests.timeoutSuite') == null) {
|
|
// override the suite timeout to 30 mins for windows, because it has the most inefficient filesystem known to man
|
|
systemProperty 'tests.timeoutSuite', '1800000!'
|
|
}
|
|
}
|
|
|
|
processTestResources {
|
|
inputs.properties(project(':distribution').restTestExpansions)
|
|
MavenFilteringHack.filter(it, project(':distribution').restTestExpansions)
|
|
}
|
|
}
|
|
|
|
/*****************************************************************************
|
|
* Maven config *
|
|
*****************************************************************************/
|
|
configure(subprojects.findAll { it.name.contains('zip') }) {
|
|
// only zip distributions go to maven
|
|
BuildPlugin.configurePomGeneration(project)
|
|
apply plugin: 'nebula.info-scm'
|
|
apply plugin: 'nebula.maven-base-publish'
|
|
apply plugin: 'nebula.maven-scm'
|
|
|
|
// note: the group must be correct before applying the nexus plugin, or
|
|
// it will capture the wrong value...
|
|
String subgroup = project.name == 'integ-test-zip' ? 'integ-test-zip' : 'zip'
|
|
project.group = "org.elasticsearch.distribution.${subgroup}"
|
|
|
|
// make the pom file name use elasticsearch instead of the project name
|
|
archivesBaseName = "elasticsearch${it.name.contains('oss') ? '-oss' : ''}"
|
|
|
|
publishing {
|
|
publications {
|
|
nebula {
|
|
artifactId archivesBaseName
|
|
artifact buildDist
|
|
}
|
|
/*
|
|
* HUGE HACK: the underlying maven publication library refuses to
|
|
* deploy any attached artifacts when the packaging type is set to
|
|
* 'pom'. But Sonatype's OSS repositories require source files for
|
|
* artifacts that are of type 'zip'. We already publish the source
|
|
* and javadoc for Elasticsearch under the various other subprojects.
|
|
* So here we create another publication using the same name that
|
|
* has the "real" pom, and rely on the fact that gradle will execute
|
|
* the publish tasks in alphabetical order. This lets us publish the
|
|
* zip file and even though the pom says the type is 'pom' instead of
|
|
* 'zip'. We cannot setup a dependency between the tasks because the
|
|
* publishing tasks are created *extremely* late in the configuration
|
|
* phase, so that we cannot get ahold of the actual task. Furthermore,
|
|
* this entire hack only exists so we can make publishing to maven
|
|
* local work, since we publish to maven central externally.
|
|
*/
|
|
nebulaRealPom(MavenPublication) {
|
|
artifactId archivesBaseName
|
|
pom.packaging = 'pom'
|
|
pom.withXml { XmlProvider xml ->
|
|
Node root = xml.asNode()
|
|
root.appendNode('name', 'Elasticsearch')
|
|
root.appendNode('description', 'A Distributed RESTful Search Engine')
|
|
root.appendNode('url', PluginBuildPlugin.urlFromOrigin(project.scminfo.origin))
|
|
Node scmNode = root.appendNode('scm')
|
|
scmNode.appendNode('url', project.scminfo.origin)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|