Merge branch 'master' into index-lifecycle

This commit is contained in:
Colin Goodheart-Smithe 2018-05-29 09:47:43 +01:00
commit 2ccb4df005
No known key found for this signature in database
GPG Key ID: F975E7BDD739B3C7
95 changed files with 1855 additions and 1171 deletions

View File

@ -70,31 +70,44 @@ public class RestIntegTestTask extends DefaultTask {
runner.parallelism = '1' runner.parallelism = '1'
runner.include('**/*IT.class') runner.include('**/*IT.class')
runner.systemProperty('tests.rest.load_packaged', 'false') runner.systemProperty('tests.rest.load_packaged', 'false')
// we pass all nodes to the rest cluster to allow the clients to round-robin between them
// this is more realistic than just talking to a single node
runner.systemProperty('tests.rest.cluster', "${-> nodes.collect{it.httpUri()}.join(",")}")
runner.systemProperty('tests.config.dir', "${-> nodes[0].pathConf}")
// TODO: our "client" qa tests currently use the rest-test plugin. instead they should have their own plugin
// that sets up the test cluster and passes this transport uri instead of http uri. Until then, we pass
// both as separate sysprops
runner.systemProperty('tests.cluster', "${-> nodes[0].transportUri()}")
// dump errors and warnings from cluster log on failure if (System.getProperty("tests.rest.cluster") == null) {
TaskExecutionAdapter logDumpListener = new TaskExecutionAdapter() { if (System.getProperty("tests.cluster") != null) {
@Override throw new IllegalArgumentException("tests.rest.cluster and tests.cluster must both be null or non-null")
void afterExecute(Task task, TaskState state) { }
if (state.failure != null) { // we pass all nodes to the rest cluster to allow the clients to round-robin between them
for (NodeInfo nodeInfo : nodes) { // this is more realistic than just talking to a single node
printLogExcerpt(nodeInfo) runner.systemProperty('tests.rest.cluster', "${-> nodes.collect{it.httpUri()}.join(",")}")
runner.systemProperty('tests.config.dir', "${-> nodes[0].pathConf}")
// TODO: our "client" qa tests currently use the rest-test plugin. instead they should have their own plugin
// that sets up the test cluster and passes this transport uri instead of http uri. Until then, we pass
// both as separate sysprops
runner.systemProperty('tests.cluster', "${-> nodes[0].transportUri()}")
// dump errors and warnings from cluster log on failure
TaskExecutionAdapter logDumpListener = new TaskExecutionAdapter() {
@Override
void afterExecute(Task task, TaskState state) {
if (state.failure != null) {
for (NodeInfo nodeInfo : nodes) {
printLogExcerpt(nodeInfo)
}
} }
} }
} }
} runner.doFirst {
runner.doFirst { project.gradle.addListener(logDumpListener)
project.gradle.addListener(logDumpListener) }
} runner.doLast {
runner.doLast { project.gradle.removeListener(logDumpListener)
project.gradle.removeListener(logDumpListener) }
} else {
if (System.getProperty("tests.cluster") == null) {
throw new IllegalArgumentException("tests.rest.cluster and tests.cluster must both be null or non-null")
}
// an external cluster was specified and all responsibility for cluster configuration is taken by the user
runner.systemProperty('tests.rest.cluster', System.getProperty("tests.rest.cluster"))
runner.systemProperty('test.cluster', System.getProperty("tests.cluster"))
} }
// copy the rest spec/tests into the test resources // copy the rest spec/tests into the test resources
@ -109,7 +122,10 @@ public class RestIntegTestTask extends DefaultTask {
clusterInit.enabled = false clusterInit.enabled = false
return // no need to add cluster formation tasks if the task won't run! return // no need to add cluster formation tasks if the task won't run!
} }
nodes = ClusterFormationTasks.setup(project, "${name}Cluster", runner, clusterConfig) // only create the cluster if needed as otherwise an external cluster to use was specified
if (System.getProperty("tests.rest.cluster") == null) {
nodes = ClusterFormationTasks.setup(project, "${name}Cluster", runner, clusterConfig)
}
super.dependsOn(runner.finalizedBy) super.dependsOn(runner.finalizedBy)
} }
} }

View File

@ -49,7 +49,9 @@ task createPluginsDir(type: EmptyDirTask) {
CopySpec archiveFiles(CopySpec modulesFiles, String distributionType, boolean oss) { CopySpec archiveFiles(CopySpec modulesFiles, String distributionType, boolean oss) {
return copySpec { return copySpec {
into("elasticsearch-${version}") { into("elasticsearch-${version}") {
with libFiles into('lib') {
with libFiles
}
into('config') { into('config') {
dirMode 0750 dirMode 0750
fileMode 0660 fileMode 0660

View File

@ -227,13 +227,15 @@ configure(subprojects.findAll { ['archives', 'packages'].contains(it.name) }) {
* Common files in all distributions * * Common files in all distributions *
*****************************************************************************/ *****************************************************************************/
libFiles = copySpec { libFiles = copySpec {
into 'lib' // delay by using closures, since they have not yet been configured, so no jar task exists yet
from { project(':server').jar } from { project(':server').jar }
from { project(':server').configurations.runtime } from { project(':server').configurations.runtime }
from { project(':libs:plugin-classloader').jar } from { project(':libs:plugin-classloader').jar }
// delay add tools using closures, since they have not yet been configured, so no jar task exists yet
from { project(':distribution:tools:launchers').jar } from { project(':distribution:tools:launchers').jar }
from { project(':distribution:tools:plugin-cli').jar } into('tools/plugin-cli') {
from { project(':distribution:tools:plugin-cli').jar }
from { project(':distribution:tools:plugin-cli').configurations.runtime }
}
} }
modulesFiles = { oss -> modulesFiles = { oss ->

View File

@ -124,13 +124,23 @@ Closure commonPackageConfig(String type, boolean oss) {
include 'README.textile' include 'README.textile'
fileMode 0644 fileMode 0644
} }
into('lib') {
with copySpec {
with libFiles
// we need to specify every intermediate directory so we iterate through the parents; duplicate calls with the same part are fine
eachFile { FileCopyDetails fcp ->
String[] segments = fcp.relativePath.segments
for (int i = segments.length - 2; i > 0 && segments[i] != 'lib'; --i) {
directory('/' + segments[0..i].join('/'), 0755)
}
fcp.mode = 0644
}
}
}
into('modules') { into('modules') {
with copySpec { with copySpec {
with modulesFiles(oss) with modulesFiles(oss)
// we need to specify every intermediate directory, but modules could have sub directories // we need to specify every intermediate directory so we iterate through the parents; duplicate calls with the same part are fine
// and there might not be any files as direct children of intermediates (eg platform)
// so we must iterate through the parents, but duplicate calls with the same path
// are ok (they don't show up in the built packages)
eachFile { FileCopyDetails fcp -> eachFile { FileCopyDetails fcp ->
String[] segments = fcp.relativePath.segments String[] segments = fcp.relativePath.segments
for (int i = segments.length - 2; i > 0 && segments[i] != 'modules'; --i) { for (int i = segments.length - 2; i > 0 && segments[i] != 'modules'; --i) {
@ -251,8 +261,8 @@ ospackage {
signingKeyId = project.hasProperty('signing.keyId') ? project.property('signing.keyId') : 'D88E42B4' signingKeyId = project.hasProperty('signing.keyId') ? project.property('signing.keyId') : 'D88E42B4'
signingKeyPassphrase = project.property('signing.password') signingKeyPassphrase = project.property('signing.password')
signingKeyRingFile = project.hasProperty('signing.secretKeyRingFile') ? signingKeyRingFile = project.hasProperty('signing.secretKeyRingFile') ?
project.file(project.property('signing.secretKeyRingFile')) : project.file(project.property('signing.secretKeyRingFile')) :
new File(new File(System.getProperty('user.home'), '.gnupg'), 'secring.gpg') new File(new File(System.getProperty('user.home'), '.gnupg'), 'secring.gpg')
} }
requires('coreutils') requires('coreutils')
@ -263,7 +273,6 @@ ospackage {
permissionGroup 'root' permissionGroup 'root'
into '/usr/share/elasticsearch' into '/usr/share/elasticsearch'
with libFiles
with noticeFile with noticeFile
} }

View File

@ -10,6 +10,12 @@ do
source "`dirname "$0"`"/$additional_source source "`dirname "$0"`"/$additional_source
done done
IFS=';' read -r -a additional_classpath_directories <<< "$ES_ADDITIONAL_CLASSPATH_DIRECTORIES"
for additional_classpath_directory in "${additional_classpath_directories[@]}"
do
ES_CLASSPATH="$ES_CLASSPATH:$ES_HOME/$additional_classpath_directory/*"
done
exec \ exec \
"$JAVA" \ "$JAVA" \
$ES_JAVA_OPTS \ $ES_JAVA_OPTS \

View File

@ -11,6 +11,12 @@ for /f "tokens=1*" %%a in ("%*") do (
set arguments=%%b set arguments=%%b
) )
if defined ES_ADDITIONAL_CLASSPATH_DIRECTORIES (
for %%a in ("%ES_ADDITIONAL_CLASSPATH_DIRECTORIES:;=","%") do (
set ES_CLASSPATH=!ES_CLASSPATH!;!ES_HOME!/%%a/*
)
)
%JAVA% ^ %JAVA% ^
%ES_JAVA_OPTS% ^ %ES_JAVA_OPTS% ^
-Des.path.home="%ES_HOME%" ^ -Des.path.home="%ES_HOME%" ^

View File

@ -1,5 +1,6 @@
#!/bin/bash #!/bin/bash
"`dirname "$0"`"/elasticsearch-cli \ ES_ADDITIONAL_CLASSPATH_DIRECTORIES=lib/tools/plugin-cli \
"`dirname "$0"`"/elasticsearch-cli \
org.elasticsearch.plugins.PluginCli \ org.elasticsearch.plugins.PluginCli \
"$@" "$@"

View File

@ -3,6 +3,7 @@
setlocal enabledelayedexpansion setlocal enabledelayedexpansion
setlocal enableextensions setlocal enableextensions
set ES_ADDITIONAL_CLASSPATH_DIRECTORIES=lib/tools/plugin-cli
call "%~dp0elasticsearch-cli.bat" ^ call "%~dp0elasticsearch-cli.bat" ^
org.elasticsearch.plugins.PluginCli ^ org.elasticsearch.plugins.PluginCli ^
%* ^ %* ^

View File

@ -19,14 +19,22 @@
apply plugin: 'elasticsearch.build' apply plugin: 'elasticsearch.build'
archivesBaseName = 'elasticsearch-plugin-cli'
dependencies { dependencies {
compileOnly "org.elasticsearch:elasticsearch:${version}" compileOnly "org.elasticsearch:elasticsearch:${version}"
compileOnly "org.elasticsearch:elasticsearch-cli:${version}" compileOnly "org.elasticsearch:elasticsearch-cli:${version}"
compile "org.bouncycastle:bcpg-jdk15on:1.59"
compile "org.bouncycastle:bcprov-jdk15on:1.59"
testCompile "org.elasticsearch.test:framework:${version}" testCompile "org.elasticsearch.test:framework:${version}"
testCompile 'com.google.jimfs:jimfs:1.1' testCompile 'com.google.jimfs:jimfs:1.1'
testCompile 'com.google.guava:guava:18.0' testCompile 'com.google.guava:guava:18.0'
} }
dependencyLicenses {
mapping from: /bc.*/, to: 'bouncycastle'
}
test { test {
// TODO: find a way to add permissions for the tests in this module // TODO: find a way to add permissions for the tests in this module
systemProperty 'tests.security.manager', 'false' systemProperty 'tests.security.manager', 'false'

View File

@ -0,0 +1 @@
ee93e5376bb6cf0a15c027b5f5e4393f2738e709

View File

@ -0,0 +1 @@
2507204241ab450456bdb8e8c0a8f986e418bd99

View File

@ -0,0 +1,17 @@
Copyright (c) 2000-2015 The Legion of the Bouncy Castle Inc. (http://www.bouncycastle.org)
Permission is hereby granted, free of charge, to any person obtaining a copy of this software
and associated documentation files (the "Software"), to deal in the Software without restriction,
including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense,
and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so,
subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial
portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED,
INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR
PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
DEALINGS IN THE SOFTWARE.

View File

@ -23,6 +23,16 @@ import joptsimple.OptionSet;
import joptsimple.OptionSpec; import joptsimple.OptionSpec;
import org.apache.lucene.search.spell.LevensteinDistance; import org.apache.lucene.search.spell.LevensteinDistance;
import org.apache.lucene.util.CollectionUtil; import org.apache.lucene.util.CollectionUtil;
import org.bouncycastle.jce.provider.BouncyCastleProvider;
import org.bouncycastle.openpgp.PGPException;
import org.bouncycastle.openpgp.PGPPublicKey;
import org.bouncycastle.openpgp.PGPPublicKeyRingCollection;
import org.bouncycastle.openpgp.PGPSignature;
import org.bouncycastle.openpgp.PGPSignatureList;
import org.bouncycastle.openpgp.PGPUtil;
import org.bouncycastle.openpgp.jcajce.JcaPGPObjectFactory;
import org.bouncycastle.openpgp.operator.jcajce.JcaKeyFingerprintCalculator;
import org.bouncycastle.openpgp.operator.jcajce.JcaPGPContentVerifierBuilderProvider;
import org.elasticsearch.Build; import org.elasticsearch.Build;
import org.elasticsearch.Version; import org.elasticsearch.Version;
import org.elasticsearch.bootstrap.JarHell; import org.elasticsearch.bootstrap.JarHell;
@ -37,12 +47,14 @@ import org.elasticsearch.core.internal.io.IOUtils;
import org.elasticsearch.env.Environment; import org.elasticsearch.env.Environment;
import java.io.BufferedReader; import java.io.BufferedReader;
import java.io.ByteArrayInputStream;
import java.io.IOException; import java.io.IOException;
import java.io.InputStream; import java.io.InputStream;
import java.io.InputStreamReader; import java.io.InputStreamReader;
import java.io.OutputStream; import java.io.OutputStream;
import java.net.HttpURLConnection; import java.net.HttpURLConnection;
import java.net.URI; import java.net.URI;
import java.net.URISyntaxException;
import java.net.URL; import java.net.URL;
import java.net.URLConnection; import java.net.URLConnection;
import java.net.URLDecoder; import java.net.URLDecoder;
@ -59,8 +71,10 @@ import java.nio.file.attribute.PosixFileAttributes;
import java.nio.file.attribute.PosixFilePermission; import java.nio.file.attribute.PosixFilePermission;
import java.nio.file.attribute.PosixFilePermissions; import java.nio.file.attribute.PosixFilePermissions;
import java.security.MessageDigest; import java.security.MessageDigest;
import java.security.NoSuchAlgorithmException;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Arrays; import java.util.Arrays;
import java.util.Base64;
import java.util.Collections; import java.util.Collections;
import java.util.HashMap; import java.util.HashMap;
import java.util.HashSet; import java.util.HashSet;
@ -116,7 +130,6 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
/** The plugin zip is not properly structured. */ /** The plugin zip is not properly structured. */
static final int PLUGIN_MALFORMED = 2; static final int PLUGIN_MALFORMED = 2;
/** The builtin modules, which are plugins, but cannot be installed or removed. */ /** The builtin modules, which are plugins, but cannot be installed or removed. */
static final Set<String> MODULES; static final Set<String> MODULES;
static { static {
@ -241,7 +254,7 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
if (OFFICIAL_PLUGINS.contains(pluginId)) { if (OFFICIAL_PLUGINS.contains(pluginId)) {
final String url = getElasticUrl(terminal, getStagingHash(), Version.CURRENT, isSnapshot(), pluginId, Platforms.PLATFORM_NAME); final String url = getElasticUrl(terminal, getStagingHash(), Version.CURRENT, isSnapshot(), pluginId, Platforms.PLATFORM_NAME);
terminal.println("-> Downloading " + pluginId + " from elastic"); terminal.println("-> Downloading " + pluginId + " from elastic");
return downloadZipAndChecksum(terminal, url, tmpDir, false); return downloadAndValidate(terminal, url, tmpDir, true);
} }
// now try as maven coordinates, a valid URL would only have a colon and slash // now try as maven coordinates, a valid URL would only have a colon and slash
@ -249,7 +262,7 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
if (coordinates.length == 3 && pluginId.contains("/") == false && pluginId.startsWith("file:") == false) { if (coordinates.length == 3 && pluginId.contains("/") == false && pluginId.startsWith("file:") == false) {
String mavenUrl = getMavenUrl(terminal, coordinates, Platforms.PLATFORM_NAME); String mavenUrl = getMavenUrl(terminal, coordinates, Platforms.PLATFORM_NAME);
terminal.println("-> Downloading " + pluginId + " from maven central"); terminal.println("-> Downloading " + pluginId + " from maven central");
return downloadZipAndChecksum(terminal, mavenUrl, tmpDir, true); return downloadAndValidate(terminal, mavenUrl, tmpDir, false);
} }
// fall back to plain old URL // fall back to plain old URL
@ -406,16 +419,44 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
} }
} }
/** Downloads a zip from the url, as well as a SHA512 (or SHA1) checksum, and checks the checksum. */ @SuppressForbidden(reason = "URL#openStream")
// pkg private for tests private InputStream urlOpenStream(final URL url) throws IOException {
@SuppressForbidden(reason = "We use openStream to download plugins") return url.openStream();
private Path downloadZipAndChecksum(Terminal terminal, String urlString, Path tmpDir, boolean allowSha1) throws Exception { }
/**
* Downloads a ZIP from the URL. This method also validates the downloaded plugin ZIP via the following means:
* <ul>
* <li>
* For an official plugin we download the SHA-512 checksum and validate the integrity of the downloaded ZIP. We also download the
* armored signature and validate the authenticity of the downloaded ZIP.
* </li>
* <li>
* For a non-official plugin we download the SHA-512 checksum and fallback to the SHA-1 checksum and validate the integrity of the
* downloaded ZIP.
* </li>
* </ul>
*
* @param terminal a terminal to log messages to
* @param urlString the URL of the plugin ZIP
* @param tmpDir a temporary directory to write downloaded files to
* @param officialPlugin true if the plugin is an official plugin
* @return the path to the downloaded plugin ZIP
* @throws IOException if an I/O exception occurs download or reading files and resources
* @throws PGPException if an exception occurs verifying the downloaded ZIP signature
* @throws UserException if checksum validation fails
*/
private Path downloadAndValidate(
final Terminal terminal,
final String urlString,
final Path tmpDir,
final boolean officialPlugin) throws IOException, PGPException, UserException {
Path zip = downloadZip(terminal, urlString, tmpDir); Path zip = downloadZip(terminal, urlString, tmpDir);
pathsToDeleteOnShutdown.add(zip); pathsToDeleteOnShutdown.add(zip);
String checksumUrlString = urlString + ".sha512"; String checksumUrlString = urlString + ".sha512";
URL checksumUrl = openUrl(checksumUrlString); URL checksumUrl = openUrl(checksumUrlString);
String digestAlgo = "SHA-512"; String digestAlgo = "SHA-512";
if (checksumUrl == null && allowSha1) { if (checksumUrl == null && officialPlugin == false) {
// fallback to sha1, until 7.0, but with warning // fallback to sha1, until 7.0, but with warning
terminal.println("Warning: sha512 not found, falling back to sha1. This behavior is deprecated and will be removed in a " + terminal.println("Warning: sha512 not found, falling back to sha1. This behavior is deprecated and will be removed in a " +
"future release. Please update the plugin to use a sha512 checksum."); "future release. Please update the plugin to use a sha512 checksum.");
@ -427,7 +468,7 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
throw new UserException(ExitCodes.IO_ERROR, "Plugin checksum missing: " + checksumUrlString); throw new UserException(ExitCodes.IO_ERROR, "Plugin checksum missing: " + checksumUrlString);
} }
final String expectedChecksum; final String expectedChecksum;
try (InputStream in = checksumUrl.openStream()) { try (InputStream in = urlOpenStream(checksumUrl)) {
/* /*
* The supported format of the SHA-1 files is a single-line file containing the SHA-1. The supported format of the SHA-512 files * The supported format of the SHA-1 files is a single-line file containing the SHA-1. The supported format of the SHA-512 files
* is a single-line file containing the SHA-512 and the filename, separated by two spaces. For SHA-1, we verify that the hash * is a single-line file containing the SHA-512 and the filename, separated by two spaces. For SHA-1, we verify that the hash
@ -465,23 +506,119 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
} }
} }
byte[] zipbytes = Files.readAllBytes(zip); try {
String gotChecksum = MessageDigests.toHexString(MessageDigest.getInstance(digestAlgo).digest(zipbytes)); final byte[] zipBytes = Files.readAllBytes(zip);
if (expectedChecksum.equals(gotChecksum) == false) { final String actualChecksum = MessageDigests.toHexString(MessageDigest.getInstance(digestAlgo).digest(zipBytes));
throw new UserException(ExitCodes.IO_ERROR, if (expectedChecksum.equals(actualChecksum) == false) {
digestAlgo + " mismatch, expected " + expectedChecksum + " but got " + gotChecksum); throw new UserException(
ExitCodes.IO_ERROR,
digestAlgo + " mismatch, expected " + expectedChecksum + " but got " + actualChecksum);
}
} catch (final NoSuchAlgorithmException e) {
// this should never happen as we are using SHA-1 and SHA-512 here
throw new AssertionError(e);
}
if (officialPlugin) {
verifySignature(zip, urlString);
} }
return zip; return zip;
} }
/**
* Verify the signature of the downloaded plugin ZIP. The signature is obtained from the source of the downloaded plugin by appending
* ".asc" to the URL. It is expected that the plugin is signed with the Elastic signing key with ID D27D666CD88E42B4.
*
* @param zip the path to the downloaded plugin ZIP
* @param urlString the URL source of the downloade plugin ZIP
* @throws IOException if an I/O exception occurs reading from various input streams
* @throws PGPException if the PGP implementation throws an internal exception during verification
*/
void verifySignature(final Path zip, final String urlString) throws IOException, PGPException {
final String ascUrlString = urlString + ".asc";
final URL ascUrl = openUrl(ascUrlString);
try (
// fin is a file stream over the downloaded plugin zip whose signature to verify
InputStream fin = pluginZipInputStream(zip);
// sin is a URL stream to the signature corresponding to the downloaded plugin zip
InputStream sin = urlOpenStream(ascUrl);
// pin is a input stream to the public key in ASCII-Armor format (RFC4880); the Armor data is in RFC2045 format
InputStream pin = getPublicKey()) {
final JcaPGPObjectFactory factory = new JcaPGPObjectFactory(PGPUtil.getDecoderStream(sin));
final PGPSignature signature = ((PGPSignatureList) factory.nextObject()).get(0);
// validate the signature has key ID matching our public key ID
final String keyId = Long.toHexString(signature.getKeyID()).toUpperCase(Locale.ROOT);
if (getPublicKeyId().equals(keyId) == false) {
throw new IllegalStateException("key id [" + keyId + "] does not match expected key id [" + getPublicKeyId() + "]");
}
// compute the signature of the downloaded plugin zip
final List<String> lines =
new BufferedReader(new InputStreamReader(pin, StandardCharsets.UTF_8)).lines().collect(Collectors.toList());
// skip armor headers and possible blank line
int index = 1;
for (; index < lines.size(); index++) {
if (lines.get(index).matches(".*: .*") == false && lines.get(index).matches("\\s*") == false) {
break;
}
}
final byte[] armoredData =
lines.subList(index, lines.size() - 1).stream().collect(Collectors.joining("\n")).getBytes(StandardCharsets.UTF_8);
final InputStream ain = Base64.getMimeDecoder().wrap(new ByteArrayInputStream(armoredData));
final PGPPublicKeyRingCollection collection = new PGPPublicKeyRingCollection(ain, new JcaKeyFingerprintCalculator());
final PGPPublicKey key = collection.getPublicKey(signature.getKeyID());
signature.init(new JcaPGPContentVerifierBuilderProvider().setProvider(new BouncyCastleProvider()), key);
final byte[] buffer = new byte[1024];
int read;
while ((read = fin.read(buffer)) != -1) {
signature.update(buffer, 0, read);
}
// finally we verify the signature of the downloaded plugin zip matches the expected signature
if (signature.verify() == false) {
throw new IllegalStateException("signature verification for [" + urlString + "] failed");
}
}
}
/**
* An input stream to the raw bytes of the plugin ZIP.
*
* @param zip the path to the downloaded plugin ZIP
* @return an input stream to the raw bytes of the plugin ZIP.
* @throws IOException if an I/O exception occurs preparing the input stream
*/
InputStream pluginZipInputStream(final Path zip) throws IOException {
return Files.newInputStream(zip);
}
/**
* Return the public key ID of the signing key that is expected to have signed the official plugin.
*
* @return the public key ID
*/
String getPublicKeyId() {
return "D27D666CD88E42B4";
}
/**
* An input stream to the public key of the signing key.
*
* @return an input stream to the public key
*/
InputStream getPublicKey() {
return InstallPluginCommand.class.getResourceAsStream("/public_key.asc");
}
/** /**
* Creates a URL and opens a connection. * Creates a URL and opens a connection.
* *
* If the URL returns a 404, {@code null} is returned, otherwise the open URL opject is returned. * If the URL returns a 404, {@code null} is returned, otherwise the open URL opject is returned.
*/ */
// pkg private for tests // pkg private for tests
URL openUrl(String urlString) throws Exception { URL openUrl(String urlString) throws IOException {
URL checksumUrl = new URL(urlString); URL checksumUrl = new URL(urlString);
HttpURLConnection connection = (HttpURLConnection)checksumUrl.openConnection(); HttpURLConnection connection = (HttpURLConnection)checksumUrl.openConnection();
if (connection.getResponseCode() == 404) { if (connection.getResponseCode() == 404) {
@ -605,11 +742,27 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
return info; return info;
} }
private static final String LIB_TOOLS_PLUGIN_CLI_CLASSPATH_JAR;
static {
LIB_TOOLS_PLUGIN_CLI_CLASSPATH_JAR =
String.format(Locale.ROOT, ".+%1$slib%1$stools%1$splugin-cli%1$s[^%1$s]+\\.jar", "(/|\\\\)");
}
/** check a candidate plugin for jar hell before installing it */ /** check a candidate plugin for jar hell before installing it */
void jarHellCheck(PluginInfo candidateInfo, Path candidateDir, Path pluginsDir, Path modulesDir) throws Exception { void jarHellCheck(PluginInfo candidateInfo, Path candidateDir, Path pluginsDir, Path modulesDir) throws Exception {
// create list of current jars in classpath // create list of current jars in classpath
final Set<URL> jars = new HashSet<>(JarHell.parseClassPath()); final Set<URL> classpath =
JarHell.parseClassPath()
.stream()
.filter(url -> {
try {
return url.toURI().getPath().matches(LIB_TOOLS_PLUGIN_CLI_CLASSPATH_JAR) == false;
} catch (final URISyntaxException e) {
throw new AssertionError(e);
}
})
.collect(Collectors.toSet());
// read existing bundles. this does some checks on the installation too. // read existing bundles. this does some checks on the installation too.
Set<PluginsService.Bundle> bundles = new HashSet<>(PluginsService.getPluginBundles(pluginsDir)); Set<PluginsService.Bundle> bundles = new HashSet<>(PluginsService.getPluginBundles(pluginsDir));
@ -621,7 +774,7 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
// TODO: optimize to skip any bundles not connected to the candidate plugin? // TODO: optimize to skip any bundles not connected to the candidate plugin?
Map<String, Set<URL>> transitiveUrls = new HashMap<>(); Map<String, Set<URL>> transitiveUrls = new HashMap<>();
for (PluginsService.Bundle bundle : sortedBundles) { for (PluginsService.Bundle bundle : sortedBundles) {
PluginsService.checkBundleJarHell(bundle, transitiveUrls); PluginsService.checkBundleJarHell(classpath, bundle, transitiveUrls);
} }
// TODO: no jars should be an error // TODO: no jars should be an error

View File

@ -0,0 +1,29 @@
-----BEGIN PGP PUBLIC KEY BLOCK-----
Version: SKS 1.1.6
Comment: Hostname: pgp.mit.edu
mQENBFI3HsoBCADXDtbNJnxbPqB1vDNtCsqhe49vFYsZN9IOZsZXgp7aHjh6CJBDA+bGFOwy
hbd7at35jQjWAw1O3cfYsKAmFy+Ar3LHCMkV3oZspJACTIgCrwnkic/9CUliQe324qvObU2Q
RtP4Fl0zWcfb/S8UYzWXWIFuJqMvE9MaRY1bwUBvzoqavLGZj3SF1SPO+TB5QrHkrQHBsmX+
Jda6d4Ylt8/t6CvMwgQNlrlzIO9WT+YN6zS+sqHd1YK/aY5qhoLNhp9G/HxhcSVCkLq8SStj
1ZZ1S9juBPoXV1ZWNbxFNGwOh/NYGldD2kmBf3YgCqeLzHahsAEpvAm8TBa7Q9W21C8vABEB
AAG0RUVsYXN0aWNzZWFyY2ggKEVsYXN0aWNzZWFyY2ggU2lnbmluZyBLZXkpIDxkZXZfb3Bz
QGVsYXN0aWNzZWFyY2gub3JnPokBOAQTAQIAIgUCUjceygIbAwYLCQgHAwIGFQgCCQoLBBYC
AwECHgECF4AACgkQ0n1mbNiOQrRzjAgAlTUQ1mgo3nK6BGXbj4XAJvuZDG0HILiUt+pPnz75
nsf0NWhqR4yGFlmpuctgCmTD+HzYtV9fp9qW/bwVuJCNtKXk3sdzYABY+Yl0Cez/7C2GuGCO
lbn0luCNT9BxJnh4mC9h/cKI3y5jvZ7wavwe41teqG14V+EoFSn3NPKmTxcDTFrV7SmVPxCB
cQze00cJhprKxkuZMPPVqpBS+JfDQtzUQD/LSFfhHj9eD+Xe8d7sw+XvxB2aN4gnTlRzjL1n
TRp0h2/IOGkqYfIG9rWmSLNlxhB2t+c0RsjdGM4/eRlPWylFbVMc5pmDpItrkWSnzBfkmXL3
vO2X3WvwmSFiQbkBDQRSNx7KAQgA5JUlzcMW5/cuyZR8alSacKqhSbvoSqqbzHKcUQZmlzNM
KGTABFG1yRx9r+wa/fvqP6OTRzRDvVS/cycws8YX7Ddum7x8uI95b9ye1/Xy5noPEm8cD+hp
lnpU+PBQZJ5XJ2I+1l9Nixx47wPGXeClLqcdn0ayd+v+Rwf3/XUJrvccG2YZUiQ4jWZkoxsA
07xx7Bj+Lt8/FKG7sHRFvePFU0ZS6JFx9GJqjSBbHRRkam+4emW3uWgVfZxuwcUCn1ayNgRt
KiFv9jQrg2TIWEvzYx9tywTCxc+FFMWAlbCzi+m4WD+QUWWfDQ009U/WM0ks0KwwEwSk/UDu
ToxGnKU2dQARAQABiQEfBBgBAgAJBQJSNx7KAhsMAAoJENJ9ZmzYjkK0c3MIAIE9hAR20mqJ
WLcsxLtrRs6uNF1VrpB+4n/55QU7oxA1iVBO6IFu4qgsF12JTavnJ5MLaETlggXY+zDef9sy
TPXoQctpzcaNVDmedwo1SiL03uMoblOvWpMR/Y0j6rm7IgrMWUDXDPvoPGjMl2q1iTeyHkMZ
EyUJ8SKsaHh4jV9wp9KmC8C+9CwMukL7vM5w8cgvJoAwsp3Fn59AxWthN3XJYcnMfStkIuWg
R7U2r+a210W6vnUxU4oN0PmMcursYPyeV0NX/KQeUeNMwGTFB6QHS/anRaGQewijkrYYoTNt
fllxIu9XYmiBERQ/qPDlGRlOgVTd9xUfHFkzB52c70E=
=92oX
-----END PGP PUBLIC KEY BLOCK-----

View File

@ -23,6 +23,26 @@ import com.carrotsearch.randomizedtesting.annotations.ParametersFactory;
import com.google.common.jimfs.Configuration; import com.google.common.jimfs.Configuration;
import com.google.common.jimfs.Jimfs; import com.google.common.jimfs.Jimfs;
import org.apache.lucene.util.LuceneTestCase; import org.apache.lucene.util.LuceneTestCase;
import org.apache.lucene.util.LuceneTestCase.AwaitsFix;
import org.bouncycastle.bcpg.ArmoredOutputStream;
import org.bouncycastle.bcpg.BCPGOutputStream;
import org.bouncycastle.bcpg.HashAlgorithmTags;
import org.bouncycastle.jce.provider.BouncyCastleProvider;
import org.bouncycastle.openpgp.PGPEncryptedData;
import org.bouncycastle.openpgp.PGPException;
import org.bouncycastle.openpgp.PGPKeyPair;
import org.bouncycastle.openpgp.PGPPrivateKey;
import org.bouncycastle.openpgp.PGPPublicKey;
import org.bouncycastle.openpgp.PGPSecretKey;
import org.bouncycastle.openpgp.PGPSignature;
import org.bouncycastle.openpgp.PGPSignatureGenerator;
import org.bouncycastle.openpgp.operator.PGPDigestCalculator;
import org.bouncycastle.openpgp.operator.bc.BcPBESecretKeyDecryptorBuilder;
import org.bouncycastle.openpgp.operator.bc.BcPGPContentSignerBuilder;
import org.bouncycastle.openpgp.operator.jcajce.JcaPGPContentSignerBuilder;
import org.bouncycastle.openpgp.operator.jcajce.JcaPGPDigestCalculatorProviderBuilder;
import org.bouncycastle.openpgp.operator.jcajce.JcaPGPKeyPair;
import org.bouncycastle.openpgp.operator.jcajce.JcePBESecretKeyEncryptorBuilder;
import org.elasticsearch.Build; import org.elasticsearch.Build;
import org.elasticsearch.Version; import org.elasticsearch.Version;
import org.elasticsearch.cli.ExitCodes; import org.elasticsearch.cli.ExitCodes;
@ -44,6 +64,8 @@ import org.junit.After;
import org.junit.Before; import org.junit.Before;
import java.io.BufferedReader; import java.io.BufferedReader;
import java.io.ByteArrayInputStream;
import java.io.ByteArrayOutputStream;
import java.io.IOException; import java.io.IOException;
import java.io.InputStream; import java.io.InputStream;
import java.io.StringReader; import java.io.StringReader;
@ -66,13 +88,19 @@ import java.nio.file.attribute.PosixFileAttributeView;
import java.nio.file.attribute.PosixFileAttributes; import java.nio.file.attribute.PosixFileAttributes;
import java.nio.file.attribute.PosixFilePermission; import java.nio.file.attribute.PosixFilePermission;
import java.nio.file.attribute.UserPrincipal; import java.nio.file.attribute.UserPrincipal;
import java.security.KeyPair;
import java.security.KeyPairGenerator;
import java.security.MessageDigest; import java.security.MessageDigest;
import java.security.NoSuchAlgorithmException;
import java.security.NoSuchProviderException;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Arrays; import java.util.Arrays;
import java.util.Date;
import java.util.HashSet; import java.util.HashSet;
import java.util.List; import java.util.List;
import java.util.Locale; import java.util.Locale;
import java.util.Set; import java.util.Set;
import java.util.function.BiFunction;
import java.util.function.Function; import java.util.function.Function;
import java.util.stream.Collectors; import java.util.stream.Collectors;
import java.util.stream.Stream; import java.util.stream.Stream;
@ -88,6 +116,7 @@ import static org.hamcrest.Matchers.hasToString;
import static org.hamcrest.Matchers.not; import static org.hamcrest.Matchers.not;
@LuceneTestCase.SuppressFileSystems("*") @LuceneTestCase.SuppressFileSystems("*")
@AwaitsFix(bugUrl = "https://github.com/elastic/elasticsearch/issues/30900")
public class InstallPluginCommandTests extends ESTestCase { public class InstallPluginCommandTests extends ESTestCase {
private InstallPluginCommand skipJarHellCommand; private InstallPluginCommand skipJarHellCommand;
@ -800,8 +829,16 @@ public class InstallPluginCommandTests extends ESTestCase {
skipJarHellCommand.execute(terminal, pluginZip, isBatch, env.v2()); skipJarHellCommand.execute(terminal, pluginZip, isBatch, env.v2());
} }
void assertInstallPluginFromUrl(String pluginId, String name, String url, String stagingHash, boolean isSnapshot, void assertInstallPluginFromUrl(
String shaExtension, Function<byte[], String> shaCalculator) throws Exception { final String pluginId,
final String name,
final String url,
final String stagingHash,
final boolean isSnapshot,
final String shaExtension,
final Function<byte[], String> shaCalculator,
final PGPSecretKey secretKey,
final BiFunction<byte[], PGPSecretKey, String> signature) throws Exception {
Tuple<Path, Environment> env = createEnv(fs, temp); Tuple<Path, Environment> env = createEnv(fs, temp);
Path pluginDir = createPluginDir(temp); Path pluginDir = createPluginDir(temp);
Path pluginZip = createPlugin(name, pluginDir); Path pluginZip = createPlugin(name, pluginDir);
@ -814,18 +851,56 @@ public class InstallPluginCommandTests extends ESTestCase {
return downloadedPath; return downloadedPath;
} }
@Override @Override
URL openUrl(String urlString) throws Exception { URL openUrl(String urlString) throws IOException {
String expectedUrl = url + shaExtension; if ((url + shaExtension).equals(urlString)) {
if (expectedUrl.equals(urlString)) {
// calc sha an return file URL to it // calc sha an return file URL to it
Path shaFile = temp.apply("shas").resolve("downloaded.zip" + shaExtension); Path shaFile = temp.apply("shas").resolve("downloaded.zip" + shaExtension);
byte[] zipbytes = Files.readAllBytes(pluginZip); byte[] zipbytes = Files.readAllBytes(pluginZip);
String checksum = shaCalculator.apply(zipbytes); String checksum = shaCalculator.apply(zipbytes);
Files.write(shaFile, checksum.getBytes(StandardCharsets.UTF_8)); Files.write(shaFile, checksum.getBytes(StandardCharsets.UTF_8));
return shaFile.toUri().toURL(); return shaFile.toUri().toURL();
} else if ((url + ".asc").equals(urlString)) {
final Path ascFile = temp.apply("asc").resolve("downloaded.zip" + ".asc");
final byte[] zipBytes = Files.readAllBytes(pluginZip);
final String asc = signature.apply(zipBytes, secretKey);
Files.write(ascFile, asc.getBytes(StandardCharsets.UTF_8));
return ascFile.toUri().toURL();
} }
return null; return null;
} }
@Override
void verifySignature(Path zip, String urlString) throws IOException, PGPException {
if (InstallPluginCommand.OFFICIAL_PLUGINS.contains(name)) {
super.verifySignature(zip, urlString);
} else {
throw new UnsupportedOperationException("verify signature should not be called for unofficial plugins");
}
}
@Override
InputStream pluginZipInputStream(Path zip) throws IOException {
return new ByteArrayInputStream(Files.readAllBytes(zip));
}
@Override
String getPublicKeyId() {
return Long.toHexString(secretKey.getKeyID()).toUpperCase(Locale.ROOT);
}
@Override
InputStream getPublicKey() {
try {
final ByteArrayOutputStream output = new ByteArrayOutputStream();
final ArmoredOutputStream armored = new ArmoredOutputStream(output);
secretKey.getPublicKey().encode(armored);
armored.close();
return new ByteArrayInputStream(output.toByteArray());
} catch (final IOException e) {
throw new AssertionError(e);
}
}
@Override @Override
boolean urlExists(Terminal terminal, String urlString) throws IOException { boolean urlExists(Terminal terminal, String urlString) throws IOException {
return urlString.equals(url); return urlString.equals(url);
@ -851,11 +926,12 @@ public class InstallPluginCommandTests extends ESTestCase {
public void assertInstallPluginFromUrl( public void assertInstallPluginFromUrl(
final String pluginId, final String name, final String url, final String stagingHash, boolean isSnapshot) throws Exception { final String pluginId, final String name, final String url, final String stagingHash, boolean isSnapshot) throws Exception {
MessageDigest digest = MessageDigest.getInstance("SHA-512"); final MessageDigest digest = MessageDigest.getInstance("SHA-512");
assertInstallPluginFromUrl(pluginId, name, url, stagingHash, isSnapshot, ".sha512", checksumAndFilename(digest, url)); assertInstallPluginFromUrl(
pluginId, name, url, stagingHash, isSnapshot, ".sha512", checksumAndFilename(digest, url), newSecretKey(), this::signature);
} }
public void testOfficalPlugin() throws Exception { public void testOfficialPlugin() throws Exception {
String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Version.CURRENT + ".zip"; String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Version.CURRENT + ".zip";
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false); assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false);
} }
@ -883,13 +959,13 @@ public class InstallPluginCommandTests extends ESTestCase {
e, hasToString(containsString("attempted to install release build of official plugin on snapshot build of Elasticsearch"))); e, hasToString(containsString("attempted to install release build of official plugin on snapshot build of Elasticsearch")));
} }
public void testOfficalPluginStaging() throws Exception { public void testOfficialPluginStaging() throws Exception {
String url = "https://staging.elastic.co/" + Version.CURRENT + "-abc123/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" String url = "https://staging.elastic.co/" + Version.CURRENT + "-abc123/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-"
+ Version.CURRENT + ".zip"; + Version.CURRENT + ".zip";
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, "abc123", false); assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, "abc123", false);
} }
public void testOfficalPlatformPlugin() throws Exception { public void testOfficialPlatformPlugin() throws Exception {
String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Platforms.PLATFORM_NAME + String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Platforms.PLATFORM_NAME +
"-" + Version.CURRENT + ".zip"; "-" + Version.CURRENT + ".zip";
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false); assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false);
@ -905,7 +981,7 @@ public class InstallPluginCommandTests extends ESTestCase {
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, "abc123", true); assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, "abc123", true);
} }
public void testOfficalPlatformPluginStaging() throws Exception { public void testOfficialPlatformPluginStaging() throws Exception {
String url = "https://staging.elastic.co/" + Version.CURRENT + "-abc123/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" String url = "https://staging.elastic.co/" + Version.CURRENT + "-abc123/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-"
+ Platforms.PLATFORM_NAME + "-"+ Version.CURRENT + ".zip"; + Platforms.PLATFORM_NAME + "-"+ Version.CURRENT + ".zip";
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, "abc123", false); assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, "abc123", false);
@ -924,7 +1000,7 @@ public class InstallPluginCommandTests extends ESTestCase {
public void testMavenSha1Backcompat() throws Exception { public void testMavenSha1Backcompat() throws Exception {
String url = "https://repo1.maven.org/maven2/mygroup/myplugin/1.0.0/myplugin-1.0.0.zip"; String url = "https://repo1.maven.org/maven2/mygroup/myplugin/1.0.0/myplugin-1.0.0.zip";
MessageDigest digest = MessageDigest.getInstance("SHA-1"); MessageDigest digest = MessageDigest.getInstance("SHA-1");
assertInstallPluginFromUrl("mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".sha1", checksum(digest)); assertInstallPluginFromUrl("mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".sha1", checksum(digest), null, (b, p) -> null);
assertTrue(terminal.getOutput(), terminal.getOutput().contains("sha512 not found, falling back to sha1")); assertTrue(terminal.getOutput(), terminal.getOutput().contains("sha512 not found, falling back to sha1"));
} }
@ -932,7 +1008,7 @@ public class InstallPluginCommandTests extends ESTestCase {
String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Version.CURRENT + ".zip"; String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Version.CURRENT + ".zip";
MessageDigest digest = MessageDigest.getInstance("SHA-1"); MessageDigest digest = MessageDigest.getInstance("SHA-1");
UserException e = expectThrows(UserException.class, () -> UserException e = expectThrows(UserException.class, () ->
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false, ".sha1", checksum(digest))); assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false, ".sha1", checksum(digest), null, (b, p) -> null));
assertEquals(ExitCodes.IO_ERROR, e.exitCode); assertEquals(ExitCodes.IO_ERROR, e.exitCode);
assertEquals("Plugin checksum missing: " + url + ".sha512", e.getMessage()); assertEquals("Plugin checksum missing: " + url + ".sha512", e.getMessage());
} }
@ -940,7 +1016,8 @@ public class InstallPluginCommandTests extends ESTestCase {
public void testMavenShaMissing() throws Exception { public void testMavenShaMissing() throws Exception {
String url = "https://repo1.maven.org/maven2/mygroup/myplugin/1.0.0/myplugin-1.0.0.zip"; String url = "https://repo1.maven.org/maven2/mygroup/myplugin/1.0.0/myplugin-1.0.0.zip";
UserException e = expectThrows(UserException.class, () -> UserException e = expectThrows(UserException.class, () ->
assertInstallPluginFromUrl("mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".dne", bytes -> null)); assertInstallPluginFromUrl(
"mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".dne", bytes -> null, null, (b, p) -> null));
assertEquals(ExitCodes.IO_ERROR, e.exitCode); assertEquals(ExitCodes.IO_ERROR, e.exitCode);
assertEquals("Plugin checksum missing: " + url + ".sha1", e.getMessage()); assertEquals("Plugin checksum missing: " + url + ".sha1", e.getMessage());
} }
@ -948,8 +1025,9 @@ public class InstallPluginCommandTests extends ESTestCase {
public void testInvalidShaFileMissingFilename() throws Exception { public void testInvalidShaFileMissingFilename() throws Exception {
String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Version.CURRENT + ".zip"; String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Version.CURRENT + ".zip";
MessageDigest digest = MessageDigest.getInstance("SHA-512"); MessageDigest digest = MessageDigest.getInstance("SHA-512");
UserException e = expectThrows(UserException.class, () -> UserException e = expectThrows(UserException.class,
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false, ".sha512", checksum(digest))); () -> assertInstallPluginFromUrl(
"analysis-icu", "analysis-icu", url, null, false, ".sha512", checksum(digest), null, (b, p) -> null));
assertEquals(ExitCodes.IO_ERROR, e.exitCode); assertEquals(ExitCodes.IO_ERROR, e.exitCode);
assertTrue(e.getMessage(), e.getMessage().startsWith("Invalid checksum file")); assertTrue(e.getMessage(), e.getMessage().startsWith("Invalid checksum file"));
} }
@ -965,7 +1043,9 @@ public class InstallPluginCommandTests extends ESTestCase {
null, null,
false, false,
".sha512", ".sha512",
checksumAndString(digest, " repository-s3-" + Version.CURRENT + ".zip"))); checksumAndString(digest, " repository-s3-" + Version.CURRENT + ".zip"),
null,
(b, p) -> null));
assertEquals(ExitCodes.IO_ERROR, e.exitCode); assertEquals(ExitCodes.IO_ERROR, e.exitCode);
assertThat(e, hasToString(matches("checksum file at \\[.*\\] is not for this plugin"))); assertThat(e, hasToString(matches("checksum file at \\[.*\\] is not for this plugin")));
} }
@ -981,7 +1061,9 @@ public class InstallPluginCommandTests extends ESTestCase {
null, null,
false, false,
".sha512", ".sha512",
checksumAndString(digest, " analysis-icu-" + Version.CURRENT + ".zip\nfoobar"))); checksumAndString(digest, " analysis-icu-" + Version.CURRENT + ".zip\nfoobar"),
null,
(b, p) -> null));
assertEquals(ExitCodes.IO_ERROR, e.exitCode); assertEquals(ExitCodes.IO_ERROR, e.exitCode);
assertTrue(e.getMessage(), e.getMessage().startsWith("Invalid checksum file")); assertTrue(e.getMessage(), e.getMessage().startsWith("Invalid checksum file"));
} }
@ -996,7 +1078,9 @@ public class InstallPluginCommandTests extends ESTestCase {
null, null,
false, false,
".sha512", ".sha512",
bytes -> "foobar analysis-icu-" + Version.CURRENT + ".zip")); bytes -> "foobar analysis-icu-" + Version.CURRENT + ".zip",
null,
(b, p) -> null));
assertEquals(ExitCodes.IO_ERROR, e.exitCode); assertEquals(ExitCodes.IO_ERROR, e.exitCode);
assertTrue(e.getMessage(), e.getMessage().contains("SHA-512 mismatch, expected foobar")); assertTrue(e.getMessage(), e.getMessage().contains("SHA-512 mismatch, expected foobar"));
} }
@ -1004,11 +1088,77 @@ public class InstallPluginCommandTests extends ESTestCase {
public void testSha1Mismatch() throws Exception { public void testSha1Mismatch() throws Exception {
String url = "https://repo1.maven.org/maven2/mygroup/myplugin/1.0.0/myplugin-1.0.0.zip"; String url = "https://repo1.maven.org/maven2/mygroup/myplugin/1.0.0/myplugin-1.0.0.zip";
UserException e = expectThrows(UserException.class, () -> UserException e = expectThrows(UserException.class, () ->
assertInstallPluginFromUrl("mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".sha1", bytes -> "foobar")); assertInstallPluginFromUrl(
"mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".sha1", bytes -> "foobar", null, (b, p) -> null));
assertEquals(ExitCodes.IO_ERROR, e.exitCode); assertEquals(ExitCodes.IO_ERROR, e.exitCode);
assertTrue(e.getMessage(), e.getMessage().contains("SHA-1 mismatch, expected foobar")); assertTrue(e.getMessage(), e.getMessage().contains("SHA-1 mismatch, expected foobar"));
} }
public void testPublicKeyIdMismatchToExpectedPublicKeyId() throws Exception {
final String icu = "analysis-icu";
final String url =
"https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/" + icu + "-" + Version.CURRENT + ".zip";
final MessageDigest digest = MessageDigest.getInstance("SHA-512");
/*
* To setup a situation where the expected public key ID does not match the public key ID used for signing, we generate a new public
* key at the moment of signing (see the signature invocation). Note that this key will not match the key that we push down to the
* install plugin command.
*/
final PGPSecretKey signingKey = newSecretKey(); // the actual key used for signing
final String actualID = Long.toHexString(signingKey.getKeyID()).toUpperCase(Locale.ROOT);
final BiFunction<byte[], PGPSecretKey, String> signature = (b, p) -> signature(b, signingKey);
final PGPSecretKey verifyingKey = newSecretKey(); // the expected key used for signing
final String expectedID = Long.toHexString(verifyingKey.getKeyID()).toUpperCase(Locale.ROOT);
final IllegalStateException e = expectThrows(
IllegalStateException.class,
() ->
assertInstallPluginFromUrl(
icu, icu, url, null, false, ".sha512", checksumAndFilename(digest, url), verifyingKey, signature));
assertThat(e, hasToString(containsString("key id [" + actualID + "] does not match expected key id [" + expectedID + "]")));
}
public void testFailedSignatureVerification() throws Exception {
final String icu = "analysis-icu";
final String url =
"https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/" + icu + "-" + Version.CURRENT + ".zip";
final MessageDigest digest = MessageDigest.getInstance("SHA-512");
/*
* To setup a situation where signature verification fails, we will mutate the input byte array by modifying a single byte to some
* random byte value other than the actual value. This is enough to change the signature and cause verification to intentionally
* fail.
*/
final BiFunction<byte[], PGPSecretKey, String> signature = (b, p) -> {
final byte[] bytes = Arrays.copyOf(b, b.length);
bytes[0] = randomValueOtherThan(b[0], ESTestCase::randomByte);
return signature(bytes, p);
};
final IllegalStateException e = expectThrows(
IllegalStateException.class,
() ->
assertInstallPluginFromUrl(
icu, icu, url, null, false, ".sha512", checksumAndFilename(digest, url), newSecretKey(), signature));
assertThat(e, hasToString(equalTo("java.lang.IllegalStateException: signature verification for [" + url + "] failed")));
}
public PGPSecretKey newSecretKey() throws NoSuchAlgorithmException, NoSuchProviderException, PGPException {
final KeyPairGenerator kpg = KeyPairGenerator.getInstance("RSA");
kpg.initialize(2048);
final KeyPair pair = kpg.generateKeyPair();
final PGPDigestCalculator sha1Calc = new JcaPGPDigestCalculatorProviderBuilder().build().get(HashAlgorithmTags.SHA1);
final PGPKeyPair pkp = new JcaPGPKeyPair(PGPPublicKey.RSA_GENERAL, pair, new Date());
return new PGPSecretKey(
PGPSignature.DEFAULT_CERTIFICATION,
pkp,
"example@example.com",
sha1Calc,
null,
null,
new JcaPGPContentSignerBuilder(pkp.getPublicKey().getAlgorithm(), HashAlgorithmTags.SHA1),
new JcePBESecretKeyEncryptorBuilder(PGPEncryptedData.CAST5, sha1Calc)
.setProvider(new BouncyCastleProvider())
.build("passphrase".toCharArray()));
}
private Function<byte[], String> checksum(final MessageDigest digest) { private Function<byte[], String> checksum(final MessageDigest digest) {
return checksumAndString(digest, ""); return checksumAndString(digest, "");
} }
@ -1022,6 +1172,32 @@ public class InstallPluginCommandTests extends ESTestCase {
return bytes -> MessageDigests.toHexString(digest.digest(bytes)) + s; return bytes -> MessageDigests.toHexString(digest.digest(bytes)) + s;
} }
private String signature(final byte[] bytes, final PGPSecretKey secretKey) {
try {
final PGPPrivateKey privateKey
= secretKey.extractPrivateKey(
new BcPBESecretKeyDecryptorBuilder(
new JcaPGPDigestCalculatorProviderBuilder().build()).build("passphrase".toCharArray()));
final PGPSignatureGenerator generator =
new PGPSignatureGenerator(
new BcPGPContentSignerBuilder(privateKey.getPublicKeyPacket().getAlgorithm(), HashAlgorithmTags.SHA512));
generator.init(PGPSignature.BINARY_DOCUMENT, privateKey);
final ByteArrayOutputStream output = new ByteArrayOutputStream();
try (BCPGOutputStream pout = new BCPGOutputStream(new ArmoredOutputStream(output));
InputStream is = new ByteArrayInputStream(bytes)) {
final byte[] buffer = new byte[1024];
int read;
while ((read = is.read(buffer)) != -1) {
generator.update(buffer, 0, read);
}
generator.generate().encode(pout);
}
return new String(output.toByteArray(), "UTF-8");
} catch (IOException | PGPException e) {
throw new RuntimeException(e);
}
}
// checks the plugin requires a policy confirmation, and does not install when that is rejected by the user // checks the plugin requires a policy confirmation, and does not install when that is rejected by the user
// the plugin is installed after this method completes // the plugin is installed after this method completes
private void assertPolicyConfirmation(Tuple<Path, Environment> env, String pluginZip, String... warnings) throws Exception { private void assertPolicyConfirmation(Tuple<Path, Environment> env, String pluginZip, String... warnings) throws Exception {

View File

@ -432,7 +432,15 @@ Remote hosts have to be explicitly whitelisted in elasticsearch.yaml using the
`reindex.remote.whitelist` property. It can be set to a comma delimited list `reindex.remote.whitelist` property. It can be set to a comma delimited list
of allowed remote `host` and `port` combinations (e.g. of allowed remote `host` and `port` combinations (e.g.
`otherhost:9200, another:9200, 127.0.10.*:9200, localhost:*`). Scheme is `otherhost:9200, another:9200, 127.0.10.*:9200, localhost:*`). Scheme is
ignored by the whitelist - only host and port are used. ignored by the whitelist - only host and port are used, for example:
[source,yaml]
--------------------------------------------------
reindex.remote.whitelist: "otherhost:9200, another:9200, 127.0.10.*:9200, localhost:*"
--------------------------------------------------
The whitelist must be configured on any nodes that will coordinate the reindex.
This feature should work with remote clusters of any version of Elasticsearch This feature should work with remote clusters of any version of Elasticsearch
you are likely to find. This should allow you to upgrade from any version of you are likely to find. This should allow you to upgrade from any version of

View File

@ -1,16 +1,39 @@
[[indices-create-index]] [[indices-create-index]]
== Create Index == Create Index
The create index API allows to instantiate an index. Elasticsearch The Create Index API is used to manually create an index in Elasticsearch. All documents in Elasticsearch
provides support for multiple indices, including executing operations are stored inside of one index or another.
across several indices.
The most basic command is the following:
[source,js]
--------------------------------------------------
PUT twitter
--------------------------------------------------
// CONSOLE
This create an index named `twitter` with all default setting.
[NOTE]
.Index name limitations
======================================================
There are several limitations to what you can name your index. The complete list of limitations are:
- Lowercase only
- Cannot include `\`, `/`, `*`, `?`, `"`, `<`, `>`, `|`, ` ` (space character), `,`, `#`
- Indices prior to 7.0 could contain a colon (`:`), but that's been deprecated and won't be supported in 7.0+
- Cannot start with `-`, `_`, `+`
- Cannot be `.` or ``..`
- Cannot be longer than 255 bytes (note it is bytes, so multi-byte characters will count towards the 255 limit faster)
======================================================
[float] [float]
[[create-index-settings]] [[create-index-settings]]
=== Index Settings === Index Settings
Each index created can have specific settings Each index created can have specific settings
associated with it. associated with it, defined in the body:
[source,js] [source,js]
-------------------------------------------------- --------------------------------------------------
@ -28,25 +51,6 @@ PUT twitter
<1> Default for `number_of_shards` is 1 <1> Default for `number_of_shards` is 1
<2> Default for `number_of_replicas` is 1 (ie one replica for each primary shard) <2> Default for `number_of_replicas` is 1 (ie one replica for each primary shard)
The above second curl example shows how an index called `twitter` can be
created with specific settings for it using http://www.yaml.org[YAML].
In this case, creating an index with 3 shards, each with 2 replicas. The
index settings can also be defined with http://www.json.org[JSON]:
[source,js]
--------------------------------------------------
PUT twitter
{
"settings" : {
"index" : {
"number_of_shards" : 3,
"number_of_replicas" : 2
}
}
}
--------------------------------------------------
// CONSOLE
or more simplified or more simplified
[source,js] [source,js]

View File

@ -35,7 +35,7 @@ Elasticsearch 6.x in order to be readable by Elasticsearch 7.x.
* <<breaking_70_java_changes>> * <<breaking_70_java_changes>>
* <<breaking_70_settings_changes>> * <<breaking_70_settings_changes>>
* <<breaking_70_scripting_changes>> * <<breaking_70_scripting_changes>>
* <<breaking_70_snapshotstats_changes>>
include::migrate_7_0/aggregations.asciidoc[] include::migrate_7_0/aggregations.asciidoc[]
include::migrate_7_0/analysis.asciidoc[] include::migrate_7_0/analysis.asciidoc[]
@ -49,3 +49,4 @@ include::migrate_7_0/api.asciidoc[]
include::migrate_7_0/java.asciidoc[] include::migrate_7_0/java.asciidoc[]
include::migrate_7_0/settings.asciidoc[] include::migrate_7_0/settings.asciidoc[]
include::migrate_7_0/scripting.asciidoc[] include::migrate_7_0/scripting.asciidoc[]
include::migrate_7_0/snapshotstats.asciidoc[]

View File

@ -0,0 +1,13 @@
[[breaking_70_snapshotstats_changes]]
=== Snapshot stats changes
Snapshot stats details are provided in a new structured way:
* `total` section for all the files that are referenced by the snapshot.
* `incremental` section for those files that actually needed to be copied over as part of the incremental snapshotting.
* In case of a snapshot that's still in progress, there's also a `processed` section for files that are in the process of being copied.
==== Deprecated `number_of_files`, `processed_files`, `total_size_in_bytes` and `processed_size_in_bytes` snapshot stats properties have been removed
* Properties `number_of_files` and `total_size_in_bytes` are removed and should be replaced by values of nested object `total`.
* Properties `processed_files` and `processed_size_in_bytes` are removed and should be replaced by values of nested object `processed`.

View File

@ -563,6 +563,54 @@ GET /_snapshot/my_backup/snapshot_1/_status
// CONSOLE // CONSOLE
// TEST[continued] // TEST[continued]
The output looks similar to the following:
[source,js]
--------------------------------------------------
{
"snapshots": [
{
"snapshot": "snapshot_1",
"repository": "my_backup",
"uuid": "XuBo4l4ISYiVg0nYUen9zg",
"state": "SUCCESS",
"include_global_state": true,
"shards_stats": {
"initializing": 0,
"started": 0,
"finalizing": 0,
"done": 5,
"failed": 0,
"total": 5
},
"stats": {
"incremental": {
"file_count": 8,
"size_in_bytes": 4704
},
"processed": {
"file_count": 7,
"size_in_bytes": 4254
},
"total": {
"file_count": 8,
"size_in_bytes": 4704
},
"start_time_in_millis": 1526280280355,
"time_in_millis": 358
}
}
]
}
--------------------------------------------------
// TESTRESPONSE
The output is composed of different sections. The `stats` sub-object provides details on the number and size of files that were
snapshotted. As snapshots are incremental, copying only the Lucene segments that are not already in the repository,
the `stats` object contains a `total` section for all the files that are referenced by the snapshot, as well as an `incremental` section
for those files that actually needed to be copied over as part of the incremental snapshotting. In case of a snapshot that's still
in progress, there's also a `processed` section that contains information about the files that are in the process of being copied.
Multiple ids are also supported: Multiple ids are also supported:
[source,sh] [source,sh]

View File

@ -1,8 +1,10 @@
When you shut down a node, the allocation process waits for one minute When you shut down a node, the allocation process waits for
before starting to replicate the shards on that node to other nodes `index.unassigned.node_left.delayed_timeout` (by default, one minute) before
in the cluster, causing a lot of wasted I/O. You can avoid racing the clock starting to replicate the shards on that node to other nodes in the cluster,
by disabling allocation before shutting down the node: which can involve a lot of I/O. Since the node is shortly going to be
restarted, this I/O is unnecessary. You can avoid racing the clock by disabling
allocation before shutting down the node:
[source,js] [source,js]
-------------------------------------------------- --------------------------------------------------

View File

@ -23,8 +23,10 @@ import org.apache.lucene.expressions.Expression;
import org.apache.lucene.expressions.SimpleBindings; import org.apache.lucene.expressions.SimpleBindings;
import org.apache.lucene.expressions.js.JavascriptCompiler; import org.apache.lucene.expressions.js.JavascriptCompiler;
import org.apache.lucene.expressions.js.VariableContext; import org.apache.lucene.expressions.js.VariableContext;
import org.apache.lucene.index.LeafReaderContext;
import org.apache.lucene.queries.function.ValueSource; import org.apache.lucene.queries.function.ValueSource;
import org.apache.lucene.queries.function.valuesource.DoubleConstValueSource; import org.apache.lucene.queries.function.valuesource.DoubleConstValueSource;
import org.apache.lucene.search.Scorer;
import org.apache.lucene.search.SortField; import org.apache.lucene.search.SortField;
import org.elasticsearch.SpecialPermission; import org.elasticsearch.SpecialPermission;
import org.elasticsearch.common.Nullable; import org.elasticsearch.common.Nullable;
@ -39,12 +41,14 @@ import org.elasticsearch.index.mapper.MapperService;
import org.elasticsearch.script.ClassPermission; import org.elasticsearch.script.ClassPermission;
import org.elasticsearch.script.ExecutableScript; import org.elasticsearch.script.ExecutableScript;
import org.elasticsearch.script.FilterScript; import org.elasticsearch.script.FilterScript;
import org.elasticsearch.script.ScoreScript;
import org.elasticsearch.script.ScriptContext; import org.elasticsearch.script.ScriptContext;
import org.elasticsearch.script.ScriptEngine; import org.elasticsearch.script.ScriptEngine;
import org.elasticsearch.script.ScriptException; import org.elasticsearch.script.ScriptException;
import org.elasticsearch.script.SearchScript; import org.elasticsearch.script.SearchScript;
import org.elasticsearch.search.lookup.SearchLookup; import org.elasticsearch.search.lookup.SearchLookup;
import java.io.IOException;
import java.security.AccessControlContext; import java.security.AccessControlContext;
import java.security.AccessController; import java.security.AccessController;
import java.security.PrivilegedAction; import java.security.PrivilegedAction;
@ -111,6 +115,9 @@ public class ExpressionScriptEngine extends AbstractComponent implements ScriptE
} else if (context.instanceClazz.equals(FilterScript.class)) { } else if (context.instanceClazz.equals(FilterScript.class)) {
FilterScript.Factory factory = (p, lookup) -> newFilterScript(expr, lookup, p); FilterScript.Factory factory = (p, lookup) -> newFilterScript(expr, lookup, p);
return context.factoryClazz.cast(factory); return context.factoryClazz.cast(factory);
} else if (context.instanceClazz.equals(ScoreScript.class)) {
ScoreScript.Factory factory = (p, lookup) -> newScoreScript(expr, lookup, p);
return context.factoryClazz.cast(factory);
} }
throw new IllegalArgumentException("expression engine does not know how to handle script context [" + context.name + "]"); throw new IllegalArgumentException("expression engine does not know how to handle script context [" + context.name + "]");
} }
@ -260,6 +267,42 @@ public class ExpressionScriptEngine extends AbstractComponent implements ScriptE
}; };
}; };
} }
private ScoreScript.LeafFactory newScoreScript(Expression expr, SearchLookup lookup, @Nullable Map<String, Object> vars) {
SearchScript.LeafFactory searchLeafFactory = newSearchScript(expr, lookup, vars);
return new ScoreScript.LeafFactory() {
@Override
public boolean needs_score() {
return searchLeafFactory.needs_score();
}
@Override
public ScoreScript newInstance(LeafReaderContext ctx) throws IOException {
SearchScript script = searchLeafFactory.newInstance(ctx);
return new ScoreScript(vars, lookup, ctx) {
@Override
public double execute() {
return script.runAsDouble();
}
@Override
public void setDocument(int docid) {
script.setDocument(docid);
}
@Override
public void setScorer(Scorer scorer) {
script.setScorer(scorer);
}
@Override
public double get_score() {
return script.getScore();
}
};
}
};
}
/** /**
* converts a ParseException at compile-time or link-time to a ScriptException * converts a ParseException at compile-time or link-time to a ScriptException

View File

@ -30,9 +30,9 @@ import org.apache.lucene.index.Term;
import org.elasticsearch.common.settings.Settings; import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.plugins.Plugin; import org.elasticsearch.plugins.Plugin;
import org.elasticsearch.plugins.ScriptPlugin; import org.elasticsearch.plugins.ScriptPlugin;
import org.elasticsearch.script.ScoreScript;
import org.elasticsearch.script.ScriptContext; import org.elasticsearch.script.ScriptContext;
import org.elasticsearch.script.ScriptEngine; import org.elasticsearch.script.ScriptEngine;
import org.elasticsearch.script.SearchScript;
/** /**
* An example script plugin that adds a {@link ScriptEngine} implementing expert scoring. * An example script plugin that adds a {@link ScriptEngine} implementing expert scoring.
@ -54,12 +54,12 @@ public class ExpertScriptPlugin extends Plugin implements ScriptPlugin {
@Override @Override
public <T> T compile(String scriptName, String scriptSource, ScriptContext<T> context, Map<String, String> params) { public <T> T compile(String scriptName, String scriptSource, ScriptContext<T> context, Map<String, String> params) {
if (context.equals(SearchScript.SCRIPT_SCORE_CONTEXT) == false) { if (context.equals(ScoreScript.CONTEXT) == false) {
throw new IllegalArgumentException(getType() + " scripts cannot be used for context [" + context.name + "]"); throw new IllegalArgumentException(getType() + " scripts cannot be used for context [" + context.name + "]");
} }
// we use the script "source" as the script identifier // we use the script "source" as the script identifier
if ("pure_df".equals(scriptSource)) { if ("pure_df".equals(scriptSource)) {
SearchScript.Factory factory = (p, lookup) -> new SearchScript.LeafFactory() { ScoreScript.Factory factory = (p, lookup) -> new ScoreScript.LeafFactory() {
final String field; final String field;
final String term; final String term;
{ {
@ -74,18 +74,18 @@ public class ExpertScriptPlugin extends Plugin implements ScriptPlugin {
} }
@Override @Override
public SearchScript newInstance(LeafReaderContext context) throws IOException { public ScoreScript newInstance(LeafReaderContext context) throws IOException {
PostingsEnum postings = context.reader().postings(new Term(field, term)); PostingsEnum postings = context.reader().postings(new Term(field, term));
if (postings == null) { if (postings == null) {
// the field and/or term don't exist in this segment, so always return 0 // the field and/or term don't exist in this segment, so always return 0
return new SearchScript(p, lookup, context) { return new ScoreScript(p, lookup, context) {
@Override @Override
public double runAsDouble() { public double execute() {
return 0.0d; return 0.0d;
} }
}; };
} }
return new SearchScript(p, lookup, context) { return new ScoreScript(p, lookup, context) {
int currentDocid = -1; int currentDocid = -1;
@Override @Override
public void setDocument(int docid) { public void setDocument(int docid) {
@ -100,7 +100,7 @@ public class ExpertScriptPlugin extends Plugin implements ScriptPlugin {
currentDocid = docid; currentDocid = docid;
} }
@Override @Override
public double runAsDouble() { public double execute() {
if (postings.docID() != currentDocid) { if (postings.docID() != currentDocid) {
// advance moved past the current doc, so this doc has no occurrences of the term // advance moved past the current doc, so this doc has no occurrences of the term
return 0.0d; return 0.0d;

View File

@ -25,9 +25,3 @@ dependencies {
testCompile project(path: ':modules:rank-eval', configuration: 'runtime') testCompile project(path: ':modules:rank-eval', configuration: 'runtime')
testCompile project(path: ':modules:lang-mustache', configuration: 'runtime') testCompile project(path: ':modules:lang-mustache', configuration: 'runtime')
} }
/*
* One of the integration tests doesn't work with the zip distribution
* and will be fixed later.
* Tracked by https://github.com/elastic/elasticsearch/issues/30628
*/

View File

@ -0,0 +1,28 @@
---
setup:
- do:
indices.create:
body:
settings:
index:
number_of_shards: 1
number_of_replicas: 1
index: test-index
---
Test retrieval of default settings:
- skip:
version: " - 6.3.99"
reason: include_defaults will not work in mixed-mode clusters containing nodes pre-6.4
- do:
indices.get_settings:
flat_settings: true
index: test-index
- is_false:
test-index.settings.index\.refresh_interval
- do:
indices.get_settings:
include_defaults: true
flat_settings: true
index: test-index
- match:
test-index.defaults.index\.refresh_interval: "1s"

View File

@ -23,15 +23,3 @@ Test reset index settings:
indices.get_settings: indices.get_settings:
flat_settings: false flat_settings: false
- is_false: test-index.settings.index\.refresh_interval - is_false: test-index.settings.index\.refresh_interval
# Disabled until https://github.com/elastic/elasticsearch/pull/29229 is back-ported
# That PR changed the execution path of index settings default to be on the master
# until the PR is back-ported the old master will not return default settings.
#
# - do:
# indices.get_settings:
# include_defaults: true
# flat_settings: true
# index: test-index
# - match:
# test-index.defaults.index\.refresh_interval: "1s"

View File

@ -52,7 +52,7 @@ setup:
--- ---
"Verify created repository": "Verify created repository":
- skip: - skip:
version: " - 6.99.99" version: "all"
reason: AwaitsFix for https://github.com/elastic/elasticsearch/issues/30807 reason: AwaitsFix for https://github.com/elastic/elasticsearch/issues/30807
- do: - do:
snapshot.verify_repository: snapshot.verify_repository:

View File

@ -11,7 +11,6 @@ setup:
--- ---
"Get snapshot status": "Get snapshot status":
- do: - do:
indices.create: indices.create:
index: test_index index: test_index
@ -32,6 +31,14 @@ setup:
snapshot: test_snapshot snapshot: test_snapshot
- is_true: snapshots - is_true: snapshots
- match: { snapshots.0.snapshot: test_snapshot }
- match: { snapshots.0.state: SUCCESS }
- gt: { snapshots.0.stats.incremental.file_count: 0 }
- gt: { snapshots.0.stats.incremental.size_in_bytes: 0 }
- gt: { snapshots.0.stats.total.file_count: 0 }
- gt: { snapshots.0.stats.total.size_in_bytes: 0 }
- is_true: snapshots.0.stats.start_time_in_millis
- is_true: snapshots.0.stats.time_in_millis
--- ---
"Get missing snapshot status throws an exception": "Get missing snapshot status throws an exception":

View File

@ -140,7 +140,7 @@ public class VerifyRepositoryResponse extends ActionResponse implements ToXConte
@Override @Override
public void readFrom(StreamInput in) throws IOException { public void readFrom(StreamInput in) throws IOException {
super.readFrom(in); super.readFrom(in);
if (in.getVersion().onOrAfter(Version.V_7_0_0_alpha1)) { if (in.getVersion().onOrAfter(Version.V_6_4_0)) {
this.nodes = in.readList(NodeView::new).stream().map(n -> n.convertToDiscoveryNode()).collect(Collectors.toList()); this.nodes = in.readList(NodeView::new).stream().map(n -> n.convertToDiscoveryNode()).collect(Collectors.toList());
} else { } else {
clusterName = new ClusterName(in); clusterName = new ClusterName(in);
@ -151,7 +151,7 @@ public class VerifyRepositoryResponse extends ActionResponse implements ToXConte
@Override @Override
public void writeTo(StreamOutput out) throws IOException { public void writeTo(StreamOutput out) throws IOException {
super.writeTo(out); super.writeTo(out);
if (Version.CURRENT.onOrAfter(Version.V_7_0_0_alpha1)) { if (out.getVersion().onOrAfter(Version.V_6_4_0)) {
out.writeList(getNodes()); out.writeList(getNodes());
} else { } else {
clusterName.writeTo(out); clusterName.writeTo(out);

View File

@ -74,8 +74,8 @@ public class SnapshotIndexShardStatus extends BroadcastShardResponse implements
throw new IllegalArgumentException("Unknown stage type " + indexShardStatus.getStage()); throw new IllegalArgumentException("Unknown stage type " + indexShardStatus.getStage());
} }
this.stats = new SnapshotStats(indexShardStatus.getStartTime(), indexShardStatus.getTotalTime(), this.stats = new SnapshotStats(indexShardStatus.getStartTime(), indexShardStatus.getTotalTime(),
indexShardStatus.getNumberOfFiles(), indexShardStatus.getProcessedFiles(), indexShardStatus.getIncrementalFileCount(), indexShardStatus.getTotalFileCount(), indexShardStatus.getProcessedFileCount(),
indexShardStatus.getTotalSize(), indexShardStatus.getProcessedSize()); indexShardStatus.getIncrementalSize(), indexShardStatus.getTotalSize(), indexShardStatus.getProcessedSize());
this.failure = indexShardStatus.getFailure(); this.failure = indexShardStatus.getFailure();
this.nodeId = nodeId; this.nodeId = nodeId;
} }

View File

@ -19,6 +19,7 @@
package org.elasticsearch.action.admin.cluster.snapshots.status; package org.elasticsearch.action.admin.cluster.snapshots.status;
import org.elasticsearch.Version;
import org.elasticsearch.common.io.stream.StreamInput; import org.elasticsearch.common.io.stream.StreamInput;
import org.elasticsearch.common.io.stream.StreamOutput; import org.elasticsearch.common.io.stream.StreamOutput;
import org.elasticsearch.common.io.stream.Streamable; import org.elasticsearch.common.io.stream.Streamable;
@ -34,19 +35,25 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
private long startTime; private long startTime;
private long time; private long time;
private int numberOfFiles; private int incrementalFileCount;
private int processedFiles; private int totalFileCount;
private int processedFileCount;
private long incrementalSize;
private long totalSize; private long totalSize;
private long processedSize; private long processedSize;
SnapshotStats() { SnapshotStats() {
} }
SnapshotStats(long startTime, long time, int numberOfFiles, int processedFiles, long totalSize, long processedSize) { SnapshotStats(long startTime, long time,
int incrementalFileCount, int totalFileCount, int processedFileCount,
long incrementalSize, long totalSize, long processedSize) {
this.startTime = startTime; this.startTime = startTime;
this.time = time; this.time = time;
this.numberOfFiles = numberOfFiles; this.incrementalFileCount = incrementalFileCount;
this.processedFiles = processedFiles; this.totalFileCount = totalFileCount;
this.processedFileCount = processedFileCount;
this.incrementalSize = incrementalSize;
this.totalSize = totalSize; this.totalSize = totalSize;
this.processedSize = processedSize; this.processedSize = processedSize;
} }
@ -66,17 +73,31 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
} }
/** /**
* Returns number of files in the snapshot * Returns incremental file count of the snapshot
*/ */
public int getNumberOfFiles() { public int getIncrementalFileCount() {
return numberOfFiles; return incrementalFileCount;
}
/**
* Returns total number of files in the snapshot
*/
public int getTotalFileCount() {
return totalFileCount;
} }
/** /**
* Returns number of files in the snapshot that were processed so far * Returns number of files in the snapshot that were processed so far
*/ */
public int getProcessedFiles() { public int getProcessedFileCount() {
return processedFiles; return processedFileCount;
}
/**
* Return incremental files size of the snapshot
*/
public long getIncrementalSize() {
return incrementalSize;
} }
/** /**
@ -105,11 +126,16 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
out.writeVLong(startTime); out.writeVLong(startTime);
out.writeVLong(time); out.writeVLong(time);
out.writeVInt(numberOfFiles); out.writeVInt(incrementalFileCount);
out.writeVInt(processedFiles); out.writeVInt(processedFileCount);
out.writeVLong(totalSize); out.writeVLong(incrementalSize);
out.writeVLong(processedSize); out.writeVLong(processedSize);
if (out.getVersion().onOrAfter(Version.V_6_4_0)) {
out.writeVInt(totalFileCount);
out.writeVLong(totalSize);
}
} }
@Override @Override
@ -117,21 +143,32 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
startTime = in.readVLong(); startTime = in.readVLong();
time = in.readVLong(); time = in.readVLong();
numberOfFiles = in.readVInt(); incrementalFileCount = in.readVInt();
processedFiles = in.readVInt(); processedFileCount = in.readVInt();
totalSize = in.readVLong(); incrementalSize = in.readVLong();
processedSize = in.readVLong(); processedSize = in.readVLong();
if (in.getVersion().onOrAfter(Version.V_6_4_0)) {
totalFileCount = in.readVInt();
totalSize = in.readVLong();
} else {
totalFileCount = incrementalFileCount;
totalSize = incrementalSize;
}
} }
static final class Fields { static final class Fields {
static final String STATS = "stats"; static final String STATS = "stats";
static final String NUMBER_OF_FILES = "number_of_files";
static final String PROCESSED_FILES = "processed_files"; static final String INCREMENTAL = "incremental";
static final String TOTAL_SIZE_IN_BYTES = "total_size_in_bytes"; static final String PROCESSED = "processed";
static final String TOTAL_SIZE = "total_size"; static final String TOTAL = "total";
static final String PROCESSED_SIZE_IN_BYTES = "processed_size_in_bytes";
static final String PROCESSED_SIZE = "processed_size"; static final String FILE_COUNT = "file_count";
static final String SIZE = "size";
static final String SIZE_IN_BYTES = "size_in_bytes";
static final String START_TIME_IN_MILLIS = "start_time_in_millis"; static final String START_TIME_IN_MILLIS = "start_time_in_millis";
static final String TIME_IN_MILLIS = "time_in_millis"; static final String TIME_IN_MILLIS = "time_in_millis";
static final String TIME = "time"; static final String TIME = "time";
@ -139,25 +176,44 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
@Override @Override
public XContentBuilder toXContent(XContentBuilder builder, ToXContent.Params params) throws IOException { public XContentBuilder toXContent(XContentBuilder builder, ToXContent.Params params) throws IOException {
builder.startObject(Fields.STATS); builder.startObject(Fields.STATS)
builder.field(Fields.NUMBER_OF_FILES, getNumberOfFiles()); // incremental starts
builder.field(Fields.PROCESSED_FILES, getProcessedFiles()); .startObject(Fields.INCREMENTAL)
builder.humanReadableField(Fields.TOTAL_SIZE_IN_BYTES, Fields.TOTAL_SIZE, new ByteSizeValue(getTotalSize())); .field(Fields.FILE_COUNT, getIncrementalFileCount())
builder.humanReadableField(Fields.PROCESSED_SIZE_IN_BYTES, Fields.PROCESSED_SIZE, new ByteSizeValue(getProcessedSize())); .humanReadableField(Fields.SIZE_IN_BYTES, Fields.SIZE, new ByteSizeValue(getIncrementalSize()))
builder.field(Fields.START_TIME_IN_MILLIS, getStartTime()); // incremental ends
builder.humanReadableField(Fields.TIME_IN_MILLIS, Fields.TIME, new TimeValue(getTime())); .endObject();
builder.endObject();
return builder; if (getProcessedFileCount() != getIncrementalFileCount()) {
// processed starts
builder.startObject(Fields.PROCESSED)
.field(Fields.FILE_COUNT, getProcessedFileCount())
.humanReadableField(Fields.SIZE_IN_BYTES, Fields.SIZE, new ByteSizeValue(getProcessedSize()))
// processed ends
.endObject();
}
// total starts
builder.startObject(Fields.TOTAL)
.field(Fields.FILE_COUNT, getTotalFileCount())
.humanReadableField(Fields.SIZE_IN_BYTES, Fields.SIZE, new ByteSizeValue(getTotalSize()))
// total ends
.endObject();
// timings stats
builder.field(Fields.START_TIME_IN_MILLIS, getStartTime())
.humanReadableField(Fields.TIME_IN_MILLIS, Fields.TIME, new TimeValue(getTime()));
return builder.endObject();
} }
void add(SnapshotStats stats) { void add(SnapshotStats stats) {
numberOfFiles += stats.numberOfFiles; incrementalFileCount += stats.incrementalFileCount;
processedFiles += stats.processedFiles; totalFileCount += stats.totalFileCount;
processedFileCount += stats.processedFileCount;
incrementalSize += stats.incrementalSize;
totalSize += stats.totalSize; totalSize += stats.totalSize;
processedSize += stats.processedSize; processedSize += stats.processedSize;
if (startTime == 0) { if (startTime == 0) {
// First time here // First time here
startTime = stats.startTime; startTime = stats.startTime;

View File

@ -18,12 +18,9 @@
*/ */
package org.elasticsearch.common.geo.parsers; package org.elasticsearch.common.geo.parsers;
import org.locationtech.jts.geom.Coordinate;
import org.elasticsearch.ElasticsearchParseException; import org.elasticsearch.ElasticsearchParseException;
import org.elasticsearch.common.geo.GeoPoint; import org.elasticsearch.common.geo.GeoPoint;
import org.elasticsearch.common.geo.GeoShapeType; import org.elasticsearch.common.geo.GeoShapeType;
import java.io.StringReader;
import org.elasticsearch.common.geo.builders.CoordinatesBuilder; import org.elasticsearch.common.geo.builders.CoordinatesBuilder;
import org.elasticsearch.common.geo.builders.EnvelopeBuilder; import org.elasticsearch.common.geo.builders.EnvelopeBuilder;
import org.elasticsearch.common.geo.builders.GeometryCollectionBuilder; import org.elasticsearch.common.geo.builders.GeometryCollectionBuilder;
@ -37,9 +34,11 @@ import org.elasticsearch.common.geo.builders.ShapeBuilder;
import org.elasticsearch.common.logging.Loggers; import org.elasticsearch.common.logging.Loggers;
import org.elasticsearch.common.xcontent.XContentParser; import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.index.mapper.GeoShapeFieldMapper; import org.elasticsearch.index.mapper.GeoShapeFieldMapper;
import org.locationtech.jts.geom.Coordinate;
import java.io.IOException; import java.io.IOException;
import java.io.StreamTokenizer; import java.io.StreamTokenizer;
import java.io.StringReader;
import java.util.List; import java.util.List;
/** /**
@ -77,8 +76,7 @@ public class GeoWKTParser {
public static ShapeBuilder parseExpectedType(XContentParser parser, final GeoShapeType shapeType, public static ShapeBuilder parseExpectedType(XContentParser parser, final GeoShapeType shapeType,
final GeoShapeFieldMapper shapeMapper) final GeoShapeFieldMapper shapeMapper)
throws IOException, ElasticsearchParseException { throws IOException, ElasticsearchParseException {
StringReader reader = new StringReader(parser.text()); try (StringReader reader = new StringReader(parser.text())) {
try {
boolean ignoreZValue = (shapeMapper != null && shapeMapper.ignoreZValue().value() == true); boolean ignoreZValue = (shapeMapper != null && shapeMapper.ignoreZValue().value() == true);
// setup the tokenizer; configured to read words w/o numbers // setup the tokenizer; configured to read words w/o numbers
StreamTokenizer tokenizer = new StreamTokenizer(reader); StreamTokenizer tokenizer = new StreamTokenizer(reader);
@ -95,8 +93,6 @@ public class GeoWKTParser {
ShapeBuilder builder = parseGeometry(tokenizer, shapeType, ignoreZValue); ShapeBuilder builder = parseGeometry(tokenizer, shapeType, ignoreZValue);
checkEOF(tokenizer); checkEOF(tokenizer);
return builder; return builder;
} finally {
reader.close();
} }
} }

View File

@ -24,8 +24,8 @@ import org.apache.lucene.search.DocIdSetIterator;
import org.apache.lucene.search.Explanation; import org.apache.lucene.search.Explanation;
import org.apache.lucene.search.Scorer; import org.apache.lucene.search.Scorer;
import org.elasticsearch.script.ExplainableSearchScript; import org.elasticsearch.script.ExplainableSearchScript;
import org.elasticsearch.script.ScoreScript;
import org.elasticsearch.script.Script; import org.elasticsearch.script.Script;
import org.elasticsearch.script.SearchScript;
import java.io.IOException; import java.io.IOException;
import java.util.Objects; import java.util.Objects;
@ -58,10 +58,10 @@ public class ScriptScoreFunction extends ScoreFunction {
private final Script sScript; private final Script sScript;
private final SearchScript.LeafFactory script; private final ScoreScript.LeafFactory script;
public ScriptScoreFunction(Script sScript, SearchScript.LeafFactory script) { public ScriptScoreFunction(Script sScript, ScoreScript.LeafFactory script) {
super(CombineFunction.REPLACE); super(CombineFunction.REPLACE);
this.sScript = sScript; this.sScript = sScript;
this.script = script; this.script = script;
@ -69,7 +69,7 @@ public class ScriptScoreFunction extends ScoreFunction {
@Override @Override
public LeafScoreFunction getLeafScoreFunction(LeafReaderContext ctx) throws IOException { public LeafScoreFunction getLeafScoreFunction(LeafReaderContext ctx) throws IOException {
final SearchScript leafScript = script.newInstance(ctx); final ScoreScript leafScript = script.newInstance(ctx);
final CannedScorer scorer = new CannedScorer(); final CannedScorer scorer = new CannedScorer();
leafScript.setScorer(scorer); leafScript.setScorer(scorer);
return new LeafScoreFunction() { return new LeafScoreFunction() {
@ -78,7 +78,7 @@ public class ScriptScoreFunction extends ScoreFunction {
leafScript.setDocument(docId); leafScript.setDocument(docId);
scorer.docid = docId; scorer.docid = docId;
scorer.score = subQueryScore; scorer.score = subQueryScore;
double result = leafScript.runAsDouble(); double result = leafScript.execute();
return result; return result;
} }

View File

@ -234,8 +234,8 @@ public class Analysis {
final Path path = env.configFile().resolve(wordListPath); final Path path = env.configFile().resolve(wordListPath);
try (BufferedReader reader = Files.newBufferedReader(path, StandardCharsets.UTF_8)) { try {
return loadWordList(reader, "#"); return loadWordList(path, "#");
} catch (CharacterCodingException ex) { } catch (CharacterCodingException ex) {
String message = String.format(Locale.ROOT, String message = String.format(Locale.ROOT,
"Unsupported character encoding detected while reading %s_path: %s - files must be UTF-8 encoded", "Unsupported character encoding detected while reading %s_path: %s - files must be UTF-8 encoded",
@ -247,15 +247,9 @@ public class Analysis {
} }
} }
public static List<String> loadWordList(Reader reader, String comment) throws IOException { private static List<String> loadWordList(Path path, String comment) throws IOException {
final List<String> result = new ArrayList<>(); final List<String> result = new ArrayList<>();
BufferedReader br = null; try (BufferedReader br = Files.newBufferedReader(path, StandardCharsets.UTF_8)) {
try {
if (reader instanceof BufferedReader) {
br = (BufferedReader) reader;
} else {
br = new BufferedReader(reader);
}
String word; String word;
while ((word = br.readLine()) != null) { while ((word = br.readLine()) != null) {
if (!Strings.hasText(word)) { if (!Strings.hasText(word)) {
@ -265,9 +259,6 @@ public class Analysis {
result.add(word.trim()); result.add(word.trim());
} }
} }
} finally {
if (br != null)
br.close();
} }
return result; return result;
} }

View File

@ -1424,10 +1424,6 @@ public abstract class Engine implements Closeable {
@Override @Override
public void close() { public void close() {
release();
}
public void release() {
Releasables.close(searcher); Releasables.close(searcher);
} }
} }

View File

@ -159,7 +159,7 @@ public final class ShardGetService extends AbstractIndexShardComponent {
get = indexShard.get(new Engine.Get(realtime, readFromTranslog, type, id, uidTerm) get = indexShard.get(new Engine.Get(realtime, readFromTranslog, type, id, uidTerm)
.version(version).versionType(versionType)); .version(version).versionType(versionType));
if (get.exists() == false) { if (get.exists() == false) {
get.release(); get.close();
} }
} }
} }
@ -172,7 +172,7 @@ public final class ShardGetService extends AbstractIndexShardComponent {
// break between having loaded it from translog (so we only have _source), and having a document to load // break between having loaded it from translog (so we only have _source), and having a document to load
return innerGetLoadFromStoredFields(type, id, gFields, fetchSourceContext, get, mapperService); return innerGetLoadFromStoredFields(type, id, gFields, fetchSourceContext, get, mapperService);
} finally { } finally {
get.release(); get.close();
} }
} }

View File

@ -28,6 +28,7 @@ import org.elasticsearch.common.xcontent.XContentBuilder;
import org.elasticsearch.common.xcontent.XContentParser; import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.index.query.QueryShardContext; import org.elasticsearch.index.query.QueryShardContext;
import org.elasticsearch.index.query.QueryShardException; import org.elasticsearch.index.query.QueryShardException;
import org.elasticsearch.script.ScoreScript;
import org.elasticsearch.script.Script; import org.elasticsearch.script.Script;
import org.elasticsearch.script.SearchScript; import org.elasticsearch.script.SearchScript;
@ -92,8 +93,8 @@ public class ScriptScoreFunctionBuilder extends ScoreFunctionBuilder<ScriptScore
@Override @Override
protected ScoreFunction doToFunction(QueryShardContext context) { protected ScoreFunction doToFunction(QueryShardContext context) {
try { try {
SearchScript.Factory factory = context.getScriptService().compile(script, SearchScript.SCRIPT_SCORE_CONTEXT); ScoreScript.Factory factory = context.getScriptService().compile(script, ScoreScript.CONTEXT);
SearchScript.LeafFactory searchScript = factory.newFactory(script.getParams(), context.lookup()); ScoreScript.LeafFactory searchScript = factory.newFactory(script.getParams(), context.lookup());
return new ScriptScoreFunction(script, searchScript); return new ScriptScoreFunction(script, searchScript);
} catch (Exception e) { } catch (Exception e) {
throw new QueryShardException(context, "script_score: the script could not be loaded", e); throw new QueryShardException(context, "script_score: the script could not be loaded", e);

View File

@ -60,31 +60,39 @@ public class IndexShardSnapshotStatus {
private final AtomicReference<Stage> stage; private final AtomicReference<Stage> stage;
private long startTime; private long startTime;
private long totalTime; private long totalTime;
private int numberOfFiles; private int incrementalFileCount;
private int processedFiles; private int totalFileCount;
private int processedFileCount;
private long totalSize; private long totalSize;
private long incrementalSize;
private long processedSize; private long processedSize;
private long indexVersion; private long indexVersion;
private String failure; private String failure;
private IndexShardSnapshotStatus(final Stage stage, final long startTime, final long totalTime, private IndexShardSnapshotStatus(final Stage stage, final long startTime, final long totalTime,
final int numberOfFiles, final int processedFiles, final long totalSize, final long processedSize, final int incrementalFileCount, final int totalFileCount, final int processedFileCount,
final long incrementalSize, final long totalSize, final long processedSize,
final long indexVersion, final String failure) { final long indexVersion, final String failure) {
this.stage = new AtomicReference<>(Objects.requireNonNull(stage)); this.stage = new AtomicReference<>(Objects.requireNonNull(stage));
this.startTime = startTime; this.startTime = startTime;
this.totalTime = totalTime; this.totalTime = totalTime;
this.numberOfFiles = numberOfFiles; this.incrementalFileCount = incrementalFileCount;
this.processedFiles = processedFiles; this.totalFileCount = totalFileCount;
this.processedFileCount = processedFileCount;
this.totalSize = totalSize; this.totalSize = totalSize;
this.processedSize = processedSize; this.processedSize = processedSize;
this.incrementalSize = incrementalSize;
this.indexVersion = indexVersion; this.indexVersion = indexVersion;
this.failure = failure; this.failure = failure;
} }
public synchronized Copy moveToStarted(final long startTime, final int numberOfFiles, final long totalSize) { public synchronized Copy moveToStarted(final long startTime, final int incrementalFileCount, final int totalFileCount,
final long incrementalSize, final long totalSize) {
if (stage.compareAndSet(Stage.INIT, Stage.STARTED)) { if (stage.compareAndSet(Stage.INIT, Stage.STARTED)) {
this.startTime = startTime; this.startTime = startTime;
this.numberOfFiles = numberOfFiles; this.incrementalFileCount = incrementalFileCount;
this.totalFileCount = totalFileCount;
this.incrementalSize = incrementalSize;
this.totalSize = totalSize; this.totalSize = totalSize;
} else { } else {
throw new IllegalStateException("Unable to move the shard snapshot status to [STARTED]: " + throw new IllegalStateException("Unable to move the shard snapshot status to [STARTED]: " +
@ -135,7 +143,7 @@ public class IndexShardSnapshotStatus {
* Increments number of processed files * Increments number of processed files
*/ */
public synchronized void addProcessedFile(long size) { public synchronized void addProcessedFile(long size) {
processedFiles++; processedFileCount++;
processedSize += size; processedSize += size;
} }
@ -146,12 +154,14 @@ public class IndexShardSnapshotStatus {
* @return a {@link IndexShardSnapshotStatus.Copy} * @return a {@link IndexShardSnapshotStatus.Copy}
*/ */
public synchronized IndexShardSnapshotStatus.Copy asCopy() { public synchronized IndexShardSnapshotStatus.Copy asCopy() {
return new IndexShardSnapshotStatus.Copy(stage.get(), startTime, totalTime, numberOfFiles, processedFiles, totalSize, processedSize, return new IndexShardSnapshotStatus.Copy(stage.get(), startTime, totalTime,
indexVersion, failure); incrementalFileCount, totalFileCount, processedFileCount,
incrementalSize, totalSize, processedSize,
indexVersion, failure);
} }
public static IndexShardSnapshotStatus newInitializing() { public static IndexShardSnapshotStatus newInitializing() {
return new IndexShardSnapshotStatus(Stage.INIT, 0L, 0L, 0, 0, 0, 0, 0, null); return new IndexShardSnapshotStatus(Stage.INIT, 0L, 0L, 0, 0, 0, 0, 0, 0, 0, null);
} }
public static IndexShardSnapshotStatus newFailed(final String failure) { public static IndexShardSnapshotStatus newFailed(final String failure) {
@ -159,12 +169,15 @@ public class IndexShardSnapshotStatus {
if (failure == null) { if (failure == null) {
throw new IllegalArgumentException("A failure description is required for a failed IndexShardSnapshotStatus"); throw new IllegalArgumentException("A failure description is required for a failed IndexShardSnapshotStatus");
} }
return new IndexShardSnapshotStatus(Stage.FAILURE, 0L, 0L, 0, 0, 0, 0, 0, failure); return new IndexShardSnapshotStatus(Stage.FAILURE, 0L, 0L, 0, 0, 0, 0, 0, 0, 0, failure);
} }
public static IndexShardSnapshotStatus newDone(final long startTime, final long totalTime, final int files, final long size) { public static IndexShardSnapshotStatus newDone(final long startTime, final long totalTime,
final int incrementalFileCount, final int fileCount,
final long incrementalSize, final long size) {
// The snapshot is done which means the number of processed files is the same as total // The snapshot is done which means the number of processed files is the same as total
return new IndexShardSnapshotStatus(Stage.DONE, startTime, totalTime, files, files, size, size, 0, null); return new IndexShardSnapshotStatus(Stage.DONE, startTime, totalTime, incrementalFileCount, fileCount, incrementalFileCount,
incrementalSize, size, incrementalSize, 0, null);
} }
/** /**
@ -175,23 +188,28 @@ public class IndexShardSnapshotStatus {
private final Stage stage; private final Stage stage;
private final long startTime; private final long startTime;
private final long totalTime; private final long totalTime;
private final int numberOfFiles; private final int incrementalFileCount;
private final int processedFiles; private final int totalFileCount;
private final int processedFileCount;
private final long totalSize; private final long totalSize;
private final long processedSize; private final long processedSize;
private final long incrementalSize;
private final long indexVersion; private final long indexVersion;
private final String failure; private final String failure;
public Copy(final Stage stage, final long startTime, final long totalTime, public Copy(final Stage stage, final long startTime, final long totalTime,
final int numberOfFiles, final int processedFiles, final long totalSize, final long processedSize, final int incrementalFileCount, final int totalFileCount, final int processedFileCount,
final long incrementalSize, final long totalSize, final long processedSize,
final long indexVersion, final String failure) { final long indexVersion, final String failure) {
this.stage = stage; this.stage = stage;
this.startTime = startTime; this.startTime = startTime;
this.totalTime = totalTime; this.totalTime = totalTime;
this.numberOfFiles = numberOfFiles; this.incrementalFileCount = incrementalFileCount;
this.processedFiles = processedFiles; this.totalFileCount = totalFileCount;
this.processedFileCount = processedFileCount;
this.totalSize = totalSize; this.totalSize = totalSize;
this.processedSize = processedSize; this.processedSize = processedSize;
this.incrementalSize = incrementalSize;
this.indexVersion = indexVersion; this.indexVersion = indexVersion;
this.failure = failure; this.failure = failure;
} }
@ -208,12 +226,20 @@ public class IndexShardSnapshotStatus {
return totalTime; return totalTime;
} }
public int getNumberOfFiles() { public int getIncrementalFileCount() {
return numberOfFiles; return incrementalFileCount;
} }
public int getProcessedFiles() { public int getTotalFileCount() {
return processedFiles; return totalFileCount;
}
public int getProcessedFileCount() {
return processedFileCount;
}
public long getIncrementalSize() {
return incrementalSize;
} }
public long getTotalSize() { public long getTotalSize() {
@ -238,8 +264,10 @@ public class IndexShardSnapshotStatus {
"stage=" + stage + "stage=" + stage +
", startTime=" + startTime + ", startTime=" + startTime +
", totalTime=" + totalTime + ", totalTime=" + totalTime +
", numberOfFiles=" + numberOfFiles + ", incrementalFileCount=" + incrementalFileCount +
", processedFiles=" + processedFiles + ", totalFileCount=" + totalFileCount +
", processedFileCount=" + processedFileCount +
", incrementalSize=" + incrementalSize +
", totalSize=" + totalSize + ", totalSize=" + totalSize +
", processedSize=" + processedSize + ", processedSize=" + processedSize +
", indexVersion=" + indexVersion + ", indexVersion=" + indexVersion +

View File

@ -356,25 +356,28 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
private final long time; private final long time;
private final int numberOfFiles; private final int incrementalFileCount;
private final long totalSize; private final long incrementalSize;
private final List<FileInfo> indexFiles; private final List<FileInfo> indexFiles;
/** /**
* Constructs new shard snapshot metadata from snapshot metadata * Constructs new shard snapshot metadata from snapshot metadata
* *
* @param snapshot snapshot id * @param snapshot snapshot id
* @param indexVersion index version * @param indexVersion index version
* @param indexFiles list of files in the shard * @param indexFiles list of files in the shard
* @param startTime snapshot start time * @param startTime snapshot start time
* @param time snapshot running time * @param time snapshot running time
* @param numberOfFiles number of files that where snapshotted * @param incrementalFileCount incremental of files that were snapshotted
* @param totalSize total size of all files snapshotted * @param incrementalSize incremental size of snapshot
*/ */
public BlobStoreIndexShardSnapshot(String snapshot, long indexVersion, List<FileInfo> indexFiles, long startTime, long time, public BlobStoreIndexShardSnapshot(String snapshot, long indexVersion, List<FileInfo> indexFiles,
int numberOfFiles, long totalSize) { long startTime, long time,
int incrementalFileCount,
long incrementalSize
) {
assert snapshot != null; assert snapshot != null;
assert indexVersion >= 0; assert indexVersion >= 0;
this.snapshot = snapshot; this.snapshot = snapshot;
@ -382,8 +385,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
this.indexFiles = Collections.unmodifiableList(new ArrayList<>(indexFiles)); this.indexFiles = Collections.unmodifiableList(new ArrayList<>(indexFiles));
this.startTime = startTime; this.startTime = startTime;
this.time = time; this.time = time;
this.numberOfFiles = numberOfFiles; this.incrementalFileCount = incrementalFileCount;
this.totalSize = totalSize; this.incrementalSize = incrementalSize;
} }
/** /**
@ -395,8 +398,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
this.indexFiles = Collections.emptyList(); this.indexFiles = Collections.emptyList();
this.startTime = 0; this.startTime = 0;
this.time = 0; this.time = 0;
this.numberOfFiles = 0; this.incrementalFileCount = 0;
this.totalSize = 0; this.incrementalSize = 0;
} }
/** /**
@ -441,34 +444,51 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
} }
/** /**
* Returns number of files that where snapshotted * Returns incremental of files that were snapshotted
*/ */
public int numberOfFiles() { public int incrementalFileCount() {
return numberOfFiles; return incrementalFileCount;
}
/**
* Returns total number of files that are referenced by this snapshot
*/
public int totalFileCount() {
return indexFiles.size();
}
/**
* Returns incremental of files size that were snapshotted
*/
public long incrementalSize() {
return incrementalSize;
} }
/** /**
* Returns total size of all files that where snapshotted * Returns total size of all files that where snapshotted
*/ */
public long totalSize() { public long totalSize() {
return totalSize; return indexFiles.stream().mapToLong(fi -> fi.metadata().length()).sum();
} }
private static final String NAME = "name"; private static final String NAME = "name";
private static final String INDEX_VERSION = "index_version"; private static final String INDEX_VERSION = "index_version";
private static final String START_TIME = "start_time"; private static final String START_TIME = "start_time";
private static final String TIME = "time"; private static final String TIME = "time";
private static final String NUMBER_OF_FILES = "number_of_files";
private static final String TOTAL_SIZE = "total_size";
private static final String FILES = "files"; private static final String FILES = "files";
// for the sake of BWC keep the actual property names as in 6.x
// + there is a constraint in #fromXContent() that leads to ElasticsearchParseException("unknown parameter [incremental_file_count]");
private static final String INCREMENTAL_FILE_COUNT = "number_of_files";
private static final String INCREMENTAL_SIZE = "total_size";
private static final ParseField PARSE_NAME = new ParseField("name");
private static final ParseField PARSE_INDEX_VERSION = new ParseField("index_version", "index-version"); private static final ParseField PARSE_NAME = new ParseField(NAME);
private static final ParseField PARSE_START_TIME = new ParseField("start_time"); private static final ParseField PARSE_INDEX_VERSION = new ParseField(INDEX_VERSION, "index-version");
private static final ParseField PARSE_TIME = new ParseField("time"); private static final ParseField PARSE_START_TIME = new ParseField(START_TIME);
private static final ParseField PARSE_NUMBER_OF_FILES = new ParseField("number_of_files"); private static final ParseField PARSE_TIME = new ParseField(TIME);
private static final ParseField PARSE_TOTAL_SIZE = new ParseField("total_size"); private static final ParseField PARSE_INCREMENTAL_FILE_COUNT = new ParseField(INCREMENTAL_FILE_COUNT);
private static final ParseField PARSE_FILES = new ParseField("files"); private static final ParseField PARSE_INCREMENTAL_SIZE = new ParseField(INCREMENTAL_SIZE);
private static final ParseField PARSE_FILES = new ParseField(FILES);
/** /**
* Serializes shard snapshot metadata info into JSON * Serializes shard snapshot metadata info into JSON
@ -482,8 +502,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
builder.field(INDEX_VERSION, indexVersion); builder.field(INDEX_VERSION, indexVersion);
builder.field(START_TIME, startTime); builder.field(START_TIME, startTime);
builder.field(TIME, time); builder.field(TIME, time);
builder.field(NUMBER_OF_FILES, numberOfFiles); builder.field(INCREMENTAL_FILE_COUNT, incrementalFileCount);
builder.field(TOTAL_SIZE, totalSize); builder.field(INCREMENTAL_SIZE, incrementalSize);
builder.startArray(FILES); builder.startArray(FILES);
for (FileInfo fileInfo : indexFiles) { for (FileInfo fileInfo : indexFiles) {
FileInfo.toXContent(fileInfo, builder, params); FileInfo.toXContent(fileInfo, builder, params);
@ -503,8 +523,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
long indexVersion = -1; long indexVersion = -1;
long startTime = 0; long startTime = 0;
long time = 0; long time = 0;
int numberOfFiles = 0; int incrementalFileCount = 0;
long totalSize = 0; long incrementalSize = 0;
List<FileInfo> indexFiles = new ArrayList<>(); List<FileInfo> indexFiles = new ArrayList<>();
if (parser.currentToken() == null) { // fresh parser? move to the first token if (parser.currentToken() == null) { // fresh parser? move to the first token
@ -526,10 +546,10 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
startTime = parser.longValue(); startTime = parser.longValue();
} else if (PARSE_TIME.match(currentFieldName, parser.getDeprecationHandler())) { } else if (PARSE_TIME.match(currentFieldName, parser.getDeprecationHandler())) {
time = parser.longValue(); time = parser.longValue();
} else if (PARSE_NUMBER_OF_FILES.match(currentFieldName, parser.getDeprecationHandler())) { } else if (PARSE_INCREMENTAL_FILE_COUNT.match(currentFieldName, parser.getDeprecationHandler())) {
numberOfFiles = parser.intValue(); incrementalFileCount = parser.intValue();
} else if (PARSE_TOTAL_SIZE.match(currentFieldName, parser.getDeprecationHandler())) { } else if (PARSE_INCREMENTAL_SIZE.match(currentFieldName, parser.getDeprecationHandler())) {
totalSize = parser.longValue(); incrementalSize = parser.longValue();
} else { } else {
throw new ElasticsearchParseException("unknown parameter [{}]", currentFieldName); throw new ElasticsearchParseException("unknown parameter [{}]", currentFieldName);
} }
@ -549,7 +569,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
} }
} }
} }
return new BlobStoreIndexShardSnapshot(snapshot, indexVersion, Collections.unmodifiableList(indexFiles), return new BlobStoreIndexShardSnapshot(snapshot, indexVersion, Collections.unmodifiableList(indexFiles),
startTime, time, numberOfFiles, totalSize); startTime, time, incrementalFileCount, incrementalSize);
} }
} }

View File

@ -85,8 +85,6 @@ public class TermVectorsService {
termVectorsResponse.setExists(false); termVectorsResponse.setExists(false);
return termVectorsResponse; return termVectorsResponse;
} }
Engine.GetResult get = indexShard.get(new Engine.Get(request.realtime(), false, request.type(), request.id(), uidTerm)
.version(request.version()).versionType(request.versionType()));
Fields termVectorsByField = null; Fields termVectorsByField = null;
AggregatedDfs dfs = null; AggregatedDfs dfs = null;
@ -97,8 +95,9 @@ public class TermVectorsService {
handleFieldWildcards(indexShard, request); handleFieldWildcards(indexShard, request);
} }
final Engine.Searcher searcher = indexShard.acquireSearcher("term_vector"); try (Engine.GetResult get = indexShard.get(new Engine.Get(request.realtime(), false, request.type(), request.id(), uidTerm)
try { .version(request.version()).versionType(request.versionType()));
Engine.Searcher searcher = indexShard.acquireSearcher("term_vector")) {
Fields topLevelFields = MultiFields.getFields(get.searcher() != null ? get.searcher().reader() : searcher.reader()); Fields topLevelFields = MultiFields.getFields(get.searcher() != null ? get.searcher().reader() : searcher.reader());
DocIdAndVersion docIdAndVersion = get.docIdAndVersion(); DocIdAndVersion docIdAndVersion = get.docIdAndVersion();
/* from an artificial document */ /* from an artificial document */
@ -143,14 +142,12 @@ public class TermVectorsService {
} }
} }
// write term vectors // write term vectors
termVectorsResponse.setFields(termVectorsByField, request.selectedFields(), request.getFlags(), topLevelFields, dfs, termVectorsFilter); termVectorsResponse.setFields(termVectorsByField, request.selectedFields(), request.getFlags(), topLevelFields, dfs,
termVectorsFilter);
} }
termVectorsResponse.setTookInMillis(TimeUnit.NANOSECONDS.toMillis(nanoTimeSupplier.getAsLong() - startTime)); termVectorsResponse.setTookInMillis(TimeUnit.NANOSECONDS.toMillis(nanoTimeSupplier.getAsLong() - startTime));
} catch (Exception ex) { } catch (Exception ex) {
throw new ElasticsearchException("failed to execute term vector request", ex); throw new ElasticsearchException("failed to execute term vector request", ex);
} finally {
searcher.close();
get.release();
} }
return termVectorsResponse; return termVectorsResponse;
} }

View File

@ -26,7 +26,6 @@ import org.elasticsearch.common.Nullable;
import org.elasticsearch.common.unit.TimeValue; import org.elasticsearch.common.unit.TimeValue;
import org.elasticsearch.tasks.CancellableTask; import org.elasticsearch.tasks.CancellableTask;
import org.elasticsearch.tasks.Task; import org.elasticsearch.tasks.Task;
import org.elasticsearch.tasks.TaskCancelledException;
import org.elasticsearch.tasks.TaskId; import org.elasticsearch.tasks.TaskId;
import org.elasticsearch.tasks.TaskManager; import org.elasticsearch.tasks.TaskManager;
@ -38,18 +37,16 @@ import java.util.function.Predicate;
* Represents a executor node operation that corresponds to a persistent task * Represents a executor node operation that corresponds to a persistent task
*/ */
public class AllocatedPersistentTask extends CancellableTask { public class AllocatedPersistentTask extends CancellableTask {
private volatile String persistentTaskId;
private volatile long allocationId;
private final AtomicReference<State> state; private final AtomicReference<State> state;
@Nullable
private volatile Exception failure;
private volatile String persistentTaskId;
private volatile long allocationId;
private volatile @Nullable Exception failure;
private volatile PersistentTasksService persistentTasksService; private volatile PersistentTasksService persistentTasksService;
private volatile Logger logger; private volatile Logger logger;
private volatile TaskManager taskManager; private volatile TaskManager taskManager;
public AllocatedPersistentTask(long id, String type, String action, String description, TaskId parentTask, public AllocatedPersistentTask(long id, String type, String action, String description, TaskId parentTask,
Map<String, String> headers) { Map<String, String> headers) {
super(id, type, action, description, parentTask, headers); super(id, type, action, description, parentTask, headers);
@ -101,24 +98,10 @@ public class AllocatedPersistentTask extends CancellableTask {
return failure; return failure;
} }
boolean markAsCancelled() {
return state.compareAndSet(AllocatedPersistentTask.State.STARTED, AllocatedPersistentTask.State.PENDING_CANCEL);
}
public State getState() {
return state.get();
}
public long getAllocationId() { public long getAllocationId() {
return allocationId; return allocationId;
} }
public enum State {
STARTED, // the task is currently running
PENDING_CANCEL, // the task is cancelled on master, cancelling it locally
COMPLETED // the task is done running and trying to notify caller
}
/** /**
* Waits for this persistent task to have the desired state. * Waits for this persistent task to have the desired state.
*/ */
@ -128,6 +111,14 @@ public class AllocatedPersistentTask extends CancellableTask {
persistentTasksService.waitForPersistentTaskStatus(persistentTaskId, predicate, timeout, listener); persistentTasksService.waitForPersistentTaskStatus(persistentTaskId, predicate, timeout, listener);
} }
final boolean isCompleted() {
return state.get() == State.COMPLETED;
}
boolean markAsCancelled() {
return state.compareAndSet(State.STARTED, State.PENDING_CANCEL);
}
public void markAsCompleted() { public void markAsCompleted() {
completeAndNotifyIfNeeded(null); completeAndNotifyIfNeeded(null);
} }
@ -138,11 +129,10 @@ public class AllocatedPersistentTask extends CancellableTask {
} else { } else {
completeAndNotifyIfNeeded(e); completeAndNotifyIfNeeded(e);
} }
} }
private void completeAndNotifyIfNeeded(@Nullable Exception failure) { private void completeAndNotifyIfNeeded(@Nullable Exception failure) {
State prevState = state.getAndSet(AllocatedPersistentTask.State.COMPLETED); final State prevState = state.getAndSet(State.COMPLETED);
if (prevState == State.COMPLETED) { if (prevState == State.COMPLETED) {
logger.warn("attempt to complete task [{}] with id [{}] in the [{}] state", getAction(), getPersistentTaskId(), prevState); logger.warn("attempt to complete task [{}] with id [{}] in the [{}] state", getAction(), getPersistentTaskId(), prevState);
} else { } else {
@ -173,4 +163,10 @@ public class AllocatedPersistentTask extends CancellableTask {
} }
} }
} }
public enum State {
STARTED, // the task is currently running
PENDING_CANCEL, // the task is cancelled on master, cancelling it locally
COMPLETED // the task is done running and trying to notify caller
}
} }

View File

@ -123,7 +123,7 @@ public class PersistentTasksNodeService extends AbstractComponent implements Clu
for (Long id : notVisitedTasks) { for (Long id : notVisitedTasks) {
AllocatedPersistentTask task = runningTasks.get(id); AllocatedPersistentTask task = runningTasks.get(id);
if (task.getState() == AllocatedPersistentTask.State.COMPLETED) { if (task.isCompleted()) {
// Result was sent to the caller and the caller acknowledged acceptance of the result // Result was sent to the caller and the caller acknowledged acceptance of the result
logger.trace("Found completed persistent task [{}] with id [{}] and allocation id [{}] - removing", logger.trace("Found completed persistent task [{}] with id [{}] and allocation id [{}] - removing",
task.getAction(), task.getPersistentTaskId(), task.getAllocationId()); task.getAction(), task.getPersistentTaskId(), task.getAllocationId());

View File

@ -440,7 +440,7 @@ public class PluginsService extends AbstractComponent {
List<Bundle> sortedBundles = sortBundles(bundles); List<Bundle> sortedBundles = sortBundles(bundles);
for (Bundle bundle : sortedBundles) { for (Bundle bundle : sortedBundles) {
checkBundleJarHell(bundle, transitiveUrls); checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveUrls);
final Plugin plugin = loadBundle(bundle, loaded); final Plugin plugin = loadBundle(bundle, loaded);
plugins.add(new Tuple<>(bundle.plugin, plugin)); plugins.add(new Tuple<>(bundle.plugin, plugin));
@ -451,7 +451,7 @@ public class PluginsService extends AbstractComponent {
// jar-hell check the bundle against the parent classloader and extended plugins // jar-hell check the bundle against the parent classloader and extended plugins
// the plugin cli does it, but we do it again, in case lusers mess with jar files manually // the plugin cli does it, but we do it again, in case lusers mess with jar files manually
static void checkBundleJarHell(Bundle bundle, Map<String, Set<URL>> transitiveUrls) { static void checkBundleJarHell(Set<URL> classpath, Bundle bundle, Map<String, Set<URL>> transitiveUrls) {
// invariant: any plugins this plugin bundle extends have already been added to transitiveUrls // invariant: any plugins this plugin bundle extends have already been added to transitiveUrls
List<String> exts = bundle.plugin.getExtendedPlugins(); List<String> exts = bundle.plugin.getExtendedPlugins();
@ -484,7 +484,6 @@ public class PluginsService extends AbstractComponent {
JarHell.checkJarHell(urls, logger::debug); // check jarhell of each extended plugin against this plugin JarHell.checkJarHell(urls, logger::debug); // check jarhell of each extended plugin against this plugin
transitiveUrls.put(bundle.plugin.getName(), urls); transitiveUrls.put(bundle.plugin.getName(), urls);
Set<URL> classpath = JarHell.parseClassPath();
// check we don't have conflicting codebases with core // check we don't have conflicting codebases with core
Set<URL> intersection = new HashSet<>(classpath); Set<URL> intersection = new HashSet<>(classpath);
intersection.retainAll(bundle.urls); intersection.retainAll(bundle.urls);

View File

@ -818,7 +818,9 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
public IndexShardSnapshotStatus getShardSnapshotStatus(SnapshotId snapshotId, Version version, IndexId indexId, ShardId shardId) { public IndexShardSnapshotStatus getShardSnapshotStatus(SnapshotId snapshotId, Version version, IndexId indexId, ShardId shardId) {
Context context = new Context(snapshotId, version, indexId, shardId); Context context = new Context(snapshotId, version, indexId, shardId);
BlobStoreIndexShardSnapshot snapshot = context.loadSnapshot(); BlobStoreIndexShardSnapshot snapshot = context.loadSnapshot();
return IndexShardSnapshotStatus.newDone(snapshot.startTime(), snapshot.time(), snapshot.numberOfFiles(), snapshot.totalSize()); return IndexShardSnapshotStatus.newDone(snapshot.startTime(), snapshot.time(),
snapshot.incrementalFileCount(), snapshot.totalFileCount(),
snapshot.incrementalSize(), snapshot.totalSize());
} }
@Override @Override
@ -1139,9 +1141,11 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
final List<BlobStoreIndexShardSnapshot.FileInfo> indexCommitPointFiles = new ArrayList<>(); final List<BlobStoreIndexShardSnapshot.FileInfo> indexCommitPointFiles = new ArrayList<>();
store.incRef(); store.incRef();
int indexIncrementalFileCount = 0;
int indexTotalNumberOfFiles = 0;
long indexIncrementalSize = 0;
long indexTotalFileCount = 0;
try { try {
int indexNumberOfFiles = 0;
long indexTotalFilesSize = 0;
ArrayList<BlobStoreIndexShardSnapshot.FileInfo> filesToSnapshot = new ArrayList<>(); ArrayList<BlobStoreIndexShardSnapshot.FileInfo> filesToSnapshot = new ArrayList<>();
final Store.MetadataSnapshot metadata; final Store.MetadataSnapshot metadata;
// TODO apparently we don't use the MetadataSnapshot#.recoveryDiff(...) here but we should // TODO apparently we don't use the MetadataSnapshot#.recoveryDiff(...) here but we should
@ -1182,9 +1186,13 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
} }
} }
} }
indexTotalFileCount += md.length();
indexTotalNumberOfFiles++;
if (existingFileInfo == null) { if (existingFileInfo == null) {
indexNumberOfFiles++; indexIncrementalFileCount++;
indexTotalFilesSize += md.length(); indexIncrementalSize += md.length();
// create a new FileInfo // create a new FileInfo
BlobStoreIndexShardSnapshot.FileInfo snapshotFileInfo = new BlobStoreIndexShardSnapshot.FileInfo(fileNameFromGeneration(++generation), md, chunkSize()); BlobStoreIndexShardSnapshot.FileInfo snapshotFileInfo = new BlobStoreIndexShardSnapshot.FileInfo(fileNameFromGeneration(++generation), md, chunkSize());
indexCommitPointFiles.add(snapshotFileInfo); indexCommitPointFiles.add(snapshotFileInfo);
@ -1194,7 +1202,8 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
} }
} }
snapshotStatus.moveToStarted(startTime, indexNumberOfFiles, indexTotalFilesSize); snapshotStatus.moveToStarted(startTime, indexIncrementalFileCount,
indexTotalNumberOfFiles, indexIncrementalSize, indexTotalFileCount);
for (BlobStoreIndexShardSnapshot.FileInfo snapshotFileInfo : filesToSnapshot) { for (BlobStoreIndexShardSnapshot.FileInfo snapshotFileInfo : filesToSnapshot) {
try { try {
@ -1217,8 +1226,9 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
// snapshotStatus.startTime() is assigned on the same machine, // snapshotStatus.startTime() is assigned on the same machine,
// so it's safe to use with VLong // so it's safe to use with VLong
System.currentTimeMillis() - lastSnapshotStatus.getStartTime(), System.currentTimeMillis() - lastSnapshotStatus.getStartTime(),
lastSnapshotStatus.getNumberOfFiles(), lastSnapshotStatus.getIncrementalFileCount(),
lastSnapshotStatus.getTotalSize()); lastSnapshotStatus.getIncrementalSize()
);
//TODO: The time stored in snapshot doesn't include cleanup time. //TODO: The time stored in snapshot doesn't include cleanup time.
logger.trace("[{}] [{}] writing shard snapshot file", shardId, snapshotId); logger.trace("[{}] [{}] writing shard snapshot file", shardId, snapshotId);

View File

@ -0,0 +1,102 @@
/*
* Licensed to Elasticsearch under one or more contributor
* license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright
* ownership. Elasticsearch licenses this file to you under
* the Apache License, Version 2.0 (the "License"); you may
* not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.elasticsearch.script;
import org.apache.lucene.index.LeafReaderContext;
import org.apache.lucene.search.Scorer;
import org.elasticsearch.index.fielddata.ScriptDocValues;
import org.elasticsearch.search.lookup.LeafSearchLookup;
import org.elasticsearch.search.lookup.SearchLookup;
import java.io.IOException;
import java.io.UncheckedIOException;
import java.util.Map;
import java.util.function.DoubleSupplier;
/**
* A script used for adjusting the score on a per document basis.
*/
public abstract class ScoreScript {
public static final String[] PARAMETERS = new String[]{};
/** The generic runtime parameters for the script. */
private final Map<String, Object> params;
/** A leaf lookup for the bound segment this script will operate on. */
private final LeafSearchLookup leafLookup;
private DoubleSupplier scoreSupplier = () -> 0.0;
public ScoreScript(Map<String, Object> params, SearchLookup lookup, LeafReaderContext leafContext) {
this.params = params;
this.leafLookup = lookup.getLeafSearchLookup(leafContext);
}
public abstract double execute();
/** Return the parameters for this script. */
public Map<String, Object> getParams() {
return params;
}
/** The doc lookup for the Lucene segment this script was created for. */
public final Map<String, ScriptDocValues<?>> getDoc() {
return leafLookup.doc();
}
/** Set the current document to run the script on next. */
public void setDocument(int docid) {
leafLookup.setDocument(docid);
}
public void setScorer(Scorer scorer) {
this.scoreSupplier = () -> {
try {
return scorer.score();
} catch (IOException e) {
throw new UncheckedIOException(e);
}
};
}
public double get_score() {
return scoreSupplier.getAsDouble();
}
/** A factory to construct {@link ScoreScript} instances. */
public interface LeafFactory {
/**
* Return {@code true} if the script needs {@code _score} calculated, or {@code false} otherwise.
*/
boolean needs_score();
ScoreScript newInstance(LeafReaderContext ctx) throws IOException;
}
/** A factory to construct stateful {@link ScoreScript} factories for a specific index. */
public interface Factory {
ScoreScript.LeafFactory newFactory(Map<String, Object> params, SearchLookup lookup);
}
public static final ScriptContext<ScoreScript.Factory> CONTEXT = new ScriptContext<>("score", ScoreScript.Factory.class);
}

View File

@ -42,7 +42,7 @@ public class ScriptModule {
CORE_CONTEXTS = Stream.of( CORE_CONTEXTS = Stream.of(
SearchScript.CONTEXT, SearchScript.CONTEXT,
SearchScript.AGGS_CONTEXT, SearchScript.AGGS_CONTEXT,
SearchScript.SCRIPT_SCORE_CONTEXT, ScoreScript.CONTEXT,
SearchScript.SCRIPT_SORT_CONTEXT, SearchScript.SCRIPT_SORT_CONTEXT,
SearchScript.TERMS_SET_QUERY_CONTEXT, SearchScript.TERMS_SET_QUERY_CONTEXT,
ExecutableScript.CONTEXT, ExecutableScript.CONTEXT,

View File

@ -162,8 +162,6 @@ public abstract class SearchScript implements ScorerAware, ExecutableScript {
public static final ScriptContext<Factory> AGGS_CONTEXT = new ScriptContext<>("aggs", Factory.class); public static final ScriptContext<Factory> AGGS_CONTEXT = new ScriptContext<>("aggs", Factory.class);
// Can return a double. (For ScriptSortType#NUMBER only, for ScriptSortType#STRING normal CONTEXT should be used) // Can return a double. (For ScriptSortType#NUMBER only, for ScriptSortType#STRING normal CONTEXT should be used)
public static final ScriptContext<Factory> SCRIPT_SORT_CONTEXT = new ScriptContext<>("sort", Factory.class); public static final ScriptContext<Factory> SCRIPT_SORT_CONTEXT = new ScriptContext<>("sort", Factory.class);
// Can return a float
public static final ScriptContext<Factory> SCRIPT_SCORE_CONTEXT = new ScriptContext<>("score", Factory.class);
// Can return a long // Can return a long
public static final ScriptContext<Factory> TERMS_SET_QUERY_CONTEXT = new ScriptContext<>("terms_set", Factory.class); public static final ScriptContext<Factory> TERMS_SET_QUERY_CONTEXT = new ScriptContext<>("terms_set", Factory.class);
} }

View File

@ -21,7 +21,6 @@ package org.elasticsearch.search;
import org.apache.lucene.search.FieldDoc; import org.apache.lucene.search.FieldDoc;
import org.apache.lucene.search.TopDocs; import org.apache.lucene.search.TopDocs;
import org.elasticsearch.core.internal.io.IOUtils;
import org.elasticsearch.ElasticsearchException; import org.elasticsearch.ElasticsearchException;
import org.elasticsearch.ExceptionsHelper; import org.elasticsearch.ExceptionsHelper;
import org.elasticsearch.action.ActionListener; import org.elasticsearch.action.ActionListener;
@ -39,6 +38,7 @@ import org.elasticsearch.common.util.BigArrays;
import org.elasticsearch.common.util.concurrent.AbstractRunnable; import org.elasticsearch.common.util.concurrent.AbstractRunnable;
import org.elasticsearch.common.util.concurrent.ConcurrentCollections; import org.elasticsearch.common.util.concurrent.ConcurrentCollections;
import org.elasticsearch.common.util.concurrent.ConcurrentMapLong; import org.elasticsearch.common.util.concurrent.ConcurrentMapLong;
import org.elasticsearch.core.internal.io.IOUtils;
import org.elasticsearch.index.Index; import org.elasticsearch.index.Index;
import org.elasticsearch.index.IndexService; import org.elasticsearch.index.IndexService;
import org.elasticsearch.index.IndexSettings; import org.elasticsearch.index.IndexSettings;
@ -92,8 +92,8 @@ import org.elasticsearch.search.sort.SortAndFormats;
import org.elasticsearch.search.sort.SortBuilder; import org.elasticsearch.search.sort.SortBuilder;
import org.elasticsearch.search.suggest.Suggest; import org.elasticsearch.search.suggest.Suggest;
import org.elasticsearch.search.suggest.completion.CompletionSuggestion; import org.elasticsearch.search.suggest.completion.CompletionSuggestion;
import org.elasticsearch.threadpool.ThreadPool;
import org.elasticsearch.threadpool.Scheduler.Cancellable; import org.elasticsearch.threadpool.Scheduler.Cancellable;
import org.elasticsearch.threadpool.ThreadPool;
import org.elasticsearch.threadpool.ThreadPool.Names; import org.elasticsearch.threadpool.ThreadPool.Names;
import org.elasticsearch.transport.TransportRequest; import org.elasticsearch.transport.TransportRequest;
@ -646,20 +646,17 @@ public class SearchService extends AbstractLifecycleComponent implements IndexEv
public boolean freeContext(long id) { public boolean freeContext(long id) {
final SearchContext context = removeContext(id); try (SearchContext context = removeContext(id)) {
if (context != null) { if (context != null) {
assert context.refCount() > 0 : " refCount must be > 0: " + context.refCount(); assert context.refCount() > 0 : " refCount must be > 0: " + context.refCount();
try {
context.indexShard().getSearchOperationListener().onFreeContext(context); context.indexShard().getSearchOperationListener().onFreeContext(context);
if (context.scrollContext() != null) { if (context.scrollContext() != null) {
context.indexShard().getSearchOperationListener().onFreeScrollContext(context); context.indexShard().getSearchOperationListener().onFreeScrollContext(context);
} }
} finally { return true;
context.close();
} }
return true; return false;
} }
return false;
} }
public void freeAllScrollContexts() { public void freeAllScrollContexts() {

View File

@ -65,11 +65,11 @@ public class SnapshotBlocksIT extends ESIntegTestCase {
client().prepareIndex(OTHER_INDEX_NAME, "type").setSource("test", "init").execute().actionGet(); client().prepareIndex(OTHER_INDEX_NAME, "type").setSource("test", "init").execute().actionGet();
} }
logger.info("--> register a repository"); logger.info("--> register a repository");
assertAcked(client().admin().cluster().preparePutRepository(REPOSITORY_NAME) assertAcked(client().admin().cluster().preparePutRepository(REPOSITORY_NAME)
.setType("fs") .setType("fs")
.setSettings(Settings.builder().put("location", randomRepoPath()))); .setSettings(Settings.builder().put("location", randomRepoPath())));
logger.info("--> verify the repository"); logger.info("--> verify the repository");
VerifyRepositoryResponse verifyResponse = client().admin().cluster().prepareVerifyRepository(REPOSITORY_NAME).get(); VerifyRepositoryResponse verifyResponse = client().admin().cluster().prepareVerifyRepository(REPOSITORY_NAME).get();

View File

@ -91,10 +91,14 @@ public class SnapshotStatusTests extends ESTestCase {
" \"total\" : " + totalShards + "\n" + " \"total\" : " + totalShards + "\n" +
" },\n" + " },\n" +
" \"stats\" : {\n" + " \"stats\" : {\n" +
" \"number_of_files\" : 0,\n" + " \"incremental\" : {\n" +
" \"processed_files\" : 0,\n" + " \"file_count\" : 0,\n" +
" \"total_size_in_bytes\" : 0,\n" + " \"size_in_bytes\" : 0\n" +
" \"processed_size_in_bytes\" : 0,\n" + " },\n" +
" \"total\" : {\n" +
" \"file_count\" : 0,\n" +
" \"size_in_bytes\" : 0\n" +
" },\n" +
" \"start_time_in_millis\" : 0,\n" + " \"start_time_in_millis\" : 0,\n" +
" \"time_in_millis\" : 0\n" + " \"time_in_millis\" : 0\n" +
" },\n" + " },\n" +
@ -109,10 +113,14 @@ public class SnapshotStatusTests extends ESTestCase {
" \"total\" : " + totalShards + "\n" + " \"total\" : " + totalShards + "\n" +
" },\n" + " },\n" +
" \"stats\" : {\n" + " \"stats\" : {\n" +
" \"number_of_files\" : 0,\n" + " \"incremental\" : {\n" +
" \"processed_files\" : 0,\n" + " \"file_count\" : 0,\n" +
" \"total_size_in_bytes\" : 0,\n" + " \"size_in_bytes\" : 0\n" +
" \"processed_size_in_bytes\" : 0,\n" + " },\n" +
" \"total\" : {\n" +
" \"file_count\" : 0,\n" +
" \"size_in_bytes\" : 0\n" +
" },\n" +
" \"start_time_in_millis\" : 0,\n" + " \"start_time_in_millis\" : 0,\n" +
" \"time_in_millis\" : 0\n" + " \"time_in_millis\" : 0\n" +
" },\n" + " },\n" +
@ -120,10 +128,14 @@ public class SnapshotStatusTests extends ESTestCase {
" \"" + shardId + "\" : {\n" + " \"" + shardId + "\" : {\n" +
" \"stage\" : \"" + shardStage.toString() + "\",\n" + " \"stage\" : \"" + shardStage.toString() + "\",\n" +
" \"stats\" : {\n" + " \"stats\" : {\n" +
" \"number_of_files\" : 0,\n" + " \"incremental\" : {\n" +
" \"processed_files\" : 0,\n" + " \"file_count\" : 0,\n" +
" \"total_size_in_bytes\" : 0,\n" + " \"size_in_bytes\" : 0\n" +
" \"processed_size_in_bytes\" : 0,\n" + " },\n" +
" \"total\" : {\n" +
" \"file_count\" : 0,\n" +
" \"size_in_bytes\" : 0\n" +
" },\n" +
" \"start_time_in_millis\" : 0,\n" + " \"start_time_in_millis\" : 0,\n" +
" \"time_in_millis\" : 0\n" + " \"time_in_millis\" : 0\n" +
" }\n" + " }\n" +

View File

@ -23,6 +23,7 @@ import org.apache.lucene.index.DirectoryReader;
import org.apache.lucene.index.Fields; import org.apache.lucene.index.Fields;
import org.apache.lucene.index.Terms; import org.apache.lucene.index.Terms;
import org.apache.lucene.index.TermsEnum; import org.apache.lucene.index.TermsEnum;
import org.elasticsearch.ElasticsearchException;
import org.elasticsearch.action.admin.indices.alias.Alias; import org.elasticsearch.action.admin.indices.alias.Alias;
import org.elasticsearch.common.lucene.uid.Versions; import org.elasticsearch.common.lucene.uid.Versions;
import org.elasticsearch.common.settings.Settings; import org.elasticsearch.common.settings.Settings;
@ -111,7 +112,8 @@ public class MultiTermVectorsIT extends AbstractTermVectorsTestCase {
checkTermTexts(response.getResponses()[1].getResponse().getFields().terms("field"), new String[]{"value1"}); checkTermTexts(response.getResponses()[1].getResponse().getFields().terms("field"), new String[]{"value1"});
assertThat(response.getResponses()[2].getFailure(), notNullValue()); assertThat(response.getResponses()[2].getFailure(), notNullValue());
assertThat(response.getResponses()[2].getFailure().getId(), equalTo("1")); assertThat(response.getResponses()[2].getFailure().getId(), equalTo("1"));
assertThat(response.getResponses()[2].getFailure().getCause(), instanceOf(VersionConflictEngineException.class)); assertThat(response.getResponses()[2].getFailure().getCause(), instanceOf(ElasticsearchException.class));
assertThat(response.getResponses()[2].getFailure().getCause().getCause(), instanceOf(VersionConflictEngineException.class));
//Version from Lucene index //Version from Lucene index
refresh(); refresh();
@ -132,7 +134,8 @@ public class MultiTermVectorsIT extends AbstractTermVectorsTestCase {
checkTermTexts(response.getResponses()[1].getResponse().getFields().terms("field"), new String[]{"value1"}); checkTermTexts(response.getResponses()[1].getResponse().getFields().terms("field"), new String[]{"value1"});
assertThat(response.getResponses()[2].getFailure(), notNullValue()); assertThat(response.getResponses()[2].getFailure(), notNullValue());
assertThat(response.getResponses()[2].getFailure().getId(), equalTo("1")); assertThat(response.getResponses()[2].getFailure().getId(), equalTo("1"));
assertThat(response.getResponses()[2].getFailure().getCause(), instanceOf(VersionConflictEngineException.class)); assertThat(response.getResponses()[2].getFailure().getCause(), instanceOf(ElasticsearchException.class));
assertThat(response.getResponses()[2].getFailure().getCause().getCause(), instanceOf(VersionConflictEngineException.class));
for (int i = 0; i < 3; i++) { for (int i = 0; i < 3; i++) {
@ -155,7 +158,8 @@ public class MultiTermVectorsIT extends AbstractTermVectorsTestCase {
assertThat(response.getResponses()[1].getFailure(), notNullValue()); assertThat(response.getResponses()[1].getFailure(), notNullValue());
assertThat(response.getResponses()[1].getFailure().getId(), equalTo("2")); assertThat(response.getResponses()[1].getFailure().getId(), equalTo("2"));
assertThat(response.getResponses()[1].getIndex(), equalTo("test")); assertThat(response.getResponses()[1].getIndex(), equalTo("test"));
assertThat(response.getResponses()[1].getFailure().getCause(), instanceOf(VersionConflictEngineException.class)); assertThat(response.getResponses()[1].getFailure().getCause(), instanceOf(ElasticsearchException.class));
assertThat(response.getResponses()[1].getFailure().getCause().getCause(), instanceOf(VersionConflictEngineException.class));
assertThat(response.getResponses()[2].getId(), equalTo("2")); assertThat(response.getResponses()[2].getId(), equalTo("2"));
assertThat(response.getResponses()[2].getIndex(), equalTo("test")); assertThat(response.getResponses()[2].getIndex(), equalTo("test"));
assertThat(response.getResponses()[2].getFailure(), nullValue()); assertThat(response.getResponses()[2].getFailure(), nullValue());
@ -180,7 +184,8 @@ public class MultiTermVectorsIT extends AbstractTermVectorsTestCase {
assertThat(response.getResponses()[1].getFailure(), notNullValue()); assertThat(response.getResponses()[1].getFailure(), notNullValue());
assertThat(response.getResponses()[1].getFailure().getId(), equalTo("2")); assertThat(response.getResponses()[1].getFailure().getId(), equalTo("2"));
assertThat(response.getResponses()[1].getIndex(), equalTo("test")); assertThat(response.getResponses()[1].getIndex(), equalTo("test"));
assertThat(response.getResponses()[1].getFailure().getCause(), instanceOf(VersionConflictEngineException.class)); assertThat(response.getResponses()[1].getFailure().getCause(), instanceOf(ElasticsearchException.class));
assertThat(response.getResponses()[1].getFailure().getCause().getCause(), instanceOf(VersionConflictEngineException.class));
assertThat(response.getResponses()[2].getId(), equalTo("2")); assertThat(response.getResponses()[2].getId(), equalTo("2"));
assertThat(response.getResponses()[2].getIndex(), equalTo("test")); assertThat(response.getResponses()[2].getIndex(), equalTo("test"));
assertThat(response.getResponses()[2].getFailure(), nullValue()); assertThat(response.getResponses()[2].getFailure(), nullValue());

View File

@ -471,7 +471,7 @@ public class CacheTests extends ESTestCase {
keys.add(key); keys.add(key);
} else { } else {
// invalidate with incorrect value // invalidate with incorrect value
cache.invalidate(key, Integer.toString(key * randomIntBetween(2, 10))); cache.invalidate(key, Integer.toString(key + randomIntBetween(2, 10)));
} }
} }
} }
@ -506,7 +506,7 @@ public class CacheTests extends ESTestCase {
invalidated.add(i); invalidated.add(i);
} else { } else {
// invalidate with incorrect value // invalidate with incorrect value
cache.invalidate(i, Integer.toString(i * randomIntBetween(2, 10))); cache.invalidate(i, Integer.toString(i + randomIntBetween(2, 10)));
} }
} }
} }

View File

@ -21,6 +21,7 @@ package org.elasticsearch.index.engine;
import com.carrotsearch.hppc.cursors.ObjectObjectCursor; import com.carrotsearch.hppc.cursors.ObjectObjectCursor;
import com.carrotsearch.randomizedtesting.generators.RandomNumbers; import com.carrotsearch.randomizedtesting.generators.RandomNumbers;
import org.apache.logging.log4j.Level; import org.apache.logging.log4j.Level;
import org.apache.logging.log4j.LogManager; import org.apache.logging.log4j.LogManager;
import org.apache.logging.log4j.Logger; import org.apache.logging.log4j.Logger;
@ -793,7 +794,7 @@ public class InternalEngineTests extends EngineTestCase {
while (flushFinished.get() == false) { while (flushFinished.get() == false) {
Engine.GetResult previousGetResult = latestGetResult.get(); Engine.GetResult previousGetResult = latestGetResult.get();
if (previousGetResult != null) { if (previousGetResult != null) {
previousGetResult.release(); previousGetResult.close();
} }
latestGetResult.set(engine.get(newGet(true, doc), searcherFactory)); latestGetResult.set(engine.get(newGet(true, doc), searcherFactory));
if (latestGetResult.get().exists() == false) { if (latestGetResult.get().exists() == false) {
@ -807,7 +808,7 @@ public class InternalEngineTests extends EngineTestCase {
flushFinished.set(true); flushFinished.set(true);
getThread.join(); getThread.join();
assertTrue(latestGetResult.get().exists()); assertTrue(latestGetResult.get().exists());
latestGetResult.get().release(); latestGetResult.get().close();
} }
public void testSimpleOperations() throws Exception { public void testSimpleOperations() throws Exception {
@ -830,21 +831,20 @@ public class InternalEngineTests extends EngineTestCase {
searchResult.close(); searchResult.close();
// but, not there non realtime // but, not there non realtime
Engine.GetResult getResult = engine.get(newGet(false, doc), searcherFactory); try (Engine.GetResult getResult = engine.get(newGet(false, doc), searcherFactory)) {
assertThat(getResult.exists(), equalTo(false)); assertThat(getResult.exists(), equalTo(false));
getResult.release(); }
// but, we can still get it (in realtime) // but, we can still get it (in realtime)
getResult = engine.get(newGet(true, doc), searcherFactory); try (Engine.GetResult getResult = engine.get(newGet(true, doc), searcherFactory)) {
assertThat(getResult.exists(), equalTo(true)); assertThat(getResult.exists(), equalTo(true));
assertThat(getResult.docIdAndVersion(), notNullValue()); assertThat(getResult.docIdAndVersion(), notNullValue());
getResult.release(); }
// but not real time is not yet visible // but not real time is not yet visible
getResult = engine.get(newGet(false, doc), searcherFactory); try (Engine.GetResult getResult = engine.get(newGet(false, doc), searcherFactory)) {
assertThat(getResult.exists(), equalTo(false)); assertThat(getResult.exists(), equalTo(false));
getResult.release(); }
// refresh and it should be there // refresh and it should be there
engine.refresh("test"); engine.refresh("test");
@ -856,10 +856,10 @@ public class InternalEngineTests extends EngineTestCase {
searchResult.close(); searchResult.close();
// also in non realtime // also in non realtime
getResult = engine.get(newGet(false, doc), searcherFactory); try (Engine.GetResult getResult = engine.get(newGet(false, doc), searcherFactory)) {
assertThat(getResult.exists(), equalTo(true)); assertThat(getResult.exists(), equalTo(true));
assertThat(getResult.docIdAndVersion(), notNullValue()); assertThat(getResult.docIdAndVersion(), notNullValue());
getResult.release(); }
// now do an update // now do an update
document = testDocument(); document = testDocument();
@ -876,10 +876,10 @@ public class InternalEngineTests extends EngineTestCase {
searchResult.close(); searchResult.close();
// but, we can still get it (in realtime) // but, we can still get it (in realtime)
getResult = engine.get(newGet(true, doc), searcherFactory); try (Engine.GetResult getResult = engine.get(newGet(true, doc), searcherFactory)) {
assertThat(getResult.exists(), equalTo(true)); assertThat(getResult.exists(), equalTo(true));
assertThat(getResult.docIdAndVersion(), notNullValue()); assertThat(getResult.docIdAndVersion(), notNullValue());
getResult.release(); }
// refresh and it should be updated // refresh and it should be updated
engine.refresh("test"); engine.refresh("test");
@ -901,9 +901,9 @@ public class InternalEngineTests extends EngineTestCase {
searchResult.close(); searchResult.close();
// but, get should not see it (in realtime) // but, get should not see it (in realtime)
getResult = engine.get(newGet(true, doc), searcherFactory); try (Engine.GetResult getResult = engine.get(newGet(true, doc), searcherFactory)) {
assertThat(getResult.exists(), equalTo(false)); assertThat(getResult.exists(), equalTo(false));
getResult.release(); }
// refresh and it should be deleted // refresh and it should be deleted
engine.refresh("test"); engine.refresh("test");
@ -941,10 +941,10 @@ public class InternalEngineTests extends EngineTestCase {
engine.flush(); engine.flush();
// and, verify get (in real time) // and, verify get (in real time)
getResult = engine.get(newGet(true, doc), searcherFactory); try (Engine.GetResult getResult = engine.get(newGet(true, doc), searcherFactory)) {
assertThat(getResult.exists(), equalTo(true)); assertThat(getResult.exists(), equalTo(true));
assertThat(getResult.docIdAndVersion(), notNullValue()); assertThat(getResult.docIdAndVersion(), notNullValue());
getResult.release(); }
// make sure we can still work with the engine // make sure we can still work with the engine
// now do an update // now do an update
@ -4156,7 +4156,7 @@ public class InternalEngineTests extends EngineTestCase {
new Term("_id", parsedDocument.id()), new Term("_id", parsedDocument.id()),
parsedDocument, parsedDocument,
SequenceNumbers.UNASSIGNED_SEQ_NO, SequenceNumbers.UNASSIGNED_SEQ_NO,
(long) randomIntBetween(1, 8), randomIntBetween(1, 8),
Versions.MATCH_ANY, Versions.MATCH_ANY,
VersionType.INTERNAL, VersionType.INTERNAL,
Engine.Operation.Origin.PRIMARY, Engine.Operation.Origin.PRIMARY,
@ -4172,7 +4172,7 @@ public class InternalEngineTests extends EngineTestCase {
id, id,
new Term("_id", parsedDocument.id()), new Term("_id", parsedDocument.id()),
SequenceNumbers.UNASSIGNED_SEQ_NO, SequenceNumbers.UNASSIGNED_SEQ_NO,
(long) randomIntBetween(1, 8), randomIntBetween(1, 8),
Versions.MATCH_ANY, Versions.MATCH_ANY,
VersionType.INTERNAL, VersionType.INTERNAL,
Engine.Operation.Origin.PRIMARY, Engine.Operation.Origin.PRIMARY,

View File

@ -1861,10 +1861,11 @@ public class IndexShardTests extends IndexShardTestCase {
indexDoc(shard, "_doc", "1", "{\"foobar\" : \"bar\"}"); indexDoc(shard, "_doc", "1", "{\"foobar\" : \"bar\"}");
shard.refresh("test"); shard.refresh("test");
Engine.GetResult getResult = shard.get(new Engine.Get(false, false, "test", "1", new Term(IdFieldMapper.NAME, Uid.encodeId("1")))); try (Engine.GetResult getResult = shard
assertTrue(getResult.exists()); .get(new Engine.Get(false, false, "test", "1", new Term(IdFieldMapper.NAME, Uid.encodeId("1"))))) {
assertNotNull(getResult.searcher()); assertTrue(getResult.exists());
getResult.release(); assertNotNull(getResult.searcher());
}
try (Engine.Searcher searcher = shard.acquireSearcher("test")) { try (Engine.Searcher searcher = shard.acquireSearcher("test")) {
TopDocs search = searcher.searcher().search(new TermQuery(new Term("foo", "bar")), 10); TopDocs search = searcher.searcher().search(new TermQuery(new Term("foo", "bar")), 10);
assertEquals(search.totalHits, 1); assertEquals(search.totalHits, 1);
@ -1895,11 +1896,12 @@ public class IndexShardTests extends IndexShardTestCase {
search = searcher.searcher().search(new TermQuery(new Term("foobar", "bar")), 10); search = searcher.searcher().search(new TermQuery(new Term("foobar", "bar")), 10);
assertEquals(search.totalHits, 1); assertEquals(search.totalHits, 1);
} }
getResult = newShard.get(new Engine.Get(false, false, "test", "1", new Term(IdFieldMapper.NAME, Uid.encodeId("1")))); try (Engine.GetResult getResult = newShard
assertTrue(getResult.exists()); .get(new Engine.Get(false, false, "test", "1", new Term(IdFieldMapper.NAME, Uid.encodeId("1"))))) {
assertNotNull(getResult.searcher()); // make sure get uses the wrapped reader assertTrue(getResult.exists());
assertTrue(getResult.searcher().reader() instanceof FieldMaskingReader); assertNotNull(getResult.searcher()); // make sure get uses the wrapped reader
getResult.release(); assertTrue(getResult.searcher().reader() instanceof FieldMaskingReader);
}
closeShards(newShard); closeShards(newShard);
} }

View File

@ -52,6 +52,7 @@ import java.util.concurrent.atomic.AtomicLong;
import java.util.concurrent.atomic.AtomicReference; import java.util.concurrent.atomic.AtomicReference;
import static org.hamcrest.Matchers.empty; import static org.hamcrest.Matchers.empty;
import static org.hamcrest.Matchers.is;
import static org.hamcrest.Matchers.sameInstance; import static org.hamcrest.Matchers.sameInstance;
import static org.hamcrest.core.IsEqual.equalTo; import static org.hamcrest.core.IsEqual.equalTo;
import static org.mockito.Matchers.any; import static org.mockito.Matchers.any;
@ -73,7 +74,6 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
threadPool = new TestThreadPool(getClass().getName()); threadPool = new TestThreadPool(getClass().getName());
} }
@Override @Override
@After @After
public void tearDown() throws Exception { public void tearDown() throws Exception {
@ -95,7 +95,7 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
return state.build(); return state.build();
} }
public void testStartTask() throws Exception { public void testStartTask() {
PersistentTasksService persistentTasksService = mock(PersistentTasksService.class); PersistentTasksService persistentTasksService = mock(PersistentTasksService.class);
@SuppressWarnings("unchecked") PersistentTasksExecutor<TestParams> action = mock(PersistentTasksExecutor.class); @SuppressWarnings("unchecked") PersistentTasksExecutor<TestParams> action = mock(PersistentTasksExecutor.class);
when(action.getExecutor()).thenReturn(ThreadPool.Names.SAME); when(action.getExecutor()).thenReturn(ThreadPool.Names.SAME);
@ -131,8 +131,8 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
if (added == false) { if (added == false) {
logger.info("No local node action was added"); logger.info("No local node action was added");
} }
MetaData.Builder metaData = MetaData.builder(state.metaData()); MetaData.Builder metaData = MetaData.builder(state.metaData());
metaData.putCustom(PersistentTasksCustomMetaData.TYPE, tasks.build()); metaData.putCustom(PersistentTasksCustomMetaData.TYPE, tasks.build());
ClusterState newClusterState = ClusterState.builder(state).metaData(metaData).build(); ClusterState newClusterState = ClusterState.builder(state).metaData(metaData).build();
@ -149,6 +149,7 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
// Make sure action wasn't called again // Make sure action wasn't called again
assertThat(executor.executions.size(), equalTo(1)); assertThat(executor.executions.size(), equalTo(1));
assertThat(executor.get(0).task.isCompleted(), is(false));
// Start another task on this node // Start another task on this node
state = newClusterState; state = newClusterState;
@ -157,10 +158,15 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
// Make sure action was called this time // Make sure action was called this time
assertThat(executor.size(), equalTo(2)); assertThat(executor.size(), equalTo(2));
assertThat(executor.get(1).task.isCompleted(), is(false));
// Finish both tasks // Finish both tasks
executor.get(0).task.markAsFailed(new RuntimeException()); executor.get(0).task.markAsFailed(new RuntimeException());
executor.get(1).task.markAsCompleted(); executor.get(1).task.markAsCompleted();
assertThat(executor.get(0).task.isCompleted(), is(true));
assertThat(executor.get(1).task.isCompleted(), is(true));
String failedTaskId = executor.get(0).task.getPersistentTaskId(); String failedTaskId = executor.get(0).task.getPersistentTaskId();
String finishedTaskId = executor.get(1).task.getPersistentTaskId(); String finishedTaskId = executor.get(1).task.getPersistentTaskId();
executor.clear(); executor.clear();
@ -186,7 +192,6 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
// Make sure action was only allocated on this node once // Make sure action was only allocated on this node once
assertThat(executor.size(), equalTo(1)); assertThat(executor.size(), equalTo(1));
} }
} }
public void testParamsStatusAndNodeTaskAreDelegated() throws Exception { public void testParamsStatusAndNodeTaskAreDelegated() throws Exception {
@ -300,7 +305,6 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
// Check the the task is now removed from task manager // Check the the task is now removed from task manager
assertThat(taskManager.getTasks().values(), empty()); assertThat(taskManager.getTasks().values(), empty());
} }
private <Params extends PersistentTaskParams> ClusterState addTask(ClusterState state, String action, Params params, private <Params extends PersistentTaskParams> ClusterState addTask(ClusterState state, String action, Params params,

View File

@ -23,6 +23,7 @@ import org.apache.log4j.Level;
import org.apache.lucene.util.Constants; import org.apache.lucene.util.Constants;
import org.apache.lucene.util.LuceneTestCase; import org.apache.lucene.util.LuceneTestCase;
import org.elasticsearch.Version; import org.elasticsearch.Version;
import org.elasticsearch.bootstrap.JarHell;
import org.elasticsearch.common.io.PathUtils; import org.elasticsearch.common.io.PathUtils;
import org.elasticsearch.common.settings.Settings; import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.env.Environment; import org.elasticsearch.env.Environment;
@ -443,7 +444,7 @@ public class PluginsServiceTests extends ESTestCase {
"MyPlugin", Collections.singletonList("dep"), false); "MyPlugin", Collections.singletonList("dep"), false);
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir); PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
IllegalStateException e = expectThrows(IllegalStateException.class, () -> IllegalStateException e = expectThrows(IllegalStateException.class, () ->
PluginsService.checkBundleJarHell(bundle, transitiveDeps)); PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps));
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage()); assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
assertThat(e.getCause().getMessage(), containsString("jar hell! duplicate codebases with extended plugin")); assertThat(e.getCause().getMessage(), containsString("jar hell! duplicate codebases with extended plugin"));
} }
@ -462,7 +463,7 @@ public class PluginsServiceTests extends ESTestCase {
"MyPlugin", Arrays.asList("dep1", "dep2"), false); "MyPlugin", Arrays.asList("dep1", "dep2"), false);
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir); PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
IllegalStateException e = expectThrows(IllegalStateException.class, () -> IllegalStateException e = expectThrows(IllegalStateException.class, () ->
PluginsService.checkBundleJarHell(bundle, transitiveDeps)); PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps));
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage()); assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
assertThat(e.getCause().getMessage(), containsString("jar hell!")); assertThat(e.getCause().getMessage(), containsString("jar hell!"));
assertThat(e.getCause().getMessage(), containsString("duplicate codebases")); assertThat(e.getCause().getMessage(), containsString("duplicate codebases"));
@ -479,7 +480,7 @@ public class PluginsServiceTests extends ESTestCase {
"MyPlugin", Collections.emptyList(), false); "MyPlugin", Collections.emptyList(), false);
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir); PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
IllegalStateException e = expectThrows(IllegalStateException.class, () -> IllegalStateException e = expectThrows(IllegalStateException.class, () ->
PluginsService.checkBundleJarHell(bundle, new HashMap<>())); PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, new HashMap<>()));
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage()); assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
assertThat(e.getCause().getMessage(), containsString("jar hell!")); assertThat(e.getCause().getMessage(), containsString("jar hell!"));
assertThat(e.getCause().getMessage(), containsString("Level")); assertThat(e.getCause().getMessage(), containsString("Level"));
@ -498,7 +499,7 @@ public class PluginsServiceTests extends ESTestCase {
"MyPlugin", Collections.singletonList("dep"), false); "MyPlugin", Collections.singletonList("dep"), false);
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir); PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
IllegalStateException e = expectThrows(IllegalStateException.class, () -> IllegalStateException e = expectThrows(IllegalStateException.class, () ->
PluginsService.checkBundleJarHell(bundle, transitiveDeps)); PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps));
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage()); assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
assertThat(e.getCause().getMessage(), containsString("jar hell!")); assertThat(e.getCause().getMessage(), containsString("jar hell!"));
assertThat(e.getCause().getMessage(), containsString("DummyClass1")); assertThat(e.getCause().getMessage(), containsString("DummyClass1"));
@ -521,7 +522,7 @@ public class PluginsServiceTests extends ESTestCase {
"MyPlugin", Arrays.asList("dep1", "dep2"), false); "MyPlugin", Arrays.asList("dep1", "dep2"), false);
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir); PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
IllegalStateException e = expectThrows(IllegalStateException.class, () -> IllegalStateException e = expectThrows(IllegalStateException.class, () ->
PluginsService.checkBundleJarHell(bundle, transitiveDeps)); PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps));
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage()); assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
assertThat(e.getCause().getMessage(), containsString("jar hell!")); assertThat(e.getCause().getMessage(), containsString("jar hell!"));
assertThat(e.getCause().getMessage(), containsString("DummyClass2")); assertThat(e.getCause().getMessage(), containsString("DummyClass2"));
@ -543,7 +544,7 @@ public class PluginsServiceTests extends ESTestCase {
PluginInfo info1 = new PluginInfo("myplugin", "desc", "1.0", Version.CURRENT, "1.8", PluginInfo info1 = new PluginInfo("myplugin", "desc", "1.0", Version.CURRENT, "1.8",
"MyPlugin", Arrays.asList("dep1", "dep2"), false); "MyPlugin", Arrays.asList("dep1", "dep2"), false);
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir); PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
PluginsService.checkBundleJarHell(bundle, transitiveDeps); PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps);
Set<URL> deps = transitiveDeps.get("myplugin"); Set<URL> deps = transitiveDeps.get("myplugin");
assertNotNull(deps); assertNotNull(deps);
assertThat(deps, containsInAnyOrder(pluginJar.toUri().toURL(), dep1Jar.toUri().toURL(), dep2Jar.toUri().toURL())); assertThat(deps, containsInAnyOrder(pluginJar.toUri().toURL(), dep1Jar.toUri().toURL(), dep2Jar.toUri().toURL()));

View File

@ -30,14 +30,14 @@ import org.elasticsearch.index.fielddata.ScriptDocValues;
import org.elasticsearch.plugins.Plugin; import org.elasticsearch.plugins.Plugin;
import org.elasticsearch.plugins.ScriptPlugin; import org.elasticsearch.plugins.ScriptPlugin;
import org.elasticsearch.script.ExplainableSearchScript; import org.elasticsearch.script.ExplainableSearchScript;
import org.elasticsearch.script.ScoreScript;
import org.elasticsearch.script.Script; import org.elasticsearch.script.Script;
import org.elasticsearch.script.ScriptContext; import org.elasticsearch.script.ScriptContext;
import org.elasticsearch.script.ScriptEngine; import org.elasticsearch.script.ScriptEngine;
import org.elasticsearch.script.ScriptType; import org.elasticsearch.script.ScriptType;
import org.elasticsearch.script.SearchScript;
import org.elasticsearch.search.SearchHit; import org.elasticsearch.search.SearchHit;
import org.elasticsearch.search.SearchHits; import org.elasticsearch.search.SearchHits;
import org.elasticsearch.search.lookup.LeafDocLookup; import org.elasticsearch.search.lookup.SearchLookup;
import org.elasticsearch.test.ESIntegTestCase; import org.elasticsearch.test.ESIntegTestCase;
import org.elasticsearch.test.ESIntegTestCase.ClusterScope; import org.elasticsearch.test.ESIntegTestCase.ClusterScope;
import org.elasticsearch.test.ESIntegTestCase.Scope; import org.elasticsearch.test.ESIntegTestCase.Scope;
@ -76,16 +76,17 @@ public class ExplainableScriptIT extends ESIntegTestCase {
@Override @Override
public <T> T compile(String scriptName, String scriptSource, ScriptContext<T> context, Map<String, String> params) { public <T> T compile(String scriptName, String scriptSource, ScriptContext<T> context, Map<String, String> params) {
assert scriptSource.equals("explainable_script"); assert scriptSource.equals("explainable_script");
assert context == SearchScript.SCRIPT_SCORE_CONTEXT; assert context == ScoreScript.CONTEXT;
SearchScript.Factory factory = (p, lookup) -> new SearchScript.LeafFactory() { ScoreScript.Factory factory = (params1, lookup) -> new ScoreScript.LeafFactory() {
@Override
public SearchScript newInstance(LeafReaderContext context) throws IOException {
return new MyScript(lookup.doc().getLeafDocLookup(context));
}
@Override @Override
public boolean needs_score() { public boolean needs_score() {
return false; return false;
} }
@Override
public ScoreScript newInstance(LeafReaderContext ctx) throws IOException {
return new MyScript(params1, lookup, ctx);
}
}; };
return context.factoryClazz.cast(factory); return context.factoryClazz.cast(factory);
} }
@ -93,28 +94,21 @@ public class ExplainableScriptIT extends ESIntegTestCase {
} }
} }
static class MyScript extends SearchScript implements ExplainableSearchScript { static class MyScript extends ScoreScript implements ExplainableSearchScript {
LeafDocLookup docLookup;
MyScript(LeafDocLookup docLookup) { MyScript(Map<String, Object> params, SearchLookup lookup, LeafReaderContext leafContext) {
super(null, null, null); super(params, lookup, leafContext);
this.docLookup = docLookup;
} }
@Override
public void setDocument(int doc) {
docLookup.setDocument(doc);
}
@Override @Override
public Explanation explain(Explanation subQueryScore) throws IOException { public Explanation explain(Explanation subQueryScore) throws IOException {
Explanation scoreExp = Explanation.match(subQueryScore.getValue(), "_score: ", subQueryScore); Explanation scoreExp = Explanation.match(subQueryScore.getValue(), "_score: ", subQueryScore);
return Explanation.match((float) (runAsDouble()), "This script returned " + runAsDouble(), scoreExp); return Explanation.match((float) (execute()), "This script returned " + execute(), scoreExp);
} }
@Override @Override
public double runAsDouble() { public double execute() {
return ((Number) ((ScriptDocValues) docLookup.get("number_field")).getValues().get(0)).doubleValue(); return ((Number) ((ScriptDocValues) getDoc().get("number_field")).getValues().get(0)).doubleValue();
} }
} }

View File

@ -23,10 +23,12 @@ import com.carrotsearch.hppc.IntHashSet;
import com.carrotsearch.hppc.IntSet; import com.carrotsearch.hppc.IntSet;
import org.elasticsearch.action.ActionFuture; import org.elasticsearch.action.ActionFuture;
import org.elasticsearch.action.admin.cluster.repositories.put.PutRepositoryResponse; import org.elasticsearch.action.admin.cluster.repositories.put.PutRepositoryResponse;
import org.elasticsearch.action.admin.cluster.repositories.verify.VerifyRepositoryResponse;
import org.elasticsearch.action.admin.cluster.snapshots.create.CreateSnapshotResponse; import org.elasticsearch.action.admin.cluster.snapshots.create.CreateSnapshotResponse;
import org.elasticsearch.action.admin.cluster.snapshots.delete.DeleteSnapshotResponse; import org.elasticsearch.action.admin.cluster.snapshots.delete.DeleteSnapshotResponse;
import org.elasticsearch.action.admin.cluster.snapshots.get.GetSnapshotsResponse; import org.elasticsearch.action.admin.cluster.snapshots.get.GetSnapshotsResponse;
import org.elasticsearch.action.admin.cluster.snapshots.restore.RestoreSnapshotResponse; import org.elasticsearch.action.admin.cluster.snapshots.restore.RestoreSnapshotResponse;
import org.elasticsearch.action.admin.cluster.snapshots.status.SnapshotStats;
import org.elasticsearch.action.admin.cluster.snapshots.status.SnapshotStatus; import org.elasticsearch.action.admin.cluster.snapshots.status.SnapshotStatus;
import org.elasticsearch.action.admin.cluster.snapshots.status.SnapshotsStatusResponse; import org.elasticsearch.action.admin.cluster.snapshots.status.SnapshotsStatusResponse;
import org.elasticsearch.action.admin.cluster.state.ClusterStateResponse; import org.elasticsearch.action.admin.cluster.state.ClusterStateResponse;
@ -83,7 +85,12 @@ import org.elasticsearch.test.TestCustomMetaData;
import org.elasticsearch.test.rest.FakeRestRequest; import org.elasticsearch.test.rest.FakeRestRequest;
import java.io.IOException; import java.io.IOException;
import java.io.UncheckedIOException;
import java.nio.file.FileVisitResult;
import java.nio.file.Files;
import java.nio.file.Path; import java.nio.file.Path;
import java.nio.file.SimpleFileVisitor;
import java.nio.file.attribute.BasicFileAttributes;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.Arrays; import java.util.Arrays;
import java.util.Collection; import java.util.Collection;
@ -102,6 +109,7 @@ import static org.hamcrest.Matchers.equalTo;
import static org.hamcrest.Matchers.greaterThan; import static org.hamcrest.Matchers.greaterThan;
import static org.hamcrest.Matchers.greaterThanOrEqualTo; import static org.hamcrest.Matchers.greaterThanOrEqualTo;
import static org.hamcrest.Matchers.hasSize; import static org.hamcrest.Matchers.hasSize;
import static org.hamcrest.Matchers.is;
import static org.hamcrest.Matchers.lessThan; import static org.hamcrest.Matchers.lessThan;
import static org.hamcrest.Matchers.not; import static org.hamcrest.Matchers.not;
import static org.hamcrest.Matchers.notNullValue; import static org.hamcrest.Matchers.notNullValue;
@ -1019,6 +1027,129 @@ public class DedicatedClusterSnapshotRestoreIT extends AbstractSnapshotIntegTest
assertThat(snapshots.get(0).getState().completed(), equalTo(true)); assertThat(snapshots.get(0).getState().completed(), equalTo(true));
} }
public void testSnapshotTotalAndIncrementalSizes() throws IOException {
Client client = client();
final String indexName = "test-blocks-1";
final String repositoryName = "repo-" + indexName;
final String snapshot0 = "snapshot-0";
final String snapshot1 = "snapshot-1";
createIndex(indexName);
int docs = between(10, 100);
for (int i = 0; i < docs; i++) {
client.prepareIndex(indexName, "type").setSource("test", "init").execute().actionGet();
}
logger.info("--> register a repository");
final Path repoPath = randomRepoPath();
assertAcked(client.admin().cluster().preparePutRepository(repositoryName)
.setType("fs")
.setSettings(Settings.builder().put("location", repoPath)));
logger.info("--> create a snapshot");
client.admin().cluster().prepareCreateSnapshot(repositoryName, snapshot0)
.setIncludeGlobalState(true)
.setWaitForCompletion(true)
.get();
SnapshotsStatusResponse response = client.admin().cluster().prepareSnapshotStatus(repositoryName)
.setSnapshots(snapshot0)
.get();
List<SnapshotStatus> snapshots = response.getSnapshots();
List<Path> snapshot0Files = scanSnapshotFolder(repoPath);
assertThat(snapshots, hasSize(1));
final int snapshot0FileCount = snapshot0Files.size();
final long snapshot0FileSize = calculateTotalFilesSize(snapshot0Files);
SnapshotStats stats = snapshots.get(0).getStats();
assertThat(stats.getTotalFileCount(), is(snapshot0FileCount));
assertThat(stats.getTotalSize(), is(snapshot0FileSize));
assertThat(stats.getIncrementalFileCount(), equalTo(snapshot0FileCount));
assertThat(stats.getIncrementalSize(), equalTo(snapshot0FileSize));
assertThat(stats.getIncrementalFileCount(), equalTo(stats.getProcessedFileCount()));
assertThat(stats.getIncrementalSize(), equalTo(stats.getProcessedSize()));
// add few docs - less than initially
docs = between(1, 5);
for (int i = 0; i < docs; i++) {
client.prepareIndex(indexName, "type").setSource("test", "test" + i).execute().actionGet();
}
// create another snapshot
// total size has to grow and has to be equal to files on fs
assertThat(client.admin().cluster()
.prepareCreateSnapshot(repositoryName, snapshot1)
.setWaitForCompletion(true).get().status(),
equalTo(RestStatus.OK));
// drop 1st one to avoid miscalculation as snapshot reuses some files of prev snapshot
assertTrue(client.admin().cluster()
.prepareDeleteSnapshot(repositoryName, snapshot0)
.get().isAcknowledged());
response = client.admin().cluster().prepareSnapshotStatus(repositoryName)
.setSnapshots(snapshot1)
.get();
final List<Path> snapshot1Files = scanSnapshotFolder(repoPath);
final int snapshot1FileCount = snapshot1Files.size();
final long snapshot1FileSize = calculateTotalFilesSize(snapshot1Files);
snapshots = response.getSnapshots();
SnapshotStats anotherStats = snapshots.get(0).getStats();
ArrayList<Path> snapshotFilesDiff = new ArrayList<>(snapshot1Files);
snapshotFilesDiff.removeAll(snapshot0Files);
assertThat(anotherStats.getIncrementalFileCount(), equalTo(snapshotFilesDiff.size()));
assertThat(anotherStats.getIncrementalSize(), equalTo(calculateTotalFilesSize(snapshotFilesDiff)));
assertThat(anotherStats.getIncrementalFileCount(), equalTo(anotherStats.getProcessedFileCount()));
assertThat(anotherStats.getIncrementalSize(), equalTo(anotherStats.getProcessedSize()));
assertThat(stats.getTotalSize(), lessThan(anotherStats.getTotalSize()));
assertThat(stats.getTotalFileCount(), lessThan(anotherStats.getTotalFileCount()));
assertThat(anotherStats.getTotalFileCount(), is(snapshot1FileCount));
assertThat(anotherStats.getTotalSize(), is(snapshot1FileSize));
}
private long calculateTotalFilesSize(List<Path> files) {
return files.stream().mapToLong(f -> {
try {
return Files.size(f);
} catch (IOException e) {
throw new UncheckedIOException(e);
}
}).sum();
}
private List<Path> scanSnapshotFolder(Path repoPath) throws IOException {
List<Path> files = new ArrayList<>();
Files.walkFileTree(repoPath, new SimpleFileVisitor<Path>(){
@Override
public FileVisitResult visitFile(Path file, BasicFileAttributes attrs) throws IOException {
if (file.getFileName().toString().startsWith("__")){
files.add(file);
}
return super.visitFile(file, attrs);
}
}
);
return files;
}
public static class SnapshottableMetadata extends TestCustomMetaData { public static class SnapshottableMetadata extends TestCustomMetaData {
public static final String TYPE = "test_snapshottable"; public static final String TYPE = "test_snapshottable";

View File

@ -2066,7 +2066,7 @@ public class SharedClusterSnapshotRestoreIT extends AbstractSnapshotIntegTestCas
SnapshotStatus snapshotStatus = client.admin().cluster().prepareSnapshotStatus("test-repo").setSnapshots("test").get().getSnapshots().get(0); SnapshotStatus snapshotStatus = client.admin().cluster().prepareSnapshotStatus("test-repo").setSnapshots("test").get().getSnapshots().get(0);
List<SnapshotIndexShardStatus> shards = snapshotStatus.getShards(); List<SnapshotIndexShardStatus> shards = snapshotStatus.getShards();
for (SnapshotIndexShardStatus status : shards) { for (SnapshotIndexShardStatus status : shards) {
assertThat(status.getStats().getProcessedFiles(), greaterThan(1)); assertThat(status.getStats().getProcessedFileCount(), greaterThan(1));
} }
} }
@ -2078,7 +2078,7 @@ public class SharedClusterSnapshotRestoreIT extends AbstractSnapshotIntegTestCas
SnapshotStatus snapshotStatus = client.admin().cluster().prepareSnapshotStatus("test-repo").setSnapshots("test-1").get().getSnapshots().get(0); SnapshotStatus snapshotStatus = client.admin().cluster().prepareSnapshotStatus("test-repo").setSnapshots("test-1").get().getSnapshots().get(0);
List<SnapshotIndexShardStatus> shards = snapshotStatus.getShards(); List<SnapshotIndexShardStatus> shards = snapshotStatus.getShards();
for (SnapshotIndexShardStatus status : shards) { for (SnapshotIndexShardStatus status : shards) {
assertThat(status.getStats().getProcessedFiles(), equalTo(0)); assertThat(status.getStats().getProcessedFileCount(), equalTo(0));
} }
} }
@ -2091,7 +2091,7 @@ public class SharedClusterSnapshotRestoreIT extends AbstractSnapshotIntegTestCas
SnapshotStatus snapshotStatus = client.admin().cluster().prepareSnapshotStatus("test-repo").setSnapshots("test-2").get().getSnapshots().get(0); SnapshotStatus snapshotStatus = client.admin().cluster().prepareSnapshotStatus("test-repo").setSnapshots("test-2").get().getSnapshots().get(0);
List<SnapshotIndexShardStatus> shards = snapshotStatus.getShards(); List<SnapshotIndexShardStatus> shards = snapshotStatus.getShards();
for (SnapshotIndexShardStatus status : shards) { for (SnapshotIndexShardStatus status : shards) {
assertThat(status.getStats().getProcessedFiles(), equalTo(2)); // we flush before the snapshot such that we have to process the segments_N files plus the .del file assertThat(status.getStats().getProcessedFileCount(), equalTo(2)); // we flush before the snapshot such that we have to process the segments_N files plus the .del file
} }
} }
} }

View File

@ -632,7 +632,7 @@ public abstract class IndexShardTestCase extends ESTestCase {
final IndexShardSnapshotStatus.Copy lastSnapshotStatus = snapshotStatus.asCopy(); final IndexShardSnapshotStatus.Copy lastSnapshotStatus = snapshotStatus.asCopy();
assertEquals(IndexShardSnapshotStatus.Stage.DONE, lastSnapshotStatus.getStage()); assertEquals(IndexShardSnapshotStatus.Stage.DONE, lastSnapshotStatus.getStage());
assertEquals(shard.snapshotStoreMetadata().size(), lastSnapshotStatus.getNumberOfFiles()); assertEquals(shard.snapshotStoreMetadata().size(), lastSnapshotStatus.getTotalFileCount());
assertNull(lastSnapshotStatus.getFailure()); assertNull(lastSnapshotStatus.getFailure());
} }

View File

@ -25,7 +25,6 @@ import org.elasticsearch.index.similarity.ScriptedSimilarity.Doc;
import org.elasticsearch.index.similarity.ScriptedSimilarity.Field; import org.elasticsearch.index.similarity.ScriptedSimilarity.Field;
import org.elasticsearch.index.similarity.ScriptedSimilarity.Query; import org.elasticsearch.index.similarity.ScriptedSimilarity.Query;
import org.elasticsearch.index.similarity.ScriptedSimilarity.Term; import org.elasticsearch.index.similarity.ScriptedSimilarity.Term;
import org.elasticsearch.index.similarity.SimilarityService;
import org.elasticsearch.search.aggregations.pipeline.movfn.MovingFunctionScript; import org.elasticsearch.search.aggregations.pipeline.movfn.MovingFunctionScript;
import org.elasticsearch.search.aggregations.pipeline.movfn.MovingFunctions; import org.elasticsearch.search.aggregations.pipeline.movfn.MovingFunctions;
import org.elasticsearch.search.lookup.LeafSearchLookup; import org.elasticsearch.search.lookup.LeafSearchLookup;
@ -36,7 +35,6 @@ import java.util.Collections;
import java.util.HashMap; import java.util.HashMap;
import java.util.Map; import java.util.Map;
import java.util.function.Function; import java.util.function.Function;
import java.util.function.Predicate;
import static java.util.Collections.emptyMap; import static java.util.Collections.emptyMap;
@ -114,6 +112,9 @@ public class MockScriptEngine implements ScriptEngine {
} else if (context.instanceClazz.equals(MovingFunctionScript.class)) { } else if (context.instanceClazz.equals(MovingFunctionScript.class)) {
MovingFunctionScript.Factory factory = mockCompiled::createMovingFunctionScript; MovingFunctionScript.Factory factory = mockCompiled::createMovingFunctionScript;
return context.factoryClazz.cast(factory); return context.factoryClazz.cast(factory);
} else if (context.instanceClazz.equals(ScoreScript.class)) {
ScoreScript.Factory factory = new MockScoreScript(script);
return context.factoryClazz.cast(factory);
} }
throw new IllegalArgumentException("mock script engine does not know how to handle context [" + context.name + "]"); throw new IllegalArgumentException("mock script engine does not know how to handle context [" + context.name + "]");
} }
@ -342,5 +343,45 @@ public class MockScriptEngine implements ScriptEngine {
return MovingFunctions.unweightedAvg(values); return MovingFunctions.unweightedAvg(values);
} }
} }
public class MockScoreScript implements ScoreScript.Factory {
private final Function<Map<String, Object>, Object> scripts;
MockScoreScript(Function<Map<String, Object>, Object> scripts) {
this.scripts = scripts;
}
@Override
public ScoreScript.LeafFactory newFactory(Map<String, Object> params, SearchLookup lookup) {
return new ScoreScript.LeafFactory() {
@Override
public boolean needs_score() {
return true;
}
@Override
public ScoreScript newInstance(LeafReaderContext ctx) throws IOException {
Scorer[] scorerHolder = new Scorer[1];
return new ScoreScript(params, lookup, ctx) {
@Override
public double execute() {
Map<String, Object> vars = new HashMap<>(getParams());
vars.put("doc", getDoc());
if (scorerHolder[0] != null) {
vars.put("_score", new ScoreAccessor(scorerHolder[0]));
}
return ((Number) scripts.apply(vars)).doubleValue();
}
@Override
public void setScorer(Scorer scorer) {
scorerHolder[0] = scorer;
}
};
}
};
}
}
} }

View File

@ -123,13 +123,15 @@ public class XPackInfoResponse extends ActionResponse {
@Override @Override
public XContentBuilder toXContent(XContentBuilder builder, Params params) throws IOException { public XContentBuilder toXContent(XContentBuilder builder, Params params) throws IOException {
return builder.startObject() builder.startObject()
.field("uid", uid) .field("uid", uid)
.field("type", type) .field("type", type)
.field("mode", mode) .field("mode", mode)
.field("status", status.label()) .field("status", status.label());
.timeField("expiry_date_in_millis", "expiry_date", expiryDate) if (expiryDate != LicenseService.BASIC_SELF_GENERATED_LICENSE_EXPIRATION_MILLIS) {
.endObject(); builder.timeField("expiry_date_in_millis", "expiry_date", expiryDate);
}
return builder.endObject();
} }
public void writeTo(StreamOutput out) throws IOException { public void writeTo(StreamOutput out) throws IOException {

View File

@ -52,24 +52,20 @@ dependencies {
compile (xpackProject('plugin:sql:sql-shared-client')) { compile (xpackProject('plugin:sql:sql-shared-client')) {
transitive = false transitive = false
} }
compile (xpackProject('plugin:sql:sql-proto')) { compile (xpackProject('plugin:sql:sql-shared-proto')) {
transitive = false transitive = false
} }
} else { } else {
bundled (xpackProject('plugin:sql:sql-shared-client')) { bundled (xpackProject('plugin:sql:sql-shared-client')) {
transitive = false transitive = false
} }
bundled (xpackProject('plugin:sql:sql-proto')) { bundled (xpackProject('plugin:sql:sql-shared-proto')) {
transitive = false transitive = false
} }
} }
compile (project(':server')) {
transitive = false
}
compile (project(':libs:x-content')) { compile (project(':libs:x-content')) {
transitive = false transitive = false
} }
compile "org.apache.lucene:lucene-core:${versions.lucene}"
compile 'joda-time:joda-time:2.9.9' compile 'joda-time:joda-time:2.9.9'
compile project(':libs:elasticsearch-core') compile project(':libs:elasticsearch-core')
runtime "com.fasterxml.jackson.core:jackson-core:${versions.jackson}" runtime "com.fasterxml.jackson.core:jackson-core:${versions.jackson}"
@ -80,15 +76,13 @@ dependencies {
} }
dependencyLicenses { dependencyLicenses {
mapping from: /sql-proto.*/, to: 'elasticsearch' mapping from: /sql-shared-proto.*/, to: 'elasticsearch'
mapping from: /sql-shared-client.*/, to: 'elasticsearch' mapping from: /sql-shared-client.*/, to: 'elasticsearch'
mapping from: /jackson-.*/, to: 'jackson' mapping from: /jackson-.*/, to: 'jackson'
mapping from: /lucene-.*/, to: 'lucene'
mapping from: /elasticsearch-core.*/, to: 'elasticsearch' mapping from: /elasticsearch-core.*/, to: 'elasticsearch'
ignoreSha 'sql-proto' ignoreSha 'sql-shared-proto'
ignoreSha 'sql-shared-client' ignoreSha 'sql-shared-client'
ignoreSha 'elasticsearch' ignoreSha 'elasticsearch'
ignoreSha 'elasticsearch-core'
} }
/* /*

View File

@ -1,475 +0,0 @@
Apache License
Version 2.0, January 2004
http://www.apache.org/licenses/
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
1. Definitions.
"License" shall mean the terms and conditions for use, reproduction,
and distribution as defined by Sections 1 through 9 of this document.
"Licensor" shall mean the copyright owner or entity authorized by
the copyright owner that is granting the License.
"Legal Entity" shall mean the union of the acting entity and all
other entities that control, are controlled by, or are under common
control with that entity. For the purposes of this definition,
"control" means (i) the power, direct or indirect, to cause the
direction or management of such entity, whether by contract or
otherwise, or (ii) ownership of fifty percent (50%) or more of the
outstanding shares, or (iii) beneficial ownership of such entity.
"You" (or "Your") shall mean an individual or Legal Entity
exercising permissions granted by this License.
"Source" form shall mean the preferred form for making modifications,
including but not limited to software source code, documentation
source, and configuration files.
"Object" form shall mean any form resulting from mechanical
transformation or translation of a Source form, including but
not limited to compiled object code, generated documentation,
and conversions to other media types.
"Work" shall mean the work of authorship, whether in Source or
Object form, made available under the License, as indicated by a
copyright notice that is included in or attached to the work
(an example is provided in the Appendix below).
"Derivative Works" shall mean any work, whether in Source or Object
form, that is based on (or derived from) the Work and for which the
editorial revisions, annotations, elaborations, or other modifications
represent, as a whole, an original work of authorship. For the purposes
of this License, Derivative Works shall not include works that remain
separable from, or merely link (or bind by name) to the interfaces of,
the Work and Derivative Works thereof.
"Contribution" shall mean any work of authorship, including
the original version of the Work and any modifications or additions
to that Work or Derivative Works thereof, that is intentionally
submitted to Licensor for inclusion in the Work by the copyright owner
or by an individual or Legal Entity authorized to submit on behalf of
the copyright owner. For the purposes of this definition, "submitted"
means any form of electronic, verbal, or written communication sent
to the Licensor or its representatives, including but not limited to
communication on electronic mailing lists, source code control systems,
and issue tracking systems that are managed by, or on behalf of, the
Licensor for the purpose of discussing and improving the Work, but
excluding communication that is conspicuously marked or otherwise
designated in writing by the copyright owner as "Not a Contribution."
"Contributor" shall mean Licensor and any individual or Legal Entity
on behalf of whom a Contribution has been received by Licensor and
subsequently incorporated within the Work.
2. Grant of Copyright License. Subject to the terms and conditions of
this License, each Contributor hereby grants to You a perpetual,
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
copyright license to reproduce, prepare Derivative Works of,
publicly display, publicly perform, sublicense, and distribute the
Work and such Derivative Works in Source or Object form.
3. Grant of Patent License. Subject to the terms and conditions of
this License, each Contributor hereby grants to You a perpetual,
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
(except as stated in this section) patent license to make, have made,
use, offer to sell, sell, import, and otherwise transfer the Work,
where such license applies only to those patent claims licensable
by such Contributor that are necessarily infringed by their
Contribution(s) alone or by combination of their Contribution(s)
with the Work to which such Contribution(s) was submitted. If You
institute patent litigation against any entity (including a
cross-claim or counterclaim in a lawsuit) alleging that the Work
or a Contribution incorporated within the Work constitutes direct
or contributory patent infringement, then any patent licenses
granted to You under this License for that Work shall terminate
as of the date such litigation is filed.
4. Redistribution. You may reproduce and distribute copies of the
Work or Derivative Works thereof in any medium, with or without
modifications, and in Source or Object form, provided that You
meet the following conditions:
(a) You must give any other recipients of the Work or
Derivative Works a copy of this License; and
(b) You must cause any modified files to carry prominent notices
stating that You changed the files; and
(c) You must retain, in the Source form of any Derivative Works
that You distribute, all copyright, patent, trademark, and
attribution notices from the Source form of the Work,
excluding those notices that do not pertain to any part of
the Derivative Works; and
(d) If the Work includes a "NOTICE" text file as part of its
distribution, then any Derivative Works that You distribute must
include a readable copy of the attribution notices contained
within such NOTICE file, excluding those notices that do not
pertain to any part of the Derivative Works, in at least one
of the following places: within a NOTICE text file distributed
as part of the Derivative Works; within the Source form or
documentation, if provided along with the Derivative Works; or,
within a display generated by the Derivative Works, if and
wherever such third-party notices normally appear. The contents
of the NOTICE file are for informational purposes only and
do not modify the License. You may add Your own attribution
notices within Derivative Works that You distribute, alongside
or as an addendum to the NOTICE text from the Work, provided
that such additional attribution notices cannot be construed
as modifying the License.
You may add Your own copyright statement to Your modifications and
may provide additional or different license terms and conditions
for use, reproduction, or distribution of Your modifications, or
for any such Derivative Works as a whole, provided Your use,
reproduction, and distribution of the Work otherwise complies with
the conditions stated in this License.
5. Submission of Contributions. Unless You explicitly state otherwise,
any Contribution intentionally submitted for inclusion in the Work
by You to the Licensor shall be under the terms and conditions of
this License, without any additional terms or conditions.
Notwithstanding the above, nothing herein shall supersede or modify
the terms of any separate license agreement you may have executed
with Licensor regarding such Contributions.
6. Trademarks. This License does not grant permission to use the trade
names, trademarks, service marks, or product names of the Licensor,
except as required for reasonable and customary use in describing the
origin of the Work and reproducing the content of the NOTICE file.
7. Disclaimer of Warranty. Unless required by applicable law or
agreed to in writing, Licensor provides the Work (and each
Contributor provides its Contributions) on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
implied, including, without limitation, any warranties or conditions
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
PARTICULAR PURPOSE. You are solely responsible for determining the
appropriateness of using or redistributing the Work and assume any
risks associated with Your exercise of permissions under this License.
8. Limitation of Liability. In no event and under no legal theory,
whether in tort (including negligence), contract, or otherwise,
unless required by applicable law (such as deliberate and grossly
negligent acts) or agreed to in writing, shall any Contributor be
liable to You for damages, including any direct, indirect, special,
incidental, or consequential damages of any character arising as a
result of this License or out of the use or inability to use the
Work (including but not limited to damages for loss of goodwill,
work stoppage, computer failure or malfunction, or any and all
other commercial damages or losses), even if such Contributor
has been advised of the possibility of such damages.
9. Accepting Warranty or Additional Liability. While redistributing
the Work or Derivative Works thereof, You may choose to offer,
and charge a fee for, acceptance of support, warranty, indemnity,
or other liability obligations and/or rights consistent with this
License. However, in accepting such obligations, You may act only
on Your own behalf and on Your sole responsibility, not on behalf
of any other Contributor, and only if You agree to indemnify,
defend, and hold each Contributor harmless for any liability
incurred by, or claims asserted against, such Contributor by reason
of your accepting any such warranty or additional liability.
END OF TERMS AND CONDITIONS
APPENDIX: How to apply the Apache License to your work.
To apply the Apache License to your work, attach the following
boilerplate notice, with the fields enclosed by brackets "[]"
replaced with your own identifying information. (Don't include
the brackets!) The text should be enclosed in the appropriate
comment syntax for the file format. We also recommend that a
file or class name and description of purpose be included on the
same "printed page" as the copyright notice for easier
identification within third-party archives.
Copyright [yyyy] [name of copyright owner]
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
Some code in core/src/java/org/apache/lucene/util/UnicodeUtil.java was
derived from unicode conversion examples available at
http://www.unicode.org/Public/PROGRAMS/CVTUTF. Here is the copyright
from those sources:
/*
* Copyright 2001-2004 Unicode, Inc.
*
* Disclaimer
*
* This source code is provided as is by Unicode, Inc. No claims are
* made as to fitness for any particular purpose. No warranties of any
* kind are expressed or implied. The recipient agrees to determine
* applicability of information provided. If this file has been
* purchased on magnetic or optical media from Unicode, Inc., the
* sole remedy for any claim will be exchange of defective media
* within 90 days of receipt.
*
* Limitations on Rights to Redistribute This Code
*
* Unicode, Inc. hereby grants the right to freely use the information
* supplied in this file in the creation of products supporting the
* Unicode Standard, and to make copies of this file in any form
* for internal or external distribution as long as this notice
* remains attached.
*/
Some code in core/src/java/org/apache/lucene/util/ArrayUtil.java was
derived from Python 2.4.2 sources available at
http://www.python.org. Full license is here:
http://www.python.org/download/releases/2.4.2/license/
Some code in core/src/java/org/apache/lucene/util/UnicodeUtil.java was
derived from Python 3.1.2 sources available at
http://www.python.org. Full license is here:
http://www.python.org/download/releases/3.1.2/license/
Some code in core/src/java/org/apache/lucene/util/automaton was
derived from Brics automaton sources available at
www.brics.dk/automaton/. Here is the copyright from those sources:
/*
* Copyright (c) 2001-2009 Anders Moeller
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. The name of the author may not be used to endorse or promote products
* derived from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
The levenshtein automata tables in core/src/java/org/apache/lucene/util/automaton
were automatically generated with the moman/finenight FSA package.
Here is the copyright for those sources:
# Copyright (c) 2010, Jean-Philippe Barrette-LaPierre, <jpb@rrette.com>
#
# Permission is hereby granted, free of charge, to any person
# obtaining a copy of this software and associated documentation
# files (the "Software"), to deal in the Software without
# restriction, including without limitation the rights to use,
# copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the
# Software is furnished to do so, subject to the following
# conditions:
#
# The above copyright notice and this permission notice shall be
# included in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
# OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
# NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
# HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
# OTHER DEALINGS IN THE SOFTWARE.
Some code in core/src/java/org/apache/lucene/util/UnicodeUtil.java was
derived from ICU (http://www.icu-project.org)
The full license is available here:
http://source.icu-project.org/repos/icu/icu/trunk/license.html
/*
* Copyright (C) 1999-2010, International Business Machines
* Corporation and others. All Rights Reserved.
*
* Permission is hereby granted, free of charge, to any person obtaining a copy
* of this software and associated documentation files (the "Software"), to deal
* in the Software without restriction, including without limitation the rights
* to use, copy, modify, merge, publish, distribute, and/or sell copies of the
* Software, and to permit persons to whom the Software is furnished to do so,
* provided that the above copyright notice(s) and this permission notice appear
* in all copies of the Software and that both the above copyright notice(s) and
* this permission notice appear in supporting documentation.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
* IN NO EVENT SHALL THE COPYRIGHT HOLDER OR HOLDERS INCLUDED IN THIS NOTICE BE
* LIABLE FOR ANY CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR
* ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER
* IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT
* OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
*
* Except as contained in this notice, the name of a copyright holder shall not
* be used in advertising or otherwise to promote the sale, use or other
* dealings in this Software without prior written authorization of the
* copyright holder.
*/
The following license applies to the Snowball stemmers:
Copyright (c) 2001, Dr Martin Porter
Copyright (c) 2002, Richard Boulton
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are met:
* Redistributions of source code must retain the above copyright notice,
* this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* Neither the name of the copyright holders nor the names of its contributors
* may be used to endorse or promote products derived from this software
* without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
The following license applies to the KStemmer:
Copyright © 2003,
Center for Intelligent Information Retrieval,
University of Massachusetts, Amherst.
All rights reserved.
Redistribution and use in source and binary forms, with or without modification,
are permitted provided that the following conditions are met:
1. Redistributions of source code must retain the above copyright notice, this
list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright notice,
this list of conditions and the following disclaimer in the documentation
and/or other materials provided with the distribution.
3. The names "Center for Intelligent Information Retrieval" and
"University of Massachusetts" must not be used to endorse or promote products
derived from this software without prior written permission. To obtain
permission, contact info@ciir.cs.umass.edu.
THIS SOFTWARE IS PROVIDED BY UNIVERSITY OF MASSACHUSETTS AND OTHER CONTRIBUTORS
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
SUCH DAMAGE.
The following license applies to the Morfologik project:
Copyright (c) 2006 Dawid Weiss
Copyright (c) 2007-2011 Dawid Weiss, Marcin Miłkowski
All rights reserved.
Redistribution and use in source and binary forms, with or without modification,
are permitted provided that the following conditions are met:
* Redistributions of source code must retain the above copyright notice,
this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright notice,
this list of conditions and the following disclaimer in the documentation
and/or other materials provided with the distribution.
* Neither the name of Morfologik nor the names of its contributors
may be used to endorse or promote products derived from this software
without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR
ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON
ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
---
The dictionary comes from Morfologik project. Morfologik uses data from
Polish ispell/myspell dictionary hosted at http://www.sjp.pl/slownik/en/ and
is licenced on the terms of (inter alia) LGPL and Creative Commons
ShareAlike. The part-of-speech tags were added in Morfologik project and
are not found in the data from sjp.pl. The tagset is similar to IPI PAN
tagset.
---
The following license applies to the Morfeusz project,
used by org.apache.lucene.analysis.morfologik.
BSD-licensed dictionary of Polish (SGJP)
http://sgjp.pl/morfeusz/
Copyright © 2011 Zygmunt Saloni, Włodzimierz Gruszczyński,
Marcin Woliński, Robert Wołosz
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
1. Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in the
documentation and/or other materials provided with the
distribution.
THIS SOFTWARE IS PROVIDED BY COPYRIGHT HOLDERS “AS IS” AND ANY EXPRESS
OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
DISCLAIMED. IN NO EVENT SHALL COPYRIGHT HOLDERS OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE
OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN
IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

View File

@ -1,192 +0,0 @@
Apache Lucene
Copyright 2014 The Apache Software Foundation
This product includes software developed at
The Apache Software Foundation (http://www.apache.org/).
Includes software from other Apache Software Foundation projects,
including, but not limited to:
- Apache Ant
- Apache Jakarta Regexp
- Apache Commons
- Apache Xerces
ICU4J, (under analysis/icu) is licensed under an MIT styles license
and Copyright (c) 1995-2008 International Business Machines Corporation and others
Some data files (under analysis/icu/src/data) are derived from Unicode data such
as the Unicode Character Database. See http://unicode.org/copyright.html for more
details.
Brics Automaton (under core/src/java/org/apache/lucene/util/automaton) is
BSD-licensed, created by Anders Møller. See http://www.brics.dk/automaton/
The levenshtein automata tables (under core/src/java/org/apache/lucene/util/automaton) were
automatically generated with the moman/finenight FSA library, created by
Jean-Philippe Barrette-LaPierre. This library is available under an MIT license,
see http://sites.google.com/site/rrettesite/moman and
http://bitbucket.org/jpbarrette/moman/overview/
The class org.apache.lucene.util.WeakIdentityMap was derived from
the Apache CXF project and is Apache License 2.0.
The Google Code Prettify is Apache License 2.0.
See http://code.google.com/p/google-code-prettify/
JUnit (junit-4.10) is licensed under the Common Public License v. 1.0
See http://junit.sourceforge.net/cpl-v10.html
This product includes code (JaspellTernarySearchTrie) from Java Spelling Checkin
g Package (jaspell): http://jaspell.sourceforge.net/
License: The BSD License (http://www.opensource.org/licenses/bsd-license.php)
The snowball stemmers in
analysis/common/src/java/net/sf/snowball
were developed by Martin Porter and Richard Boulton.
The snowball stopword lists in
analysis/common/src/resources/org/apache/lucene/analysis/snowball
were developed by Martin Porter and Richard Boulton.
The full snowball package is available from
http://snowball.tartarus.org/
The KStem stemmer in
analysis/common/src/org/apache/lucene/analysis/en
was developed by Bob Krovetz and Sergio Guzman-Lara (CIIR-UMass Amherst)
under the BSD-license.
The Arabic,Persian,Romanian,Bulgarian, Hindi and Bengali analyzers (common) come with a default
stopword list that is BSD-licensed created by Jacques Savoy. These files reside in:
analysis/common/src/resources/org/apache/lucene/analysis/ar/stopwords.txt,
analysis/common/src/resources/org/apache/lucene/analysis/fa/stopwords.txt,
analysis/common/src/resources/org/apache/lucene/analysis/ro/stopwords.txt,
analysis/common/src/resources/org/apache/lucene/analysis/bg/stopwords.txt,
analysis/common/src/resources/org/apache/lucene/analysis/hi/stopwords.txt,
analysis/common/src/resources/org/apache/lucene/analysis/bn/stopwords.txt
See http://members.unine.ch/jacques.savoy/clef/index.html.
The German,Spanish,Finnish,French,Hungarian,Italian,Portuguese,Russian and Swedish light stemmers
(common) are based on BSD-licensed reference implementations created by Jacques Savoy and
Ljiljana Dolamic. These files reside in:
analysis/common/src/java/org/apache/lucene/analysis/de/GermanLightStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/de/GermanMinimalStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/es/SpanishLightStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/fi/FinnishLightStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/fr/FrenchLightStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/fr/FrenchMinimalStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/hu/HungarianLightStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/it/ItalianLightStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/pt/PortugueseLightStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/ru/RussianLightStemmer.java
analysis/common/src/java/org/apache/lucene/analysis/sv/SwedishLightStemmer.java
The Stempel analyzer (stempel) includes BSD-licensed software developed
by the Egothor project http://egothor.sf.net/, created by Leo Galambos, Martin Kvapil,
and Edmond Nolan.
The Polish analyzer (stempel) comes with a default
stopword list that is BSD-licensed created by the Carrot2 project. The file resides
in stempel/src/resources/org/apache/lucene/analysis/pl/stopwords.txt.
See http://project.carrot2.org/license.html.
The SmartChineseAnalyzer source code (smartcn) was
provided by Xiaoping Gao and copyright 2009 by www.imdict.net.
WordBreakTestUnicode_*.java (under modules/analysis/common/src/test/)
is derived from Unicode data such as the Unicode Character Database.
See http://unicode.org/copyright.html for more details.
The Morfologik analyzer (morfologik) includes BSD-licensed software
developed by Dawid Weiss and Marcin Miłkowski (http://morfologik.blogspot.com/).
Morfologik uses data from Polish ispell/myspell dictionary
(http://www.sjp.pl/slownik/en/) licenced on the terms of (inter alia)
LGPL and Creative Commons ShareAlike.
Morfologic includes data from BSD-licensed dictionary of Polish (SGJP)
(http://sgjp.pl/morfeusz/)
Servlet-api.jar and javax.servlet-*.jar are under the CDDL license, the original
source code for this can be found at http://www.eclipse.org/jetty/downloads.php
===========================================================================
Kuromoji Japanese Morphological Analyzer - Apache Lucene Integration
===========================================================================
This software includes a binary and/or source version of data from
mecab-ipadic-2.7.0-20070801
which can be obtained from
http://atilika.com/releases/mecab-ipadic/mecab-ipadic-2.7.0-20070801.tar.gz
or
http://jaist.dl.sourceforge.net/project/mecab/mecab-ipadic/2.7.0-20070801/mecab-ipadic-2.7.0-20070801.tar.gz
===========================================================================
mecab-ipadic-2.7.0-20070801 Notice
===========================================================================
Nara Institute of Science and Technology (NAIST),
the copyright holders, disclaims all warranties with regard to this
software, including all implied warranties of merchantability and
fitness, in no event shall NAIST be liable for
any special, indirect or consequential damages or any damages
whatsoever resulting from loss of use, data or profits, whether in an
action of contract, negligence or other tortuous action, arising out
of or in connection with the use or performance of this software.
A large portion of the dictionary entries
originate from ICOT Free Software. The following conditions for ICOT
Free Software applies to the current dictionary as well.
Each User may also freely distribute the Program, whether in its
original form or modified, to any third party or parties, PROVIDED
that the provisions of Section 3 ("NO WARRANTY") will ALWAYS appear
on, or be attached to, the Program, which is distributed substantially
in the same form as set out herein and that such intended
distribution, if actually made, will neither violate or otherwise
contravene any of the laws and regulations of the countries having
jurisdiction over the User or the intended distribution itself.
NO WARRANTY
The program was produced on an experimental basis in the course of the
research and development conducted during the project and is provided
to users as so produced on an experimental basis. Accordingly, the
program is provided without any warranty whatsoever, whether express,
implied, statutory or otherwise. The term "warranty" used herein
includes, but is not limited to, any warranty of the quality,
performance, merchantability and fitness for a particular purpose of
the program and the nonexistence of any infringement or violation of
any right of any third party.
Each user of the program will agree and understand, and be deemed to
have agreed and understood, that there is no warranty whatsoever for
the program and, accordingly, the entire risk arising from or
otherwise connected with the program is assumed by the user.
Therefore, neither ICOT, the copyright holder, or any other
organization that participated in or was otherwise related to the
development of the program and their respective officials, directors,
officers and other employees shall be held liable for any and all
damages, including, without limitation, general, special, incidental
and consequential damages, arising out of or otherwise in connection
with the use or inability to use the program or any product, material
or result produced or otherwise obtained by using the program,
regardless of whether they have been advised of, or otherwise had
knowledge of, the possibility of such damages at any time during the
project or thereafter. Each user will be deemed to have agreed to the
foregoing by his or her commencement of use of the program. The term
"use" as used herein includes, but is not limited to, the use,
modification, copying and distribution of the program and the
production of secondary products from the program.
In the case where the program, whether in its original form or
modified, was distributed or delivered to or received by a user from
any person, organization or entity other than ICOT, unless it makes or
grants independently of ICOT any specific warranty to the user in
writing, such person, organization or entity, will also be exempted
from and not be held liable to the user for any such damages as noted
above as far as the program is concerned.

View File

@ -1 +0,0 @@
e118e4d05070378516b9055184b74498ba528dee

View File

@ -10,9 +10,8 @@ import org.elasticsearch.common.xcontent.XContentBuilder;
import org.elasticsearch.common.xcontent.XContentHelper; import org.elasticsearch.common.xcontent.XContentHelper;
import org.elasticsearch.common.xcontent.json.JsonXContent; import org.elasticsearch.common.xcontent.json.JsonXContent;
import org.elasticsearch.test.ESTestCase; import org.elasticsearch.test.ESTestCase;
import org.elasticsearch.xpack.sql.plugin.SqlQueryResponse;
import org.elasticsearch.xpack.sql.proto.Mode;
import org.joda.time.DateTime; import org.joda.time.DateTime;
import org.joda.time.ReadableDateTime;
import java.sql.JDBCType; import java.sql.JDBCType;
@ -51,7 +50,11 @@ public class TypeConverterTests extends ESTestCase {
XContentBuilder builder = JsonXContent.contentBuilder(); XContentBuilder builder = JsonXContent.contentBuilder();
builder.startObject(); builder.startObject();
builder.field("value"); builder.field("value");
SqlQueryResponse.value(builder, Mode.JDBC, value); if (value instanceof ReadableDateTime) {
builder.value(((ReadableDateTime) value).getMillis());
} else {
builder.value(value);
}
builder.endObject(); builder.endObject();
builder.close(); builder.close();
Object copy = XContentHelper.convertToMap(BytesReference.bytes(builder), false, builder.contentType()).v2().get("value"); Object copy = XContentHelper.convertToMap(BytesReference.bytes(builder), false, builder.contentType()).v2().get("value");

View File

@ -8,7 +8,6 @@ package org.elasticsearch.xpack.sql.cli.command;
import org.elasticsearch.xpack.sql.client.HttpClient; import org.elasticsearch.xpack.sql.client.HttpClient;
import org.elasticsearch.xpack.sql.client.shared.ClientException; import org.elasticsearch.xpack.sql.client.shared.ClientException;
import org.elasticsearch.xpack.sql.client.shared.Version; import org.elasticsearch.xpack.sql.client.shared.Version;
import org.elasticsearch.xpack.sql.plugin.AbstractSqlQueryRequest;
import org.elasticsearch.xpack.sql.proto.MainResponse; import org.elasticsearch.xpack.sql.proto.MainResponse;
import org.elasticsearch.xpack.sql.proto.Protocol; import org.elasticsearch.xpack.sql.proto.Protocol;

View File

@ -28,7 +28,7 @@ public class CliSessionTests extends ESTestCase {
public void testProperConnection() throws Exception { public void testProperConnection() throws Exception {
HttpClient httpClient = mock(HttpClient.class); HttpClient httpClient = mock(HttpClient.class);
when(httpClient.serverInfo()).thenReturn(new MainResponse(randomAlphaOfLength(5), org.elasticsearch.Version.CURRENT.toString(), when(httpClient.serverInfo()).thenReturn(new MainResponse(randomAlphaOfLength(5), org.elasticsearch.Version.CURRENT.toString(),
ClusterName.DEFAULT.value(), UUIDs.randomBase64UUID(), Build.CURRENT)); ClusterName.DEFAULT.value(), UUIDs.randomBase64UUID()));
CliSession cliSession = new CliSession(httpClient); CliSession cliSession = new CliSession(httpClient);
cliSession.checkConnection(); cliSession.checkConnection();
verify(httpClient, times(1)).serverInfo(); verify(httpClient, times(1)).serverInfo();
@ -58,7 +58,7 @@ public class CliSessionTests extends ESTestCase {
} }
when(httpClient.serverInfo()).thenReturn(new MainResponse(randomAlphaOfLength(5), when(httpClient.serverInfo()).thenReturn(new MainResponse(randomAlphaOfLength(5),
org.elasticsearch.Version.fromString(major + "." + minor + ".23").toString(), org.elasticsearch.Version.fromString(major + "." + minor + ".23").toString(),
ClusterName.DEFAULT.value(), UUIDs.randomBase64UUID(), Build.CURRENT)); ClusterName.DEFAULT.value(), UUIDs.randomBase64UUID()));
CliSession cliSession = new CliSession(httpClient); CliSession cliSession = new CliSession(httpClient);
expectThrows(ClientException.class, cliSession::checkConnection); expectThrows(ClientException.class, cliSession::checkConnection);
verify(httpClient, times(1)).serverInfo(); verify(httpClient, times(1)).serverInfo();

View File

@ -36,7 +36,7 @@ public class ServerInfoCliCommandTests extends ESTestCase {
HttpClient client = mock(HttpClient.class); HttpClient client = mock(HttpClient.class);
CliSession cliSession = new CliSession(client); CliSession cliSession = new CliSession(client);
when(client.serverInfo()).thenReturn(new MainResponse("my_node", "1.2.3", when(client.serverInfo()).thenReturn(new MainResponse("my_node", "1.2.3",
new ClusterName("my_cluster").value(), UUIDs.randomBase64UUID(), Build.CURRENT)); new ClusterName("my_cluster").value(), UUIDs.randomBase64UUID()));
ServerInfoCliCommand cliCommand = new ServerInfoCliCommand(); ServerInfoCliCommand cliCommand = new ServerInfoCliCommand();
assertTrue(cliCommand.handle(testTerminal, cliSession, "info")); assertTrue(cliCommand.handle(testTerminal, cliSession, "info"));
assertEquals(testTerminal.toString(), "Node:<em>my_node</em> Cluster:<em>my_cluster</em> Version:<em>1.2.3</em>\n"); assertEquals(testTerminal.toString(), "Node:<em>my_node</em> Cluster:<em>my_cluster</em> Version:<em>1.2.3</em>\n");

View File

@ -24,6 +24,7 @@ dependencies {
compile (project(':libs:x-content')) { compile (project(':libs:x-content')) {
transitive = false transitive = false
} }
compile xpackProject('plugin:sql:sql-shared-proto')
compile "org.apache.lucene:lucene-core:${versions.lucene}" compile "org.apache.lucene:lucene-core:${versions.lucene}"
compile 'joda-time:joda-time:2.9.9' compile 'joda-time:joda-time:2.9.9'
runtime "com.fasterxml.jackson.core:jackson-core:${versions.jackson}" runtime "com.fasterxml.jackson.core:jackson-core:${versions.jackson}"

View File

@ -10,14 +10,14 @@ apply plugin: 'elasticsearch.build'
description = 'Code shared between jdbc and cli' description = 'Code shared between jdbc and cli'
dependencies { dependencies {
compile xpackProject('plugin:sql:sql-proto') compile xpackProject('plugin:sql:sql-shared-proto')
compile "com.fasterxml.jackson.core:jackson-core:${versions.jackson}" compile "com.fasterxml.jackson.core:jackson-core:${versions.jackson}"
testCompile "org.elasticsearch.test:framework:${version}" testCompile "org.elasticsearch.test:framework:${version}"
} }
dependencyLicenses { dependencyLicenses {
mapping from: /jackson-.*/, to: 'jackson' mapping from: /jackson-.*/, to: 'jackson'
mapping from: /sql-proto.*/, to: 'elasticsearch' mapping from: /sql-shared-proto.*/, to: 'elasticsearch'
mapping from: /elasticsearch-cli.*/, to: 'elasticsearch' mapping from: /elasticsearch-cli.*/, to: 'elasticsearch'
mapping from: /elasticsearch-core.*/, to: 'elasticsearch' mapping from: /elasticsearch-core.*/, to: 'elasticsearch'
mapping from: /lucene-.*/, to: 'lucene' mapping from: /lucene-.*/, to: 'lucene'

View File

@ -5,14 +5,12 @@
*/ */
package org.elasticsearch.xpack.sql.client; package org.elasticsearch.xpack.sql.client;
import org.elasticsearch.common.bytes.BytesReference;
import org.elasticsearch.common.collect.Tuple; import org.elasticsearch.common.collect.Tuple;
import org.elasticsearch.common.io.stream.BytesStreamOutput;
import org.elasticsearch.common.unit.TimeValue; import org.elasticsearch.common.unit.TimeValue;
import org.elasticsearch.common.xcontent.LoggingDeprecationHandler; import org.elasticsearch.common.xcontent.DeprecationHandler;
import org.elasticsearch.common.xcontent.NamedXContentRegistry; import org.elasticsearch.common.xcontent.NamedXContentRegistry;
import org.elasticsearch.common.xcontent.ToXContent; import org.elasticsearch.common.xcontent.ToXContent;
import org.elasticsearch.common.xcontent.XContentHelper; import org.elasticsearch.common.xcontent.XContentBuilder;
import org.elasticsearch.common.xcontent.XContentParser; import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.common.xcontent.XContentType; import org.elasticsearch.common.xcontent.XContentType;
import org.elasticsearch.core.internal.io.Streams; import org.elasticsearch.core.internal.io.Streams;
@ -30,6 +28,8 @@ import org.elasticsearch.xpack.sql.proto.SqlClearCursorResponse;
import org.elasticsearch.xpack.sql.proto.SqlQueryRequest; import org.elasticsearch.xpack.sql.proto.SqlQueryRequest;
import org.elasticsearch.xpack.sql.proto.SqlQueryResponse; import org.elasticsearch.xpack.sql.proto.SqlQueryResponse;
import java.io.ByteArrayInputStream;
import java.io.ByteArrayOutputStream;
import java.io.IOException; import java.io.IOException;
import java.io.InputStream; import java.io.InputStream;
import java.security.AccessController; import java.security.AccessController;
@ -67,8 +67,8 @@ public class HttpClient {
public SqlQueryResponse queryInit(String query, int fetchSize) throws SQLException { public SqlQueryResponse queryInit(String query, int fetchSize) throws SQLException {
// TODO allow customizing the time zone - this is what session set/reset/get should be about // TODO allow customizing the time zone - this is what session set/reset/get should be about
SqlQueryRequest sqlRequest = new SqlQueryRequest(Mode.PLAIN, query, Collections.emptyList(), null, SqlQueryRequest sqlRequest = new SqlQueryRequest(Mode.PLAIN, query, Collections.emptyList(), null,
TimeZone.getTimeZone("UTC"), fetchSize, TimeValue.timeValueMillis(cfg.queryTimeout()), TimeZone.getTimeZone("UTC"), fetchSize, TimeValue.timeValueMillis(cfg.queryTimeout()),
TimeValue.timeValueMillis(cfg.pageTimeout())); TimeValue.timeValueMillis(cfg.pageTimeout()));
return query(sqlRequest); return query(sqlRequest);
} }
@ -84,83 +84,92 @@ public class HttpClient {
public boolean queryClose(String cursor) throws SQLException { public boolean queryClose(String cursor) throws SQLException {
SqlClearCursorResponse response = post(Protocol.CLEAR_CURSOR_REST_ENDPOINT, SqlClearCursorResponse response = post(Protocol.CLEAR_CURSOR_REST_ENDPOINT,
new SqlClearCursorRequest(Mode.PLAIN, cursor), new SqlClearCursorRequest(Mode.PLAIN, cursor),
SqlClearCursorResponse::fromXContent); SqlClearCursorResponse::fromXContent);
return response.isSucceeded(); return response.isSucceeded();
} }
private <Request extends AbstractSqlRequest, Response> Response post(String path, Request request, private <Request extends AbstractSqlRequest, Response> Response post(String path, Request request,
CheckedFunction<XContentParser, Response, IOException> responseParser) CheckedFunction<XContentParser, Response, IOException> responseParser)
throws SQLException { throws SQLException {
BytesReference requestBytes = toXContent(request); byte[] requestBytes = toXContent(request);
String query = "error_trace&mode=" + request.mode(); String query = "error_trace&mode=" + request.mode();
Tuple<XContentType, BytesReference> response = Tuple<XContentType, byte[]> response =
AccessController.doPrivileged((PrivilegedAction<ResponseOrException<Tuple<XContentType, BytesReference>>>) () -> AccessController.doPrivileged((PrivilegedAction<ResponseOrException<Tuple<XContentType, byte[]>>>) () ->
JreHttpUrlConnection.http(path, query, cfg, con -> JreHttpUrlConnection.http(path, query, cfg, con ->
con.request( con.request(
requestBytes::writeTo, (out) -> out.write(requestBytes),
this::readFrom, this::readFrom,
"POST" "POST"
) )
)).getResponseOrThrowException(); )).getResponseOrThrowException();
return fromXContent(response.v1(), response.v2(), responseParser); return fromXContent(response.v1(), response.v2(), responseParser);
} }
private boolean head(String path, long timeoutInMs) throws SQLException { private boolean head(String path, long timeoutInMs) throws SQLException {
ConnectionConfiguration pingCfg = new ConnectionConfiguration(cfg.baseUri(), cfg.connectionString(), ConnectionConfiguration pingCfg = new ConnectionConfiguration(cfg.baseUri(), cfg.connectionString(),
cfg.connectTimeout(), timeoutInMs, cfg.queryTimeout(), cfg.pageTimeout(), cfg.pageSize(), cfg.connectTimeout(), timeoutInMs, cfg.queryTimeout(), cfg.pageTimeout(), cfg.pageSize(),
cfg.authUser(), cfg.authPass(), cfg.sslConfig(), cfg.proxyConfig()); cfg.authUser(), cfg.authPass(), cfg.sslConfig(), cfg.proxyConfig());
try { try {
return AccessController.doPrivileged((PrivilegedAction<Boolean>) () -> return AccessController.doPrivileged((PrivilegedAction<Boolean>) () ->
JreHttpUrlConnection.http(path, "error_trace", pingCfg, JreHttpUrlConnection::head)); JreHttpUrlConnection.http(path, "error_trace", pingCfg, JreHttpUrlConnection::head));
} catch (ClientException ex) { } catch (ClientException ex) {
throw new SQLException("Cannot ping server", ex); throw new SQLException("Cannot ping server", ex);
} }
} }
private <Response> Response get(String path, CheckedFunction<XContentParser, Response, IOException> responseParser) private <Response> Response get(String path, CheckedFunction<XContentParser, Response, IOException> responseParser)
throws SQLException { throws SQLException {
Tuple<XContentType, BytesReference> response = Tuple<XContentType, byte[]> response =
AccessController.doPrivileged((PrivilegedAction<ResponseOrException<Tuple<XContentType, BytesReference>>>) () -> AccessController.doPrivileged((PrivilegedAction<ResponseOrException<Tuple<XContentType, byte[]>>>) () ->
JreHttpUrlConnection.http(path, "error_trace", cfg, con -> JreHttpUrlConnection.http(path, "error_trace", cfg, con ->
con.request( con.request(
null, null,
this::readFrom, this::readFrom,
"GET" "GET"
) )
)).getResponseOrThrowException(); )).getResponseOrThrowException();
return fromXContent(response.v1(), response.v2(), responseParser); return fromXContent(response.v1(), response.v2(), responseParser);
} }
private static <Request extends ToXContent> BytesReference toXContent(Request xContent) { private static <Request extends ToXContent> byte[] toXContent(Request xContent) {
try { try(ByteArrayOutputStream buffer = new ByteArrayOutputStream()) {
return XContentHelper.toXContent(xContent, REQUEST_BODY_CONTENT_TYPE, false); try (XContentBuilder xContentBuilder = new XContentBuilder(REQUEST_BODY_CONTENT_TYPE.xContent(), buffer)) {
if (xContent.isFragment()) {
xContentBuilder.startObject();
}
xContent.toXContent(xContentBuilder, ToXContent.EMPTY_PARAMS);
if (xContent.isFragment()) {
xContentBuilder.endObject();
}
}
return buffer.toByteArray();
} catch (IOException ex) { } catch (IOException ex) {
throw new ClientException("Cannot serialize request", ex); throw new ClientException("Cannot serialize request", ex);
} }
} }
private Tuple<XContentType, BytesReference> readFrom(InputStream inputStream, Function<String, String> headers) { private Tuple<XContentType, byte[]> readFrom(InputStream inputStream, Function<String, String> headers) {
String contentType = headers.apply("Content-Type"); String contentType = headers.apply("Content-Type");
XContentType xContentType = XContentType.fromMediaTypeOrFormat(contentType); XContentType xContentType = XContentType.fromMediaTypeOrFormat(contentType);
if (xContentType == null) { if (xContentType == null) {
throw new IllegalStateException("Unsupported Content-Type: " + contentType); throw new IllegalStateException("Unsupported Content-Type: " + contentType);
} }
BytesStreamOutput out = new BytesStreamOutput(); ByteArrayOutputStream out = new ByteArrayOutputStream();
try { try {
Streams.copy(inputStream, out); Streams.copy(inputStream, out);
} catch (IOException ex) { } catch (IOException ex) {
throw new ClientException("Cannot deserialize response", ex); throw new ClientException("Cannot deserialize response", ex);
} }
return new Tuple<>(xContentType, out.bytes()); return new Tuple<>(xContentType, out.toByteArray());
} }
private <Response> Response fromXContent(XContentType xContentType, BytesReference bytesReference, private <Response> Response fromXContent(XContentType xContentType, byte[] bytesReference,
CheckedFunction<XContentParser, Response, IOException> responseParser) { CheckedFunction<XContentParser, Response, IOException> responseParser) {
try (InputStream stream = bytesReference.streamInput(); try (InputStream stream = new ByteArrayInputStream(bytesReference);
XContentParser parser = xContentType.xContent().createParser(registry, XContentParser parser = xContentType.xContent().createParser(registry,
LoggingDeprecationHandler.INSTANCE, stream)) { DeprecationHandler.THROW_UNSUPPORTED_OPERATION, stream)) {
return responseParser.apply(parser); return responseParser.apply(parser);
} catch (IOException ex) { } catch (IOException ex) {
throw new ClientException("Cannot parse response", ex); throw new ClientException("Cannot parse response", ex);

View File

@ -0,0 +1,35 @@
/*
* This project contains XContent protocol classes shared between server and http client
*/
import org.elasticsearch.gradle.precommit.PrecommitTasks
apply plugin: 'elasticsearch.build'
description = 'Request and response objects shared by the cli, jdbc ' +
'and the Elasticsearch plugin'
dependencies {
compile (project(':libs:elasticsearch-core')) {
transitive = false
}
compile (project(':libs:x-content')) {
transitive = false
}
compile 'joda-time:joda-time:2.9.9'
runtime "com.fasterxml.jackson.core:jackson-core:${versions.jackson}"
testCompile "org.elasticsearch.test:framework:${version}"
}
forbiddenApisMain {
//sql does not depend on server, so only jdk signatures should be checked
signaturesURLs = [PrecommitTasks.getResource('/forbidden/jdk-signatures.txt')]
}
dependencyLicenses {
mapping from: /elasticsearch-core.*/, to: 'elasticsearch'
mapping from: /jackson-.*/, to: 'jackson'
ignoreSha 'elasticsearch-core'
}

View File

@ -0,0 +1,8 @@
This copy of Jackson JSON processor streaming parser/generator is licensed under the
Apache (Software) License, version 2.0 ("the License").
See the License for details about distribution rights, and the
specific rights regarding derivate works.
You may obtain a copy of the License at:
http://www.apache.org/licenses/LICENSE-2.0

View File

@ -0,0 +1,20 @@
# Jackson JSON processor
Jackson is a high-performance, Free/Open Source JSON processing library.
It was originally written by Tatu Saloranta (tatu.saloranta@iki.fi), and has
been in development since 2007.
It is currently developed by a community of developers, as well as supported
commercially by FasterXML.com.
## Licensing
Jackson core and extension components may licensed under different licenses.
To find the details that apply to this artifact see the accompanying LICENSE file.
For more information, including possible other licensing options, contact
FasterXML.com (http://fasterxml.com).
## Credits
A list of contributors may be found from CREDITS file, which is included
in some artifacts (usually source distributions); but is always available
from the source code management (SCM) system project uses.

View File

@ -0,0 +1 @@
eb21a035c66ad307e66ec8fce37f5d50fd62d039

View File

@ -0,0 +1 @@
f7b520c458572890807d143670c9b24f4de90897

View File

@ -0,0 +1,202 @@
Apache License
Version 2.0, January 2004
http://www.apache.org/licenses/
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
1. Definitions.
"License" shall mean the terms and conditions for use, reproduction,
and distribution as defined by Sections 1 through 9 of this document.
"Licensor" shall mean the copyright owner or entity authorized by
the copyright owner that is granting the License.
"Legal Entity" shall mean the union of the acting entity and all
other entities that control, are controlled by, or are under common
control with that entity. For the purposes of this definition,
"control" means (i) the power, direct or indirect, to cause the
direction or management of such entity, whether by contract or
otherwise, or (ii) ownership of fifty percent (50%) or more of the
outstanding shares, or (iii) beneficial ownership of such entity.
"You" (or "Your") shall mean an individual or Legal Entity
exercising permissions granted by this License.
"Source" form shall mean the preferred form for making modifications,
including but not limited to software source code, documentation
source, and configuration files.
"Object" form shall mean any form resulting from mechanical
transformation or translation of a Source form, including but
not limited to compiled object code, generated documentation,
and conversions to other media types.
"Work" shall mean the work of authorship, whether in Source or
Object form, made available under the License, as indicated by a
copyright notice that is included in or attached to the work
(an example is provided in the Appendix below).
"Derivative Works" shall mean any work, whether in Source or Object
form, that is based on (or derived from) the Work and for which the
editorial revisions, annotations, elaborations, or other modifications
represent, as a whole, an original work of authorship. For the purposes
of this License, Derivative Works shall not include works that remain
separable from, or merely link (or bind by name) to the interfaces of,
the Work and Derivative Works thereof.
"Contribution" shall mean any work of authorship, including
the original version of the Work and any modifications or additions
to that Work or Derivative Works thereof, that is intentionally
submitted to Licensor for inclusion in the Work by the copyright owner
or by an individual or Legal Entity authorized to submit on behalf of
the copyright owner. For the purposes of this definition, "submitted"
means any form of electronic, verbal, or written communication sent
to the Licensor or its representatives, including but not limited to
communication on electronic mailing lists, source code control systems,
and issue tracking systems that are managed by, or on behalf of, the
Licensor for the purpose of discussing and improving the Work, but
excluding communication that is conspicuously marked or otherwise
designated in writing by the copyright owner as "Not a Contribution."
"Contributor" shall mean Licensor and any individual or Legal Entity
on behalf of whom a Contribution has been received by Licensor and
subsequently incorporated within the Work.
2. Grant of Copyright License. Subject to the terms and conditions of
this License, each Contributor hereby grants to You a perpetual,
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
copyright license to reproduce, prepare Derivative Works of,
publicly display, publicly perform, sublicense, and distribute the
Work and such Derivative Works in Source or Object form.
3. Grant of Patent License. Subject to the terms and conditions of
this License, each Contributor hereby grants to You a perpetual,
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
(except as stated in this section) patent license to make, have made,
use, offer to sell, sell, import, and otherwise transfer the Work,
where such license applies only to those patent claims licensable
by such Contributor that are necessarily infringed by their
Contribution(s) alone or by combination of their Contribution(s)
with the Work to which such Contribution(s) was submitted. If You
institute patent litigation against any entity (including a
cross-claim or counterclaim in a lawsuit) alleging that the Work
or a Contribution incorporated within the Work constitutes direct
or contributory patent infringement, then any patent licenses
granted to You under this License for that Work shall terminate
as of the date such litigation is filed.
4. Redistribution. You may reproduce and distribute copies of the
Work or Derivative Works thereof in any medium, with or without
modifications, and in Source or Object form, provided that You
meet the following conditions:
(a) You must give any other recipients of the Work or
Derivative Works a copy of this License; and
(b) You must cause any modified files to carry prominent notices
stating that You changed the files; and
(c) You must retain, in the Source form of any Derivative Works
that You distribute, all copyright, patent, trademark, and
attribution notices from the Source form of the Work,
excluding those notices that do not pertain to any part of
the Derivative Works; and
(d) If the Work includes a "NOTICE" text file as part of its
distribution, then any Derivative Works that You distribute must
include a readable copy of the attribution notices contained
within such NOTICE file, excluding those notices that do not
pertain to any part of the Derivative Works, in at least one
of the following places: within a NOTICE text file distributed
as part of the Derivative Works; within the Source form or
documentation, if provided along with the Derivative Works; or,
within a display generated by the Derivative Works, if and
wherever such third-party notices normally appear. The contents
of the NOTICE file are for informational purposes only and
do not modify the License. You may add Your own attribution
notices within Derivative Works that You distribute, alongside
or as an addendum to the NOTICE text from the Work, provided
that such additional attribution notices cannot be construed
as modifying the License.
You may add Your own copyright statement to Your modifications and
may provide additional or different license terms and conditions
for use, reproduction, or distribution of Your modifications, or
for any such Derivative Works as a whole, provided Your use,
reproduction, and distribution of the Work otherwise complies with
the conditions stated in this License.
5. Submission of Contributions. Unless You explicitly state otherwise,
any Contribution intentionally submitted for inclusion in the Work
by You to the Licensor shall be under the terms and conditions of
this License, without any additional terms or conditions.
Notwithstanding the above, nothing herein shall supersede or modify
the terms of any separate license agreement you may have executed
with Licensor regarding such Contributions.
6. Trademarks. This License does not grant permission to use the trade
names, trademarks, service marks, or product names of the Licensor,
except as required for reasonable and customary use in describing the
origin of the Work and reproducing the content of the NOTICE file.
7. Disclaimer of Warranty. Unless required by applicable law or
agreed to in writing, Licensor provides the Work (and each
Contributor provides its Contributions) on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
implied, including, without limitation, any warranties or conditions
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
PARTICULAR PURPOSE. You are solely responsible for determining the
appropriateness of using or redistributing the Work and assume any
risks associated with Your exercise of permissions under this License.
8. Limitation of Liability. In no event and under no legal theory,
whether in tort (including negligence), contract, or otherwise,
unless required by applicable law (such as deliberate and grossly
negligent acts) or agreed to in writing, shall any Contributor be
liable to You for damages, including any direct, indirect, special,
incidental, or consequential damages of any character arising as a
result of this License or out of the use or inability to use the
Work (including but not limited to damages for loss of goodwill,
work stoppage, computer failure or malfunction, or any and all
other commercial damages or losses), even if such Contributor
has been advised of the possibility of such damages.
9. Accepting Warranty or Additional Liability. While redistributing
the Work or Derivative Works thereof, You may choose to offer,
and charge a fee for, acceptance of support, warranty, indemnity,
or other liability obligations and/or rights consistent with this
License. However, in accepting such obligations, You may act only
on Your own behalf and on Your sole responsibility, not on behalf
of any other Contributor, and only if You agree to indemnify,
defend, and hold each Contributor harmless for any liability
incurred by, or claims asserted against, such Contributor by reason
of your accepting any such warranty or additional liability.
END OF TERMS AND CONDITIONS
APPENDIX: How to apply the Apache License to your work.
To apply the Apache License to your work, attach the following
boilerplate notice, with the fields enclosed by brackets "[]"
replaced with your own identifying information. (Don't include
the brackets!) The text should be enclosed in the appropriate
comment syntax for the file format. We also recommend that a
file or class name and description of purpose be included on the
same "printed page" as the copyright notice for easier
identification within third-party archives.
Copyright [yyyy] [name of copyright owner]
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.

View File

@ -0,0 +1,5 @@
=============================================================================
= NOTICE file corresponding to section 4d of the Apache License Version 2.0 =
=============================================================================
This product includes software developed by
Joda.org (http://www.joda.org/).

View File

@ -7,7 +7,6 @@ package org.elasticsearch.xpack.sql.proto;
import org.elasticsearch.common.Nullable; import org.elasticsearch.common.Nullable;
import org.elasticsearch.common.ParseField; import org.elasticsearch.common.ParseField;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.xcontent.ConstructingObjectParser; import org.elasticsearch.common.xcontent.ConstructingObjectParser;
import org.elasticsearch.common.xcontent.ToXContentObject; import org.elasticsearch.common.xcontent.ToXContentObject;
import org.elasticsearch.common.xcontent.XContentBuilder; import org.elasticsearch.common.xcontent.XContentBuilder;
@ -74,7 +73,7 @@ public class ColumnInfo implements ToXContentObject {
@Override @Override
public XContentBuilder toXContent(XContentBuilder builder, Params params) throws IOException { public XContentBuilder toXContent(XContentBuilder builder, Params params) throws IOException {
builder.startObject(); builder.startObject();
if (Strings.hasText(table)) { if (table != null && table.isEmpty() == false) {
builder.field("table", table); builder.field("table", table);
} }
builder.field("name", name); builder.field("name", name);
@ -146,6 +145,6 @@ public class ColumnInfo implements ToXContentObject {
@Override @Override
public String toString() { public String toString() {
return Strings.toString(this); return ProtoUtils.toString(this);
} }
} }

View File

@ -6,7 +6,6 @@
package org.elasticsearch.xpack.sql.proto; package org.elasticsearch.xpack.sql.proto;
import org.elasticsearch.Build;
import org.elasticsearch.common.ParseField; import org.elasticsearch.common.ParseField;
import org.elasticsearch.common.xcontent.ObjectParser; import org.elasticsearch.common.xcontent.ObjectParser;
import org.elasticsearch.common.xcontent.XContentParser; import org.elasticsearch.common.xcontent.XContentParser;
@ -21,18 +20,16 @@ public class MainResponse {
private String version; private String version;
private String clusterName; private String clusterName;
private String clusterUuid; private String clusterUuid;
// TODO: Add parser for Build
private Build build;
private MainResponse() { private MainResponse() {
} }
public MainResponse(String nodeName, String version, String clusterName, String clusterUuid, Build build) { public MainResponse(String nodeName, String version, String clusterName, String clusterUuid) {
this.nodeName = nodeName; this.nodeName = nodeName;
this.version = version; this.version = version;
this.clusterName = clusterName; this.clusterName = clusterName;
this.clusterUuid = clusterUuid; this.clusterUuid = clusterUuid;
this.build = build;
} }
public String getNodeName() { public String getNodeName() {
@ -51,10 +48,6 @@ public class MainResponse {
return clusterUuid; return clusterUuid;
} }
public Build getBuild() {
return build;
}
private static final ObjectParser<MainResponse, Void> PARSER = new ObjectParser<>(MainResponse.class.getName(), true, private static final ObjectParser<MainResponse, Void> PARSER = new ObjectParser<>(MainResponse.class.getName(), true,
MainResponse::new); MainResponse::new);
@ -65,15 +58,6 @@ public class MainResponse {
PARSER.declareString((response, value) -> { PARSER.declareString((response, value) -> {
}, new ParseField("tagline")); }, new ParseField("tagline"));
PARSER.declareObject((response, value) -> { PARSER.declareObject((response, value) -> {
final String buildFlavor = (String) value.get("build_flavor");
final String buildType = (String) value.get("build_type");
response.build =
new Build(
buildFlavor == null ? Build.Flavor.UNKNOWN : Build.Flavor.fromDisplayName(buildFlavor),
buildType == null ? Build.Type.UNKNOWN : Build.Type.fromDisplayName(buildType),
(String) value.get("build_hash"),
(String) value.get("build_date"),
(boolean) value.get("build_snapshot"));
response.version = (String) value.get("number"); response.version = (String) value.get("number");
}, (parser, context) -> parser.map(), new ParseField("version")); }, (parser, context) -> parser.map(), new ParseField("version"));
} }
@ -94,12 +78,11 @@ public class MainResponse {
return Objects.equals(nodeName, other.nodeName) && return Objects.equals(nodeName, other.nodeName) &&
Objects.equals(version, other.version) && Objects.equals(version, other.version) &&
Objects.equals(clusterUuid, other.clusterUuid) && Objects.equals(clusterUuid, other.clusterUuid) &&
Objects.equals(build, other.build) &&
Objects.equals(clusterName, other.clusterName); Objects.equals(clusterName, other.clusterName);
} }
@Override @Override
public int hashCode() { public int hashCode() {
return Objects.hash(nodeName, version, clusterUuid, build, clusterName); return Objects.hash(nodeName, version, clusterUuid, clusterName);
} }
} }

View File

@ -0,0 +1,85 @@
/*
* Copyright Elasticsearch B.V. and/or licensed to Elasticsearch B.V. under one
* or more contributor license agreements. Licensed under the Elastic License;
* you may not use this file except in compliance with the Elastic License.
*/
package org.elasticsearch.xpack.sql.proto;
import org.elasticsearch.common.xcontent.ToXContent;
import org.elasticsearch.common.xcontent.XContentBuilder;
import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.common.xcontent.json.JsonXContent;
import java.io.ByteArrayOutputStream;
import java.io.IOException;
import java.nio.charset.StandardCharsets;
import java.util.Locale;
public final class ProtoUtils {
private ProtoUtils() {
}
/**
* Parses a generic value from the XContent stream
*/
public static Object parseFieldsValue(XContentParser parser) throws IOException {
XContentParser.Token token = parser.currentToken();
if (token == XContentParser.Token.VALUE_STRING) {
//binary values will be parsed back and returned as base64 strings when reading from json and yaml
return parser.text();
} else if (token == XContentParser.Token.VALUE_NUMBER) {
return parser.numberValue();
} else if (token == XContentParser.Token.VALUE_BOOLEAN) {
return parser.booleanValue();
} else if (token == XContentParser.Token.VALUE_NULL) {
return null;
} else if (token == XContentParser.Token.START_OBJECT) {
return parser.mapOrdered();
} else if (token == XContentParser.Token.START_ARRAY) {
return parser.listOrderedMap();
} else {
String message = "Failed to parse object: unexpected token [%s] found";
throw new IllegalStateException(String.format(Locale.ROOT, message, token));
}
}
/**
* Returns a string representation of the builder (only applicable for text based xcontent).
*
* @param xContentBuilder builder containing an object to converted to a string
*/
public static String toString(XContentBuilder xContentBuilder) {
byte[] byteArray = ((ByteArrayOutputStream) xContentBuilder.getOutputStream()).toByteArray();
return new String(byteArray, StandardCharsets.UTF_8);
}
public static String toString(ToXContent toXContent) {
try {
XContentBuilder builder = JsonXContent.contentBuilder();
if (toXContent.isFragment()) {
builder.startObject();
}
toXContent.toXContent(builder, ToXContent.EMPTY_PARAMS);
if (toXContent.isFragment()) {
builder.endObject();
}
builder.close();
return toString(builder);
} catch (IOException e) {
try {
XContentBuilder builder = JsonXContent.contentBuilder();
builder.startObject();
builder.field("error", "error building toString out of XContent: " + e.getMessage());
builder.endObject();
builder.close();
return toString(builder);
} catch (IOException e2) {
throw new IllegalArgumentException("cannot generate error message for deserialization", e);
}
}
}
}

View File

@ -18,7 +18,7 @@ import java.util.Objects;
import static org.elasticsearch.common.xcontent.ConstructingObjectParser.constructorArg; import static org.elasticsearch.common.xcontent.ConstructingObjectParser.constructorArg;
import static org.elasticsearch.common.xcontent.ConstructingObjectParser.optionalConstructorArg; import static org.elasticsearch.common.xcontent.ConstructingObjectParser.optionalConstructorArg;
import static org.elasticsearch.common.xcontent.XContentParserUtils.parseFieldsValue; import static org.elasticsearch.xpack.sql.proto.ProtoUtils.parseFieldsValue;
/** /**
* Response to perform an sql query for JDBC/CLI client * Response to perform an sql query for JDBC/CLI client

View File

@ -11,13 +11,13 @@ import org.elasticsearch.common.xcontent.ObjectParser;
import org.elasticsearch.common.xcontent.ToXContentObject; import org.elasticsearch.common.xcontent.ToXContentObject;
import org.elasticsearch.common.xcontent.XContentBuilder; import org.elasticsearch.common.xcontent.XContentBuilder;
import org.elasticsearch.common.xcontent.XContentParser; import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.common.xcontent.XContentParserUtils;
import org.elasticsearch.xpack.sql.type.DataType; import org.elasticsearch.xpack.sql.type.DataType;
import java.io.IOException; import java.io.IOException;
import java.util.Objects; import java.util.Objects;
import static org.elasticsearch.common.xcontent.ConstructingObjectParser.constructorArg; import static org.elasticsearch.common.xcontent.ConstructingObjectParser.constructorArg;
import static org.elasticsearch.xpack.sql.proto.ProtoUtils.parseFieldsValue;
/** /**
* Represent a strongly typed parameter value * Represent a strongly typed parameter value
@ -33,7 +33,7 @@ public class SqlTypedParamValue implements ToXContentObject {
private static final ParseField TYPE = new ParseField("type"); private static final ParseField TYPE = new ParseField("type");
static { static {
PARSER.declareField(constructorArg(), (p, c) -> XContentParserUtils.parseFieldsValue(p), VALUE, ObjectParser.ValueType.VALUE); PARSER.declareField(constructorArg(), (p, c) -> parseFieldsValue(p), VALUE, ObjectParser.ValueType.VALUE);
PARSER.declareString(constructorArg(), TYPE); PARSER.declareString(constructorArg(), TYPE);
} }

View File

@ -0,0 +1,59 @@
/*
* Copyright Elasticsearch B.V. and/or licensed to Elasticsearch B.V. under one
* or more contributor license agreements. Licensed under the Elastic License;
* you may not use this file except in compliance with the Elastic License.
*/
package org.elasticsearch.xpack.sql.proto;
import org.elasticsearch.common.xcontent.DeprecationHandler;
import org.elasticsearch.common.xcontent.NamedXContentRegistry;
import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.common.xcontent.json.JsonXContent;
import org.elasticsearch.test.ESTestCase;
import java.io.IOException;
public class ProtoUtilsTests extends ESTestCase {
public void testGenericValueParsing() throws IOException {
String json = ProtoUtils.toString((builder, params) -> {
builder.field("int", 42);
builder.field("double", 42.5);
builder.field("string", "foobar");
builder.nullField("null");
return builder;
});
XContentParser parser =
JsonXContent.jsonXContent.createParser(NamedXContentRegistry.EMPTY, DeprecationHandler.THROW_UNSUPPORTED_OPERATION, json);
assertEquals(XContentParser.Token.START_OBJECT, parser.nextToken());
while (parser.nextToken() != XContentParser.Token.END_OBJECT) {
assertEquals(XContentParser.Token.FIELD_NAME, parser.currentToken());
String fieldName = parser.currentName();
parser.nextToken();
Object val = ProtoUtils.parseFieldsValue(parser);
switch (fieldName) {
case "int":
assertEquals(42, val);
break;
case "double":
assertEquals(42.5, val);
break;
case "string":
assertEquals("foobar", val);
break;
case "null":
assertNull(val);
break;
default:
fail("Unexpected value " + fieldName);
}
}
assertNull(parser.nextToken());
}
}

View File

@ -96,6 +96,9 @@ subprojects {
// CLI testing dependencies // CLI testing dependencies
testRuntime project(path: xpackModule('sql:sql-cli'), configuration: 'nodeps') testRuntime project(path: xpackModule('sql:sql-cli'), configuration: 'nodeps')
testRuntime (xpackProject('plugin:sql:sql-proto')) {
transitive = false
}
testRuntime "org.jline:jline:3.6.0" testRuntime "org.jline:jline:3.6.0"
} }