Merge branch 'master' into index-lifecycle
This commit is contained in:
commit
2ccb4df005
|
@ -70,31 +70,44 @@ public class RestIntegTestTask extends DefaultTask {
|
|||
runner.parallelism = '1'
|
||||
runner.include('**/*IT.class')
|
||||
runner.systemProperty('tests.rest.load_packaged', 'false')
|
||||
// we pass all nodes to the rest cluster to allow the clients to round-robin between them
|
||||
// this is more realistic than just talking to a single node
|
||||
runner.systemProperty('tests.rest.cluster', "${-> nodes.collect{it.httpUri()}.join(",")}")
|
||||
runner.systemProperty('tests.config.dir', "${-> nodes[0].pathConf}")
|
||||
// TODO: our "client" qa tests currently use the rest-test plugin. instead they should have their own plugin
|
||||
// that sets up the test cluster and passes this transport uri instead of http uri. Until then, we pass
|
||||
// both as separate sysprops
|
||||
runner.systemProperty('tests.cluster', "${-> nodes[0].transportUri()}")
|
||||
|
||||
// dump errors and warnings from cluster log on failure
|
||||
TaskExecutionAdapter logDumpListener = new TaskExecutionAdapter() {
|
||||
@Override
|
||||
void afterExecute(Task task, TaskState state) {
|
||||
if (state.failure != null) {
|
||||
for (NodeInfo nodeInfo : nodes) {
|
||||
printLogExcerpt(nodeInfo)
|
||||
if (System.getProperty("tests.rest.cluster") == null) {
|
||||
if (System.getProperty("tests.cluster") != null) {
|
||||
throw new IllegalArgumentException("tests.rest.cluster and tests.cluster must both be null or non-null")
|
||||
}
|
||||
// we pass all nodes to the rest cluster to allow the clients to round-robin between them
|
||||
// this is more realistic than just talking to a single node
|
||||
runner.systemProperty('tests.rest.cluster', "${-> nodes.collect{it.httpUri()}.join(",")}")
|
||||
runner.systemProperty('tests.config.dir', "${-> nodes[0].pathConf}")
|
||||
// TODO: our "client" qa tests currently use the rest-test plugin. instead they should have their own plugin
|
||||
// that sets up the test cluster and passes this transport uri instead of http uri. Until then, we pass
|
||||
// both as separate sysprops
|
||||
runner.systemProperty('tests.cluster', "${-> nodes[0].transportUri()}")
|
||||
|
||||
// dump errors and warnings from cluster log on failure
|
||||
TaskExecutionAdapter logDumpListener = new TaskExecutionAdapter() {
|
||||
@Override
|
||||
void afterExecute(Task task, TaskState state) {
|
||||
if (state.failure != null) {
|
||||
for (NodeInfo nodeInfo : nodes) {
|
||||
printLogExcerpt(nodeInfo)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
runner.doFirst {
|
||||
project.gradle.addListener(logDumpListener)
|
||||
}
|
||||
runner.doLast {
|
||||
project.gradle.removeListener(logDumpListener)
|
||||
runner.doFirst {
|
||||
project.gradle.addListener(logDumpListener)
|
||||
}
|
||||
runner.doLast {
|
||||
project.gradle.removeListener(logDumpListener)
|
||||
}
|
||||
} else {
|
||||
if (System.getProperty("tests.cluster") == null) {
|
||||
throw new IllegalArgumentException("tests.rest.cluster and tests.cluster must both be null or non-null")
|
||||
}
|
||||
// an external cluster was specified and all responsibility for cluster configuration is taken by the user
|
||||
runner.systemProperty('tests.rest.cluster', System.getProperty("tests.rest.cluster"))
|
||||
runner.systemProperty('test.cluster', System.getProperty("tests.cluster"))
|
||||
}
|
||||
|
||||
// copy the rest spec/tests into the test resources
|
||||
|
@ -109,7 +122,10 @@ public class RestIntegTestTask extends DefaultTask {
|
|||
clusterInit.enabled = false
|
||||
return // no need to add cluster formation tasks if the task won't run!
|
||||
}
|
||||
nodes = ClusterFormationTasks.setup(project, "${name}Cluster", runner, clusterConfig)
|
||||
// only create the cluster if needed as otherwise an external cluster to use was specified
|
||||
if (System.getProperty("tests.rest.cluster") == null) {
|
||||
nodes = ClusterFormationTasks.setup(project, "${name}Cluster", runner, clusterConfig)
|
||||
}
|
||||
super.dependsOn(runner.finalizedBy)
|
||||
}
|
||||
}
|
||||
|
|
|
@ -49,7 +49,9 @@ task createPluginsDir(type: EmptyDirTask) {
|
|||
CopySpec archiveFiles(CopySpec modulesFiles, String distributionType, boolean oss) {
|
||||
return copySpec {
|
||||
into("elasticsearch-${version}") {
|
||||
with libFiles
|
||||
into('lib') {
|
||||
with libFiles
|
||||
}
|
||||
into('config') {
|
||||
dirMode 0750
|
||||
fileMode 0660
|
||||
|
|
|
@ -227,13 +227,15 @@ configure(subprojects.findAll { ['archives', 'packages'].contains(it.name) }) {
|
|||
* Common files in all distributions *
|
||||
*****************************************************************************/
|
||||
libFiles = copySpec {
|
||||
into 'lib'
|
||||
// delay by using closures, since they have not yet been configured, so no jar task exists yet
|
||||
from { project(':server').jar }
|
||||
from { project(':server').configurations.runtime }
|
||||
from { project(':libs:plugin-classloader').jar }
|
||||
// delay add tools using closures, since they have not yet been configured, so no jar task exists yet
|
||||
from { project(':distribution:tools:launchers').jar }
|
||||
from { project(':distribution:tools:plugin-cli').jar }
|
||||
into('tools/plugin-cli') {
|
||||
from { project(':distribution:tools:plugin-cli').jar }
|
||||
from { project(':distribution:tools:plugin-cli').configurations.runtime }
|
||||
}
|
||||
}
|
||||
|
||||
modulesFiles = { oss ->
|
||||
|
|
|
@ -124,13 +124,23 @@ Closure commonPackageConfig(String type, boolean oss) {
|
|||
include 'README.textile'
|
||||
fileMode 0644
|
||||
}
|
||||
into('lib') {
|
||||
with copySpec {
|
||||
with libFiles
|
||||
// we need to specify every intermediate directory so we iterate through the parents; duplicate calls with the same part are fine
|
||||
eachFile { FileCopyDetails fcp ->
|
||||
String[] segments = fcp.relativePath.segments
|
||||
for (int i = segments.length - 2; i > 0 && segments[i] != 'lib'; --i) {
|
||||
directory('/' + segments[0..i].join('/'), 0755)
|
||||
}
|
||||
fcp.mode = 0644
|
||||
}
|
||||
}
|
||||
}
|
||||
into('modules') {
|
||||
with copySpec {
|
||||
with modulesFiles(oss)
|
||||
// we need to specify every intermediate directory, but modules could have sub directories
|
||||
// and there might not be any files as direct children of intermediates (eg platform)
|
||||
// so we must iterate through the parents, but duplicate calls with the same path
|
||||
// are ok (they don't show up in the built packages)
|
||||
// we need to specify every intermediate directory so we iterate through the parents; duplicate calls with the same part are fine
|
||||
eachFile { FileCopyDetails fcp ->
|
||||
String[] segments = fcp.relativePath.segments
|
||||
for (int i = segments.length - 2; i > 0 && segments[i] != 'modules'; --i) {
|
||||
|
@ -251,8 +261,8 @@ ospackage {
|
|||
signingKeyId = project.hasProperty('signing.keyId') ? project.property('signing.keyId') : 'D88E42B4'
|
||||
signingKeyPassphrase = project.property('signing.password')
|
||||
signingKeyRingFile = project.hasProperty('signing.secretKeyRingFile') ?
|
||||
project.file(project.property('signing.secretKeyRingFile')) :
|
||||
new File(new File(System.getProperty('user.home'), '.gnupg'), 'secring.gpg')
|
||||
project.file(project.property('signing.secretKeyRingFile')) :
|
||||
new File(new File(System.getProperty('user.home'), '.gnupg'), 'secring.gpg')
|
||||
}
|
||||
|
||||
requires('coreutils')
|
||||
|
@ -263,7 +273,6 @@ ospackage {
|
|||
permissionGroup 'root'
|
||||
|
||||
into '/usr/share/elasticsearch'
|
||||
with libFiles
|
||||
with noticeFile
|
||||
}
|
||||
|
||||
|
|
|
@ -10,6 +10,12 @@ do
|
|||
source "`dirname "$0"`"/$additional_source
|
||||
done
|
||||
|
||||
IFS=';' read -r -a additional_classpath_directories <<< "$ES_ADDITIONAL_CLASSPATH_DIRECTORIES"
|
||||
for additional_classpath_directory in "${additional_classpath_directories[@]}"
|
||||
do
|
||||
ES_CLASSPATH="$ES_CLASSPATH:$ES_HOME/$additional_classpath_directory/*"
|
||||
done
|
||||
|
||||
exec \
|
||||
"$JAVA" \
|
||||
$ES_JAVA_OPTS \
|
||||
|
|
|
@ -11,6 +11,12 @@ for /f "tokens=1*" %%a in ("%*") do (
|
|||
set arguments=%%b
|
||||
)
|
||||
|
||||
if defined ES_ADDITIONAL_CLASSPATH_DIRECTORIES (
|
||||
for %%a in ("%ES_ADDITIONAL_CLASSPATH_DIRECTORIES:;=","%") do (
|
||||
set ES_CLASSPATH=!ES_CLASSPATH!;!ES_HOME!/%%a/*
|
||||
)
|
||||
)
|
||||
|
||||
%JAVA% ^
|
||||
%ES_JAVA_OPTS% ^
|
||||
-Des.path.home="%ES_HOME%" ^
|
||||
|
|
|
@ -1,5 +1,6 @@
|
|||
#!/bin/bash
|
||||
|
||||
"`dirname "$0"`"/elasticsearch-cli \
|
||||
ES_ADDITIONAL_CLASSPATH_DIRECTORIES=lib/tools/plugin-cli \
|
||||
"`dirname "$0"`"/elasticsearch-cli \
|
||||
org.elasticsearch.plugins.PluginCli \
|
||||
"$@"
|
||||
|
|
|
@ -3,6 +3,7 @@
|
|||
setlocal enabledelayedexpansion
|
||||
setlocal enableextensions
|
||||
|
||||
set ES_ADDITIONAL_CLASSPATH_DIRECTORIES=lib/tools/plugin-cli
|
||||
call "%~dp0elasticsearch-cli.bat" ^
|
||||
org.elasticsearch.plugins.PluginCli ^
|
||||
%* ^
|
||||
|
|
|
@ -19,14 +19,22 @@
|
|||
|
||||
apply plugin: 'elasticsearch.build'
|
||||
|
||||
archivesBaseName = 'elasticsearch-plugin-cli'
|
||||
|
||||
dependencies {
|
||||
compileOnly "org.elasticsearch:elasticsearch:${version}"
|
||||
compileOnly "org.elasticsearch:elasticsearch-cli:${version}"
|
||||
compile "org.bouncycastle:bcpg-jdk15on:1.59"
|
||||
compile "org.bouncycastle:bcprov-jdk15on:1.59"
|
||||
testCompile "org.elasticsearch.test:framework:${version}"
|
||||
testCompile 'com.google.jimfs:jimfs:1.1'
|
||||
testCompile 'com.google.guava:guava:18.0'
|
||||
}
|
||||
|
||||
dependencyLicenses {
|
||||
mapping from: /bc.*/, to: 'bouncycastle'
|
||||
}
|
||||
|
||||
test {
|
||||
// TODO: find a way to add permissions for the tests in this module
|
||||
systemProperty 'tests.security.manager', 'false'
|
||||
|
|
|
@ -0,0 +1 @@
|
|||
ee93e5376bb6cf0a15c027b5f5e4393f2738e709
|
|
@ -0,0 +1 @@
|
|||
2507204241ab450456bdb8e8c0a8f986e418bd99
|
|
@ -0,0 +1,17 @@
|
|||
Copyright (c) 2000-2015 The Legion of the Bouncy Castle Inc. (http://www.bouncycastle.org)
|
||||
|
||||
Permission is hereby granted, free of charge, to any person obtaining a copy of this software
|
||||
and associated documentation files (the "Software"), to deal in the Software without restriction,
|
||||
including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||||
and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so,
|
||||
subject to the following conditions:
|
||||
|
||||
The above copyright notice and this permission notice shall be included in all copies or substantial
|
||||
portions of the Software.
|
||||
|
||||
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED,
|
||||
INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR
|
||||
PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
|
||||
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
|
||||
OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
|
||||
DEALINGS IN THE SOFTWARE.
|
|
@ -0,0 +1 @@
|
|||
|
|
@ -23,6 +23,16 @@ import joptsimple.OptionSet;
|
|||
import joptsimple.OptionSpec;
|
||||
import org.apache.lucene.search.spell.LevensteinDistance;
|
||||
import org.apache.lucene.util.CollectionUtil;
|
||||
import org.bouncycastle.jce.provider.BouncyCastleProvider;
|
||||
import org.bouncycastle.openpgp.PGPException;
|
||||
import org.bouncycastle.openpgp.PGPPublicKey;
|
||||
import org.bouncycastle.openpgp.PGPPublicKeyRingCollection;
|
||||
import org.bouncycastle.openpgp.PGPSignature;
|
||||
import org.bouncycastle.openpgp.PGPSignatureList;
|
||||
import org.bouncycastle.openpgp.PGPUtil;
|
||||
import org.bouncycastle.openpgp.jcajce.JcaPGPObjectFactory;
|
||||
import org.bouncycastle.openpgp.operator.jcajce.JcaKeyFingerprintCalculator;
|
||||
import org.bouncycastle.openpgp.operator.jcajce.JcaPGPContentVerifierBuilderProvider;
|
||||
import org.elasticsearch.Build;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.bootstrap.JarHell;
|
||||
|
@ -37,12 +47,14 @@ import org.elasticsearch.core.internal.io.IOUtils;
|
|||
import org.elasticsearch.env.Environment;
|
||||
|
||||
import java.io.BufferedReader;
|
||||
import java.io.ByteArrayInputStream;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStream;
|
||||
import java.io.InputStreamReader;
|
||||
import java.io.OutputStream;
|
||||
import java.net.HttpURLConnection;
|
||||
import java.net.URI;
|
||||
import java.net.URISyntaxException;
|
||||
import java.net.URL;
|
||||
import java.net.URLConnection;
|
||||
import java.net.URLDecoder;
|
||||
|
@ -59,8 +71,10 @@ import java.nio.file.attribute.PosixFileAttributes;
|
|||
import java.nio.file.attribute.PosixFilePermission;
|
||||
import java.nio.file.attribute.PosixFilePermissions;
|
||||
import java.security.MessageDigest;
|
||||
import java.security.NoSuchAlgorithmException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Base64;
|
||||
import java.util.Collections;
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
|
@ -116,7 +130,6 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
|
|||
/** The plugin zip is not properly structured. */
|
||||
static final int PLUGIN_MALFORMED = 2;
|
||||
|
||||
|
||||
/** The builtin modules, which are plugins, but cannot be installed or removed. */
|
||||
static final Set<String> MODULES;
|
||||
static {
|
||||
|
@ -241,7 +254,7 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
|
|||
if (OFFICIAL_PLUGINS.contains(pluginId)) {
|
||||
final String url = getElasticUrl(terminal, getStagingHash(), Version.CURRENT, isSnapshot(), pluginId, Platforms.PLATFORM_NAME);
|
||||
terminal.println("-> Downloading " + pluginId + " from elastic");
|
||||
return downloadZipAndChecksum(terminal, url, tmpDir, false);
|
||||
return downloadAndValidate(terminal, url, tmpDir, true);
|
||||
}
|
||||
|
||||
// now try as maven coordinates, a valid URL would only have a colon and slash
|
||||
|
@ -249,7 +262,7 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
|
|||
if (coordinates.length == 3 && pluginId.contains("/") == false && pluginId.startsWith("file:") == false) {
|
||||
String mavenUrl = getMavenUrl(terminal, coordinates, Platforms.PLATFORM_NAME);
|
||||
terminal.println("-> Downloading " + pluginId + " from maven central");
|
||||
return downloadZipAndChecksum(terminal, mavenUrl, tmpDir, true);
|
||||
return downloadAndValidate(terminal, mavenUrl, tmpDir, false);
|
||||
}
|
||||
|
||||
// fall back to plain old URL
|
||||
|
@ -406,16 +419,44 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
|
|||
}
|
||||
}
|
||||
|
||||
/** Downloads a zip from the url, as well as a SHA512 (or SHA1) checksum, and checks the checksum. */
|
||||
// pkg private for tests
|
||||
@SuppressForbidden(reason = "We use openStream to download plugins")
|
||||
private Path downloadZipAndChecksum(Terminal terminal, String urlString, Path tmpDir, boolean allowSha1) throws Exception {
|
||||
@SuppressForbidden(reason = "URL#openStream")
|
||||
private InputStream urlOpenStream(final URL url) throws IOException {
|
||||
return url.openStream();
|
||||
}
|
||||
|
||||
/**
|
||||
* Downloads a ZIP from the URL. This method also validates the downloaded plugin ZIP via the following means:
|
||||
* <ul>
|
||||
* <li>
|
||||
* For an official plugin we download the SHA-512 checksum and validate the integrity of the downloaded ZIP. We also download the
|
||||
* armored signature and validate the authenticity of the downloaded ZIP.
|
||||
* </li>
|
||||
* <li>
|
||||
* For a non-official plugin we download the SHA-512 checksum and fallback to the SHA-1 checksum and validate the integrity of the
|
||||
* downloaded ZIP.
|
||||
* </li>
|
||||
* </ul>
|
||||
*
|
||||
* @param terminal a terminal to log messages to
|
||||
* @param urlString the URL of the plugin ZIP
|
||||
* @param tmpDir a temporary directory to write downloaded files to
|
||||
* @param officialPlugin true if the plugin is an official plugin
|
||||
* @return the path to the downloaded plugin ZIP
|
||||
* @throws IOException if an I/O exception occurs download or reading files and resources
|
||||
* @throws PGPException if an exception occurs verifying the downloaded ZIP signature
|
||||
* @throws UserException if checksum validation fails
|
||||
*/
|
||||
private Path downloadAndValidate(
|
||||
final Terminal terminal,
|
||||
final String urlString,
|
||||
final Path tmpDir,
|
||||
final boolean officialPlugin) throws IOException, PGPException, UserException {
|
||||
Path zip = downloadZip(terminal, urlString, tmpDir);
|
||||
pathsToDeleteOnShutdown.add(zip);
|
||||
String checksumUrlString = urlString + ".sha512";
|
||||
URL checksumUrl = openUrl(checksumUrlString);
|
||||
String digestAlgo = "SHA-512";
|
||||
if (checksumUrl == null && allowSha1) {
|
||||
if (checksumUrl == null && officialPlugin == false) {
|
||||
// fallback to sha1, until 7.0, but with warning
|
||||
terminal.println("Warning: sha512 not found, falling back to sha1. This behavior is deprecated and will be removed in a " +
|
||||
"future release. Please update the plugin to use a sha512 checksum.");
|
||||
|
@ -427,7 +468,7 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
|
|||
throw new UserException(ExitCodes.IO_ERROR, "Plugin checksum missing: " + checksumUrlString);
|
||||
}
|
||||
final String expectedChecksum;
|
||||
try (InputStream in = checksumUrl.openStream()) {
|
||||
try (InputStream in = urlOpenStream(checksumUrl)) {
|
||||
/*
|
||||
* The supported format of the SHA-1 files is a single-line file containing the SHA-1. The supported format of the SHA-512 files
|
||||
* is a single-line file containing the SHA-512 and the filename, separated by two spaces. For SHA-1, we verify that the hash
|
||||
|
@ -465,23 +506,119 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
|
|||
}
|
||||
}
|
||||
|
||||
byte[] zipbytes = Files.readAllBytes(zip);
|
||||
String gotChecksum = MessageDigests.toHexString(MessageDigest.getInstance(digestAlgo).digest(zipbytes));
|
||||
if (expectedChecksum.equals(gotChecksum) == false) {
|
||||
throw new UserException(ExitCodes.IO_ERROR,
|
||||
digestAlgo + " mismatch, expected " + expectedChecksum + " but got " + gotChecksum);
|
||||
try {
|
||||
final byte[] zipBytes = Files.readAllBytes(zip);
|
||||
final String actualChecksum = MessageDigests.toHexString(MessageDigest.getInstance(digestAlgo).digest(zipBytes));
|
||||
if (expectedChecksum.equals(actualChecksum) == false) {
|
||||
throw new UserException(
|
||||
ExitCodes.IO_ERROR,
|
||||
digestAlgo + " mismatch, expected " + expectedChecksum + " but got " + actualChecksum);
|
||||
}
|
||||
} catch (final NoSuchAlgorithmException e) {
|
||||
// this should never happen as we are using SHA-1 and SHA-512 here
|
||||
throw new AssertionError(e);
|
||||
}
|
||||
|
||||
if (officialPlugin) {
|
||||
verifySignature(zip, urlString);
|
||||
}
|
||||
|
||||
return zip;
|
||||
}
|
||||
|
||||
/**
|
||||
* Verify the signature of the downloaded plugin ZIP. The signature is obtained from the source of the downloaded plugin by appending
|
||||
* ".asc" to the URL. It is expected that the plugin is signed with the Elastic signing key with ID D27D666CD88E42B4.
|
||||
*
|
||||
* @param zip the path to the downloaded plugin ZIP
|
||||
* @param urlString the URL source of the downloade plugin ZIP
|
||||
* @throws IOException if an I/O exception occurs reading from various input streams
|
||||
* @throws PGPException if the PGP implementation throws an internal exception during verification
|
||||
*/
|
||||
void verifySignature(final Path zip, final String urlString) throws IOException, PGPException {
|
||||
final String ascUrlString = urlString + ".asc";
|
||||
final URL ascUrl = openUrl(ascUrlString);
|
||||
try (
|
||||
// fin is a file stream over the downloaded plugin zip whose signature to verify
|
||||
InputStream fin = pluginZipInputStream(zip);
|
||||
// sin is a URL stream to the signature corresponding to the downloaded plugin zip
|
||||
InputStream sin = urlOpenStream(ascUrl);
|
||||
// pin is a input stream to the public key in ASCII-Armor format (RFC4880); the Armor data is in RFC2045 format
|
||||
InputStream pin = getPublicKey()) {
|
||||
final JcaPGPObjectFactory factory = new JcaPGPObjectFactory(PGPUtil.getDecoderStream(sin));
|
||||
final PGPSignature signature = ((PGPSignatureList) factory.nextObject()).get(0);
|
||||
|
||||
// validate the signature has key ID matching our public key ID
|
||||
final String keyId = Long.toHexString(signature.getKeyID()).toUpperCase(Locale.ROOT);
|
||||
if (getPublicKeyId().equals(keyId) == false) {
|
||||
throw new IllegalStateException("key id [" + keyId + "] does not match expected key id [" + getPublicKeyId() + "]");
|
||||
}
|
||||
|
||||
// compute the signature of the downloaded plugin zip
|
||||
final List<String> lines =
|
||||
new BufferedReader(new InputStreamReader(pin, StandardCharsets.UTF_8)).lines().collect(Collectors.toList());
|
||||
// skip armor headers and possible blank line
|
||||
int index = 1;
|
||||
for (; index < lines.size(); index++) {
|
||||
if (lines.get(index).matches(".*: .*") == false && lines.get(index).matches("\\s*") == false) {
|
||||
break;
|
||||
}
|
||||
}
|
||||
final byte[] armoredData =
|
||||
lines.subList(index, lines.size() - 1).stream().collect(Collectors.joining("\n")).getBytes(StandardCharsets.UTF_8);
|
||||
final InputStream ain = Base64.getMimeDecoder().wrap(new ByteArrayInputStream(armoredData));
|
||||
final PGPPublicKeyRingCollection collection = new PGPPublicKeyRingCollection(ain, new JcaKeyFingerprintCalculator());
|
||||
final PGPPublicKey key = collection.getPublicKey(signature.getKeyID());
|
||||
signature.init(new JcaPGPContentVerifierBuilderProvider().setProvider(new BouncyCastleProvider()), key);
|
||||
final byte[] buffer = new byte[1024];
|
||||
int read;
|
||||
while ((read = fin.read(buffer)) != -1) {
|
||||
signature.update(buffer, 0, read);
|
||||
}
|
||||
|
||||
// finally we verify the signature of the downloaded plugin zip matches the expected signature
|
||||
if (signature.verify() == false) {
|
||||
throw new IllegalStateException("signature verification for [" + urlString + "] failed");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* An input stream to the raw bytes of the plugin ZIP.
|
||||
*
|
||||
* @param zip the path to the downloaded plugin ZIP
|
||||
* @return an input stream to the raw bytes of the plugin ZIP.
|
||||
* @throws IOException if an I/O exception occurs preparing the input stream
|
||||
*/
|
||||
InputStream pluginZipInputStream(final Path zip) throws IOException {
|
||||
return Files.newInputStream(zip);
|
||||
}
|
||||
|
||||
/**
|
||||
* Return the public key ID of the signing key that is expected to have signed the official plugin.
|
||||
*
|
||||
* @return the public key ID
|
||||
*/
|
||||
String getPublicKeyId() {
|
||||
return "D27D666CD88E42B4";
|
||||
}
|
||||
|
||||
/**
|
||||
* An input stream to the public key of the signing key.
|
||||
*
|
||||
* @return an input stream to the public key
|
||||
*/
|
||||
InputStream getPublicKey() {
|
||||
return InstallPluginCommand.class.getResourceAsStream("/public_key.asc");
|
||||
}
|
||||
|
||||
/**
|
||||
* Creates a URL and opens a connection.
|
||||
*
|
||||
* If the URL returns a 404, {@code null} is returned, otherwise the open URL opject is returned.
|
||||
*/
|
||||
// pkg private for tests
|
||||
URL openUrl(String urlString) throws Exception {
|
||||
URL openUrl(String urlString) throws IOException {
|
||||
URL checksumUrl = new URL(urlString);
|
||||
HttpURLConnection connection = (HttpURLConnection)checksumUrl.openConnection();
|
||||
if (connection.getResponseCode() == 404) {
|
||||
|
@ -605,11 +742,27 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
|
|||
return info;
|
||||
}
|
||||
|
||||
private static final String LIB_TOOLS_PLUGIN_CLI_CLASSPATH_JAR;
|
||||
|
||||
static {
|
||||
LIB_TOOLS_PLUGIN_CLI_CLASSPATH_JAR =
|
||||
String.format(Locale.ROOT, ".+%1$slib%1$stools%1$splugin-cli%1$s[^%1$s]+\\.jar", "(/|\\\\)");
|
||||
}
|
||||
|
||||
/** check a candidate plugin for jar hell before installing it */
|
||||
void jarHellCheck(PluginInfo candidateInfo, Path candidateDir, Path pluginsDir, Path modulesDir) throws Exception {
|
||||
// create list of current jars in classpath
|
||||
final Set<URL> jars = new HashSet<>(JarHell.parseClassPath());
|
||||
|
||||
final Set<URL> classpath =
|
||||
JarHell.parseClassPath()
|
||||
.stream()
|
||||
.filter(url -> {
|
||||
try {
|
||||
return url.toURI().getPath().matches(LIB_TOOLS_PLUGIN_CLI_CLASSPATH_JAR) == false;
|
||||
} catch (final URISyntaxException e) {
|
||||
throw new AssertionError(e);
|
||||
}
|
||||
})
|
||||
.collect(Collectors.toSet());
|
||||
|
||||
// read existing bundles. this does some checks on the installation too.
|
||||
Set<PluginsService.Bundle> bundles = new HashSet<>(PluginsService.getPluginBundles(pluginsDir));
|
||||
|
@ -621,7 +774,7 @@ class InstallPluginCommand extends EnvironmentAwareCommand {
|
|||
// TODO: optimize to skip any bundles not connected to the candidate plugin?
|
||||
Map<String, Set<URL>> transitiveUrls = new HashMap<>();
|
||||
for (PluginsService.Bundle bundle : sortedBundles) {
|
||||
PluginsService.checkBundleJarHell(bundle, transitiveUrls);
|
||||
PluginsService.checkBundleJarHell(classpath, bundle, transitiveUrls);
|
||||
}
|
||||
|
||||
// TODO: no jars should be an error
|
||||
|
|
|
@ -0,0 +1,29 @@
|
|||
-----BEGIN PGP PUBLIC KEY BLOCK-----
|
||||
Version: SKS 1.1.6
|
||||
Comment: Hostname: pgp.mit.edu
|
||||
|
||||
mQENBFI3HsoBCADXDtbNJnxbPqB1vDNtCsqhe49vFYsZN9IOZsZXgp7aHjh6CJBDA+bGFOwy
|
||||
hbd7at35jQjWAw1O3cfYsKAmFy+Ar3LHCMkV3oZspJACTIgCrwnkic/9CUliQe324qvObU2Q
|
||||
RtP4Fl0zWcfb/S8UYzWXWIFuJqMvE9MaRY1bwUBvzoqavLGZj3SF1SPO+TB5QrHkrQHBsmX+
|
||||
Jda6d4Ylt8/t6CvMwgQNlrlzIO9WT+YN6zS+sqHd1YK/aY5qhoLNhp9G/HxhcSVCkLq8SStj
|
||||
1ZZ1S9juBPoXV1ZWNbxFNGwOh/NYGldD2kmBf3YgCqeLzHahsAEpvAm8TBa7Q9W21C8vABEB
|
||||
AAG0RUVsYXN0aWNzZWFyY2ggKEVsYXN0aWNzZWFyY2ggU2lnbmluZyBLZXkpIDxkZXZfb3Bz
|
||||
QGVsYXN0aWNzZWFyY2gub3JnPokBOAQTAQIAIgUCUjceygIbAwYLCQgHAwIGFQgCCQoLBBYC
|
||||
AwECHgECF4AACgkQ0n1mbNiOQrRzjAgAlTUQ1mgo3nK6BGXbj4XAJvuZDG0HILiUt+pPnz75
|
||||
nsf0NWhqR4yGFlmpuctgCmTD+HzYtV9fp9qW/bwVuJCNtKXk3sdzYABY+Yl0Cez/7C2GuGCO
|
||||
lbn0luCNT9BxJnh4mC9h/cKI3y5jvZ7wavwe41teqG14V+EoFSn3NPKmTxcDTFrV7SmVPxCB
|
||||
cQze00cJhprKxkuZMPPVqpBS+JfDQtzUQD/LSFfhHj9eD+Xe8d7sw+XvxB2aN4gnTlRzjL1n
|
||||
TRp0h2/IOGkqYfIG9rWmSLNlxhB2t+c0RsjdGM4/eRlPWylFbVMc5pmDpItrkWSnzBfkmXL3
|
||||
vO2X3WvwmSFiQbkBDQRSNx7KAQgA5JUlzcMW5/cuyZR8alSacKqhSbvoSqqbzHKcUQZmlzNM
|
||||
KGTABFG1yRx9r+wa/fvqP6OTRzRDvVS/cycws8YX7Ddum7x8uI95b9ye1/Xy5noPEm8cD+hp
|
||||
lnpU+PBQZJ5XJ2I+1l9Nixx47wPGXeClLqcdn0ayd+v+Rwf3/XUJrvccG2YZUiQ4jWZkoxsA
|
||||
07xx7Bj+Lt8/FKG7sHRFvePFU0ZS6JFx9GJqjSBbHRRkam+4emW3uWgVfZxuwcUCn1ayNgRt
|
||||
KiFv9jQrg2TIWEvzYx9tywTCxc+FFMWAlbCzi+m4WD+QUWWfDQ009U/WM0ks0KwwEwSk/UDu
|
||||
ToxGnKU2dQARAQABiQEfBBgBAgAJBQJSNx7KAhsMAAoJENJ9ZmzYjkK0c3MIAIE9hAR20mqJ
|
||||
WLcsxLtrRs6uNF1VrpB+4n/55QU7oxA1iVBO6IFu4qgsF12JTavnJ5MLaETlggXY+zDef9sy
|
||||
TPXoQctpzcaNVDmedwo1SiL03uMoblOvWpMR/Y0j6rm7IgrMWUDXDPvoPGjMl2q1iTeyHkMZ
|
||||
EyUJ8SKsaHh4jV9wp9KmC8C+9CwMukL7vM5w8cgvJoAwsp3Fn59AxWthN3XJYcnMfStkIuWg
|
||||
R7U2r+a210W6vnUxU4oN0PmMcursYPyeV0NX/KQeUeNMwGTFB6QHS/anRaGQewijkrYYoTNt
|
||||
fllxIu9XYmiBERQ/qPDlGRlOgVTd9xUfHFkzB52c70E=
|
||||
=92oX
|
||||
-----END PGP PUBLIC KEY BLOCK-----
|
|
@ -23,6 +23,26 @@ import com.carrotsearch.randomizedtesting.annotations.ParametersFactory;
|
|||
import com.google.common.jimfs.Configuration;
|
||||
import com.google.common.jimfs.Jimfs;
|
||||
import org.apache.lucene.util.LuceneTestCase;
|
||||
import org.apache.lucene.util.LuceneTestCase.AwaitsFix;
|
||||
import org.bouncycastle.bcpg.ArmoredOutputStream;
|
||||
import org.bouncycastle.bcpg.BCPGOutputStream;
|
||||
import org.bouncycastle.bcpg.HashAlgorithmTags;
|
||||
import org.bouncycastle.jce.provider.BouncyCastleProvider;
|
||||
import org.bouncycastle.openpgp.PGPEncryptedData;
|
||||
import org.bouncycastle.openpgp.PGPException;
|
||||
import org.bouncycastle.openpgp.PGPKeyPair;
|
||||
import org.bouncycastle.openpgp.PGPPrivateKey;
|
||||
import org.bouncycastle.openpgp.PGPPublicKey;
|
||||
import org.bouncycastle.openpgp.PGPSecretKey;
|
||||
import org.bouncycastle.openpgp.PGPSignature;
|
||||
import org.bouncycastle.openpgp.PGPSignatureGenerator;
|
||||
import org.bouncycastle.openpgp.operator.PGPDigestCalculator;
|
||||
import org.bouncycastle.openpgp.operator.bc.BcPBESecretKeyDecryptorBuilder;
|
||||
import org.bouncycastle.openpgp.operator.bc.BcPGPContentSignerBuilder;
|
||||
import org.bouncycastle.openpgp.operator.jcajce.JcaPGPContentSignerBuilder;
|
||||
import org.bouncycastle.openpgp.operator.jcajce.JcaPGPDigestCalculatorProviderBuilder;
|
||||
import org.bouncycastle.openpgp.operator.jcajce.JcaPGPKeyPair;
|
||||
import org.bouncycastle.openpgp.operator.jcajce.JcePBESecretKeyEncryptorBuilder;
|
||||
import org.elasticsearch.Build;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.cli.ExitCodes;
|
||||
|
@ -44,6 +64,8 @@ import org.junit.After;
|
|||
import org.junit.Before;
|
||||
|
||||
import java.io.BufferedReader;
|
||||
import java.io.ByteArrayInputStream;
|
||||
import java.io.ByteArrayOutputStream;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStream;
|
||||
import java.io.StringReader;
|
||||
|
@ -66,13 +88,19 @@ import java.nio.file.attribute.PosixFileAttributeView;
|
|||
import java.nio.file.attribute.PosixFileAttributes;
|
||||
import java.nio.file.attribute.PosixFilePermission;
|
||||
import java.nio.file.attribute.UserPrincipal;
|
||||
import java.security.KeyPair;
|
||||
import java.security.KeyPairGenerator;
|
||||
import java.security.MessageDigest;
|
||||
import java.security.NoSuchAlgorithmException;
|
||||
import java.security.NoSuchProviderException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Date;
|
||||
import java.util.HashSet;
|
||||
import java.util.List;
|
||||
import java.util.Locale;
|
||||
import java.util.Set;
|
||||
import java.util.function.BiFunction;
|
||||
import java.util.function.Function;
|
||||
import java.util.stream.Collectors;
|
||||
import java.util.stream.Stream;
|
||||
|
@ -88,6 +116,7 @@ import static org.hamcrest.Matchers.hasToString;
|
|||
import static org.hamcrest.Matchers.not;
|
||||
|
||||
@LuceneTestCase.SuppressFileSystems("*")
|
||||
@AwaitsFix(bugUrl = "https://github.com/elastic/elasticsearch/issues/30900")
|
||||
public class InstallPluginCommandTests extends ESTestCase {
|
||||
|
||||
private InstallPluginCommand skipJarHellCommand;
|
||||
|
@ -800,8 +829,16 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
skipJarHellCommand.execute(terminal, pluginZip, isBatch, env.v2());
|
||||
}
|
||||
|
||||
void assertInstallPluginFromUrl(String pluginId, String name, String url, String stagingHash, boolean isSnapshot,
|
||||
String shaExtension, Function<byte[], String> shaCalculator) throws Exception {
|
||||
void assertInstallPluginFromUrl(
|
||||
final String pluginId,
|
||||
final String name,
|
||||
final String url,
|
||||
final String stagingHash,
|
||||
final boolean isSnapshot,
|
||||
final String shaExtension,
|
||||
final Function<byte[], String> shaCalculator,
|
||||
final PGPSecretKey secretKey,
|
||||
final BiFunction<byte[], PGPSecretKey, String> signature) throws Exception {
|
||||
Tuple<Path, Environment> env = createEnv(fs, temp);
|
||||
Path pluginDir = createPluginDir(temp);
|
||||
Path pluginZip = createPlugin(name, pluginDir);
|
||||
|
@ -814,18 +851,56 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
return downloadedPath;
|
||||
}
|
||||
@Override
|
||||
URL openUrl(String urlString) throws Exception {
|
||||
String expectedUrl = url + shaExtension;
|
||||
if (expectedUrl.equals(urlString)) {
|
||||
URL openUrl(String urlString) throws IOException {
|
||||
if ((url + shaExtension).equals(urlString)) {
|
||||
// calc sha an return file URL to it
|
||||
Path shaFile = temp.apply("shas").resolve("downloaded.zip" + shaExtension);
|
||||
byte[] zipbytes = Files.readAllBytes(pluginZip);
|
||||
String checksum = shaCalculator.apply(zipbytes);
|
||||
Files.write(shaFile, checksum.getBytes(StandardCharsets.UTF_8));
|
||||
return shaFile.toUri().toURL();
|
||||
} else if ((url + ".asc").equals(urlString)) {
|
||||
final Path ascFile = temp.apply("asc").resolve("downloaded.zip" + ".asc");
|
||||
final byte[] zipBytes = Files.readAllBytes(pluginZip);
|
||||
final String asc = signature.apply(zipBytes, secretKey);
|
||||
Files.write(ascFile, asc.getBytes(StandardCharsets.UTF_8));
|
||||
return ascFile.toUri().toURL();
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
@Override
|
||||
void verifySignature(Path zip, String urlString) throws IOException, PGPException {
|
||||
if (InstallPluginCommand.OFFICIAL_PLUGINS.contains(name)) {
|
||||
super.verifySignature(zip, urlString);
|
||||
} else {
|
||||
throw new UnsupportedOperationException("verify signature should not be called for unofficial plugins");
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
InputStream pluginZipInputStream(Path zip) throws IOException {
|
||||
return new ByteArrayInputStream(Files.readAllBytes(zip));
|
||||
}
|
||||
|
||||
@Override
|
||||
String getPublicKeyId() {
|
||||
return Long.toHexString(secretKey.getKeyID()).toUpperCase(Locale.ROOT);
|
||||
}
|
||||
|
||||
@Override
|
||||
InputStream getPublicKey() {
|
||||
try {
|
||||
final ByteArrayOutputStream output = new ByteArrayOutputStream();
|
||||
final ArmoredOutputStream armored = new ArmoredOutputStream(output);
|
||||
secretKey.getPublicKey().encode(armored);
|
||||
armored.close();
|
||||
return new ByteArrayInputStream(output.toByteArray());
|
||||
} catch (final IOException e) {
|
||||
throw new AssertionError(e);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
boolean urlExists(Terminal terminal, String urlString) throws IOException {
|
||||
return urlString.equals(url);
|
||||
|
@ -851,11 +926,12 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
|
||||
public void assertInstallPluginFromUrl(
|
||||
final String pluginId, final String name, final String url, final String stagingHash, boolean isSnapshot) throws Exception {
|
||||
MessageDigest digest = MessageDigest.getInstance("SHA-512");
|
||||
assertInstallPluginFromUrl(pluginId, name, url, stagingHash, isSnapshot, ".sha512", checksumAndFilename(digest, url));
|
||||
final MessageDigest digest = MessageDigest.getInstance("SHA-512");
|
||||
assertInstallPluginFromUrl(
|
||||
pluginId, name, url, stagingHash, isSnapshot, ".sha512", checksumAndFilename(digest, url), newSecretKey(), this::signature);
|
||||
}
|
||||
|
||||
public void testOfficalPlugin() throws Exception {
|
||||
public void testOfficialPlugin() throws Exception {
|
||||
String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Version.CURRENT + ".zip";
|
||||
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false);
|
||||
}
|
||||
|
@ -883,13 +959,13 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
e, hasToString(containsString("attempted to install release build of official plugin on snapshot build of Elasticsearch")));
|
||||
}
|
||||
|
||||
public void testOfficalPluginStaging() throws Exception {
|
||||
public void testOfficialPluginStaging() throws Exception {
|
||||
String url = "https://staging.elastic.co/" + Version.CURRENT + "-abc123/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-"
|
||||
+ Version.CURRENT + ".zip";
|
||||
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, "abc123", false);
|
||||
}
|
||||
|
||||
public void testOfficalPlatformPlugin() throws Exception {
|
||||
public void testOfficialPlatformPlugin() throws Exception {
|
||||
String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Platforms.PLATFORM_NAME +
|
||||
"-" + Version.CURRENT + ".zip";
|
||||
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false);
|
||||
|
@ -905,7 +981,7 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, "abc123", true);
|
||||
}
|
||||
|
||||
public void testOfficalPlatformPluginStaging() throws Exception {
|
||||
public void testOfficialPlatformPluginStaging() throws Exception {
|
||||
String url = "https://staging.elastic.co/" + Version.CURRENT + "-abc123/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-"
|
||||
+ Platforms.PLATFORM_NAME + "-"+ Version.CURRENT + ".zip";
|
||||
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, "abc123", false);
|
||||
|
@ -924,7 +1000,7 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
public void testMavenSha1Backcompat() throws Exception {
|
||||
String url = "https://repo1.maven.org/maven2/mygroup/myplugin/1.0.0/myplugin-1.0.0.zip";
|
||||
MessageDigest digest = MessageDigest.getInstance("SHA-1");
|
||||
assertInstallPluginFromUrl("mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".sha1", checksum(digest));
|
||||
assertInstallPluginFromUrl("mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".sha1", checksum(digest), null, (b, p) -> null);
|
||||
assertTrue(terminal.getOutput(), terminal.getOutput().contains("sha512 not found, falling back to sha1"));
|
||||
}
|
||||
|
||||
|
@ -932,7 +1008,7 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Version.CURRENT + ".zip";
|
||||
MessageDigest digest = MessageDigest.getInstance("SHA-1");
|
||||
UserException e = expectThrows(UserException.class, () ->
|
||||
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false, ".sha1", checksum(digest)));
|
||||
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false, ".sha1", checksum(digest), null, (b, p) -> null));
|
||||
assertEquals(ExitCodes.IO_ERROR, e.exitCode);
|
||||
assertEquals("Plugin checksum missing: " + url + ".sha512", e.getMessage());
|
||||
}
|
||||
|
@ -940,7 +1016,8 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
public void testMavenShaMissing() throws Exception {
|
||||
String url = "https://repo1.maven.org/maven2/mygroup/myplugin/1.0.0/myplugin-1.0.0.zip";
|
||||
UserException e = expectThrows(UserException.class, () ->
|
||||
assertInstallPluginFromUrl("mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".dne", bytes -> null));
|
||||
assertInstallPluginFromUrl(
|
||||
"mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".dne", bytes -> null, null, (b, p) -> null));
|
||||
assertEquals(ExitCodes.IO_ERROR, e.exitCode);
|
||||
assertEquals("Plugin checksum missing: " + url + ".sha1", e.getMessage());
|
||||
}
|
||||
|
@ -948,8 +1025,9 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
public void testInvalidShaFileMissingFilename() throws Exception {
|
||||
String url = "https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/analysis-icu-" + Version.CURRENT + ".zip";
|
||||
MessageDigest digest = MessageDigest.getInstance("SHA-512");
|
||||
UserException e = expectThrows(UserException.class, () ->
|
||||
assertInstallPluginFromUrl("analysis-icu", "analysis-icu", url, null, false, ".sha512", checksum(digest)));
|
||||
UserException e = expectThrows(UserException.class,
|
||||
() -> assertInstallPluginFromUrl(
|
||||
"analysis-icu", "analysis-icu", url, null, false, ".sha512", checksum(digest), null, (b, p) -> null));
|
||||
assertEquals(ExitCodes.IO_ERROR, e.exitCode);
|
||||
assertTrue(e.getMessage(), e.getMessage().startsWith("Invalid checksum file"));
|
||||
}
|
||||
|
@ -965,7 +1043,9 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
null,
|
||||
false,
|
||||
".sha512",
|
||||
checksumAndString(digest, " repository-s3-" + Version.CURRENT + ".zip")));
|
||||
checksumAndString(digest, " repository-s3-" + Version.CURRENT + ".zip"),
|
||||
null,
|
||||
(b, p) -> null));
|
||||
assertEquals(ExitCodes.IO_ERROR, e.exitCode);
|
||||
assertThat(e, hasToString(matches("checksum file at \\[.*\\] is not for this plugin")));
|
||||
}
|
||||
|
@ -981,7 +1061,9 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
null,
|
||||
false,
|
||||
".sha512",
|
||||
checksumAndString(digest, " analysis-icu-" + Version.CURRENT + ".zip\nfoobar")));
|
||||
checksumAndString(digest, " analysis-icu-" + Version.CURRENT + ".zip\nfoobar"),
|
||||
null,
|
||||
(b, p) -> null));
|
||||
assertEquals(ExitCodes.IO_ERROR, e.exitCode);
|
||||
assertTrue(e.getMessage(), e.getMessage().startsWith("Invalid checksum file"));
|
||||
}
|
||||
|
@ -996,7 +1078,9 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
null,
|
||||
false,
|
||||
".sha512",
|
||||
bytes -> "foobar analysis-icu-" + Version.CURRENT + ".zip"));
|
||||
bytes -> "foobar analysis-icu-" + Version.CURRENT + ".zip",
|
||||
null,
|
||||
(b, p) -> null));
|
||||
assertEquals(ExitCodes.IO_ERROR, e.exitCode);
|
||||
assertTrue(e.getMessage(), e.getMessage().contains("SHA-512 mismatch, expected foobar"));
|
||||
}
|
||||
|
@ -1004,11 +1088,77 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
public void testSha1Mismatch() throws Exception {
|
||||
String url = "https://repo1.maven.org/maven2/mygroup/myplugin/1.0.0/myplugin-1.0.0.zip";
|
||||
UserException e = expectThrows(UserException.class, () ->
|
||||
assertInstallPluginFromUrl("mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".sha1", bytes -> "foobar"));
|
||||
assertInstallPluginFromUrl(
|
||||
"mygroup:myplugin:1.0.0", "myplugin", url, null, false, ".sha1", bytes -> "foobar", null, (b, p) -> null));
|
||||
assertEquals(ExitCodes.IO_ERROR, e.exitCode);
|
||||
assertTrue(e.getMessage(), e.getMessage().contains("SHA-1 mismatch, expected foobar"));
|
||||
}
|
||||
|
||||
public void testPublicKeyIdMismatchToExpectedPublicKeyId() throws Exception {
|
||||
final String icu = "analysis-icu";
|
||||
final String url =
|
||||
"https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/" + icu + "-" + Version.CURRENT + ".zip";
|
||||
final MessageDigest digest = MessageDigest.getInstance("SHA-512");
|
||||
/*
|
||||
* To setup a situation where the expected public key ID does not match the public key ID used for signing, we generate a new public
|
||||
* key at the moment of signing (see the signature invocation). Note that this key will not match the key that we push down to the
|
||||
* install plugin command.
|
||||
*/
|
||||
final PGPSecretKey signingKey = newSecretKey(); // the actual key used for signing
|
||||
final String actualID = Long.toHexString(signingKey.getKeyID()).toUpperCase(Locale.ROOT);
|
||||
final BiFunction<byte[], PGPSecretKey, String> signature = (b, p) -> signature(b, signingKey);
|
||||
final PGPSecretKey verifyingKey = newSecretKey(); // the expected key used for signing
|
||||
final String expectedID = Long.toHexString(verifyingKey.getKeyID()).toUpperCase(Locale.ROOT);
|
||||
final IllegalStateException e = expectThrows(
|
||||
IllegalStateException.class,
|
||||
() ->
|
||||
assertInstallPluginFromUrl(
|
||||
icu, icu, url, null, false, ".sha512", checksumAndFilename(digest, url), verifyingKey, signature));
|
||||
assertThat(e, hasToString(containsString("key id [" + actualID + "] does not match expected key id [" + expectedID + "]")));
|
||||
}
|
||||
|
||||
public void testFailedSignatureVerification() throws Exception {
|
||||
final String icu = "analysis-icu";
|
||||
final String url =
|
||||
"https://artifacts.elastic.co/downloads/elasticsearch-plugins/analysis-icu/" + icu + "-" + Version.CURRENT + ".zip";
|
||||
final MessageDigest digest = MessageDigest.getInstance("SHA-512");
|
||||
/*
|
||||
* To setup a situation where signature verification fails, we will mutate the input byte array by modifying a single byte to some
|
||||
* random byte value other than the actual value. This is enough to change the signature and cause verification to intentionally
|
||||
* fail.
|
||||
*/
|
||||
final BiFunction<byte[], PGPSecretKey, String> signature = (b, p) -> {
|
||||
final byte[] bytes = Arrays.copyOf(b, b.length);
|
||||
bytes[0] = randomValueOtherThan(b[0], ESTestCase::randomByte);
|
||||
return signature(bytes, p);
|
||||
};
|
||||
final IllegalStateException e = expectThrows(
|
||||
IllegalStateException.class,
|
||||
() ->
|
||||
assertInstallPluginFromUrl(
|
||||
icu, icu, url, null, false, ".sha512", checksumAndFilename(digest, url), newSecretKey(), signature));
|
||||
assertThat(e, hasToString(equalTo("java.lang.IllegalStateException: signature verification for [" + url + "] failed")));
|
||||
}
|
||||
|
||||
public PGPSecretKey newSecretKey() throws NoSuchAlgorithmException, NoSuchProviderException, PGPException {
|
||||
final KeyPairGenerator kpg = KeyPairGenerator.getInstance("RSA");
|
||||
kpg.initialize(2048);
|
||||
final KeyPair pair = kpg.generateKeyPair();
|
||||
final PGPDigestCalculator sha1Calc = new JcaPGPDigestCalculatorProviderBuilder().build().get(HashAlgorithmTags.SHA1);
|
||||
final PGPKeyPair pkp = new JcaPGPKeyPair(PGPPublicKey.RSA_GENERAL, pair, new Date());
|
||||
return new PGPSecretKey(
|
||||
PGPSignature.DEFAULT_CERTIFICATION,
|
||||
pkp,
|
||||
"example@example.com",
|
||||
sha1Calc,
|
||||
null,
|
||||
null,
|
||||
new JcaPGPContentSignerBuilder(pkp.getPublicKey().getAlgorithm(), HashAlgorithmTags.SHA1),
|
||||
new JcePBESecretKeyEncryptorBuilder(PGPEncryptedData.CAST5, sha1Calc)
|
||||
.setProvider(new BouncyCastleProvider())
|
||||
.build("passphrase".toCharArray()));
|
||||
}
|
||||
|
||||
private Function<byte[], String> checksum(final MessageDigest digest) {
|
||||
return checksumAndString(digest, "");
|
||||
}
|
||||
|
@ -1022,6 +1172,32 @@ public class InstallPluginCommandTests extends ESTestCase {
|
|||
return bytes -> MessageDigests.toHexString(digest.digest(bytes)) + s;
|
||||
}
|
||||
|
||||
private String signature(final byte[] bytes, final PGPSecretKey secretKey) {
|
||||
try {
|
||||
final PGPPrivateKey privateKey
|
||||
= secretKey.extractPrivateKey(
|
||||
new BcPBESecretKeyDecryptorBuilder(
|
||||
new JcaPGPDigestCalculatorProviderBuilder().build()).build("passphrase".toCharArray()));
|
||||
final PGPSignatureGenerator generator =
|
||||
new PGPSignatureGenerator(
|
||||
new BcPGPContentSignerBuilder(privateKey.getPublicKeyPacket().getAlgorithm(), HashAlgorithmTags.SHA512));
|
||||
generator.init(PGPSignature.BINARY_DOCUMENT, privateKey);
|
||||
final ByteArrayOutputStream output = new ByteArrayOutputStream();
|
||||
try (BCPGOutputStream pout = new BCPGOutputStream(new ArmoredOutputStream(output));
|
||||
InputStream is = new ByteArrayInputStream(bytes)) {
|
||||
final byte[] buffer = new byte[1024];
|
||||
int read;
|
||||
while ((read = is.read(buffer)) != -1) {
|
||||
generator.update(buffer, 0, read);
|
||||
}
|
||||
generator.generate().encode(pout);
|
||||
}
|
||||
return new String(output.toByteArray(), "UTF-8");
|
||||
} catch (IOException | PGPException e) {
|
||||
throw new RuntimeException(e);
|
||||
}
|
||||
}
|
||||
|
||||
// checks the plugin requires a policy confirmation, and does not install when that is rejected by the user
|
||||
// the plugin is installed after this method completes
|
||||
private void assertPolicyConfirmation(Tuple<Path, Environment> env, String pluginZip, String... warnings) throws Exception {
|
||||
|
|
|
@ -432,7 +432,15 @@ Remote hosts have to be explicitly whitelisted in elasticsearch.yaml using the
|
|||
`reindex.remote.whitelist` property. It can be set to a comma delimited list
|
||||
of allowed remote `host` and `port` combinations (e.g.
|
||||
`otherhost:9200, another:9200, 127.0.10.*:9200, localhost:*`). Scheme is
|
||||
ignored by the whitelist - only host and port are used.
|
||||
ignored by the whitelist - only host and port are used, for example:
|
||||
|
||||
|
||||
[source,yaml]
|
||||
--------------------------------------------------
|
||||
reindex.remote.whitelist: "otherhost:9200, another:9200, 127.0.10.*:9200, localhost:*"
|
||||
--------------------------------------------------
|
||||
|
||||
The whitelist must be configured on any nodes that will coordinate the reindex.
|
||||
|
||||
This feature should work with remote clusters of any version of Elasticsearch
|
||||
you are likely to find. This should allow you to upgrade from any version of
|
||||
|
|
|
@ -1,16 +1,39 @@
|
|||
[[indices-create-index]]
|
||||
== Create Index
|
||||
|
||||
The create index API allows to instantiate an index. Elasticsearch
|
||||
provides support for multiple indices, including executing operations
|
||||
across several indices.
|
||||
The Create Index API is used to manually create an index in Elasticsearch. All documents in Elasticsearch
|
||||
are stored inside of one index or another.
|
||||
|
||||
The most basic command is the following:
|
||||
|
||||
[source,js]
|
||||
--------------------------------------------------
|
||||
PUT twitter
|
||||
--------------------------------------------------
|
||||
// CONSOLE
|
||||
|
||||
This create an index named `twitter` with all default setting.
|
||||
|
||||
[NOTE]
|
||||
.Index name limitations
|
||||
======================================================
|
||||
There are several limitations to what you can name your index. The complete list of limitations are:
|
||||
|
||||
- Lowercase only
|
||||
- Cannot include `\`, `/`, `*`, `?`, `"`, `<`, `>`, `|`, ` ` (space character), `,`, `#`
|
||||
- Indices prior to 7.0 could contain a colon (`:`), but that's been deprecated and won't be supported in 7.0+
|
||||
- Cannot start with `-`, `_`, `+`
|
||||
- Cannot be `.` or ``..`
|
||||
- Cannot be longer than 255 bytes (note it is bytes, so multi-byte characters will count towards the 255 limit faster)
|
||||
|
||||
======================================================
|
||||
|
||||
[float]
|
||||
[[create-index-settings]]
|
||||
=== Index Settings
|
||||
|
||||
Each index created can have specific settings
|
||||
associated with it.
|
||||
associated with it, defined in the body:
|
||||
|
||||
[source,js]
|
||||
--------------------------------------------------
|
||||
|
@ -28,25 +51,6 @@ PUT twitter
|
|||
<1> Default for `number_of_shards` is 1
|
||||
<2> Default for `number_of_replicas` is 1 (ie one replica for each primary shard)
|
||||
|
||||
The above second curl example shows how an index called `twitter` can be
|
||||
created with specific settings for it using http://www.yaml.org[YAML].
|
||||
In this case, creating an index with 3 shards, each with 2 replicas. The
|
||||
index settings can also be defined with http://www.json.org[JSON]:
|
||||
|
||||
[source,js]
|
||||
--------------------------------------------------
|
||||
PUT twitter
|
||||
{
|
||||
"settings" : {
|
||||
"index" : {
|
||||
"number_of_shards" : 3,
|
||||
"number_of_replicas" : 2
|
||||
}
|
||||
}
|
||||
}
|
||||
--------------------------------------------------
|
||||
// CONSOLE
|
||||
|
||||
or more simplified
|
||||
|
||||
[source,js]
|
||||
|
|
|
@ -35,7 +35,7 @@ Elasticsearch 6.x in order to be readable by Elasticsearch 7.x.
|
|||
* <<breaking_70_java_changes>>
|
||||
* <<breaking_70_settings_changes>>
|
||||
* <<breaking_70_scripting_changes>>
|
||||
|
||||
* <<breaking_70_snapshotstats_changes>>
|
||||
|
||||
include::migrate_7_0/aggregations.asciidoc[]
|
||||
include::migrate_7_0/analysis.asciidoc[]
|
||||
|
@ -49,3 +49,4 @@ include::migrate_7_0/api.asciidoc[]
|
|||
include::migrate_7_0/java.asciidoc[]
|
||||
include::migrate_7_0/settings.asciidoc[]
|
||||
include::migrate_7_0/scripting.asciidoc[]
|
||||
include::migrate_7_0/snapshotstats.asciidoc[]
|
|
@ -0,0 +1,13 @@
|
|||
[[breaking_70_snapshotstats_changes]]
|
||||
=== Snapshot stats changes
|
||||
|
||||
Snapshot stats details are provided in a new structured way:
|
||||
|
||||
* `total` section for all the files that are referenced by the snapshot.
|
||||
* `incremental` section for those files that actually needed to be copied over as part of the incremental snapshotting.
|
||||
* In case of a snapshot that's still in progress, there's also a `processed` section for files that are in the process of being copied.
|
||||
|
||||
==== Deprecated `number_of_files`, `processed_files`, `total_size_in_bytes` and `processed_size_in_bytes` snapshot stats properties have been removed
|
||||
|
||||
* Properties `number_of_files` and `total_size_in_bytes` are removed and should be replaced by values of nested object `total`.
|
||||
* Properties `processed_files` and `processed_size_in_bytes` are removed and should be replaced by values of nested object `processed`.
|
|
@ -563,6 +563,54 @@ GET /_snapshot/my_backup/snapshot_1/_status
|
|||
// CONSOLE
|
||||
// TEST[continued]
|
||||
|
||||
The output looks similar to the following:
|
||||
|
||||
[source,js]
|
||||
--------------------------------------------------
|
||||
{
|
||||
"snapshots": [
|
||||
{
|
||||
"snapshot": "snapshot_1",
|
||||
"repository": "my_backup",
|
||||
"uuid": "XuBo4l4ISYiVg0nYUen9zg",
|
||||
"state": "SUCCESS",
|
||||
"include_global_state": true,
|
||||
"shards_stats": {
|
||||
"initializing": 0,
|
||||
"started": 0,
|
||||
"finalizing": 0,
|
||||
"done": 5,
|
||||
"failed": 0,
|
||||
"total": 5
|
||||
},
|
||||
"stats": {
|
||||
"incremental": {
|
||||
"file_count": 8,
|
||||
"size_in_bytes": 4704
|
||||
},
|
||||
"processed": {
|
||||
"file_count": 7,
|
||||
"size_in_bytes": 4254
|
||||
},
|
||||
"total": {
|
||||
"file_count": 8,
|
||||
"size_in_bytes": 4704
|
||||
},
|
||||
"start_time_in_millis": 1526280280355,
|
||||
"time_in_millis": 358
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
--------------------------------------------------
|
||||
// TESTRESPONSE
|
||||
|
||||
The output is composed of different sections. The `stats` sub-object provides details on the number and size of files that were
|
||||
snapshotted. As snapshots are incremental, copying only the Lucene segments that are not already in the repository,
|
||||
the `stats` object contains a `total` section for all the files that are referenced by the snapshot, as well as an `incremental` section
|
||||
for those files that actually needed to be copied over as part of the incremental snapshotting. In case of a snapshot that's still
|
||||
in progress, there's also a `processed` section that contains information about the files that are in the process of being copied.
|
||||
|
||||
Multiple ids are also supported:
|
||||
|
||||
[source,sh]
|
||||
|
|
|
@ -1,8 +1,10 @@
|
|||
|
||||
When you shut down a node, the allocation process waits for one minute
|
||||
before starting to replicate the shards on that node to other nodes
|
||||
in the cluster, causing a lot of wasted I/O. You can avoid racing the clock
|
||||
by disabling allocation before shutting down the node:
|
||||
When you shut down a node, the allocation process waits for
|
||||
`index.unassigned.node_left.delayed_timeout` (by default, one minute) before
|
||||
starting to replicate the shards on that node to other nodes in the cluster,
|
||||
which can involve a lot of I/O. Since the node is shortly going to be
|
||||
restarted, this I/O is unnecessary. You can avoid racing the clock by disabling
|
||||
allocation before shutting down the node:
|
||||
|
||||
[source,js]
|
||||
--------------------------------------------------
|
||||
|
|
|
@ -23,8 +23,10 @@ import org.apache.lucene.expressions.Expression;
|
|||
import org.apache.lucene.expressions.SimpleBindings;
|
||||
import org.apache.lucene.expressions.js.JavascriptCompiler;
|
||||
import org.apache.lucene.expressions.js.VariableContext;
|
||||
import org.apache.lucene.index.LeafReaderContext;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.queries.function.valuesource.DoubleConstValueSource;
|
||||
import org.apache.lucene.search.Scorer;
|
||||
import org.apache.lucene.search.SortField;
|
||||
import org.elasticsearch.SpecialPermission;
|
||||
import org.elasticsearch.common.Nullable;
|
||||
|
@ -39,12 +41,14 @@ import org.elasticsearch.index.mapper.MapperService;
|
|||
import org.elasticsearch.script.ClassPermission;
|
||||
import org.elasticsearch.script.ExecutableScript;
|
||||
import org.elasticsearch.script.FilterScript;
|
||||
import org.elasticsearch.script.ScoreScript;
|
||||
import org.elasticsearch.script.ScriptContext;
|
||||
import org.elasticsearch.script.ScriptEngine;
|
||||
import org.elasticsearch.script.ScriptException;
|
||||
import org.elasticsearch.script.SearchScript;
|
||||
import org.elasticsearch.search.lookup.SearchLookup;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.security.AccessControlContext;
|
||||
import java.security.AccessController;
|
||||
import java.security.PrivilegedAction;
|
||||
|
@ -111,6 +115,9 @@ public class ExpressionScriptEngine extends AbstractComponent implements ScriptE
|
|||
} else if (context.instanceClazz.equals(FilterScript.class)) {
|
||||
FilterScript.Factory factory = (p, lookup) -> newFilterScript(expr, lookup, p);
|
||||
return context.factoryClazz.cast(factory);
|
||||
} else if (context.instanceClazz.equals(ScoreScript.class)) {
|
||||
ScoreScript.Factory factory = (p, lookup) -> newScoreScript(expr, lookup, p);
|
||||
return context.factoryClazz.cast(factory);
|
||||
}
|
||||
throw new IllegalArgumentException("expression engine does not know how to handle script context [" + context.name + "]");
|
||||
}
|
||||
|
@ -260,6 +267,42 @@ public class ExpressionScriptEngine extends AbstractComponent implements ScriptE
|
|||
};
|
||||
};
|
||||
}
|
||||
|
||||
private ScoreScript.LeafFactory newScoreScript(Expression expr, SearchLookup lookup, @Nullable Map<String, Object> vars) {
|
||||
SearchScript.LeafFactory searchLeafFactory = newSearchScript(expr, lookup, vars);
|
||||
return new ScoreScript.LeafFactory() {
|
||||
@Override
|
||||
public boolean needs_score() {
|
||||
return searchLeafFactory.needs_score();
|
||||
}
|
||||
|
||||
@Override
|
||||
public ScoreScript newInstance(LeafReaderContext ctx) throws IOException {
|
||||
SearchScript script = searchLeafFactory.newInstance(ctx);
|
||||
return new ScoreScript(vars, lookup, ctx) {
|
||||
@Override
|
||||
public double execute() {
|
||||
return script.runAsDouble();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void setDocument(int docid) {
|
||||
script.setDocument(docid);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void setScorer(Scorer scorer) {
|
||||
script.setScorer(scorer);
|
||||
}
|
||||
|
||||
@Override
|
||||
public double get_score() {
|
||||
return script.getScore();
|
||||
}
|
||||
};
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* converts a ParseException at compile-time or link-time to a ScriptException
|
||||
|
|
|
@ -30,9 +30,9 @@ import org.apache.lucene.index.Term;
|
|||
import org.elasticsearch.common.settings.Settings;
|
||||
import org.elasticsearch.plugins.Plugin;
|
||||
import org.elasticsearch.plugins.ScriptPlugin;
|
||||
import org.elasticsearch.script.ScoreScript;
|
||||
import org.elasticsearch.script.ScriptContext;
|
||||
import org.elasticsearch.script.ScriptEngine;
|
||||
import org.elasticsearch.script.SearchScript;
|
||||
|
||||
/**
|
||||
* An example script plugin that adds a {@link ScriptEngine} implementing expert scoring.
|
||||
|
@ -54,12 +54,12 @@ public class ExpertScriptPlugin extends Plugin implements ScriptPlugin {
|
|||
|
||||
@Override
|
||||
public <T> T compile(String scriptName, String scriptSource, ScriptContext<T> context, Map<String, String> params) {
|
||||
if (context.equals(SearchScript.SCRIPT_SCORE_CONTEXT) == false) {
|
||||
if (context.equals(ScoreScript.CONTEXT) == false) {
|
||||
throw new IllegalArgumentException(getType() + " scripts cannot be used for context [" + context.name + "]");
|
||||
}
|
||||
// we use the script "source" as the script identifier
|
||||
if ("pure_df".equals(scriptSource)) {
|
||||
SearchScript.Factory factory = (p, lookup) -> new SearchScript.LeafFactory() {
|
||||
ScoreScript.Factory factory = (p, lookup) -> new ScoreScript.LeafFactory() {
|
||||
final String field;
|
||||
final String term;
|
||||
{
|
||||
|
@ -74,18 +74,18 @@ public class ExpertScriptPlugin extends Plugin implements ScriptPlugin {
|
|||
}
|
||||
|
||||
@Override
|
||||
public SearchScript newInstance(LeafReaderContext context) throws IOException {
|
||||
public ScoreScript newInstance(LeafReaderContext context) throws IOException {
|
||||
PostingsEnum postings = context.reader().postings(new Term(field, term));
|
||||
if (postings == null) {
|
||||
// the field and/or term don't exist in this segment, so always return 0
|
||||
return new SearchScript(p, lookup, context) {
|
||||
return new ScoreScript(p, lookup, context) {
|
||||
@Override
|
||||
public double runAsDouble() {
|
||||
public double execute() {
|
||||
return 0.0d;
|
||||
}
|
||||
};
|
||||
}
|
||||
return new SearchScript(p, lookup, context) {
|
||||
return new ScoreScript(p, lookup, context) {
|
||||
int currentDocid = -1;
|
||||
@Override
|
||||
public void setDocument(int docid) {
|
||||
|
@ -100,7 +100,7 @@ public class ExpertScriptPlugin extends Plugin implements ScriptPlugin {
|
|||
currentDocid = docid;
|
||||
}
|
||||
@Override
|
||||
public double runAsDouble() {
|
||||
public double execute() {
|
||||
if (postings.docID() != currentDocid) {
|
||||
// advance moved past the current doc, so this doc has no occurrences of the term
|
||||
return 0.0d;
|
||||
|
|
|
@ -25,9 +25,3 @@ dependencies {
|
|||
testCompile project(path: ':modules:rank-eval', configuration: 'runtime')
|
||||
testCompile project(path: ':modules:lang-mustache', configuration: 'runtime')
|
||||
}
|
||||
|
||||
/*
|
||||
* One of the integration tests doesn't work with the zip distribution
|
||||
* and will be fixed later.
|
||||
* Tracked by https://github.com/elastic/elasticsearch/issues/30628
|
||||
*/
|
||||
|
|
|
@ -0,0 +1,28 @@
|
|||
---
|
||||
setup:
|
||||
- do:
|
||||
indices.create:
|
||||
body:
|
||||
settings:
|
||||
index:
|
||||
number_of_shards: 1
|
||||
number_of_replicas: 1
|
||||
index: test-index
|
||||
---
|
||||
Test retrieval of default settings:
|
||||
- skip:
|
||||
version: " - 6.3.99"
|
||||
reason: include_defaults will not work in mixed-mode clusters containing nodes pre-6.4
|
||||
- do:
|
||||
indices.get_settings:
|
||||
flat_settings: true
|
||||
index: test-index
|
||||
- is_false:
|
||||
test-index.settings.index\.refresh_interval
|
||||
- do:
|
||||
indices.get_settings:
|
||||
include_defaults: true
|
||||
flat_settings: true
|
||||
index: test-index
|
||||
- match:
|
||||
test-index.defaults.index\.refresh_interval: "1s"
|
|
@ -23,15 +23,3 @@ Test reset index settings:
|
|||
indices.get_settings:
|
||||
flat_settings: false
|
||||
- is_false: test-index.settings.index\.refresh_interval
|
||||
|
||||
# Disabled until https://github.com/elastic/elasticsearch/pull/29229 is back-ported
|
||||
# That PR changed the execution path of index settings default to be on the master
|
||||
# until the PR is back-ported the old master will not return default settings.
|
||||
#
|
||||
# - do:
|
||||
# indices.get_settings:
|
||||
# include_defaults: true
|
||||
# flat_settings: true
|
||||
# index: test-index
|
||||
# - match:
|
||||
# test-index.defaults.index\.refresh_interval: "1s"
|
||||
|
|
|
@ -52,7 +52,7 @@ setup:
|
|||
---
|
||||
"Verify created repository":
|
||||
- skip:
|
||||
version: " - 6.99.99"
|
||||
version: "all"
|
||||
reason: AwaitsFix for https://github.com/elastic/elasticsearch/issues/30807
|
||||
- do:
|
||||
snapshot.verify_repository:
|
||||
|
|
|
@ -11,7 +11,6 @@ setup:
|
|||
|
||||
---
|
||||
"Get snapshot status":
|
||||
|
||||
- do:
|
||||
indices.create:
|
||||
index: test_index
|
||||
|
@ -32,6 +31,14 @@ setup:
|
|||
snapshot: test_snapshot
|
||||
|
||||
- is_true: snapshots
|
||||
- match: { snapshots.0.snapshot: test_snapshot }
|
||||
- match: { snapshots.0.state: SUCCESS }
|
||||
- gt: { snapshots.0.stats.incremental.file_count: 0 }
|
||||
- gt: { snapshots.0.stats.incremental.size_in_bytes: 0 }
|
||||
- gt: { snapshots.0.stats.total.file_count: 0 }
|
||||
- gt: { snapshots.0.stats.total.size_in_bytes: 0 }
|
||||
- is_true: snapshots.0.stats.start_time_in_millis
|
||||
- is_true: snapshots.0.stats.time_in_millis
|
||||
|
||||
---
|
||||
"Get missing snapshot status throws an exception":
|
||||
|
|
|
@ -140,7 +140,7 @@ public class VerifyRepositoryResponse extends ActionResponse implements ToXConte
|
|||
@Override
|
||||
public void readFrom(StreamInput in) throws IOException {
|
||||
super.readFrom(in);
|
||||
if (in.getVersion().onOrAfter(Version.V_7_0_0_alpha1)) {
|
||||
if (in.getVersion().onOrAfter(Version.V_6_4_0)) {
|
||||
this.nodes = in.readList(NodeView::new).stream().map(n -> n.convertToDiscoveryNode()).collect(Collectors.toList());
|
||||
} else {
|
||||
clusterName = new ClusterName(in);
|
||||
|
@ -151,7 +151,7 @@ public class VerifyRepositoryResponse extends ActionResponse implements ToXConte
|
|||
@Override
|
||||
public void writeTo(StreamOutput out) throws IOException {
|
||||
super.writeTo(out);
|
||||
if (Version.CURRENT.onOrAfter(Version.V_7_0_0_alpha1)) {
|
||||
if (out.getVersion().onOrAfter(Version.V_6_4_0)) {
|
||||
out.writeList(getNodes());
|
||||
} else {
|
||||
clusterName.writeTo(out);
|
||||
|
|
|
@ -74,8 +74,8 @@ public class SnapshotIndexShardStatus extends BroadcastShardResponse implements
|
|||
throw new IllegalArgumentException("Unknown stage type " + indexShardStatus.getStage());
|
||||
}
|
||||
this.stats = new SnapshotStats(indexShardStatus.getStartTime(), indexShardStatus.getTotalTime(),
|
||||
indexShardStatus.getNumberOfFiles(), indexShardStatus.getProcessedFiles(),
|
||||
indexShardStatus.getTotalSize(), indexShardStatus.getProcessedSize());
|
||||
indexShardStatus.getIncrementalFileCount(), indexShardStatus.getTotalFileCount(), indexShardStatus.getProcessedFileCount(),
|
||||
indexShardStatus.getIncrementalSize(), indexShardStatus.getTotalSize(), indexShardStatus.getProcessedSize());
|
||||
this.failure = indexShardStatus.getFailure();
|
||||
this.nodeId = nodeId;
|
||||
}
|
||||
|
|
|
@ -19,6 +19,7 @@
|
|||
|
||||
package org.elasticsearch.action.admin.cluster.snapshots.status;
|
||||
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.common.io.stream.StreamInput;
|
||||
import org.elasticsearch.common.io.stream.StreamOutput;
|
||||
import org.elasticsearch.common.io.stream.Streamable;
|
||||
|
@ -34,19 +35,25 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
|
|||
|
||||
private long startTime;
|
||||
private long time;
|
||||
private int numberOfFiles;
|
||||
private int processedFiles;
|
||||
private int incrementalFileCount;
|
||||
private int totalFileCount;
|
||||
private int processedFileCount;
|
||||
private long incrementalSize;
|
||||
private long totalSize;
|
||||
private long processedSize;
|
||||
|
||||
SnapshotStats() {
|
||||
}
|
||||
|
||||
SnapshotStats(long startTime, long time, int numberOfFiles, int processedFiles, long totalSize, long processedSize) {
|
||||
SnapshotStats(long startTime, long time,
|
||||
int incrementalFileCount, int totalFileCount, int processedFileCount,
|
||||
long incrementalSize, long totalSize, long processedSize) {
|
||||
this.startTime = startTime;
|
||||
this.time = time;
|
||||
this.numberOfFiles = numberOfFiles;
|
||||
this.processedFiles = processedFiles;
|
||||
this.incrementalFileCount = incrementalFileCount;
|
||||
this.totalFileCount = totalFileCount;
|
||||
this.processedFileCount = processedFileCount;
|
||||
this.incrementalSize = incrementalSize;
|
||||
this.totalSize = totalSize;
|
||||
this.processedSize = processedSize;
|
||||
}
|
||||
|
@ -66,17 +73,31 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
|
|||
}
|
||||
|
||||
/**
|
||||
* Returns number of files in the snapshot
|
||||
* Returns incremental file count of the snapshot
|
||||
*/
|
||||
public int getNumberOfFiles() {
|
||||
return numberOfFiles;
|
||||
public int getIncrementalFileCount() {
|
||||
return incrementalFileCount;
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns total number of files in the snapshot
|
||||
*/
|
||||
public int getTotalFileCount() {
|
||||
return totalFileCount;
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns number of files in the snapshot that were processed so far
|
||||
*/
|
||||
public int getProcessedFiles() {
|
||||
return processedFiles;
|
||||
public int getProcessedFileCount() {
|
||||
return processedFileCount;
|
||||
}
|
||||
|
||||
/**
|
||||
* Return incremental files size of the snapshot
|
||||
*/
|
||||
public long getIncrementalSize() {
|
||||
return incrementalSize;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -105,11 +126,16 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
|
|||
out.writeVLong(startTime);
|
||||
out.writeVLong(time);
|
||||
|
||||
out.writeVInt(numberOfFiles);
|
||||
out.writeVInt(processedFiles);
|
||||
out.writeVInt(incrementalFileCount);
|
||||
out.writeVInt(processedFileCount);
|
||||
|
||||
out.writeVLong(totalSize);
|
||||
out.writeVLong(incrementalSize);
|
||||
out.writeVLong(processedSize);
|
||||
|
||||
if (out.getVersion().onOrAfter(Version.V_6_4_0)) {
|
||||
out.writeVInt(totalFileCount);
|
||||
out.writeVLong(totalSize);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -117,21 +143,32 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
|
|||
startTime = in.readVLong();
|
||||
time = in.readVLong();
|
||||
|
||||
numberOfFiles = in.readVInt();
|
||||
processedFiles = in.readVInt();
|
||||
incrementalFileCount = in.readVInt();
|
||||
processedFileCount = in.readVInt();
|
||||
|
||||
totalSize = in.readVLong();
|
||||
incrementalSize = in.readVLong();
|
||||
processedSize = in.readVLong();
|
||||
|
||||
if (in.getVersion().onOrAfter(Version.V_6_4_0)) {
|
||||
totalFileCount = in.readVInt();
|
||||
totalSize = in.readVLong();
|
||||
} else {
|
||||
totalFileCount = incrementalFileCount;
|
||||
totalSize = incrementalSize;
|
||||
}
|
||||
}
|
||||
|
||||
static final class Fields {
|
||||
static final String STATS = "stats";
|
||||
static final String NUMBER_OF_FILES = "number_of_files";
|
||||
static final String PROCESSED_FILES = "processed_files";
|
||||
static final String TOTAL_SIZE_IN_BYTES = "total_size_in_bytes";
|
||||
static final String TOTAL_SIZE = "total_size";
|
||||
static final String PROCESSED_SIZE_IN_BYTES = "processed_size_in_bytes";
|
||||
static final String PROCESSED_SIZE = "processed_size";
|
||||
|
||||
static final String INCREMENTAL = "incremental";
|
||||
static final String PROCESSED = "processed";
|
||||
static final String TOTAL = "total";
|
||||
|
||||
static final String FILE_COUNT = "file_count";
|
||||
static final String SIZE = "size";
|
||||
static final String SIZE_IN_BYTES = "size_in_bytes";
|
||||
|
||||
static final String START_TIME_IN_MILLIS = "start_time_in_millis";
|
||||
static final String TIME_IN_MILLIS = "time_in_millis";
|
||||
static final String TIME = "time";
|
||||
|
@ -139,25 +176,44 @@ public class SnapshotStats implements Streamable, ToXContentFragment {
|
|||
|
||||
@Override
|
||||
public XContentBuilder toXContent(XContentBuilder builder, ToXContent.Params params) throws IOException {
|
||||
builder.startObject(Fields.STATS);
|
||||
builder.field(Fields.NUMBER_OF_FILES, getNumberOfFiles());
|
||||
builder.field(Fields.PROCESSED_FILES, getProcessedFiles());
|
||||
builder.humanReadableField(Fields.TOTAL_SIZE_IN_BYTES, Fields.TOTAL_SIZE, new ByteSizeValue(getTotalSize()));
|
||||
builder.humanReadableField(Fields.PROCESSED_SIZE_IN_BYTES, Fields.PROCESSED_SIZE, new ByteSizeValue(getProcessedSize()));
|
||||
builder.field(Fields.START_TIME_IN_MILLIS, getStartTime());
|
||||
builder.humanReadableField(Fields.TIME_IN_MILLIS, Fields.TIME, new TimeValue(getTime()));
|
||||
builder.endObject();
|
||||
return builder;
|
||||
builder.startObject(Fields.STATS)
|
||||
// incremental starts
|
||||
.startObject(Fields.INCREMENTAL)
|
||||
.field(Fields.FILE_COUNT, getIncrementalFileCount())
|
||||
.humanReadableField(Fields.SIZE_IN_BYTES, Fields.SIZE, new ByteSizeValue(getIncrementalSize()))
|
||||
// incremental ends
|
||||
.endObject();
|
||||
|
||||
if (getProcessedFileCount() != getIncrementalFileCount()) {
|
||||
// processed starts
|
||||
builder.startObject(Fields.PROCESSED)
|
||||
.field(Fields.FILE_COUNT, getProcessedFileCount())
|
||||
.humanReadableField(Fields.SIZE_IN_BYTES, Fields.SIZE, new ByteSizeValue(getProcessedSize()))
|
||||
// processed ends
|
||||
.endObject();
|
||||
}
|
||||
// total starts
|
||||
builder.startObject(Fields.TOTAL)
|
||||
.field(Fields.FILE_COUNT, getTotalFileCount())
|
||||
.humanReadableField(Fields.SIZE_IN_BYTES, Fields.SIZE, new ByteSizeValue(getTotalSize()))
|
||||
// total ends
|
||||
.endObject();
|
||||
// timings stats
|
||||
builder.field(Fields.START_TIME_IN_MILLIS, getStartTime())
|
||||
.humanReadableField(Fields.TIME_IN_MILLIS, Fields.TIME, new TimeValue(getTime()));
|
||||
|
||||
return builder.endObject();
|
||||
}
|
||||
|
||||
void add(SnapshotStats stats) {
|
||||
numberOfFiles += stats.numberOfFiles;
|
||||
processedFiles += stats.processedFiles;
|
||||
incrementalFileCount += stats.incrementalFileCount;
|
||||
totalFileCount += stats.totalFileCount;
|
||||
processedFileCount += stats.processedFileCount;
|
||||
|
||||
incrementalSize += stats.incrementalSize;
|
||||
totalSize += stats.totalSize;
|
||||
processedSize += stats.processedSize;
|
||||
|
||||
|
||||
if (startTime == 0) {
|
||||
// First time here
|
||||
startTime = stats.startTime;
|
||||
|
|
|
@ -18,12 +18,9 @@
|
|||
*/
|
||||
package org.elasticsearch.common.geo.parsers;
|
||||
|
||||
import org.locationtech.jts.geom.Coordinate;
|
||||
import org.elasticsearch.ElasticsearchParseException;
|
||||
import org.elasticsearch.common.geo.GeoPoint;
|
||||
import org.elasticsearch.common.geo.GeoShapeType;
|
||||
|
||||
import java.io.StringReader;
|
||||
import org.elasticsearch.common.geo.builders.CoordinatesBuilder;
|
||||
import org.elasticsearch.common.geo.builders.EnvelopeBuilder;
|
||||
import org.elasticsearch.common.geo.builders.GeometryCollectionBuilder;
|
||||
|
@ -37,9 +34,11 @@ import org.elasticsearch.common.geo.builders.ShapeBuilder;
|
|||
import org.elasticsearch.common.logging.Loggers;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.index.mapper.GeoShapeFieldMapper;
|
||||
import org.locationtech.jts.geom.Coordinate;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.StreamTokenizer;
|
||||
import java.io.StringReader;
|
||||
import java.util.List;
|
||||
|
||||
/**
|
||||
|
@ -77,8 +76,7 @@ public class GeoWKTParser {
|
|||
public static ShapeBuilder parseExpectedType(XContentParser parser, final GeoShapeType shapeType,
|
||||
final GeoShapeFieldMapper shapeMapper)
|
||||
throws IOException, ElasticsearchParseException {
|
||||
StringReader reader = new StringReader(parser.text());
|
||||
try {
|
||||
try (StringReader reader = new StringReader(parser.text())) {
|
||||
boolean ignoreZValue = (shapeMapper != null && shapeMapper.ignoreZValue().value() == true);
|
||||
// setup the tokenizer; configured to read words w/o numbers
|
||||
StreamTokenizer tokenizer = new StreamTokenizer(reader);
|
||||
|
@ -95,8 +93,6 @@ public class GeoWKTParser {
|
|||
ShapeBuilder builder = parseGeometry(tokenizer, shapeType, ignoreZValue);
|
||||
checkEOF(tokenizer);
|
||||
return builder;
|
||||
} finally {
|
||||
reader.close();
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -24,8 +24,8 @@ import org.apache.lucene.search.DocIdSetIterator;
|
|||
import org.apache.lucene.search.Explanation;
|
||||
import org.apache.lucene.search.Scorer;
|
||||
import org.elasticsearch.script.ExplainableSearchScript;
|
||||
import org.elasticsearch.script.ScoreScript;
|
||||
import org.elasticsearch.script.Script;
|
||||
import org.elasticsearch.script.SearchScript;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Objects;
|
||||
|
@ -58,10 +58,10 @@ public class ScriptScoreFunction extends ScoreFunction {
|
|||
|
||||
private final Script sScript;
|
||||
|
||||
private final SearchScript.LeafFactory script;
|
||||
private final ScoreScript.LeafFactory script;
|
||||
|
||||
|
||||
public ScriptScoreFunction(Script sScript, SearchScript.LeafFactory script) {
|
||||
public ScriptScoreFunction(Script sScript, ScoreScript.LeafFactory script) {
|
||||
super(CombineFunction.REPLACE);
|
||||
this.sScript = sScript;
|
||||
this.script = script;
|
||||
|
@ -69,7 +69,7 @@ public class ScriptScoreFunction extends ScoreFunction {
|
|||
|
||||
@Override
|
||||
public LeafScoreFunction getLeafScoreFunction(LeafReaderContext ctx) throws IOException {
|
||||
final SearchScript leafScript = script.newInstance(ctx);
|
||||
final ScoreScript leafScript = script.newInstance(ctx);
|
||||
final CannedScorer scorer = new CannedScorer();
|
||||
leafScript.setScorer(scorer);
|
||||
return new LeafScoreFunction() {
|
||||
|
@ -78,7 +78,7 @@ public class ScriptScoreFunction extends ScoreFunction {
|
|||
leafScript.setDocument(docId);
|
||||
scorer.docid = docId;
|
||||
scorer.score = subQueryScore;
|
||||
double result = leafScript.runAsDouble();
|
||||
double result = leafScript.execute();
|
||||
return result;
|
||||
}
|
||||
|
||||
|
|
|
@ -234,8 +234,8 @@ public class Analysis {
|
|||
|
||||
final Path path = env.configFile().resolve(wordListPath);
|
||||
|
||||
try (BufferedReader reader = Files.newBufferedReader(path, StandardCharsets.UTF_8)) {
|
||||
return loadWordList(reader, "#");
|
||||
try {
|
||||
return loadWordList(path, "#");
|
||||
} catch (CharacterCodingException ex) {
|
||||
String message = String.format(Locale.ROOT,
|
||||
"Unsupported character encoding detected while reading %s_path: %s - files must be UTF-8 encoded",
|
||||
|
@ -247,15 +247,9 @@ public class Analysis {
|
|||
}
|
||||
}
|
||||
|
||||
public static List<String> loadWordList(Reader reader, String comment) throws IOException {
|
||||
private static List<String> loadWordList(Path path, String comment) throws IOException {
|
||||
final List<String> result = new ArrayList<>();
|
||||
BufferedReader br = null;
|
||||
try {
|
||||
if (reader instanceof BufferedReader) {
|
||||
br = (BufferedReader) reader;
|
||||
} else {
|
||||
br = new BufferedReader(reader);
|
||||
}
|
||||
try (BufferedReader br = Files.newBufferedReader(path, StandardCharsets.UTF_8)) {
|
||||
String word;
|
||||
while ((word = br.readLine()) != null) {
|
||||
if (!Strings.hasText(word)) {
|
||||
|
@ -265,9 +259,6 @@ public class Analysis {
|
|||
result.add(word.trim());
|
||||
}
|
||||
}
|
||||
} finally {
|
||||
if (br != null)
|
||||
br.close();
|
||||
}
|
||||
return result;
|
||||
}
|
||||
|
|
|
@ -1424,10 +1424,6 @@ public abstract class Engine implements Closeable {
|
|||
|
||||
@Override
|
||||
public void close() {
|
||||
release();
|
||||
}
|
||||
|
||||
public void release() {
|
||||
Releasables.close(searcher);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -159,7 +159,7 @@ public final class ShardGetService extends AbstractIndexShardComponent {
|
|||
get = indexShard.get(new Engine.Get(realtime, readFromTranslog, type, id, uidTerm)
|
||||
.version(version).versionType(versionType));
|
||||
if (get.exists() == false) {
|
||||
get.release();
|
||||
get.close();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -172,7 +172,7 @@ public final class ShardGetService extends AbstractIndexShardComponent {
|
|||
// break between having loaded it from translog (so we only have _source), and having a document to load
|
||||
return innerGetLoadFromStoredFields(type, id, gFields, fetchSourceContext, get, mapperService);
|
||||
} finally {
|
||||
get.release();
|
||||
get.close();
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -28,6 +28,7 @@ import org.elasticsearch.common.xcontent.XContentBuilder;
|
|||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.index.query.QueryShardContext;
|
||||
import org.elasticsearch.index.query.QueryShardException;
|
||||
import org.elasticsearch.script.ScoreScript;
|
||||
import org.elasticsearch.script.Script;
|
||||
import org.elasticsearch.script.SearchScript;
|
||||
|
||||
|
@ -92,8 +93,8 @@ public class ScriptScoreFunctionBuilder extends ScoreFunctionBuilder<ScriptScore
|
|||
@Override
|
||||
protected ScoreFunction doToFunction(QueryShardContext context) {
|
||||
try {
|
||||
SearchScript.Factory factory = context.getScriptService().compile(script, SearchScript.SCRIPT_SCORE_CONTEXT);
|
||||
SearchScript.LeafFactory searchScript = factory.newFactory(script.getParams(), context.lookup());
|
||||
ScoreScript.Factory factory = context.getScriptService().compile(script, ScoreScript.CONTEXT);
|
||||
ScoreScript.LeafFactory searchScript = factory.newFactory(script.getParams(), context.lookup());
|
||||
return new ScriptScoreFunction(script, searchScript);
|
||||
} catch (Exception e) {
|
||||
throw new QueryShardException(context, "script_score: the script could not be loaded", e);
|
||||
|
|
|
@ -60,31 +60,39 @@ public class IndexShardSnapshotStatus {
|
|||
private final AtomicReference<Stage> stage;
|
||||
private long startTime;
|
||||
private long totalTime;
|
||||
private int numberOfFiles;
|
||||
private int processedFiles;
|
||||
private int incrementalFileCount;
|
||||
private int totalFileCount;
|
||||
private int processedFileCount;
|
||||
private long totalSize;
|
||||
private long incrementalSize;
|
||||
private long processedSize;
|
||||
private long indexVersion;
|
||||
private String failure;
|
||||
|
||||
private IndexShardSnapshotStatus(final Stage stage, final long startTime, final long totalTime,
|
||||
final int numberOfFiles, final int processedFiles, final long totalSize, final long processedSize,
|
||||
final int incrementalFileCount, final int totalFileCount, final int processedFileCount,
|
||||
final long incrementalSize, final long totalSize, final long processedSize,
|
||||
final long indexVersion, final String failure) {
|
||||
this.stage = new AtomicReference<>(Objects.requireNonNull(stage));
|
||||
this.startTime = startTime;
|
||||
this.totalTime = totalTime;
|
||||
this.numberOfFiles = numberOfFiles;
|
||||
this.processedFiles = processedFiles;
|
||||
this.incrementalFileCount = incrementalFileCount;
|
||||
this.totalFileCount = totalFileCount;
|
||||
this.processedFileCount = processedFileCount;
|
||||
this.totalSize = totalSize;
|
||||
this.processedSize = processedSize;
|
||||
this.incrementalSize = incrementalSize;
|
||||
this.indexVersion = indexVersion;
|
||||
this.failure = failure;
|
||||
}
|
||||
|
||||
public synchronized Copy moveToStarted(final long startTime, final int numberOfFiles, final long totalSize) {
|
||||
public synchronized Copy moveToStarted(final long startTime, final int incrementalFileCount, final int totalFileCount,
|
||||
final long incrementalSize, final long totalSize) {
|
||||
if (stage.compareAndSet(Stage.INIT, Stage.STARTED)) {
|
||||
this.startTime = startTime;
|
||||
this.numberOfFiles = numberOfFiles;
|
||||
this.incrementalFileCount = incrementalFileCount;
|
||||
this.totalFileCount = totalFileCount;
|
||||
this.incrementalSize = incrementalSize;
|
||||
this.totalSize = totalSize;
|
||||
} else {
|
||||
throw new IllegalStateException("Unable to move the shard snapshot status to [STARTED]: " +
|
||||
|
@ -135,7 +143,7 @@ public class IndexShardSnapshotStatus {
|
|||
* Increments number of processed files
|
||||
*/
|
||||
public synchronized void addProcessedFile(long size) {
|
||||
processedFiles++;
|
||||
processedFileCount++;
|
||||
processedSize += size;
|
||||
}
|
||||
|
||||
|
@ -146,12 +154,14 @@ public class IndexShardSnapshotStatus {
|
|||
* @return a {@link IndexShardSnapshotStatus.Copy}
|
||||
*/
|
||||
public synchronized IndexShardSnapshotStatus.Copy asCopy() {
|
||||
return new IndexShardSnapshotStatus.Copy(stage.get(), startTime, totalTime, numberOfFiles, processedFiles, totalSize, processedSize,
|
||||
indexVersion, failure);
|
||||
return new IndexShardSnapshotStatus.Copy(stage.get(), startTime, totalTime,
|
||||
incrementalFileCount, totalFileCount, processedFileCount,
|
||||
incrementalSize, totalSize, processedSize,
|
||||
indexVersion, failure);
|
||||
}
|
||||
|
||||
public static IndexShardSnapshotStatus newInitializing() {
|
||||
return new IndexShardSnapshotStatus(Stage.INIT, 0L, 0L, 0, 0, 0, 0, 0, null);
|
||||
return new IndexShardSnapshotStatus(Stage.INIT, 0L, 0L, 0, 0, 0, 0, 0, 0, 0, null);
|
||||
}
|
||||
|
||||
public static IndexShardSnapshotStatus newFailed(final String failure) {
|
||||
|
@ -159,12 +169,15 @@ public class IndexShardSnapshotStatus {
|
|||
if (failure == null) {
|
||||
throw new IllegalArgumentException("A failure description is required for a failed IndexShardSnapshotStatus");
|
||||
}
|
||||
return new IndexShardSnapshotStatus(Stage.FAILURE, 0L, 0L, 0, 0, 0, 0, 0, failure);
|
||||
return new IndexShardSnapshotStatus(Stage.FAILURE, 0L, 0L, 0, 0, 0, 0, 0, 0, 0, failure);
|
||||
}
|
||||
|
||||
public static IndexShardSnapshotStatus newDone(final long startTime, final long totalTime, final int files, final long size) {
|
||||
public static IndexShardSnapshotStatus newDone(final long startTime, final long totalTime,
|
||||
final int incrementalFileCount, final int fileCount,
|
||||
final long incrementalSize, final long size) {
|
||||
// The snapshot is done which means the number of processed files is the same as total
|
||||
return new IndexShardSnapshotStatus(Stage.DONE, startTime, totalTime, files, files, size, size, 0, null);
|
||||
return new IndexShardSnapshotStatus(Stage.DONE, startTime, totalTime, incrementalFileCount, fileCount, incrementalFileCount,
|
||||
incrementalSize, size, incrementalSize, 0, null);
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -175,23 +188,28 @@ public class IndexShardSnapshotStatus {
|
|||
private final Stage stage;
|
||||
private final long startTime;
|
||||
private final long totalTime;
|
||||
private final int numberOfFiles;
|
||||
private final int processedFiles;
|
||||
private final int incrementalFileCount;
|
||||
private final int totalFileCount;
|
||||
private final int processedFileCount;
|
||||
private final long totalSize;
|
||||
private final long processedSize;
|
||||
private final long incrementalSize;
|
||||
private final long indexVersion;
|
||||
private final String failure;
|
||||
|
||||
public Copy(final Stage stage, final long startTime, final long totalTime,
|
||||
final int numberOfFiles, final int processedFiles, final long totalSize, final long processedSize,
|
||||
final int incrementalFileCount, final int totalFileCount, final int processedFileCount,
|
||||
final long incrementalSize, final long totalSize, final long processedSize,
|
||||
final long indexVersion, final String failure) {
|
||||
this.stage = stage;
|
||||
this.startTime = startTime;
|
||||
this.totalTime = totalTime;
|
||||
this.numberOfFiles = numberOfFiles;
|
||||
this.processedFiles = processedFiles;
|
||||
this.incrementalFileCount = incrementalFileCount;
|
||||
this.totalFileCount = totalFileCount;
|
||||
this.processedFileCount = processedFileCount;
|
||||
this.totalSize = totalSize;
|
||||
this.processedSize = processedSize;
|
||||
this.incrementalSize = incrementalSize;
|
||||
this.indexVersion = indexVersion;
|
||||
this.failure = failure;
|
||||
}
|
||||
|
@ -208,12 +226,20 @@ public class IndexShardSnapshotStatus {
|
|||
return totalTime;
|
||||
}
|
||||
|
||||
public int getNumberOfFiles() {
|
||||
return numberOfFiles;
|
||||
public int getIncrementalFileCount() {
|
||||
return incrementalFileCount;
|
||||
}
|
||||
|
||||
public int getProcessedFiles() {
|
||||
return processedFiles;
|
||||
public int getTotalFileCount() {
|
||||
return totalFileCount;
|
||||
}
|
||||
|
||||
public int getProcessedFileCount() {
|
||||
return processedFileCount;
|
||||
}
|
||||
|
||||
public long getIncrementalSize() {
|
||||
return incrementalSize;
|
||||
}
|
||||
|
||||
public long getTotalSize() {
|
||||
|
@ -238,8 +264,10 @@ public class IndexShardSnapshotStatus {
|
|||
"stage=" + stage +
|
||||
", startTime=" + startTime +
|
||||
", totalTime=" + totalTime +
|
||||
", numberOfFiles=" + numberOfFiles +
|
||||
", processedFiles=" + processedFiles +
|
||||
", incrementalFileCount=" + incrementalFileCount +
|
||||
", totalFileCount=" + totalFileCount +
|
||||
", processedFileCount=" + processedFileCount +
|
||||
", incrementalSize=" + incrementalSize +
|
||||
", totalSize=" + totalSize +
|
||||
", processedSize=" + processedSize +
|
||||
", indexVersion=" + indexVersion +
|
||||
|
|
|
@ -356,25 +356,28 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
|
|||
|
||||
private final long time;
|
||||
|
||||
private final int numberOfFiles;
|
||||
private final int incrementalFileCount;
|
||||
|
||||
private final long totalSize;
|
||||
private final long incrementalSize;
|
||||
|
||||
private final List<FileInfo> indexFiles;
|
||||
|
||||
/**
|
||||
* Constructs new shard snapshot metadata from snapshot metadata
|
||||
*
|
||||
* @param snapshot snapshot id
|
||||
* @param indexVersion index version
|
||||
* @param indexFiles list of files in the shard
|
||||
* @param startTime snapshot start time
|
||||
* @param time snapshot running time
|
||||
* @param numberOfFiles number of files that where snapshotted
|
||||
* @param totalSize total size of all files snapshotted
|
||||
* @param snapshot snapshot id
|
||||
* @param indexVersion index version
|
||||
* @param indexFiles list of files in the shard
|
||||
* @param startTime snapshot start time
|
||||
* @param time snapshot running time
|
||||
* @param incrementalFileCount incremental of files that were snapshotted
|
||||
* @param incrementalSize incremental size of snapshot
|
||||
*/
|
||||
public BlobStoreIndexShardSnapshot(String snapshot, long indexVersion, List<FileInfo> indexFiles, long startTime, long time,
|
||||
int numberOfFiles, long totalSize) {
|
||||
public BlobStoreIndexShardSnapshot(String snapshot, long indexVersion, List<FileInfo> indexFiles,
|
||||
long startTime, long time,
|
||||
int incrementalFileCount,
|
||||
long incrementalSize
|
||||
) {
|
||||
assert snapshot != null;
|
||||
assert indexVersion >= 0;
|
||||
this.snapshot = snapshot;
|
||||
|
@ -382,8 +385,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
|
|||
this.indexFiles = Collections.unmodifiableList(new ArrayList<>(indexFiles));
|
||||
this.startTime = startTime;
|
||||
this.time = time;
|
||||
this.numberOfFiles = numberOfFiles;
|
||||
this.totalSize = totalSize;
|
||||
this.incrementalFileCount = incrementalFileCount;
|
||||
this.incrementalSize = incrementalSize;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -395,8 +398,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
|
|||
this.indexFiles = Collections.emptyList();
|
||||
this.startTime = 0;
|
||||
this.time = 0;
|
||||
this.numberOfFiles = 0;
|
||||
this.totalSize = 0;
|
||||
this.incrementalFileCount = 0;
|
||||
this.incrementalSize = 0;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -441,34 +444,51 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
|
|||
}
|
||||
|
||||
/**
|
||||
* Returns number of files that where snapshotted
|
||||
* Returns incremental of files that were snapshotted
|
||||
*/
|
||||
public int numberOfFiles() {
|
||||
return numberOfFiles;
|
||||
public int incrementalFileCount() {
|
||||
return incrementalFileCount;
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns total number of files that are referenced by this snapshot
|
||||
*/
|
||||
public int totalFileCount() {
|
||||
return indexFiles.size();
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns incremental of files size that were snapshotted
|
||||
*/
|
||||
public long incrementalSize() {
|
||||
return incrementalSize;
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns total size of all files that where snapshotted
|
||||
*/
|
||||
public long totalSize() {
|
||||
return totalSize;
|
||||
return indexFiles.stream().mapToLong(fi -> fi.metadata().length()).sum();
|
||||
}
|
||||
|
||||
private static final String NAME = "name";
|
||||
private static final String INDEX_VERSION = "index_version";
|
||||
private static final String START_TIME = "start_time";
|
||||
private static final String TIME = "time";
|
||||
private static final String NUMBER_OF_FILES = "number_of_files";
|
||||
private static final String TOTAL_SIZE = "total_size";
|
||||
private static final String FILES = "files";
|
||||
// for the sake of BWC keep the actual property names as in 6.x
|
||||
// + there is a constraint in #fromXContent() that leads to ElasticsearchParseException("unknown parameter [incremental_file_count]");
|
||||
private static final String INCREMENTAL_FILE_COUNT = "number_of_files";
|
||||
private static final String INCREMENTAL_SIZE = "total_size";
|
||||
|
||||
private static final ParseField PARSE_NAME = new ParseField("name");
|
||||
private static final ParseField PARSE_INDEX_VERSION = new ParseField("index_version", "index-version");
|
||||
private static final ParseField PARSE_START_TIME = new ParseField("start_time");
|
||||
private static final ParseField PARSE_TIME = new ParseField("time");
|
||||
private static final ParseField PARSE_NUMBER_OF_FILES = new ParseField("number_of_files");
|
||||
private static final ParseField PARSE_TOTAL_SIZE = new ParseField("total_size");
|
||||
private static final ParseField PARSE_FILES = new ParseField("files");
|
||||
|
||||
private static final ParseField PARSE_NAME = new ParseField(NAME);
|
||||
private static final ParseField PARSE_INDEX_VERSION = new ParseField(INDEX_VERSION, "index-version");
|
||||
private static final ParseField PARSE_START_TIME = new ParseField(START_TIME);
|
||||
private static final ParseField PARSE_TIME = new ParseField(TIME);
|
||||
private static final ParseField PARSE_INCREMENTAL_FILE_COUNT = new ParseField(INCREMENTAL_FILE_COUNT);
|
||||
private static final ParseField PARSE_INCREMENTAL_SIZE = new ParseField(INCREMENTAL_SIZE);
|
||||
private static final ParseField PARSE_FILES = new ParseField(FILES);
|
||||
|
||||
/**
|
||||
* Serializes shard snapshot metadata info into JSON
|
||||
|
@ -482,8 +502,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
|
|||
builder.field(INDEX_VERSION, indexVersion);
|
||||
builder.field(START_TIME, startTime);
|
||||
builder.field(TIME, time);
|
||||
builder.field(NUMBER_OF_FILES, numberOfFiles);
|
||||
builder.field(TOTAL_SIZE, totalSize);
|
||||
builder.field(INCREMENTAL_FILE_COUNT, incrementalFileCount);
|
||||
builder.field(INCREMENTAL_SIZE, incrementalSize);
|
||||
builder.startArray(FILES);
|
||||
for (FileInfo fileInfo : indexFiles) {
|
||||
FileInfo.toXContent(fileInfo, builder, params);
|
||||
|
@ -503,8 +523,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
|
|||
long indexVersion = -1;
|
||||
long startTime = 0;
|
||||
long time = 0;
|
||||
int numberOfFiles = 0;
|
||||
long totalSize = 0;
|
||||
int incrementalFileCount = 0;
|
||||
long incrementalSize = 0;
|
||||
|
||||
List<FileInfo> indexFiles = new ArrayList<>();
|
||||
if (parser.currentToken() == null) { // fresh parser? move to the first token
|
||||
|
@ -526,10 +546,10 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
|
|||
startTime = parser.longValue();
|
||||
} else if (PARSE_TIME.match(currentFieldName, parser.getDeprecationHandler())) {
|
||||
time = parser.longValue();
|
||||
} else if (PARSE_NUMBER_OF_FILES.match(currentFieldName, parser.getDeprecationHandler())) {
|
||||
numberOfFiles = parser.intValue();
|
||||
} else if (PARSE_TOTAL_SIZE.match(currentFieldName, parser.getDeprecationHandler())) {
|
||||
totalSize = parser.longValue();
|
||||
} else if (PARSE_INCREMENTAL_FILE_COUNT.match(currentFieldName, parser.getDeprecationHandler())) {
|
||||
incrementalFileCount = parser.intValue();
|
||||
} else if (PARSE_INCREMENTAL_SIZE.match(currentFieldName, parser.getDeprecationHandler())) {
|
||||
incrementalSize = parser.longValue();
|
||||
} else {
|
||||
throw new ElasticsearchParseException("unknown parameter [{}]", currentFieldName);
|
||||
}
|
||||
|
@ -549,7 +569,8 @@ public class BlobStoreIndexShardSnapshot implements ToXContentFragment {
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
return new BlobStoreIndexShardSnapshot(snapshot, indexVersion, Collections.unmodifiableList(indexFiles),
|
||||
startTime, time, numberOfFiles, totalSize);
|
||||
startTime, time, incrementalFileCount, incrementalSize);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -85,8 +85,6 @@ public class TermVectorsService {
|
|||
termVectorsResponse.setExists(false);
|
||||
return termVectorsResponse;
|
||||
}
|
||||
Engine.GetResult get = indexShard.get(new Engine.Get(request.realtime(), false, request.type(), request.id(), uidTerm)
|
||||
.version(request.version()).versionType(request.versionType()));
|
||||
|
||||
Fields termVectorsByField = null;
|
||||
AggregatedDfs dfs = null;
|
||||
|
@ -97,8 +95,9 @@ public class TermVectorsService {
|
|||
handleFieldWildcards(indexShard, request);
|
||||
}
|
||||
|
||||
final Engine.Searcher searcher = indexShard.acquireSearcher("term_vector");
|
||||
try {
|
||||
try (Engine.GetResult get = indexShard.get(new Engine.Get(request.realtime(), false, request.type(), request.id(), uidTerm)
|
||||
.version(request.version()).versionType(request.versionType()));
|
||||
Engine.Searcher searcher = indexShard.acquireSearcher("term_vector")) {
|
||||
Fields topLevelFields = MultiFields.getFields(get.searcher() != null ? get.searcher().reader() : searcher.reader());
|
||||
DocIdAndVersion docIdAndVersion = get.docIdAndVersion();
|
||||
/* from an artificial document */
|
||||
|
@ -143,14 +142,12 @@ public class TermVectorsService {
|
|||
}
|
||||
}
|
||||
// write term vectors
|
||||
termVectorsResponse.setFields(termVectorsByField, request.selectedFields(), request.getFlags(), topLevelFields, dfs, termVectorsFilter);
|
||||
termVectorsResponse.setFields(termVectorsByField, request.selectedFields(), request.getFlags(), topLevelFields, dfs,
|
||||
termVectorsFilter);
|
||||
}
|
||||
termVectorsResponse.setTookInMillis(TimeUnit.NANOSECONDS.toMillis(nanoTimeSupplier.getAsLong() - startTime));
|
||||
} catch (Exception ex) {
|
||||
throw new ElasticsearchException("failed to execute term vector request", ex);
|
||||
} finally {
|
||||
searcher.close();
|
||||
get.release();
|
||||
}
|
||||
return termVectorsResponse;
|
||||
}
|
||||
|
|
|
@ -26,7 +26,6 @@ import org.elasticsearch.common.Nullable;
|
|||
import org.elasticsearch.common.unit.TimeValue;
|
||||
import org.elasticsearch.tasks.CancellableTask;
|
||||
import org.elasticsearch.tasks.Task;
|
||||
import org.elasticsearch.tasks.TaskCancelledException;
|
||||
import org.elasticsearch.tasks.TaskId;
|
||||
import org.elasticsearch.tasks.TaskManager;
|
||||
|
||||
|
@ -38,18 +37,16 @@ import java.util.function.Predicate;
|
|||
* Represents a executor node operation that corresponds to a persistent task
|
||||
*/
|
||||
public class AllocatedPersistentTask extends CancellableTask {
|
||||
private volatile String persistentTaskId;
|
||||
private volatile long allocationId;
|
||||
|
||||
private final AtomicReference<State> state;
|
||||
@Nullable
|
||||
private volatile Exception failure;
|
||||
|
||||
private volatile String persistentTaskId;
|
||||
private volatile long allocationId;
|
||||
private volatile @Nullable Exception failure;
|
||||
private volatile PersistentTasksService persistentTasksService;
|
||||
private volatile Logger logger;
|
||||
private volatile TaskManager taskManager;
|
||||
|
||||
|
||||
public AllocatedPersistentTask(long id, String type, String action, String description, TaskId parentTask,
|
||||
Map<String, String> headers) {
|
||||
super(id, type, action, description, parentTask, headers);
|
||||
|
@ -101,24 +98,10 @@ public class AllocatedPersistentTask extends CancellableTask {
|
|||
return failure;
|
||||
}
|
||||
|
||||
boolean markAsCancelled() {
|
||||
return state.compareAndSet(AllocatedPersistentTask.State.STARTED, AllocatedPersistentTask.State.PENDING_CANCEL);
|
||||
}
|
||||
|
||||
public State getState() {
|
||||
return state.get();
|
||||
}
|
||||
|
||||
public long getAllocationId() {
|
||||
return allocationId;
|
||||
}
|
||||
|
||||
public enum State {
|
||||
STARTED, // the task is currently running
|
||||
PENDING_CANCEL, // the task is cancelled on master, cancelling it locally
|
||||
COMPLETED // the task is done running and trying to notify caller
|
||||
}
|
||||
|
||||
/**
|
||||
* Waits for this persistent task to have the desired state.
|
||||
*/
|
||||
|
@ -128,6 +111,14 @@ public class AllocatedPersistentTask extends CancellableTask {
|
|||
persistentTasksService.waitForPersistentTaskStatus(persistentTaskId, predicate, timeout, listener);
|
||||
}
|
||||
|
||||
final boolean isCompleted() {
|
||||
return state.get() == State.COMPLETED;
|
||||
}
|
||||
|
||||
boolean markAsCancelled() {
|
||||
return state.compareAndSet(State.STARTED, State.PENDING_CANCEL);
|
||||
}
|
||||
|
||||
public void markAsCompleted() {
|
||||
completeAndNotifyIfNeeded(null);
|
||||
}
|
||||
|
@ -138,11 +129,10 @@ public class AllocatedPersistentTask extends CancellableTask {
|
|||
} else {
|
||||
completeAndNotifyIfNeeded(e);
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
private void completeAndNotifyIfNeeded(@Nullable Exception failure) {
|
||||
State prevState = state.getAndSet(AllocatedPersistentTask.State.COMPLETED);
|
||||
final State prevState = state.getAndSet(State.COMPLETED);
|
||||
if (prevState == State.COMPLETED) {
|
||||
logger.warn("attempt to complete task [{}] with id [{}] in the [{}] state", getAction(), getPersistentTaskId(), prevState);
|
||||
} else {
|
||||
|
@ -173,4 +163,10 @@ public class AllocatedPersistentTask extends CancellableTask {
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
public enum State {
|
||||
STARTED, // the task is currently running
|
||||
PENDING_CANCEL, // the task is cancelled on master, cancelling it locally
|
||||
COMPLETED // the task is done running and trying to notify caller
|
||||
}
|
||||
}
|
||||
|
|
|
@ -123,7 +123,7 @@ public class PersistentTasksNodeService extends AbstractComponent implements Clu
|
|||
|
||||
for (Long id : notVisitedTasks) {
|
||||
AllocatedPersistentTask task = runningTasks.get(id);
|
||||
if (task.getState() == AllocatedPersistentTask.State.COMPLETED) {
|
||||
if (task.isCompleted()) {
|
||||
// Result was sent to the caller and the caller acknowledged acceptance of the result
|
||||
logger.trace("Found completed persistent task [{}] with id [{}] and allocation id [{}] - removing",
|
||||
task.getAction(), task.getPersistentTaskId(), task.getAllocationId());
|
||||
|
|
|
@ -440,7 +440,7 @@ public class PluginsService extends AbstractComponent {
|
|||
List<Bundle> sortedBundles = sortBundles(bundles);
|
||||
|
||||
for (Bundle bundle : sortedBundles) {
|
||||
checkBundleJarHell(bundle, transitiveUrls);
|
||||
checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveUrls);
|
||||
|
||||
final Plugin plugin = loadBundle(bundle, loaded);
|
||||
plugins.add(new Tuple<>(bundle.plugin, plugin));
|
||||
|
@ -451,7 +451,7 @@ public class PluginsService extends AbstractComponent {
|
|||
|
||||
// jar-hell check the bundle against the parent classloader and extended plugins
|
||||
// the plugin cli does it, but we do it again, in case lusers mess with jar files manually
|
||||
static void checkBundleJarHell(Bundle bundle, Map<String, Set<URL>> transitiveUrls) {
|
||||
static void checkBundleJarHell(Set<URL> classpath, Bundle bundle, Map<String, Set<URL>> transitiveUrls) {
|
||||
// invariant: any plugins this plugin bundle extends have already been added to transitiveUrls
|
||||
List<String> exts = bundle.plugin.getExtendedPlugins();
|
||||
|
||||
|
@ -484,7 +484,6 @@ public class PluginsService extends AbstractComponent {
|
|||
JarHell.checkJarHell(urls, logger::debug); // check jarhell of each extended plugin against this plugin
|
||||
transitiveUrls.put(bundle.plugin.getName(), urls);
|
||||
|
||||
Set<URL> classpath = JarHell.parseClassPath();
|
||||
// check we don't have conflicting codebases with core
|
||||
Set<URL> intersection = new HashSet<>(classpath);
|
||||
intersection.retainAll(bundle.urls);
|
||||
|
|
|
@ -818,7 +818,9 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
|
|||
public IndexShardSnapshotStatus getShardSnapshotStatus(SnapshotId snapshotId, Version version, IndexId indexId, ShardId shardId) {
|
||||
Context context = new Context(snapshotId, version, indexId, shardId);
|
||||
BlobStoreIndexShardSnapshot snapshot = context.loadSnapshot();
|
||||
return IndexShardSnapshotStatus.newDone(snapshot.startTime(), snapshot.time(), snapshot.numberOfFiles(), snapshot.totalSize());
|
||||
return IndexShardSnapshotStatus.newDone(snapshot.startTime(), snapshot.time(),
|
||||
snapshot.incrementalFileCount(), snapshot.totalFileCount(),
|
||||
snapshot.incrementalSize(), snapshot.totalSize());
|
||||
}
|
||||
|
||||
@Override
|
||||
|
@ -1139,9 +1141,11 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
|
|||
final List<BlobStoreIndexShardSnapshot.FileInfo> indexCommitPointFiles = new ArrayList<>();
|
||||
|
||||
store.incRef();
|
||||
int indexIncrementalFileCount = 0;
|
||||
int indexTotalNumberOfFiles = 0;
|
||||
long indexIncrementalSize = 0;
|
||||
long indexTotalFileCount = 0;
|
||||
try {
|
||||
int indexNumberOfFiles = 0;
|
||||
long indexTotalFilesSize = 0;
|
||||
ArrayList<BlobStoreIndexShardSnapshot.FileInfo> filesToSnapshot = new ArrayList<>();
|
||||
final Store.MetadataSnapshot metadata;
|
||||
// TODO apparently we don't use the MetadataSnapshot#.recoveryDiff(...) here but we should
|
||||
|
@ -1182,9 +1186,13 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
indexTotalFileCount += md.length();
|
||||
indexTotalNumberOfFiles++;
|
||||
|
||||
if (existingFileInfo == null) {
|
||||
indexNumberOfFiles++;
|
||||
indexTotalFilesSize += md.length();
|
||||
indexIncrementalFileCount++;
|
||||
indexIncrementalSize += md.length();
|
||||
// create a new FileInfo
|
||||
BlobStoreIndexShardSnapshot.FileInfo snapshotFileInfo = new BlobStoreIndexShardSnapshot.FileInfo(fileNameFromGeneration(++generation), md, chunkSize());
|
||||
indexCommitPointFiles.add(snapshotFileInfo);
|
||||
|
@ -1194,7 +1202,8 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
|
|||
}
|
||||
}
|
||||
|
||||
snapshotStatus.moveToStarted(startTime, indexNumberOfFiles, indexTotalFilesSize);
|
||||
snapshotStatus.moveToStarted(startTime, indexIncrementalFileCount,
|
||||
indexTotalNumberOfFiles, indexIncrementalSize, indexTotalFileCount);
|
||||
|
||||
for (BlobStoreIndexShardSnapshot.FileInfo snapshotFileInfo : filesToSnapshot) {
|
||||
try {
|
||||
|
@ -1217,8 +1226,9 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
|
|||
// snapshotStatus.startTime() is assigned on the same machine,
|
||||
// so it's safe to use with VLong
|
||||
System.currentTimeMillis() - lastSnapshotStatus.getStartTime(),
|
||||
lastSnapshotStatus.getNumberOfFiles(),
|
||||
lastSnapshotStatus.getTotalSize());
|
||||
lastSnapshotStatus.getIncrementalFileCount(),
|
||||
lastSnapshotStatus.getIncrementalSize()
|
||||
);
|
||||
|
||||
//TODO: The time stored in snapshot doesn't include cleanup time.
|
||||
logger.trace("[{}] [{}] writing shard snapshot file", shardId, snapshotId);
|
||||
|
|
|
@ -0,0 +1,102 @@
|
|||
/*
|
||||
* Licensed to Elasticsearch under one or more contributor
|
||||
* license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright
|
||||
* ownership. Elasticsearch licenses this file to you under
|
||||
* the Apache License, Version 2.0 (the "License"); you may
|
||||
* not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing,
|
||||
* software distributed under the License is distributed on an
|
||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
* KIND, either express or implied. See the License for the
|
||||
* specific language governing permissions and limitations
|
||||
* under the License.
|
||||
*/
|
||||
package org.elasticsearch.script;
|
||||
|
||||
import org.apache.lucene.index.LeafReaderContext;
|
||||
import org.apache.lucene.search.Scorer;
|
||||
import org.elasticsearch.index.fielddata.ScriptDocValues;
|
||||
import org.elasticsearch.search.lookup.LeafSearchLookup;
|
||||
import org.elasticsearch.search.lookup.SearchLookup;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.UncheckedIOException;
|
||||
import java.util.Map;
|
||||
import java.util.function.DoubleSupplier;
|
||||
|
||||
/**
|
||||
* A script used for adjusting the score on a per document basis.
|
||||
*/
|
||||
public abstract class ScoreScript {
|
||||
|
||||
public static final String[] PARAMETERS = new String[]{};
|
||||
|
||||
/** The generic runtime parameters for the script. */
|
||||
private final Map<String, Object> params;
|
||||
|
||||
/** A leaf lookup for the bound segment this script will operate on. */
|
||||
private final LeafSearchLookup leafLookup;
|
||||
|
||||
private DoubleSupplier scoreSupplier = () -> 0.0;
|
||||
|
||||
public ScoreScript(Map<String, Object> params, SearchLookup lookup, LeafReaderContext leafContext) {
|
||||
this.params = params;
|
||||
this.leafLookup = lookup.getLeafSearchLookup(leafContext);
|
||||
}
|
||||
|
||||
public abstract double execute();
|
||||
|
||||
/** Return the parameters for this script. */
|
||||
public Map<String, Object> getParams() {
|
||||
return params;
|
||||
}
|
||||
|
||||
/** The doc lookup for the Lucene segment this script was created for. */
|
||||
public final Map<String, ScriptDocValues<?>> getDoc() {
|
||||
return leafLookup.doc();
|
||||
}
|
||||
|
||||
/** Set the current document to run the script on next. */
|
||||
public void setDocument(int docid) {
|
||||
leafLookup.setDocument(docid);
|
||||
}
|
||||
|
||||
public void setScorer(Scorer scorer) {
|
||||
this.scoreSupplier = () -> {
|
||||
try {
|
||||
return scorer.score();
|
||||
} catch (IOException e) {
|
||||
throw new UncheckedIOException(e);
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
public double get_score() {
|
||||
return scoreSupplier.getAsDouble();
|
||||
}
|
||||
|
||||
/** A factory to construct {@link ScoreScript} instances. */
|
||||
public interface LeafFactory {
|
||||
|
||||
/**
|
||||
* Return {@code true} if the script needs {@code _score} calculated, or {@code false} otherwise.
|
||||
*/
|
||||
boolean needs_score();
|
||||
|
||||
ScoreScript newInstance(LeafReaderContext ctx) throws IOException;
|
||||
}
|
||||
|
||||
/** A factory to construct stateful {@link ScoreScript} factories for a specific index. */
|
||||
public interface Factory {
|
||||
|
||||
ScoreScript.LeafFactory newFactory(Map<String, Object> params, SearchLookup lookup);
|
||||
|
||||
}
|
||||
|
||||
public static final ScriptContext<ScoreScript.Factory> CONTEXT = new ScriptContext<>("score", ScoreScript.Factory.class);
|
||||
}
|
|
@ -42,7 +42,7 @@ public class ScriptModule {
|
|||
CORE_CONTEXTS = Stream.of(
|
||||
SearchScript.CONTEXT,
|
||||
SearchScript.AGGS_CONTEXT,
|
||||
SearchScript.SCRIPT_SCORE_CONTEXT,
|
||||
ScoreScript.CONTEXT,
|
||||
SearchScript.SCRIPT_SORT_CONTEXT,
|
||||
SearchScript.TERMS_SET_QUERY_CONTEXT,
|
||||
ExecutableScript.CONTEXT,
|
||||
|
|
|
@ -162,8 +162,6 @@ public abstract class SearchScript implements ScorerAware, ExecutableScript {
|
|||
public static final ScriptContext<Factory> AGGS_CONTEXT = new ScriptContext<>("aggs", Factory.class);
|
||||
// Can return a double. (For ScriptSortType#NUMBER only, for ScriptSortType#STRING normal CONTEXT should be used)
|
||||
public static final ScriptContext<Factory> SCRIPT_SORT_CONTEXT = new ScriptContext<>("sort", Factory.class);
|
||||
// Can return a float
|
||||
public static final ScriptContext<Factory> SCRIPT_SCORE_CONTEXT = new ScriptContext<>("score", Factory.class);
|
||||
// Can return a long
|
||||
public static final ScriptContext<Factory> TERMS_SET_QUERY_CONTEXT = new ScriptContext<>("terms_set", Factory.class);
|
||||
}
|
||||
|
|
|
@ -21,7 +21,6 @@ package org.elasticsearch.search;
|
|||
|
||||
import org.apache.lucene.search.FieldDoc;
|
||||
import org.apache.lucene.search.TopDocs;
|
||||
import org.elasticsearch.core.internal.io.IOUtils;
|
||||
import org.elasticsearch.ElasticsearchException;
|
||||
import org.elasticsearch.ExceptionsHelper;
|
||||
import org.elasticsearch.action.ActionListener;
|
||||
|
@ -39,6 +38,7 @@ import org.elasticsearch.common.util.BigArrays;
|
|||
import org.elasticsearch.common.util.concurrent.AbstractRunnable;
|
||||
import org.elasticsearch.common.util.concurrent.ConcurrentCollections;
|
||||
import org.elasticsearch.common.util.concurrent.ConcurrentMapLong;
|
||||
import org.elasticsearch.core.internal.io.IOUtils;
|
||||
import org.elasticsearch.index.Index;
|
||||
import org.elasticsearch.index.IndexService;
|
||||
import org.elasticsearch.index.IndexSettings;
|
||||
|
@ -92,8 +92,8 @@ import org.elasticsearch.search.sort.SortAndFormats;
|
|||
import org.elasticsearch.search.sort.SortBuilder;
|
||||
import org.elasticsearch.search.suggest.Suggest;
|
||||
import org.elasticsearch.search.suggest.completion.CompletionSuggestion;
|
||||
import org.elasticsearch.threadpool.ThreadPool;
|
||||
import org.elasticsearch.threadpool.Scheduler.Cancellable;
|
||||
import org.elasticsearch.threadpool.ThreadPool;
|
||||
import org.elasticsearch.threadpool.ThreadPool.Names;
|
||||
import org.elasticsearch.transport.TransportRequest;
|
||||
|
||||
|
@ -646,20 +646,17 @@ public class SearchService extends AbstractLifecycleComponent implements IndexEv
|
|||
|
||||
|
||||
public boolean freeContext(long id) {
|
||||
final SearchContext context = removeContext(id);
|
||||
if (context != null) {
|
||||
assert context.refCount() > 0 : " refCount must be > 0: " + context.refCount();
|
||||
try {
|
||||
try (SearchContext context = removeContext(id)) {
|
||||
if (context != null) {
|
||||
assert context.refCount() > 0 : " refCount must be > 0: " + context.refCount();
|
||||
context.indexShard().getSearchOperationListener().onFreeContext(context);
|
||||
if (context.scrollContext() != null) {
|
||||
context.indexShard().getSearchOperationListener().onFreeScrollContext(context);
|
||||
}
|
||||
} finally {
|
||||
context.close();
|
||||
return true;
|
||||
}
|
||||
return true;
|
||||
return false;
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
public void freeAllScrollContexts() {
|
||||
|
|
|
@ -65,11 +65,11 @@ public class SnapshotBlocksIT extends ESIntegTestCase {
|
|||
client().prepareIndex(OTHER_INDEX_NAME, "type").setSource("test", "init").execute().actionGet();
|
||||
}
|
||||
|
||||
|
||||
logger.info("--> register a repository");
|
||||
|
||||
assertAcked(client().admin().cluster().preparePutRepository(REPOSITORY_NAME)
|
||||
.setType("fs")
|
||||
.setSettings(Settings.builder().put("location", randomRepoPath())));
|
||||
.setSettings(Settings.builder().put("location", randomRepoPath())));
|
||||
|
||||
logger.info("--> verify the repository");
|
||||
VerifyRepositoryResponse verifyResponse = client().admin().cluster().prepareVerifyRepository(REPOSITORY_NAME).get();
|
||||
|
|
|
@ -91,10 +91,14 @@ public class SnapshotStatusTests extends ESTestCase {
|
|||
" \"total\" : " + totalShards + "\n" +
|
||||
" },\n" +
|
||||
" \"stats\" : {\n" +
|
||||
" \"number_of_files\" : 0,\n" +
|
||||
" \"processed_files\" : 0,\n" +
|
||||
" \"total_size_in_bytes\" : 0,\n" +
|
||||
" \"processed_size_in_bytes\" : 0,\n" +
|
||||
" \"incremental\" : {\n" +
|
||||
" \"file_count\" : 0,\n" +
|
||||
" \"size_in_bytes\" : 0\n" +
|
||||
" },\n" +
|
||||
" \"total\" : {\n" +
|
||||
" \"file_count\" : 0,\n" +
|
||||
" \"size_in_bytes\" : 0\n" +
|
||||
" },\n" +
|
||||
" \"start_time_in_millis\" : 0,\n" +
|
||||
" \"time_in_millis\" : 0\n" +
|
||||
" },\n" +
|
||||
|
@ -109,10 +113,14 @@ public class SnapshotStatusTests extends ESTestCase {
|
|||
" \"total\" : " + totalShards + "\n" +
|
||||
" },\n" +
|
||||
" \"stats\" : {\n" +
|
||||
" \"number_of_files\" : 0,\n" +
|
||||
" \"processed_files\" : 0,\n" +
|
||||
" \"total_size_in_bytes\" : 0,\n" +
|
||||
" \"processed_size_in_bytes\" : 0,\n" +
|
||||
" \"incremental\" : {\n" +
|
||||
" \"file_count\" : 0,\n" +
|
||||
" \"size_in_bytes\" : 0\n" +
|
||||
" },\n" +
|
||||
" \"total\" : {\n" +
|
||||
" \"file_count\" : 0,\n" +
|
||||
" \"size_in_bytes\" : 0\n" +
|
||||
" },\n" +
|
||||
" \"start_time_in_millis\" : 0,\n" +
|
||||
" \"time_in_millis\" : 0\n" +
|
||||
" },\n" +
|
||||
|
@ -120,10 +128,14 @@ public class SnapshotStatusTests extends ESTestCase {
|
|||
" \"" + shardId + "\" : {\n" +
|
||||
" \"stage\" : \"" + shardStage.toString() + "\",\n" +
|
||||
" \"stats\" : {\n" +
|
||||
" \"number_of_files\" : 0,\n" +
|
||||
" \"processed_files\" : 0,\n" +
|
||||
" \"total_size_in_bytes\" : 0,\n" +
|
||||
" \"processed_size_in_bytes\" : 0,\n" +
|
||||
" \"incremental\" : {\n" +
|
||||
" \"file_count\" : 0,\n" +
|
||||
" \"size_in_bytes\" : 0\n" +
|
||||
" },\n" +
|
||||
" \"total\" : {\n" +
|
||||
" \"file_count\" : 0,\n" +
|
||||
" \"size_in_bytes\" : 0\n" +
|
||||
" },\n" +
|
||||
" \"start_time_in_millis\" : 0,\n" +
|
||||
" \"time_in_millis\" : 0\n" +
|
||||
" }\n" +
|
||||
|
|
|
@ -23,6 +23,7 @@ import org.apache.lucene.index.DirectoryReader;
|
|||
import org.apache.lucene.index.Fields;
|
||||
import org.apache.lucene.index.Terms;
|
||||
import org.apache.lucene.index.TermsEnum;
|
||||
import org.elasticsearch.ElasticsearchException;
|
||||
import org.elasticsearch.action.admin.indices.alias.Alias;
|
||||
import org.elasticsearch.common.lucene.uid.Versions;
|
||||
import org.elasticsearch.common.settings.Settings;
|
||||
|
@ -111,7 +112,8 @@ public class MultiTermVectorsIT extends AbstractTermVectorsTestCase {
|
|||
checkTermTexts(response.getResponses()[1].getResponse().getFields().terms("field"), new String[]{"value1"});
|
||||
assertThat(response.getResponses()[2].getFailure(), notNullValue());
|
||||
assertThat(response.getResponses()[2].getFailure().getId(), equalTo("1"));
|
||||
assertThat(response.getResponses()[2].getFailure().getCause(), instanceOf(VersionConflictEngineException.class));
|
||||
assertThat(response.getResponses()[2].getFailure().getCause(), instanceOf(ElasticsearchException.class));
|
||||
assertThat(response.getResponses()[2].getFailure().getCause().getCause(), instanceOf(VersionConflictEngineException.class));
|
||||
|
||||
//Version from Lucene index
|
||||
refresh();
|
||||
|
@ -132,7 +134,8 @@ public class MultiTermVectorsIT extends AbstractTermVectorsTestCase {
|
|||
checkTermTexts(response.getResponses()[1].getResponse().getFields().terms("field"), new String[]{"value1"});
|
||||
assertThat(response.getResponses()[2].getFailure(), notNullValue());
|
||||
assertThat(response.getResponses()[2].getFailure().getId(), equalTo("1"));
|
||||
assertThat(response.getResponses()[2].getFailure().getCause(), instanceOf(VersionConflictEngineException.class));
|
||||
assertThat(response.getResponses()[2].getFailure().getCause(), instanceOf(ElasticsearchException.class));
|
||||
assertThat(response.getResponses()[2].getFailure().getCause().getCause(), instanceOf(VersionConflictEngineException.class));
|
||||
|
||||
|
||||
for (int i = 0; i < 3; i++) {
|
||||
|
@ -155,7 +158,8 @@ public class MultiTermVectorsIT extends AbstractTermVectorsTestCase {
|
|||
assertThat(response.getResponses()[1].getFailure(), notNullValue());
|
||||
assertThat(response.getResponses()[1].getFailure().getId(), equalTo("2"));
|
||||
assertThat(response.getResponses()[1].getIndex(), equalTo("test"));
|
||||
assertThat(response.getResponses()[1].getFailure().getCause(), instanceOf(VersionConflictEngineException.class));
|
||||
assertThat(response.getResponses()[1].getFailure().getCause(), instanceOf(ElasticsearchException.class));
|
||||
assertThat(response.getResponses()[1].getFailure().getCause().getCause(), instanceOf(VersionConflictEngineException.class));
|
||||
assertThat(response.getResponses()[2].getId(), equalTo("2"));
|
||||
assertThat(response.getResponses()[2].getIndex(), equalTo("test"));
|
||||
assertThat(response.getResponses()[2].getFailure(), nullValue());
|
||||
|
@ -180,7 +184,8 @@ public class MultiTermVectorsIT extends AbstractTermVectorsTestCase {
|
|||
assertThat(response.getResponses()[1].getFailure(), notNullValue());
|
||||
assertThat(response.getResponses()[1].getFailure().getId(), equalTo("2"));
|
||||
assertThat(response.getResponses()[1].getIndex(), equalTo("test"));
|
||||
assertThat(response.getResponses()[1].getFailure().getCause(), instanceOf(VersionConflictEngineException.class));
|
||||
assertThat(response.getResponses()[1].getFailure().getCause(), instanceOf(ElasticsearchException.class));
|
||||
assertThat(response.getResponses()[1].getFailure().getCause().getCause(), instanceOf(VersionConflictEngineException.class));
|
||||
assertThat(response.getResponses()[2].getId(), equalTo("2"));
|
||||
assertThat(response.getResponses()[2].getIndex(), equalTo("test"));
|
||||
assertThat(response.getResponses()[2].getFailure(), nullValue());
|
||||
|
|
|
@ -471,7 +471,7 @@ public class CacheTests extends ESTestCase {
|
|||
keys.add(key);
|
||||
} else {
|
||||
// invalidate with incorrect value
|
||||
cache.invalidate(key, Integer.toString(key * randomIntBetween(2, 10)));
|
||||
cache.invalidate(key, Integer.toString(key + randomIntBetween(2, 10)));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -506,7 +506,7 @@ public class CacheTests extends ESTestCase {
|
|||
invalidated.add(i);
|
||||
} else {
|
||||
// invalidate with incorrect value
|
||||
cache.invalidate(i, Integer.toString(i * randomIntBetween(2, 10)));
|
||||
cache.invalidate(i, Integer.toString(i + randomIntBetween(2, 10)));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -21,6 +21,7 @@ package org.elasticsearch.index.engine;
|
|||
|
||||
import com.carrotsearch.hppc.cursors.ObjectObjectCursor;
|
||||
import com.carrotsearch.randomizedtesting.generators.RandomNumbers;
|
||||
|
||||
import org.apache.logging.log4j.Level;
|
||||
import org.apache.logging.log4j.LogManager;
|
||||
import org.apache.logging.log4j.Logger;
|
||||
|
@ -793,7 +794,7 @@ public class InternalEngineTests extends EngineTestCase {
|
|||
while (flushFinished.get() == false) {
|
||||
Engine.GetResult previousGetResult = latestGetResult.get();
|
||||
if (previousGetResult != null) {
|
||||
previousGetResult.release();
|
||||
previousGetResult.close();
|
||||
}
|
||||
latestGetResult.set(engine.get(newGet(true, doc), searcherFactory));
|
||||
if (latestGetResult.get().exists() == false) {
|
||||
|
@ -807,7 +808,7 @@ public class InternalEngineTests extends EngineTestCase {
|
|||
flushFinished.set(true);
|
||||
getThread.join();
|
||||
assertTrue(latestGetResult.get().exists());
|
||||
latestGetResult.get().release();
|
||||
latestGetResult.get().close();
|
||||
}
|
||||
|
||||
public void testSimpleOperations() throws Exception {
|
||||
|
@ -830,21 +831,20 @@ public class InternalEngineTests extends EngineTestCase {
|
|||
searchResult.close();
|
||||
|
||||
// but, not there non realtime
|
||||
Engine.GetResult getResult = engine.get(newGet(false, doc), searcherFactory);
|
||||
assertThat(getResult.exists(), equalTo(false));
|
||||
getResult.release();
|
||||
try (Engine.GetResult getResult = engine.get(newGet(false, doc), searcherFactory)) {
|
||||
assertThat(getResult.exists(), equalTo(false));
|
||||
}
|
||||
|
||||
// but, we can still get it (in realtime)
|
||||
getResult = engine.get(newGet(true, doc), searcherFactory);
|
||||
assertThat(getResult.exists(), equalTo(true));
|
||||
assertThat(getResult.docIdAndVersion(), notNullValue());
|
||||
getResult.release();
|
||||
try (Engine.GetResult getResult = engine.get(newGet(true, doc), searcherFactory)) {
|
||||
assertThat(getResult.exists(), equalTo(true));
|
||||
assertThat(getResult.docIdAndVersion(), notNullValue());
|
||||
}
|
||||
|
||||
// but not real time is not yet visible
|
||||
getResult = engine.get(newGet(false, doc), searcherFactory);
|
||||
assertThat(getResult.exists(), equalTo(false));
|
||||
getResult.release();
|
||||
|
||||
try (Engine.GetResult getResult = engine.get(newGet(false, doc), searcherFactory)) {
|
||||
assertThat(getResult.exists(), equalTo(false));
|
||||
}
|
||||
|
||||
// refresh and it should be there
|
||||
engine.refresh("test");
|
||||
|
@ -856,10 +856,10 @@ public class InternalEngineTests extends EngineTestCase {
|
|||
searchResult.close();
|
||||
|
||||
// also in non realtime
|
||||
getResult = engine.get(newGet(false, doc), searcherFactory);
|
||||
assertThat(getResult.exists(), equalTo(true));
|
||||
assertThat(getResult.docIdAndVersion(), notNullValue());
|
||||
getResult.release();
|
||||
try (Engine.GetResult getResult = engine.get(newGet(false, doc), searcherFactory)) {
|
||||
assertThat(getResult.exists(), equalTo(true));
|
||||
assertThat(getResult.docIdAndVersion(), notNullValue());
|
||||
}
|
||||
|
||||
// now do an update
|
||||
document = testDocument();
|
||||
|
@ -876,10 +876,10 @@ public class InternalEngineTests extends EngineTestCase {
|
|||
searchResult.close();
|
||||
|
||||
// but, we can still get it (in realtime)
|
||||
getResult = engine.get(newGet(true, doc), searcherFactory);
|
||||
assertThat(getResult.exists(), equalTo(true));
|
||||
assertThat(getResult.docIdAndVersion(), notNullValue());
|
||||
getResult.release();
|
||||
try (Engine.GetResult getResult = engine.get(newGet(true, doc), searcherFactory)) {
|
||||
assertThat(getResult.exists(), equalTo(true));
|
||||
assertThat(getResult.docIdAndVersion(), notNullValue());
|
||||
}
|
||||
|
||||
// refresh and it should be updated
|
||||
engine.refresh("test");
|
||||
|
@ -901,9 +901,9 @@ public class InternalEngineTests extends EngineTestCase {
|
|||
searchResult.close();
|
||||
|
||||
// but, get should not see it (in realtime)
|
||||
getResult = engine.get(newGet(true, doc), searcherFactory);
|
||||
assertThat(getResult.exists(), equalTo(false));
|
||||
getResult.release();
|
||||
try (Engine.GetResult getResult = engine.get(newGet(true, doc), searcherFactory)) {
|
||||
assertThat(getResult.exists(), equalTo(false));
|
||||
}
|
||||
|
||||
// refresh and it should be deleted
|
||||
engine.refresh("test");
|
||||
|
@ -941,10 +941,10 @@ public class InternalEngineTests extends EngineTestCase {
|
|||
engine.flush();
|
||||
|
||||
// and, verify get (in real time)
|
||||
getResult = engine.get(newGet(true, doc), searcherFactory);
|
||||
assertThat(getResult.exists(), equalTo(true));
|
||||
assertThat(getResult.docIdAndVersion(), notNullValue());
|
||||
getResult.release();
|
||||
try (Engine.GetResult getResult = engine.get(newGet(true, doc), searcherFactory)) {
|
||||
assertThat(getResult.exists(), equalTo(true));
|
||||
assertThat(getResult.docIdAndVersion(), notNullValue());
|
||||
}
|
||||
|
||||
// make sure we can still work with the engine
|
||||
// now do an update
|
||||
|
@ -4156,7 +4156,7 @@ public class InternalEngineTests extends EngineTestCase {
|
|||
new Term("_id", parsedDocument.id()),
|
||||
parsedDocument,
|
||||
SequenceNumbers.UNASSIGNED_SEQ_NO,
|
||||
(long) randomIntBetween(1, 8),
|
||||
randomIntBetween(1, 8),
|
||||
Versions.MATCH_ANY,
|
||||
VersionType.INTERNAL,
|
||||
Engine.Operation.Origin.PRIMARY,
|
||||
|
@ -4172,7 +4172,7 @@ public class InternalEngineTests extends EngineTestCase {
|
|||
id,
|
||||
new Term("_id", parsedDocument.id()),
|
||||
SequenceNumbers.UNASSIGNED_SEQ_NO,
|
||||
(long) randomIntBetween(1, 8),
|
||||
randomIntBetween(1, 8),
|
||||
Versions.MATCH_ANY,
|
||||
VersionType.INTERNAL,
|
||||
Engine.Operation.Origin.PRIMARY,
|
||||
|
|
|
@ -1861,10 +1861,11 @@ public class IndexShardTests extends IndexShardTestCase {
|
|||
indexDoc(shard, "_doc", "1", "{\"foobar\" : \"bar\"}");
|
||||
shard.refresh("test");
|
||||
|
||||
Engine.GetResult getResult = shard.get(new Engine.Get(false, false, "test", "1", new Term(IdFieldMapper.NAME, Uid.encodeId("1"))));
|
||||
assertTrue(getResult.exists());
|
||||
assertNotNull(getResult.searcher());
|
||||
getResult.release();
|
||||
try (Engine.GetResult getResult = shard
|
||||
.get(new Engine.Get(false, false, "test", "1", new Term(IdFieldMapper.NAME, Uid.encodeId("1"))))) {
|
||||
assertTrue(getResult.exists());
|
||||
assertNotNull(getResult.searcher());
|
||||
}
|
||||
try (Engine.Searcher searcher = shard.acquireSearcher("test")) {
|
||||
TopDocs search = searcher.searcher().search(new TermQuery(new Term("foo", "bar")), 10);
|
||||
assertEquals(search.totalHits, 1);
|
||||
|
@ -1895,11 +1896,12 @@ public class IndexShardTests extends IndexShardTestCase {
|
|||
search = searcher.searcher().search(new TermQuery(new Term("foobar", "bar")), 10);
|
||||
assertEquals(search.totalHits, 1);
|
||||
}
|
||||
getResult = newShard.get(new Engine.Get(false, false, "test", "1", new Term(IdFieldMapper.NAME, Uid.encodeId("1"))));
|
||||
assertTrue(getResult.exists());
|
||||
assertNotNull(getResult.searcher()); // make sure get uses the wrapped reader
|
||||
assertTrue(getResult.searcher().reader() instanceof FieldMaskingReader);
|
||||
getResult.release();
|
||||
try (Engine.GetResult getResult = newShard
|
||||
.get(new Engine.Get(false, false, "test", "1", new Term(IdFieldMapper.NAME, Uid.encodeId("1"))))) {
|
||||
assertTrue(getResult.exists());
|
||||
assertNotNull(getResult.searcher()); // make sure get uses the wrapped reader
|
||||
assertTrue(getResult.searcher().reader() instanceof FieldMaskingReader);
|
||||
}
|
||||
|
||||
closeShards(newShard);
|
||||
}
|
||||
|
|
|
@ -52,6 +52,7 @@ import java.util.concurrent.atomic.AtomicLong;
|
|||
import java.util.concurrent.atomic.AtomicReference;
|
||||
|
||||
import static org.hamcrest.Matchers.empty;
|
||||
import static org.hamcrest.Matchers.is;
|
||||
import static org.hamcrest.Matchers.sameInstance;
|
||||
import static org.hamcrest.core.IsEqual.equalTo;
|
||||
import static org.mockito.Matchers.any;
|
||||
|
@ -73,7 +74,6 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
|
|||
threadPool = new TestThreadPool(getClass().getName());
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
@After
|
||||
public void tearDown() throws Exception {
|
||||
|
@ -95,7 +95,7 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
|
|||
return state.build();
|
||||
}
|
||||
|
||||
public void testStartTask() throws Exception {
|
||||
public void testStartTask() {
|
||||
PersistentTasksService persistentTasksService = mock(PersistentTasksService.class);
|
||||
@SuppressWarnings("unchecked") PersistentTasksExecutor<TestParams> action = mock(PersistentTasksExecutor.class);
|
||||
when(action.getExecutor()).thenReturn(ThreadPool.Names.SAME);
|
||||
|
@ -131,8 +131,8 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
|
|||
|
||||
if (added == false) {
|
||||
logger.info("No local node action was added");
|
||||
|
||||
}
|
||||
|
||||
MetaData.Builder metaData = MetaData.builder(state.metaData());
|
||||
metaData.putCustom(PersistentTasksCustomMetaData.TYPE, tasks.build());
|
||||
ClusterState newClusterState = ClusterState.builder(state).metaData(metaData).build();
|
||||
|
@ -149,6 +149,7 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
|
|||
|
||||
// Make sure action wasn't called again
|
||||
assertThat(executor.executions.size(), equalTo(1));
|
||||
assertThat(executor.get(0).task.isCompleted(), is(false));
|
||||
|
||||
// Start another task on this node
|
||||
state = newClusterState;
|
||||
|
@ -157,10 +158,15 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
|
|||
|
||||
// Make sure action was called this time
|
||||
assertThat(executor.size(), equalTo(2));
|
||||
assertThat(executor.get(1).task.isCompleted(), is(false));
|
||||
|
||||
// Finish both tasks
|
||||
executor.get(0).task.markAsFailed(new RuntimeException());
|
||||
executor.get(1).task.markAsCompleted();
|
||||
|
||||
assertThat(executor.get(0).task.isCompleted(), is(true));
|
||||
assertThat(executor.get(1).task.isCompleted(), is(true));
|
||||
|
||||
String failedTaskId = executor.get(0).task.getPersistentTaskId();
|
||||
String finishedTaskId = executor.get(1).task.getPersistentTaskId();
|
||||
executor.clear();
|
||||
|
@ -186,7 +192,6 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
|
|||
// Make sure action was only allocated on this node once
|
||||
assertThat(executor.size(), equalTo(1));
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
public void testParamsStatusAndNodeTaskAreDelegated() throws Exception {
|
||||
|
@ -300,7 +305,6 @@ public class PersistentTasksNodeServiceTests extends ESTestCase {
|
|||
|
||||
// Check the the task is now removed from task manager
|
||||
assertThat(taskManager.getTasks().values(), empty());
|
||||
|
||||
}
|
||||
|
||||
private <Params extends PersistentTaskParams> ClusterState addTask(ClusterState state, String action, Params params,
|
||||
|
|
|
@ -23,6 +23,7 @@ import org.apache.log4j.Level;
|
|||
import org.apache.lucene.util.Constants;
|
||||
import org.apache.lucene.util.LuceneTestCase;
|
||||
import org.elasticsearch.Version;
|
||||
import org.elasticsearch.bootstrap.JarHell;
|
||||
import org.elasticsearch.common.io.PathUtils;
|
||||
import org.elasticsearch.common.settings.Settings;
|
||||
import org.elasticsearch.env.Environment;
|
||||
|
@ -443,7 +444,7 @@ public class PluginsServiceTests extends ESTestCase {
|
|||
"MyPlugin", Collections.singletonList("dep"), false);
|
||||
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
|
||||
IllegalStateException e = expectThrows(IllegalStateException.class, () ->
|
||||
PluginsService.checkBundleJarHell(bundle, transitiveDeps));
|
||||
PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps));
|
||||
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
|
||||
assertThat(e.getCause().getMessage(), containsString("jar hell! duplicate codebases with extended plugin"));
|
||||
}
|
||||
|
@ -462,7 +463,7 @@ public class PluginsServiceTests extends ESTestCase {
|
|||
"MyPlugin", Arrays.asList("dep1", "dep2"), false);
|
||||
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
|
||||
IllegalStateException e = expectThrows(IllegalStateException.class, () ->
|
||||
PluginsService.checkBundleJarHell(bundle, transitiveDeps));
|
||||
PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps));
|
||||
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
|
||||
assertThat(e.getCause().getMessage(), containsString("jar hell!"));
|
||||
assertThat(e.getCause().getMessage(), containsString("duplicate codebases"));
|
||||
|
@ -479,7 +480,7 @@ public class PluginsServiceTests extends ESTestCase {
|
|||
"MyPlugin", Collections.emptyList(), false);
|
||||
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
|
||||
IllegalStateException e = expectThrows(IllegalStateException.class, () ->
|
||||
PluginsService.checkBundleJarHell(bundle, new HashMap<>()));
|
||||
PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, new HashMap<>()));
|
||||
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
|
||||
assertThat(e.getCause().getMessage(), containsString("jar hell!"));
|
||||
assertThat(e.getCause().getMessage(), containsString("Level"));
|
||||
|
@ -498,7 +499,7 @@ public class PluginsServiceTests extends ESTestCase {
|
|||
"MyPlugin", Collections.singletonList("dep"), false);
|
||||
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
|
||||
IllegalStateException e = expectThrows(IllegalStateException.class, () ->
|
||||
PluginsService.checkBundleJarHell(bundle, transitiveDeps));
|
||||
PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps));
|
||||
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
|
||||
assertThat(e.getCause().getMessage(), containsString("jar hell!"));
|
||||
assertThat(e.getCause().getMessage(), containsString("DummyClass1"));
|
||||
|
@ -521,7 +522,7 @@ public class PluginsServiceTests extends ESTestCase {
|
|||
"MyPlugin", Arrays.asList("dep1", "dep2"), false);
|
||||
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
|
||||
IllegalStateException e = expectThrows(IllegalStateException.class, () ->
|
||||
PluginsService.checkBundleJarHell(bundle, transitiveDeps));
|
||||
PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps));
|
||||
assertEquals("failed to load plugin myplugin due to jar hell", e.getMessage());
|
||||
assertThat(e.getCause().getMessage(), containsString("jar hell!"));
|
||||
assertThat(e.getCause().getMessage(), containsString("DummyClass2"));
|
||||
|
@ -543,7 +544,7 @@ public class PluginsServiceTests extends ESTestCase {
|
|||
PluginInfo info1 = new PluginInfo("myplugin", "desc", "1.0", Version.CURRENT, "1.8",
|
||||
"MyPlugin", Arrays.asList("dep1", "dep2"), false);
|
||||
PluginsService.Bundle bundle = new PluginsService.Bundle(info1, pluginDir);
|
||||
PluginsService.checkBundleJarHell(bundle, transitiveDeps);
|
||||
PluginsService.checkBundleJarHell(JarHell.parseClassPath(), bundle, transitiveDeps);
|
||||
Set<URL> deps = transitiveDeps.get("myplugin");
|
||||
assertNotNull(deps);
|
||||
assertThat(deps, containsInAnyOrder(pluginJar.toUri().toURL(), dep1Jar.toUri().toURL(), dep2Jar.toUri().toURL()));
|
||||
|
|
|
@ -30,14 +30,14 @@ import org.elasticsearch.index.fielddata.ScriptDocValues;
|
|||
import org.elasticsearch.plugins.Plugin;
|
||||
import org.elasticsearch.plugins.ScriptPlugin;
|
||||
import org.elasticsearch.script.ExplainableSearchScript;
|
||||
import org.elasticsearch.script.ScoreScript;
|
||||
import org.elasticsearch.script.Script;
|
||||
import org.elasticsearch.script.ScriptContext;
|
||||
import org.elasticsearch.script.ScriptEngine;
|
||||
import org.elasticsearch.script.ScriptType;
|
||||
import org.elasticsearch.script.SearchScript;
|
||||
import org.elasticsearch.search.SearchHit;
|
||||
import org.elasticsearch.search.SearchHits;
|
||||
import org.elasticsearch.search.lookup.LeafDocLookup;
|
||||
import org.elasticsearch.search.lookup.SearchLookup;
|
||||
import org.elasticsearch.test.ESIntegTestCase;
|
||||
import org.elasticsearch.test.ESIntegTestCase.ClusterScope;
|
||||
import org.elasticsearch.test.ESIntegTestCase.Scope;
|
||||
|
@ -76,16 +76,17 @@ public class ExplainableScriptIT extends ESIntegTestCase {
|
|||
@Override
|
||||
public <T> T compile(String scriptName, String scriptSource, ScriptContext<T> context, Map<String, String> params) {
|
||||
assert scriptSource.equals("explainable_script");
|
||||
assert context == SearchScript.SCRIPT_SCORE_CONTEXT;
|
||||
SearchScript.Factory factory = (p, lookup) -> new SearchScript.LeafFactory() {
|
||||
@Override
|
||||
public SearchScript newInstance(LeafReaderContext context) throws IOException {
|
||||
return new MyScript(lookup.doc().getLeafDocLookup(context));
|
||||
}
|
||||
assert context == ScoreScript.CONTEXT;
|
||||
ScoreScript.Factory factory = (params1, lookup) -> new ScoreScript.LeafFactory() {
|
||||
@Override
|
||||
public boolean needs_score() {
|
||||
return false;
|
||||
}
|
||||
|
||||
@Override
|
||||
public ScoreScript newInstance(LeafReaderContext ctx) throws IOException {
|
||||
return new MyScript(params1, lookup, ctx);
|
||||
}
|
||||
};
|
||||
return context.factoryClazz.cast(factory);
|
||||
}
|
||||
|
@ -93,28 +94,21 @@ public class ExplainableScriptIT extends ESIntegTestCase {
|
|||
}
|
||||
}
|
||||
|
||||
static class MyScript extends SearchScript implements ExplainableSearchScript {
|
||||
LeafDocLookup docLookup;
|
||||
static class MyScript extends ScoreScript implements ExplainableSearchScript {
|
||||
|
||||
MyScript(LeafDocLookup docLookup) {
|
||||
super(null, null, null);
|
||||
this.docLookup = docLookup;
|
||||
MyScript(Map<String, Object> params, SearchLookup lookup, LeafReaderContext leafContext) {
|
||||
super(params, lookup, leafContext);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void setDocument(int doc) {
|
||||
docLookup.setDocument(doc);
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public Explanation explain(Explanation subQueryScore) throws IOException {
|
||||
Explanation scoreExp = Explanation.match(subQueryScore.getValue(), "_score: ", subQueryScore);
|
||||
return Explanation.match((float) (runAsDouble()), "This script returned " + runAsDouble(), scoreExp);
|
||||
return Explanation.match((float) (execute()), "This script returned " + execute(), scoreExp);
|
||||
}
|
||||
|
||||
@Override
|
||||
public double runAsDouble() {
|
||||
return ((Number) ((ScriptDocValues) docLookup.get("number_field")).getValues().get(0)).doubleValue();
|
||||
public double execute() {
|
||||
return ((Number) ((ScriptDocValues) getDoc().get("number_field")).getValues().get(0)).doubleValue();
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -23,10 +23,12 @@ import com.carrotsearch.hppc.IntHashSet;
|
|||
import com.carrotsearch.hppc.IntSet;
|
||||
import org.elasticsearch.action.ActionFuture;
|
||||
import org.elasticsearch.action.admin.cluster.repositories.put.PutRepositoryResponse;
|
||||
import org.elasticsearch.action.admin.cluster.repositories.verify.VerifyRepositoryResponse;
|
||||
import org.elasticsearch.action.admin.cluster.snapshots.create.CreateSnapshotResponse;
|
||||
import org.elasticsearch.action.admin.cluster.snapshots.delete.DeleteSnapshotResponse;
|
||||
import org.elasticsearch.action.admin.cluster.snapshots.get.GetSnapshotsResponse;
|
||||
import org.elasticsearch.action.admin.cluster.snapshots.restore.RestoreSnapshotResponse;
|
||||
import org.elasticsearch.action.admin.cluster.snapshots.status.SnapshotStats;
|
||||
import org.elasticsearch.action.admin.cluster.snapshots.status.SnapshotStatus;
|
||||
import org.elasticsearch.action.admin.cluster.snapshots.status.SnapshotsStatusResponse;
|
||||
import org.elasticsearch.action.admin.cluster.state.ClusterStateResponse;
|
||||
|
@ -83,7 +85,12 @@ import org.elasticsearch.test.TestCustomMetaData;
|
|||
import org.elasticsearch.test.rest.FakeRestRequest;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.UncheckedIOException;
|
||||
import java.nio.file.FileVisitResult;
|
||||
import java.nio.file.Files;
|
||||
import java.nio.file.Path;
|
||||
import java.nio.file.SimpleFileVisitor;
|
||||
import java.nio.file.attribute.BasicFileAttributes;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.Collection;
|
||||
|
@ -102,6 +109,7 @@ import static org.hamcrest.Matchers.equalTo;
|
|||
import static org.hamcrest.Matchers.greaterThan;
|
||||
import static org.hamcrest.Matchers.greaterThanOrEqualTo;
|
||||
import static org.hamcrest.Matchers.hasSize;
|
||||
import static org.hamcrest.Matchers.is;
|
||||
import static org.hamcrest.Matchers.lessThan;
|
||||
import static org.hamcrest.Matchers.not;
|
||||
import static org.hamcrest.Matchers.notNullValue;
|
||||
|
@ -1019,6 +1027,129 @@ public class DedicatedClusterSnapshotRestoreIT extends AbstractSnapshotIntegTest
|
|||
assertThat(snapshots.get(0).getState().completed(), equalTo(true));
|
||||
}
|
||||
|
||||
public void testSnapshotTotalAndIncrementalSizes() throws IOException {
|
||||
Client client = client();
|
||||
final String indexName = "test-blocks-1";
|
||||
final String repositoryName = "repo-" + indexName;
|
||||
final String snapshot0 = "snapshot-0";
|
||||
final String snapshot1 = "snapshot-1";
|
||||
|
||||
createIndex(indexName);
|
||||
|
||||
int docs = between(10, 100);
|
||||
for (int i = 0; i < docs; i++) {
|
||||
client.prepareIndex(indexName, "type").setSource("test", "init").execute().actionGet();
|
||||
}
|
||||
|
||||
logger.info("--> register a repository");
|
||||
|
||||
final Path repoPath = randomRepoPath();
|
||||
assertAcked(client.admin().cluster().preparePutRepository(repositoryName)
|
||||
.setType("fs")
|
||||
.setSettings(Settings.builder().put("location", repoPath)));
|
||||
|
||||
logger.info("--> create a snapshot");
|
||||
client.admin().cluster().prepareCreateSnapshot(repositoryName, snapshot0)
|
||||
.setIncludeGlobalState(true)
|
||||
.setWaitForCompletion(true)
|
||||
.get();
|
||||
|
||||
SnapshotsStatusResponse response = client.admin().cluster().prepareSnapshotStatus(repositoryName)
|
||||
.setSnapshots(snapshot0)
|
||||
.get();
|
||||
|
||||
List<SnapshotStatus> snapshots = response.getSnapshots();
|
||||
|
||||
List<Path> snapshot0Files = scanSnapshotFolder(repoPath);
|
||||
assertThat(snapshots, hasSize(1));
|
||||
|
||||
final int snapshot0FileCount = snapshot0Files.size();
|
||||
final long snapshot0FileSize = calculateTotalFilesSize(snapshot0Files);
|
||||
|
||||
SnapshotStats stats = snapshots.get(0).getStats();
|
||||
|
||||
assertThat(stats.getTotalFileCount(), is(snapshot0FileCount));
|
||||
assertThat(stats.getTotalSize(), is(snapshot0FileSize));
|
||||
|
||||
assertThat(stats.getIncrementalFileCount(), equalTo(snapshot0FileCount));
|
||||
assertThat(stats.getIncrementalSize(), equalTo(snapshot0FileSize));
|
||||
|
||||
assertThat(stats.getIncrementalFileCount(), equalTo(stats.getProcessedFileCount()));
|
||||
assertThat(stats.getIncrementalSize(), equalTo(stats.getProcessedSize()));
|
||||
|
||||
// add few docs - less than initially
|
||||
docs = between(1, 5);
|
||||
for (int i = 0; i < docs; i++) {
|
||||
client.prepareIndex(indexName, "type").setSource("test", "test" + i).execute().actionGet();
|
||||
}
|
||||
|
||||
// create another snapshot
|
||||
// total size has to grow and has to be equal to files on fs
|
||||
assertThat(client.admin().cluster()
|
||||
.prepareCreateSnapshot(repositoryName, snapshot1)
|
||||
.setWaitForCompletion(true).get().status(),
|
||||
equalTo(RestStatus.OK));
|
||||
|
||||
// drop 1st one to avoid miscalculation as snapshot reuses some files of prev snapshot
|
||||
assertTrue(client.admin().cluster()
|
||||
.prepareDeleteSnapshot(repositoryName, snapshot0)
|
||||
.get().isAcknowledged());
|
||||
|
||||
response = client.admin().cluster().prepareSnapshotStatus(repositoryName)
|
||||
.setSnapshots(snapshot1)
|
||||
.get();
|
||||
|
||||
final List<Path> snapshot1Files = scanSnapshotFolder(repoPath);
|
||||
|
||||
final int snapshot1FileCount = snapshot1Files.size();
|
||||
final long snapshot1FileSize = calculateTotalFilesSize(snapshot1Files);
|
||||
|
||||
snapshots = response.getSnapshots();
|
||||
|
||||
SnapshotStats anotherStats = snapshots.get(0).getStats();
|
||||
|
||||
ArrayList<Path> snapshotFilesDiff = new ArrayList<>(snapshot1Files);
|
||||
snapshotFilesDiff.removeAll(snapshot0Files);
|
||||
|
||||
assertThat(anotherStats.getIncrementalFileCount(), equalTo(snapshotFilesDiff.size()));
|
||||
assertThat(anotherStats.getIncrementalSize(), equalTo(calculateTotalFilesSize(snapshotFilesDiff)));
|
||||
|
||||
assertThat(anotherStats.getIncrementalFileCount(), equalTo(anotherStats.getProcessedFileCount()));
|
||||
assertThat(anotherStats.getIncrementalSize(), equalTo(anotherStats.getProcessedSize()));
|
||||
|
||||
assertThat(stats.getTotalSize(), lessThan(anotherStats.getTotalSize()));
|
||||
assertThat(stats.getTotalFileCount(), lessThan(anotherStats.getTotalFileCount()));
|
||||
|
||||
assertThat(anotherStats.getTotalFileCount(), is(snapshot1FileCount));
|
||||
assertThat(anotherStats.getTotalSize(), is(snapshot1FileSize));
|
||||
}
|
||||
|
||||
private long calculateTotalFilesSize(List<Path> files) {
|
||||
return files.stream().mapToLong(f -> {
|
||||
try {
|
||||
return Files.size(f);
|
||||
} catch (IOException e) {
|
||||
throw new UncheckedIOException(e);
|
||||
}
|
||||
}).sum();
|
||||
}
|
||||
|
||||
|
||||
private List<Path> scanSnapshotFolder(Path repoPath) throws IOException {
|
||||
List<Path> files = new ArrayList<>();
|
||||
Files.walkFileTree(repoPath, new SimpleFileVisitor<Path>(){
|
||||
@Override
|
||||
public FileVisitResult visitFile(Path file, BasicFileAttributes attrs) throws IOException {
|
||||
if (file.getFileName().toString().startsWith("__")){
|
||||
files.add(file);
|
||||
}
|
||||
return super.visitFile(file, attrs);
|
||||
}
|
||||
}
|
||||
);
|
||||
return files;
|
||||
}
|
||||
|
||||
public static class SnapshottableMetadata extends TestCustomMetaData {
|
||||
public static final String TYPE = "test_snapshottable";
|
||||
|
||||
|
|
|
@ -2066,7 +2066,7 @@ public class SharedClusterSnapshotRestoreIT extends AbstractSnapshotIntegTestCas
|
|||
SnapshotStatus snapshotStatus = client.admin().cluster().prepareSnapshotStatus("test-repo").setSnapshots("test").get().getSnapshots().get(0);
|
||||
List<SnapshotIndexShardStatus> shards = snapshotStatus.getShards();
|
||||
for (SnapshotIndexShardStatus status : shards) {
|
||||
assertThat(status.getStats().getProcessedFiles(), greaterThan(1));
|
||||
assertThat(status.getStats().getProcessedFileCount(), greaterThan(1));
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -2078,7 +2078,7 @@ public class SharedClusterSnapshotRestoreIT extends AbstractSnapshotIntegTestCas
|
|||
SnapshotStatus snapshotStatus = client.admin().cluster().prepareSnapshotStatus("test-repo").setSnapshots("test-1").get().getSnapshots().get(0);
|
||||
List<SnapshotIndexShardStatus> shards = snapshotStatus.getShards();
|
||||
for (SnapshotIndexShardStatus status : shards) {
|
||||
assertThat(status.getStats().getProcessedFiles(), equalTo(0));
|
||||
assertThat(status.getStats().getProcessedFileCount(), equalTo(0));
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -2091,7 +2091,7 @@ public class SharedClusterSnapshotRestoreIT extends AbstractSnapshotIntegTestCas
|
|||
SnapshotStatus snapshotStatus = client.admin().cluster().prepareSnapshotStatus("test-repo").setSnapshots("test-2").get().getSnapshots().get(0);
|
||||
List<SnapshotIndexShardStatus> shards = snapshotStatus.getShards();
|
||||
for (SnapshotIndexShardStatus status : shards) {
|
||||
assertThat(status.getStats().getProcessedFiles(), equalTo(2)); // we flush before the snapshot such that we have to process the segments_N files plus the .del file
|
||||
assertThat(status.getStats().getProcessedFileCount(), equalTo(2)); // we flush before the snapshot such that we have to process the segments_N files plus the .del file
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -632,7 +632,7 @@ public abstract class IndexShardTestCase extends ESTestCase {
|
|||
|
||||
final IndexShardSnapshotStatus.Copy lastSnapshotStatus = snapshotStatus.asCopy();
|
||||
assertEquals(IndexShardSnapshotStatus.Stage.DONE, lastSnapshotStatus.getStage());
|
||||
assertEquals(shard.snapshotStoreMetadata().size(), lastSnapshotStatus.getNumberOfFiles());
|
||||
assertEquals(shard.snapshotStoreMetadata().size(), lastSnapshotStatus.getTotalFileCount());
|
||||
assertNull(lastSnapshotStatus.getFailure());
|
||||
}
|
||||
|
||||
|
|
|
@ -25,7 +25,6 @@ import org.elasticsearch.index.similarity.ScriptedSimilarity.Doc;
|
|||
import org.elasticsearch.index.similarity.ScriptedSimilarity.Field;
|
||||
import org.elasticsearch.index.similarity.ScriptedSimilarity.Query;
|
||||
import org.elasticsearch.index.similarity.ScriptedSimilarity.Term;
|
||||
import org.elasticsearch.index.similarity.SimilarityService;
|
||||
import org.elasticsearch.search.aggregations.pipeline.movfn.MovingFunctionScript;
|
||||
import org.elasticsearch.search.aggregations.pipeline.movfn.MovingFunctions;
|
||||
import org.elasticsearch.search.lookup.LeafSearchLookup;
|
||||
|
@ -36,7 +35,6 @@ import java.util.Collections;
|
|||
import java.util.HashMap;
|
||||
import java.util.Map;
|
||||
import java.util.function.Function;
|
||||
import java.util.function.Predicate;
|
||||
|
||||
import static java.util.Collections.emptyMap;
|
||||
|
||||
|
@ -114,6 +112,9 @@ public class MockScriptEngine implements ScriptEngine {
|
|||
} else if (context.instanceClazz.equals(MovingFunctionScript.class)) {
|
||||
MovingFunctionScript.Factory factory = mockCompiled::createMovingFunctionScript;
|
||||
return context.factoryClazz.cast(factory);
|
||||
} else if (context.instanceClazz.equals(ScoreScript.class)) {
|
||||
ScoreScript.Factory factory = new MockScoreScript(script);
|
||||
return context.factoryClazz.cast(factory);
|
||||
}
|
||||
throw new IllegalArgumentException("mock script engine does not know how to handle context [" + context.name + "]");
|
||||
}
|
||||
|
@ -342,5 +343,45 @@ public class MockScriptEngine implements ScriptEngine {
|
|||
return MovingFunctions.unweightedAvg(values);
|
||||
}
|
||||
}
|
||||
|
||||
public class MockScoreScript implements ScoreScript.Factory {
|
||||
|
||||
private final Function<Map<String, Object>, Object> scripts;
|
||||
|
||||
MockScoreScript(Function<Map<String, Object>, Object> scripts) {
|
||||
this.scripts = scripts;
|
||||
}
|
||||
|
||||
@Override
|
||||
public ScoreScript.LeafFactory newFactory(Map<String, Object> params, SearchLookup lookup) {
|
||||
return new ScoreScript.LeafFactory() {
|
||||
@Override
|
||||
public boolean needs_score() {
|
||||
return true;
|
||||
}
|
||||
|
||||
@Override
|
||||
public ScoreScript newInstance(LeafReaderContext ctx) throws IOException {
|
||||
Scorer[] scorerHolder = new Scorer[1];
|
||||
return new ScoreScript(params, lookup, ctx) {
|
||||
@Override
|
||||
public double execute() {
|
||||
Map<String, Object> vars = new HashMap<>(getParams());
|
||||
vars.put("doc", getDoc());
|
||||
if (scorerHolder[0] != null) {
|
||||
vars.put("_score", new ScoreAccessor(scorerHolder[0]));
|
||||
}
|
||||
return ((Number) scripts.apply(vars)).doubleValue();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void setScorer(Scorer scorer) {
|
||||
scorerHolder[0] = scorer;
|
||||
}
|
||||
};
|
||||
}
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -123,13 +123,15 @@ public class XPackInfoResponse extends ActionResponse {
|
|||
|
||||
@Override
|
||||
public XContentBuilder toXContent(XContentBuilder builder, Params params) throws IOException {
|
||||
return builder.startObject()
|
||||
.field("uid", uid)
|
||||
.field("type", type)
|
||||
.field("mode", mode)
|
||||
.field("status", status.label())
|
||||
.timeField("expiry_date_in_millis", "expiry_date", expiryDate)
|
||||
.endObject();
|
||||
builder.startObject()
|
||||
.field("uid", uid)
|
||||
.field("type", type)
|
||||
.field("mode", mode)
|
||||
.field("status", status.label());
|
||||
if (expiryDate != LicenseService.BASIC_SELF_GENERATED_LICENSE_EXPIRATION_MILLIS) {
|
||||
builder.timeField("expiry_date_in_millis", "expiry_date", expiryDate);
|
||||
}
|
||||
return builder.endObject();
|
||||
}
|
||||
|
||||
public void writeTo(StreamOutput out) throws IOException {
|
||||
|
|
|
@ -52,24 +52,20 @@ dependencies {
|
|||
compile (xpackProject('plugin:sql:sql-shared-client')) {
|
||||
transitive = false
|
||||
}
|
||||
compile (xpackProject('plugin:sql:sql-proto')) {
|
||||
compile (xpackProject('plugin:sql:sql-shared-proto')) {
|
||||
transitive = false
|
||||
}
|
||||
} else {
|
||||
bundled (xpackProject('plugin:sql:sql-shared-client')) {
|
||||
transitive = false
|
||||
}
|
||||
bundled (xpackProject('plugin:sql:sql-proto')) {
|
||||
bundled (xpackProject('plugin:sql:sql-shared-proto')) {
|
||||
transitive = false
|
||||
}
|
||||
}
|
||||
compile (project(':server')) {
|
||||
transitive = false
|
||||
}
|
||||
compile (project(':libs:x-content')) {
|
||||
transitive = false
|
||||
}
|
||||
compile "org.apache.lucene:lucene-core:${versions.lucene}"
|
||||
compile 'joda-time:joda-time:2.9.9'
|
||||
compile project(':libs:elasticsearch-core')
|
||||
runtime "com.fasterxml.jackson.core:jackson-core:${versions.jackson}"
|
||||
|
@ -80,15 +76,13 @@ dependencies {
|
|||
}
|
||||
|
||||
dependencyLicenses {
|
||||
mapping from: /sql-proto.*/, to: 'elasticsearch'
|
||||
mapping from: /sql-shared-proto.*/, to: 'elasticsearch'
|
||||
mapping from: /sql-shared-client.*/, to: 'elasticsearch'
|
||||
mapping from: /jackson-.*/, to: 'jackson'
|
||||
mapping from: /lucene-.*/, to: 'lucene'
|
||||
mapping from: /elasticsearch-core.*/, to: 'elasticsearch'
|
||||
ignoreSha 'sql-proto'
|
||||
ignoreSha 'sql-shared-proto'
|
||||
ignoreSha 'sql-shared-client'
|
||||
ignoreSha 'elasticsearch'
|
||||
ignoreSha 'elasticsearch-core'
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
@ -1,475 +0,0 @@
|
|||
|
||||
Apache License
|
||||
Version 2.0, January 2004
|
||||
http://www.apache.org/licenses/
|
||||
|
||||
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||
|
||||
1. Definitions.
|
||||
|
||||
"License" shall mean the terms and conditions for use, reproduction,
|
||||
and distribution as defined by Sections 1 through 9 of this document.
|
||||
|
||||
"Licensor" shall mean the copyright owner or entity authorized by
|
||||
the copyright owner that is granting the License.
|
||||
|
||||
"Legal Entity" shall mean the union of the acting entity and all
|
||||
other entities that control, are controlled by, or are under common
|
||||
control with that entity. For the purposes of this definition,
|
||||
"control" means (i) the power, direct or indirect, to cause the
|
||||
direction or management of such entity, whether by contract or
|
||||
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||
|
||||
"You" (or "Your") shall mean an individual or Legal Entity
|
||||
exercising permissions granted by this License.
|
||||
|
||||
"Source" form shall mean the preferred form for making modifications,
|
||||
including but not limited to software source code, documentation
|
||||
source, and configuration files.
|
||||
|
||||
"Object" form shall mean any form resulting from mechanical
|
||||
transformation or translation of a Source form, including but
|
||||
not limited to compiled object code, generated documentation,
|
||||
and conversions to other media types.
|
||||
|
||||
"Work" shall mean the work of authorship, whether in Source or
|
||||
Object form, made available under the License, as indicated by a
|
||||
copyright notice that is included in or attached to the work
|
||||
(an example is provided in the Appendix below).
|
||||
|
||||
"Derivative Works" shall mean any work, whether in Source or Object
|
||||
form, that is based on (or derived from) the Work and for which the
|
||||
editorial revisions, annotations, elaborations, or other modifications
|
||||
represent, as a whole, an original work of authorship. For the purposes
|
||||
of this License, Derivative Works shall not include works that remain
|
||||
separable from, or merely link (or bind by name) to the interfaces of,
|
||||
the Work and Derivative Works thereof.
|
||||
|
||||
"Contribution" shall mean any work of authorship, including
|
||||
the original version of the Work and any modifications or additions
|
||||
to that Work or Derivative Works thereof, that is intentionally
|
||||
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||
or by an individual or Legal Entity authorized to submit on behalf of
|
||||
the copyright owner. For the purposes of this definition, "submitted"
|
||||
means any form of electronic, verbal, or written communication sent
|
||||
to the Licensor or its representatives, including but not limited to
|
||||
communication on electronic mailing lists, source code control systems,
|
||||
and issue tracking systems that are managed by, or on behalf of, the
|
||||
Licensor for the purpose of discussing and improving the Work, but
|
||||
excluding communication that is conspicuously marked or otherwise
|
||||
designated in writing by the copyright owner as "Not a Contribution."
|
||||
|
||||
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||
on behalf of whom a Contribution has been received by Licensor and
|
||||
subsequently incorporated within the Work.
|
||||
|
||||
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
copyright license to reproduce, prepare Derivative Works of,
|
||||
publicly display, publicly perform, sublicense, and distribute the
|
||||
Work and such Derivative Works in Source or Object form.
|
||||
|
||||
3. Grant of Patent License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
(except as stated in this section) patent license to make, have made,
|
||||
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||
where such license applies only to those patent claims licensable
|
||||
by such Contributor that are necessarily infringed by their
|
||||
Contribution(s) alone or by combination of their Contribution(s)
|
||||
with the Work to which such Contribution(s) was submitted. If You
|
||||
institute patent litigation against any entity (including a
|
||||
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||
or a Contribution incorporated within the Work constitutes direct
|
||||
or contributory patent infringement, then any patent licenses
|
||||
granted to You under this License for that Work shall terminate
|
||||
as of the date such litigation is filed.
|
||||
|
||||
4. Redistribution. You may reproduce and distribute copies of the
|
||||
Work or Derivative Works thereof in any medium, with or without
|
||||
modifications, and in Source or Object form, provided that You
|
||||
meet the following conditions:
|
||||
|
||||
(a) You must give any other recipients of the Work or
|
||||
Derivative Works a copy of this License; and
|
||||
|
||||
(b) You must cause any modified files to carry prominent notices
|
||||
stating that You changed the files; and
|
||||
|
||||
(c) You must retain, in the Source form of any Derivative Works
|
||||
that You distribute, all copyright, patent, trademark, and
|
||||
attribution notices from the Source form of the Work,
|
||||
excluding those notices that do not pertain to any part of
|
||||
the Derivative Works; and
|
||||
|
||||
(d) If the Work includes a "NOTICE" text file as part of its
|
||||
distribution, then any Derivative Works that You distribute must
|
||||
include a readable copy of the attribution notices contained
|
||||
within such NOTICE file, excluding those notices that do not
|
||||
pertain to any part of the Derivative Works, in at least one
|
||||
of the following places: within a NOTICE text file distributed
|
||||
as part of the Derivative Works; within the Source form or
|
||||
documentation, if provided along with the Derivative Works; or,
|
||||
within a display generated by the Derivative Works, if and
|
||||
wherever such third-party notices normally appear. The contents
|
||||
of the NOTICE file are for informational purposes only and
|
||||
do not modify the License. You may add Your own attribution
|
||||
notices within Derivative Works that You distribute, alongside
|
||||
or as an addendum to the NOTICE text from the Work, provided
|
||||
that such additional attribution notices cannot be construed
|
||||
as modifying the License.
|
||||
|
||||
You may add Your own copyright statement to Your modifications and
|
||||
may provide additional or different license terms and conditions
|
||||
for use, reproduction, or distribution of Your modifications, or
|
||||
for any such Derivative Works as a whole, provided Your use,
|
||||
reproduction, and distribution of the Work otherwise complies with
|
||||
the conditions stated in this License.
|
||||
|
||||
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||
any Contribution intentionally submitted for inclusion in the Work
|
||||
by You to the Licensor shall be under the terms and conditions of
|
||||
this License, without any additional terms or conditions.
|
||||
Notwithstanding the above, nothing herein shall supersede or modify
|
||||
the terms of any separate license agreement you may have executed
|
||||
with Licensor regarding such Contributions.
|
||||
|
||||
6. Trademarks. This License does not grant permission to use the trade
|
||||
names, trademarks, service marks, or product names of the Licensor,
|
||||
except as required for reasonable and customary use in describing the
|
||||
origin of the Work and reproducing the content of the NOTICE file.
|
||||
|
||||
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||
agreed to in writing, Licensor provides the Work (and each
|
||||
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
implied, including, without limitation, any warranties or conditions
|
||||
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||
appropriateness of using or redistributing the Work and assume any
|
||||
risks associated with Your exercise of permissions under this License.
|
||||
|
||||
8. Limitation of Liability. In no event and under no legal theory,
|
||||
whether in tort (including negligence), contract, or otherwise,
|
||||
unless required by applicable law (such as deliberate and grossly
|
||||
negligent acts) or agreed to in writing, shall any Contributor be
|
||||
liable to You for damages, including any direct, indirect, special,
|
||||
incidental, or consequential damages of any character arising as a
|
||||
result of this License or out of the use or inability to use the
|
||||
Work (including but not limited to damages for loss of goodwill,
|
||||
work stoppage, computer failure or malfunction, or any and all
|
||||
other commercial damages or losses), even if such Contributor
|
||||
has been advised of the possibility of such damages.
|
||||
|
||||
9. Accepting Warranty or Additional Liability. While redistributing
|
||||
the Work or Derivative Works thereof, You may choose to offer,
|
||||
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||
or other liability obligations and/or rights consistent with this
|
||||
License. However, in accepting such obligations, You may act only
|
||||
on Your own behalf and on Your sole responsibility, not on behalf
|
||||
of any other Contributor, and only if You agree to indemnify,
|
||||
defend, and hold each Contributor harmless for any liability
|
||||
incurred by, or claims asserted against, such Contributor by reason
|
||||
of your accepting any such warranty or additional liability.
|
||||
|
||||
END OF TERMS AND CONDITIONS
|
||||
|
||||
APPENDIX: How to apply the Apache License to your work.
|
||||
|
||||
To apply the Apache License to your work, attach the following
|
||||
boilerplate notice, with the fields enclosed by brackets "[]"
|
||||
replaced with your own identifying information. (Don't include
|
||||
the brackets!) The text should be enclosed in the appropriate
|
||||
comment syntax for the file format. We also recommend that a
|
||||
file or class name and description of purpose be included on the
|
||||
same "printed page" as the copyright notice for easier
|
||||
identification within third-party archives.
|
||||
|
||||
Copyright [yyyy] [name of copyright owner]
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
|
||||
|
||||
|
||||
Some code in core/src/java/org/apache/lucene/util/UnicodeUtil.java was
|
||||
derived from unicode conversion examples available at
|
||||
http://www.unicode.org/Public/PROGRAMS/CVTUTF. Here is the copyright
|
||||
from those sources:
|
||||
|
||||
/*
|
||||
* Copyright 2001-2004 Unicode, Inc.
|
||||
*
|
||||
* Disclaimer
|
||||
*
|
||||
* This source code is provided as is by Unicode, Inc. No claims are
|
||||
* made as to fitness for any particular purpose. No warranties of any
|
||||
* kind are expressed or implied. The recipient agrees to determine
|
||||
* applicability of information provided. If this file has been
|
||||
* purchased on magnetic or optical media from Unicode, Inc., the
|
||||
* sole remedy for any claim will be exchange of defective media
|
||||
* within 90 days of receipt.
|
||||
*
|
||||
* Limitations on Rights to Redistribute This Code
|
||||
*
|
||||
* Unicode, Inc. hereby grants the right to freely use the information
|
||||
* supplied in this file in the creation of products supporting the
|
||||
* Unicode Standard, and to make copies of this file in any form
|
||||
* for internal or external distribution as long as this notice
|
||||
* remains attached.
|
||||
*/
|
||||
|
||||
|
||||
Some code in core/src/java/org/apache/lucene/util/ArrayUtil.java was
|
||||
derived from Python 2.4.2 sources available at
|
||||
http://www.python.org. Full license is here:
|
||||
|
||||
http://www.python.org/download/releases/2.4.2/license/
|
||||
|
||||
Some code in core/src/java/org/apache/lucene/util/UnicodeUtil.java was
|
||||
derived from Python 3.1.2 sources available at
|
||||
http://www.python.org. Full license is here:
|
||||
|
||||
http://www.python.org/download/releases/3.1.2/license/
|
||||
|
||||
Some code in core/src/java/org/apache/lucene/util/automaton was
|
||||
derived from Brics automaton sources available at
|
||||
www.brics.dk/automaton/. Here is the copyright from those sources:
|
||||
|
||||
/*
|
||||
* Copyright (c) 2001-2009 Anders Moeller
|
||||
* All rights reserved.
|
||||
*
|
||||
* Redistribution and use in source and binary forms, with or without
|
||||
* modification, are permitted provided that the following conditions
|
||||
* are met:
|
||||
* 1. Redistributions of source code must retain the above copyright
|
||||
* notice, this list of conditions and the following disclaimer.
|
||||
* 2. Redistributions in binary form must reproduce the above copyright
|
||||
* notice, this list of conditions and the following disclaimer in the
|
||||
* documentation and/or other materials provided with the distribution.
|
||||
* 3. The name of the author may not be used to endorse or promote products
|
||||
* derived from this software without specific prior written permission.
|
||||
*
|
||||
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
|
||||
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
|
||||
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
|
||||
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
|
||||
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
|
||||
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
|
||||
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
*/
|
||||
|
||||
The levenshtein automata tables in core/src/java/org/apache/lucene/util/automaton
|
||||
were automatically generated with the moman/finenight FSA package.
|
||||
Here is the copyright for those sources:
|
||||
|
||||
# Copyright (c) 2010, Jean-Philippe Barrette-LaPierre, <jpb@rrette.com>
|
||||
#
|
||||
# Permission is hereby granted, free of charge, to any person
|
||||
# obtaining a copy of this software and associated documentation
|
||||
# files (the "Software"), to deal in the Software without
|
||||
# restriction, including without limitation the rights to use,
|
||||
# copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||
# copies of the Software, and to permit persons to whom the
|
||||
# Software is furnished to do so, subject to the following
|
||||
# conditions:
|
||||
#
|
||||
# The above copyright notice and this permission notice shall be
|
||||
# included in all copies or substantial portions of the Software.
|
||||
#
|
||||
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
||||
# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
|
||||
# OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
||||
# NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
|
||||
# HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
|
||||
# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||||
# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
||||
# OTHER DEALINGS IN THE SOFTWARE.
|
||||
|
||||
Some code in core/src/java/org/apache/lucene/util/UnicodeUtil.java was
|
||||
derived from ICU (http://www.icu-project.org)
|
||||
The full license is available here:
|
||||
http://source.icu-project.org/repos/icu/icu/trunk/license.html
|
||||
|
||||
/*
|
||||
* Copyright (C) 1999-2010, International Business Machines
|
||||
* Corporation and others. All Rights Reserved.
|
||||
*
|
||||
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||
* of this software and associated documentation files (the "Software"), to deal
|
||||
* in the Software without restriction, including without limitation the rights
|
||||
* to use, copy, modify, merge, publish, distribute, and/or sell copies of the
|
||||
* Software, and to permit persons to whom the Software is furnished to do so,
|
||||
* provided that the above copyright notice(s) and this permission notice appear
|
||||
* in all copies of the Software and that both the above copyright notice(s) and
|
||||
* this permission notice appear in supporting documentation.
|
||||
*
|
||||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OF THIRD PARTY RIGHTS.
|
||||
* IN NO EVENT SHALL THE COPYRIGHT HOLDER OR HOLDERS INCLUDED IN THIS NOTICE BE
|
||||
* LIABLE FOR ANY CLAIM, OR ANY SPECIAL INDIRECT OR CONSEQUENTIAL DAMAGES, OR
|
||||
* ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER
|
||||
* IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT
|
||||
* OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
|
||||
*
|
||||
* Except as contained in this notice, the name of a copyright holder shall not
|
||||
* be used in advertising or otherwise to promote the sale, use or other
|
||||
* dealings in this Software without prior written authorization of the
|
||||
* copyright holder.
|
||||
*/
|
||||
|
||||
The following license applies to the Snowball stemmers:
|
||||
|
||||
Copyright (c) 2001, Dr Martin Porter
|
||||
Copyright (c) 2002, Richard Boulton
|
||||
All rights reserved.
|
||||
|
||||
Redistribution and use in source and binary forms, with or without
|
||||
modification, are permitted provided that the following conditions are met:
|
||||
|
||||
* Redistributions of source code must retain the above copyright notice,
|
||||
* this list of conditions and the following disclaimer.
|
||||
* Redistributions in binary form must reproduce the above copyright
|
||||
* notice, this list of conditions and the following disclaimer in the
|
||||
* documentation and/or other materials provided with the distribution.
|
||||
* Neither the name of the copyright holders nor the names of its contributors
|
||||
* may be used to endorse or promote products derived from this software
|
||||
* without specific prior written permission.
|
||||
|
||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE
|
||||
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
|
||||
The following license applies to the KStemmer:
|
||||
|
||||
Copyright © 2003,
|
||||
Center for Intelligent Information Retrieval,
|
||||
University of Massachusetts, Amherst.
|
||||
All rights reserved.
|
||||
|
||||
Redistribution and use in source and binary forms, with or without modification,
|
||||
are permitted provided that the following conditions are met:
|
||||
|
||||
1. Redistributions of source code must retain the above copyright notice, this
|
||||
list of conditions and the following disclaimer.
|
||||
|
||||
2. Redistributions in binary form must reproduce the above copyright notice,
|
||||
this list of conditions and the following disclaimer in the documentation
|
||||
and/or other materials provided with the distribution.
|
||||
|
||||
3. The names "Center for Intelligent Information Retrieval" and
|
||||
"University of Massachusetts" must not be used to endorse or promote products
|
||||
derived from this software without prior written permission. To obtain
|
||||
permission, contact info@ciir.cs.umass.edu.
|
||||
|
||||
THIS SOFTWARE IS PROVIDED BY UNIVERSITY OF MASSACHUSETTS AND OTHER CONTRIBUTORS
|
||||
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
|
||||
THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
||||
ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE
|
||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
||||
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
|
||||
GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
||||
HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
||||
LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
||||
OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
||||
SUCH DAMAGE.
|
||||
|
||||
The following license applies to the Morfologik project:
|
||||
|
||||
Copyright (c) 2006 Dawid Weiss
|
||||
Copyright (c) 2007-2011 Dawid Weiss, Marcin Miłkowski
|
||||
All rights reserved.
|
||||
|
||||
Redistribution and use in source and binary forms, with or without modification,
|
||||
are permitted provided that the following conditions are met:
|
||||
|
||||
* Redistributions of source code must retain the above copyright notice,
|
||||
this list of conditions and the following disclaimer.
|
||||
|
||||
* Redistributions in binary form must reproduce the above copyright notice,
|
||||
this list of conditions and the following disclaimer in the documentation
|
||||
and/or other materials provided with the distribution.
|
||||
|
||||
* Neither the name of Morfologik nor the names of its contributors
|
||||
may be used to endorse or promote products derived from this software
|
||||
without specific prior written permission.
|
||||
|
||||
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
|
||||
ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR
|
||||
ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
|
||||
(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
|
||||
LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON
|
||||
ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
|
||||
SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||
|
||||
---
|
||||
|
||||
The dictionary comes from Morfologik project. Morfologik uses data from
|
||||
Polish ispell/myspell dictionary hosted at http://www.sjp.pl/slownik/en/ and
|
||||
is licenced on the terms of (inter alia) LGPL and Creative Commons
|
||||
ShareAlike. The part-of-speech tags were added in Morfologik project and
|
||||
are not found in the data from sjp.pl. The tagset is similar to IPI PAN
|
||||
tagset.
|
||||
|
||||
---
|
||||
|
||||
The following license applies to the Morfeusz project,
|
||||
used by org.apache.lucene.analysis.morfologik.
|
||||
|
||||
BSD-licensed dictionary of Polish (SGJP)
|
||||
http://sgjp.pl/morfeusz/
|
||||
|
||||
Copyright © 2011 Zygmunt Saloni, Włodzimierz Gruszczyński,
|
||||
Marcin Woliński, Robert Wołosz
|
||||
|
||||
All rights reserved.
|
||||
|
||||
Redistribution and use in source and binary forms, with or without
|
||||
modification, are permitted provided that the following conditions are
|
||||
met:
|
||||
|
||||
1. Redistributions of source code must retain the above copyright
|
||||
notice, this list of conditions and the following disclaimer.
|
||||
|
||||
2. Redistributions in binary form must reproduce the above copyright
|
||||
notice, this list of conditions and the following disclaimer in the
|
||||
documentation and/or other materials provided with the
|
||||
distribution.
|
||||
|
||||
THIS SOFTWARE IS PROVIDED BY COPYRIGHT HOLDERS “AS IS” AND ANY EXPRESS
|
||||
OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
||||
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||
DISCLAIMED. IN NO EVENT SHALL COPYRIGHT HOLDERS OR CONTRIBUTORS BE
|
||||
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
||||
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
||||
SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
|
||||
BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
|
||||
WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE
|
||||
OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN
|
||||
IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
@ -1,192 +0,0 @@
|
|||
Apache Lucene
|
||||
Copyright 2014 The Apache Software Foundation
|
||||
|
||||
This product includes software developed at
|
||||
The Apache Software Foundation (http://www.apache.org/).
|
||||
|
||||
Includes software from other Apache Software Foundation projects,
|
||||
including, but not limited to:
|
||||
- Apache Ant
|
||||
- Apache Jakarta Regexp
|
||||
- Apache Commons
|
||||
- Apache Xerces
|
||||
|
||||
ICU4J, (under analysis/icu) is licensed under an MIT styles license
|
||||
and Copyright (c) 1995-2008 International Business Machines Corporation and others
|
||||
|
||||
Some data files (under analysis/icu/src/data) are derived from Unicode data such
|
||||
as the Unicode Character Database. See http://unicode.org/copyright.html for more
|
||||
details.
|
||||
|
||||
Brics Automaton (under core/src/java/org/apache/lucene/util/automaton) is
|
||||
BSD-licensed, created by Anders Møller. See http://www.brics.dk/automaton/
|
||||
|
||||
The levenshtein automata tables (under core/src/java/org/apache/lucene/util/automaton) were
|
||||
automatically generated with the moman/finenight FSA library, created by
|
||||
Jean-Philippe Barrette-LaPierre. This library is available under an MIT license,
|
||||
see http://sites.google.com/site/rrettesite/moman and
|
||||
http://bitbucket.org/jpbarrette/moman/overview/
|
||||
|
||||
The class org.apache.lucene.util.WeakIdentityMap was derived from
|
||||
the Apache CXF project and is Apache License 2.0.
|
||||
|
||||
The Google Code Prettify is Apache License 2.0.
|
||||
See http://code.google.com/p/google-code-prettify/
|
||||
|
||||
JUnit (junit-4.10) is licensed under the Common Public License v. 1.0
|
||||
See http://junit.sourceforge.net/cpl-v10.html
|
||||
|
||||
This product includes code (JaspellTernarySearchTrie) from Java Spelling Checkin
|
||||
g Package (jaspell): http://jaspell.sourceforge.net/
|
||||
License: The BSD License (http://www.opensource.org/licenses/bsd-license.php)
|
||||
|
||||
The snowball stemmers in
|
||||
analysis/common/src/java/net/sf/snowball
|
||||
were developed by Martin Porter and Richard Boulton.
|
||||
The snowball stopword lists in
|
||||
analysis/common/src/resources/org/apache/lucene/analysis/snowball
|
||||
were developed by Martin Porter and Richard Boulton.
|
||||
The full snowball package is available from
|
||||
http://snowball.tartarus.org/
|
||||
|
||||
The KStem stemmer in
|
||||
analysis/common/src/org/apache/lucene/analysis/en
|
||||
was developed by Bob Krovetz and Sergio Guzman-Lara (CIIR-UMass Amherst)
|
||||
under the BSD-license.
|
||||
|
||||
The Arabic,Persian,Romanian,Bulgarian, Hindi and Bengali analyzers (common) come with a default
|
||||
stopword list that is BSD-licensed created by Jacques Savoy. These files reside in:
|
||||
analysis/common/src/resources/org/apache/lucene/analysis/ar/stopwords.txt,
|
||||
analysis/common/src/resources/org/apache/lucene/analysis/fa/stopwords.txt,
|
||||
analysis/common/src/resources/org/apache/lucene/analysis/ro/stopwords.txt,
|
||||
analysis/common/src/resources/org/apache/lucene/analysis/bg/stopwords.txt,
|
||||
analysis/common/src/resources/org/apache/lucene/analysis/hi/stopwords.txt,
|
||||
analysis/common/src/resources/org/apache/lucene/analysis/bn/stopwords.txt
|
||||
See http://members.unine.ch/jacques.savoy/clef/index.html.
|
||||
|
||||
The German,Spanish,Finnish,French,Hungarian,Italian,Portuguese,Russian and Swedish light stemmers
|
||||
(common) are based on BSD-licensed reference implementations created by Jacques Savoy and
|
||||
Ljiljana Dolamic. These files reside in:
|
||||
analysis/common/src/java/org/apache/lucene/analysis/de/GermanLightStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/de/GermanMinimalStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/es/SpanishLightStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/fi/FinnishLightStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/fr/FrenchLightStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/fr/FrenchMinimalStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/hu/HungarianLightStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/it/ItalianLightStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/pt/PortugueseLightStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/ru/RussianLightStemmer.java
|
||||
analysis/common/src/java/org/apache/lucene/analysis/sv/SwedishLightStemmer.java
|
||||
|
||||
The Stempel analyzer (stempel) includes BSD-licensed software developed
|
||||
by the Egothor project http://egothor.sf.net/, created by Leo Galambos, Martin Kvapil,
|
||||
and Edmond Nolan.
|
||||
|
||||
The Polish analyzer (stempel) comes with a default
|
||||
stopword list that is BSD-licensed created by the Carrot2 project. The file resides
|
||||
in stempel/src/resources/org/apache/lucene/analysis/pl/stopwords.txt.
|
||||
See http://project.carrot2.org/license.html.
|
||||
|
||||
The SmartChineseAnalyzer source code (smartcn) was
|
||||
provided by Xiaoping Gao and copyright 2009 by www.imdict.net.
|
||||
|
||||
WordBreakTestUnicode_*.java (under modules/analysis/common/src/test/)
|
||||
is derived from Unicode data such as the Unicode Character Database.
|
||||
See http://unicode.org/copyright.html for more details.
|
||||
|
||||
The Morfologik analyzer (morfologik) includes BSD-licensed software
|
||||
developed by Dawid Weiss and Marcin Miłkowski (http://morfologik.blogspot.com/).
|
||||
|
||||
Morfologik uses data from Polish ispell/myspell dictionary
|
||||
(http://www.sjp.pl/slownik/en/) licenced on the terms of (inter alia)
|
||||
LGPL and Creative Commons ShareAlike.
|
||||
|
||||
Morfologic includes data from BSD-licensed dictionary of Polish (SGJP)
|
||||
(http://sgjp.pl/morfeusz/)
|
||||
|
||||
Servlet-api.jar and javax.servlet-*.jar are under the CDDL license, the original
|
||||
source code for this can be found at http://www.eclipse.org/jetty/downloads.php
|
||||
|
||||
===========================================================================
|
||||
Kuromoji Japanese Morphological Analyzer - Apache Lucene Integration
|
||||
===========================================================================
|
||||
|
||||
This software includes a binary and/or source version of data from
|
||||
|
||||
mecab-ipadic-2.7.0-20070801
|
||||
|
||||
which can be obtained from
|
||||
|
||||
http://atilika.com/releases/mecab-ipadic/mecab-ipadic-2.7.0-20070801.tar.gz
|
||||
|
||||
or
|
||||
|
||||
http://jaist.dl.sourceforge.net/project/mecab/mecab-ipadic/2.7.0-20070801/mecab-ipadic-2.7.0-20070801.tar.gz
|
||||
|
||||
===========================================================================
|
||||
mecab-ipadic-2.7.0-20070801 Notice
|
||||
===========================================================================
|
||||
|
||||
Nara Institute of Science and Technology (NAIST),
|
||||
the copyright holders, disclaims all warranties with regard to this
|
||||
software, including all implied warranties of merchantability and
|
||||
fitness, in no event shall NAIST be liable for
|
||||
any special, indirect or consequential damages or any damages
|
||||
whatsoever resulting from loss of use, data or profits, whether in an
|
||||
action of contract, negligence or other tortuous action, arising out
|
||||
of or in connection with the use or performance of this software.
|
||||
|
||||
A large portion of the dictionary entries
|
||||
originate from ICOT Free Software. The following conditions for ICOT
|
||||
Free Software applies to the current dictionary as well.
|
||||
|
||||
Each User may also freely distribute the Program, whether in its
|
||||
original form or modified, to any third party or parties, PROVIDED
|
||||
that the provisions of Section 3 ("NO WARRANTY") will ALWAYS appear
|
||||
on, or be attached to, the Program, which is distributed substantially
|
||||
in the same form as set out herein and that such intended
|
||||
distribution, if actually made, will neither violate or otherwise
|
||||
contravene any of the laws and regulations of the countries having
|
||||
jurisdiction over the User or the intended distribution itself.
|
||||
|
||||
NO WARRANTY
|
||||
|
||||
The program was produced on an experimental basis in the course of the
|
||||
research and development conducted during the project and is provided
|
||||
to users as so produced on an experimental basis. Accordingly, the
|
||||
program is provided without any warranty whatsoever, whether express,
|
||||
implied, statutory or otherwise. The term "warranty" used herein
|
||||
includes, but is not limited to, any warranty of the quality,
|
||||
performance, merchantability and fitness for a particular purpose of
|
||||
the program and the nonexistence of any infringement or violation of
|
||||
any right of any third party.
|
||||
|
||||
Each user of the program will agree and understand, and be deemed to
|
||||
have agreed and understood, that there is no warranty whatsoever for
|
||||
the program and, accordingly, the entire risk arising from or
|
||||
otherwise connected with the program is assumed by the user.
|
||||
|
||||
Therefore, neither ICOT, the copyright holder, or any other
|
||||
organization that participated in or was otherwise related to the
|
||||
development of the program and their respective officials, directors,
|
||||
officers and other employees shall be held liable for any and all
|
||||
damages, including, without limitation, general, special, incidental
|
||||
and consequential damages, arising out of or otherwise in connection
|
||||
with the use or inability to use the program or any product, material
|
||||
or result produced or otherwise obtained by using the program,
|
||||
regardless of whether they have been advised of, or otherwise had
|
||||
knowledge of, the possibility of such damages at any time during the
|
||||
project or thereafter. Each user will be deemed to have agreed to the
|
||||
foregoing by his or her commencement of use of the program. The term
|
||||
"use" as used herein includes, but is not limited to, the use,
|
||||
modification, copying and distribution of the program and the
|
||||
production of secondary products from the program.
|
||||
|
||||
In the case where the program, whether in its original form or
|
||||
modified, was distributed or delivered to or received by a user from
|
||||
any person, organization or entity other than ICOT, unless it makes or
|
||||
grants independently of ICOT any specific warranty to the user in
|
||||
writing, such person, organization or entity, will also be exempted
|
||||
from and not be held liable to the user for any such damages as noted
|
||||
above as far as the program is concerned.
|
|
@ -1 +0,0 @@
|
|||
e118e4d05070378516b9055184b74498ba528dee
|
|
@ -10,9 +10,8 @@ import org.elasticsearch.common.xcontent.XContentBuilder;
|
|||
import org.elasticsearch.common.xcontent.XContentHelper;
|
||||
import org.elasticsearch.common.xcontent.json.JsonXContent;
|
||||
import org.elasticsearch.test.ESTestCase;
|
||||
import org.elasticsearch.xpack.sql.plugin.SqlQueryResponse;
|
||||
import org.elasticsearch.xpack.sql.proto.Mode;
|
||||
import org.joda.time.DateTime;
|
||||
import org.joda.time.ReadableDateTime;
|
||||
|
||||
import java.sql.JDBCType;
|
||||
|
||||
|
@ -51,7 +50,11 @@ public class TypeConverterTests extends ESTestCase {
|
|||
XContentBuilder builder = JsonXContent.contentBuilder();
|
||||
builder.startObject();
|
||||
builder.field("value");
|
||||
SqlQueryResponse.value(builder, Mode.JDBC, value);
|
||||
if (value instanceof ReadableDateTime) {
|
||||
builder.value(((ReadableDateTime) value).getMillis());
|
||||
} else {
|
||||
builder.value(value);
|
||||
}
|
||||
builder.endObject();
|
||||
builder.close();
|
||||
Object copy = XContentHelper.convertToMap(BytesReference.bytes(builder), false, builder.contentType()).v2().get("value");
|
||||
|
|
|
@ -8,7 +8,6 @@ package org.elasticsearch.xpack.sql.cli.command;
|
|||
import org.elasticsearch.xpack.sql.client.HttpClient;
|
||||
import org.elasticsearch.xpack.sql.client.shared.ClientException;
|
||||
import org.elasticsearch.xpack.sql.client.shared.Version;
|
||||
import org.elasticsearch.xpack.sql.plugin.AbstractSqlQueryRequest;
|
||||
import org.elasticsearch.xpack.sql.proto.MainResponse;
|
||||
import org.elasticsearch.xpack.sql.proto.Protocol;
|
||||
|
||||
|
|
|
@ -28,7 +28,7 @@ public class CliSessionTests extends ESTestCase {
|
|||
public void testProperConnection() throws Exception {
|
||||
HttpClient httpClient = mock(HttpClient.class);
|
||||
when(httpClient.serverInfo()).thenReturn(new MainResponse(randomAlphaOfLength(5), org.elasticsearch.Version.CURRENT.toString(),
|
||||
ClusterName.DEFAULT.value(), UUIDs.randomBase64UUID(), Build.CURRENT));
|
||||
ClusterName.DEFAULT.value(), UUIDs.randomBase64UUID()));
|
||||
CliSession cliSession = new CliSession(httpClient);
|
||||
cliSession.checkConnection();
|
||||
verify(httpClient, times(1)).serverInfo();
|
||||
|
@ -58,7 +58,7 @@ public class CliSessionTests extends ESTestCase {
|
|||
}
|
||||
when(httpClient.serverInfo()).thenReturn(new MainResponse(randomAlphaOfLength(5),
|
||||
org.elasticsearch.Version.fromString(major + "." + minor + ".23").toString(),
|
||||
ClusterName.DEFAULT.value(), UUIDs.randomBase64UUID(), Build.CURRENT));
|
||||
ClusterName.DEFAULT.value(), UUIDs.randomBase64UUID()));
|
||||
CliSession cliSession = new CliSession(httpClient);
|
||||
expectThrows(ClientException.class, cliSession::checkConnection);
|
||||
verify(httpClient, times(1)).serverInfo();
|
||||
|
|
|
@ -36,7 +36,7 @@ public class ServerInfoCliCommandTests extends ESTestCase {
|
|||
HttpClient client = mock(HttpClient.class);
|
||||
CliSession cliSession = new CliSession(client);
|
||||
when(client.serverInfo()).thenReturn(new MainResponse("my_node", "1.2.3",
|
||||
new ClusterName("my_cluster").value(), UUIDs.randomBase64UUID(), Build.CURRENT));
|
||||
new ClusterName("my_cluster").value(), UUIDs.randomBase64UUID()));
|
||||
ServerInfoCliCommand cliCommand = new ServerInfoCliCommand();
|
||||
assertTrue(cliCommand.handle(testTerminal, cliSession, "info"));
|
||||
assertEquals(testTerminal.toString(), "Node:<em>my_node</em> Cluster:<em>my_cluster</em> Version:<em>1.2.3</em>\n");
|
||||
|
|
|
@ -24,6 +24,7 @@ dependencies {
|
|||
compile (project(':libs:x-content')) {
|
||||
transitive = false
|
||||
}
|
||||
compile xpackProject('plugin:sql:sql-shared-proto')
|
||||
compile "org.apache.lucene:lucene-core:${versions.lucene}"
|
||||
compile 'joda-time:joda-time:2.9.9'
|
||||
runtime "com.fasterxml.jackson.core:jackson-core:${versions.jackson}"
|
||||
|
|
|
@ -10,14 +10,14 @@ apply plugin: 'elasticsearch.build'
|
|||
description = 'Code shared between jdbc and cli'
|
||||
|
||||
dependencies {
|
||||
compile xpackProject('plugin:sql:sql-proto')
|
||||
compile xpackProject('plugin:sql:sql-shared-proto')
|
||||
compile "com.fasterxml.jackson.core:jackson-core:${versions.jackson}"
|
||||
testCompile "org.elasticsearch.test:framework:${version}"
|
||||
}
|
||||
|
||||
dependencyLicenses {
|
||||
mapping from: /jackson-.*/, to: 'jackson'
|
||||
mapping from: /sql-proto.*/, to: 'elasticsearch'
|
||||
mapping from: /sql-shared-proto.*/, to: 'elasticsearch'
|
||||
mapping from: /elasticsearch-cli.*/, to: 'elasticsearch'
|
||||
mapping from: /elasticsearch-core.*/, to: 'elasticsearch'
|
||||
mapping from: /lucene-.*/, to: 'lucene'
|
||||
|
|
|
@ -5,14 +5,12 @@
|
|||
*/
|
||||
package org.elasticsearch.xpack.sql.client;
|
||||
|
||||
import org.elasticsearch.common.bytes.BytesReference;
|
||||
import org.elasticsearch.common.collect.Tuple;
|
||||
import org.elasticsearch.common.io.stream.BytesStreamOutput;
|
||||
import org.elasticsearch.common.unit.TimeValue;
|
||||
import org.elasticsearch.common.xcontent.LoggingDeprecationHandler;
|
||||
import org.elasticsearch.common.xcontent.DeprecationHandler;
|
||||
import org.elasticsearch.common.xcontent.NamedXContentRegistry;
|
||||
import org.elasticsearch.common.xcontent.ToXContent;
|
||||
import org.elasticsearch.common.xcontent.XContentHelper;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.common.xcontent.XContentType;
|
||||
import org.elasticsearch.core.internal.io.Streams;
|
||||
|
@ -30,6 +28,8 @@ import org.elasticsearch.xpack.sql.proto.SqlClearCursorResponse;
|
|||
import org.elasticsearch.xpack.sql.proto.SqlQueryRequest;
|
||||
import org.elasticsearch.xpack.sql.proto.SqlQueryResponse;
|
||||
|
||||
import java.io.ByteArrayInputStream;
|
||||
import java.io.ByteArrayOutputStream;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStream;
|
||||
import java.security.AccessController;
|
||||
|
@ -67,8 +67,8 @@ public class HttpClient {
|
|||
public SqlQueryResponse queryInit(String query, int fetchSize) throws SQLException {
|
||||
// TODO allow customizing the time zone - this is what session set/reset/get should be about
|
||||
SqlQueryRequest sqlRequest = new SqlQueryRequest(Mode.PLAIN, query, Collections.emptyList(), null,
|
||||
TimeZone.getTimeZone("UTC"), fetchSize, TimeValue.timeValueMillis(cfg.queryTimeout()),
|
||||
TimeValue.timeValueMillis(cfg.pageTimeout()));
|
||||
TimeZone.getTimeZone("UTC"), fetchSize, TimeValue.timeValueMillis(cfg.queryTimeout()),
|
||||
TimeValue.timeValueMillis(cfg.pageTimeout()));
|
||||
return query(sqlRequest);
|
||||
}
|
||||
|
||||
|
@ -84,83 +84,92 @@ public class HttpClient {
|
|||
|
||||
public boolean queryClose(String cursor) throws SQLException {
|
||||
SqlClearCursorResponse response = post(Protocol.CLEAR_CURSOR_REST_ENDPOINT,
|
||||
new SqlClearCursorRequest(Mode.PLAIN, cursor),
|
||||
SqlClearCursorResponse::fromXContent);
|
||||
new SqlClearCursorRequest(Mode.PLAIN, cursor),
|
||||
SqlClearCursorResponse::fromXContent);
|
||||
return response.isSucceeded();
|
||||
}
|
||||
|
||||
private <Request extends AbstractSqlRequest, Response> Response post(String path, Request request,
|
||||
CheckedFunction<XContentParser, Response, IOException> responseParser)
|
||||
throws SQLException {
|
||||
BytesReference requestBytes = toXContent(request);
|
||||
byte[] requestBytes = toXContent(request);
|
||||
String query = "error_trace&mode=" + request.mode();
|
||||
Tuple<XContentType, BytesReference> response =
|
||||
AccessController.doPrivileged((PrivilegedAction<ResponseOrException<Tuple<XContentType, BytesReference>>>) () ->
|
||||
Tuple<XContentType, byte[]> response =
|
||||
AccessController.doPrivileged((PrivilegedAction<ResponseOrException<Tuple<XContentType, byte[]>>>) () ->
|
||||
JreHttpUrlConnection.http(path, query, cfg, con ->
|
||||
con.request(
|
||||
requestBytes::writeTo,
|
||||
this::readFrom,
|
||||
"POST"
|
||||
)
|
||||
con.request(
|
||||
(out) -> out.write(requestBytes),
|
||||
this::readFrom,
|
||||
"POST"
|
||||
)
|
||||
)).getResponseOrThrowException();
|
||||
return fromXContent(response.v1(), response.v2(), responseParser);
|
||||
}
|
||||
|
||||
private boolean head(String path, long timeoutInMs) throws SQLException {
|
||||
ConnectionConfiguration pingCfg = new ConnectionConfiguration(cfg.baseUri(), cfg.connectionString(),
|
||||
cfg.connectTimeout(), timeoutInMs, cfg.queryTimeout(), cfg.pageTimeout(), cfg.pageSize(),
|
||||
cfg.authUser(), cfg.authPass(), cfg.sslConfig(), cfg.proxyConfig());
|
||||
cfg.connectTimeout(), timeoutInMs, cfg.queryTimeout(), cfg.pageTimeout(), cfg.pageSize(),
|
||||
cfg.authUser(), cfg.authPass(), cfg.sslConfig(), cfg.proxyConfig());
|
||||
try {
|
||||
return AccessController.doPrivileged((PrivilegedAction<Boolean>) () ->
|
||||
JreHttpUrlConnection.http(path, "error_trace", pingCfg, JreHttpUrlConnection::head));
|
||||
JreHttpUrlConnection.http(path, "error_trace", pingCfg, JreHttpUrlConnection::head));
|
||||
} catch (ClientException ex) {
|
||||
throw new SQLException("Cannot ping server", ex);
|
||||
}
|
||||
}
|
||||
|
||||
private <Response> Response get(String path, CheckedFunction<XContentParser, Response, IOException> responseParser)
|
||||
throws SQLException {
|
||||
Tuple<XContentType, BytesReference> response =
|
||||
AccessController.doPrivileged((PrivilegedAction<ResponseOrException<Tuple<XContentType, BytesReference>>>) () ->
|
||||
throws SQLException {
|
||||
Tuple<XContentType, byte[]> response =
|
||||
AccessController.doPrivileged((PrivilegedAction<ResponseOrException<Tuple<XContentType, byte[]>>>) () ->
|
||||
JreHttpUrlConnection.http(path, "error_trace", cfg, con ->
|
||||
con.request(
|
||||
null,
|
||||
this::readFrom,
|
||||
"GET"
|
||||
)
|
||||
con.request(
|
||||
null,
|
||||
this::readFrom,
|
||||
"GET"
|
||||
)
|
||||
)).getResponseOrThrowException();
|
||||
return fromXContent(response.v1(), response.v2(), responseParser);
|
||||
}
|
||||
|
||||
private static <Request extends ToXContent> BytesReference toXContent(Request xContent) {
|
||||
try {
|
||||
return XContentHelper.toXContent(xContent, REQUEST_BODY_CONTENT_TYPE, false);
|
||||
private static <Request extends ToXContent> byte[] toXContent(Request xContent) {
|
||||
try(ByteArrayOutputStream buffer = new ByteArrayOutputStream()) {
|
||||
try (XContentBuilder xContentBuilder = new XContentBuilder(REQUEST_BODY_CONTENT_TYPE.xContent(), buffer)) {
|
||||
if (xContent.isFragment()) {
|
||||
xContentBuilder.startObject();
|
||||
}
|
||||
xContent.toXContent(xContentBuilder, ToXContent.EMPTY_PARAMS);
|
||||
if (xContent.isFragment()) {
|
||||
xContentBuilder.endObject();
|
||||
}
|
||||
}
|
||||
return buffer.toByteArray();
|
||||
} catch (IOException ex) {
|
||||
throw new ClientException("Cannot serialize request", ex);
|
||||
}
|
||||
}
|
||||
|
||||
private Tuple<XContentType, BytesReference> readFrom(InputStream inputStream, Function<String, String> headers) {
|
||||
private Tuple<XContentType, byte[]> readFrom(InputStream inputStream, Function<String, String> headers) {
|
||||
String contentType = headers.apply("Content-Type");
|
||||
XContentType xContentType = XContentType.fromMediaTypeOrFormat(contentType);
|
||||
if (xContentType == null) {
|
||||
throw new IllegalStateException("Unsupported Content-Type: " + contentType);
|
||||
}
|
||||
BytesStreamOutput out = new BytesStreamOutput();
|
||||
ByteArrayOutputStream out = new ByteArrayOutputStream();
|
||||
try {
|
||||
Streams.copy(inputStream, out);
|
||||
} catch (IOException ex) {
|
||||
throw new ClientException("Cannot deserialize response", ex);
|
||||
}
|
||||
return new Tuple<>(xContentType, out.bytes());
|
||||
return new Tuple<>(xContentType, out.toByteArray());
|
||||
|
||||
}
|
||||
|
||||
private <Response> Response fromXContent(XContentType xContentType, BytesReference bytesReference,
|
||||
private <Response> Response fromXContent(XContentType xContentType, byte[] bytesReference,
|
||||
CheckedFunction<XContentParser, Response, IOException> responseParser) {
|
||||
try (InputStream stream = bytesReference.streamInput();
|
||||
try (InputStream stream = new ByteArrayInputStream(bytesReference);
|
||||
XContentParser parser = xContentType.xContent().createParser(registry,
|
||||
LoggingDeprecationHandler.INSTANCE, stream)) {
|
||||
DeprecationHandler.THROW_UNSUPPORTED_OPERATION, stream)) {
|
||||
return responseParser.apply(parser);
|
||||
} catch (IOException ex) {
|
||||
throw new ClientException("Cannot parse response", ex);
|
||||
|
|
|
@ -0,0 +1,35 @@
|
|||
|
||||
/*
|
||||
* This project contains XContent protocol classes shared between server and http client
|
||||
*/
|
||||
|
||||
import org.elasticsearch.gradle.precommit.PrecommitTasks
|
||||
|
||||
apply plugin: 'elasticsearch.build'
|
||||
|
||||
description = 'Request and response objects shared by the cli, jdbc ' +
|
||||
'and the Elasticsearch plugin'
|
||||
|
||||
dependencies {
|
||||
compile (project(':libs:elasticsearch-core')) {
|
||||
transitive = false
|
||||
}
|
||||
compile (project(':libs:x-content')) {
|
||||
transitive = false
|
||||
}
|
||||
compile 'joda-time:joda-time:2.9.9'
|
||||
runtime "com.fasterxml.jackson.core:jackson-core:${versions.jackson}"
|
||||
|
||||
testCompile "org.elasticsearch.test:framework:${version}"
|
||||
}
|
||||
|
||||
forbiddenApisMain {
|
||||
//sql does not depend on server, so only jdk signatures should be checked
|
||||
signaturesURLs = [PrecommitTasks.getResource('/forbidden/jdk-signatures.txt')]
|
||||
}
|
||||
|
||||
dependencyLicenses {
|
||||
mapping from: /elasticsearch-core.*/, to: 'elasticsearch'
|
||||
mapping from: /jackson-.*/, to: 'jackson'
|
||||
ignoreSha 'elasticsearch-core'
|
||||
}
|
|
@ -0,0 +1,8 @@
|
|||
This copy of Jackson JSON processor streaming parser/generator is licensed under the
|
||||
Apache (Software) License, version 2.0 ("the License").
|
||||
See the License for details about distribution rights, and the
|
||||
specific rights regarding derivate works.
|
||||
|
||||
You may obtain a copy of the License at:
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
|
@ -0,0 +1,20 @@
|
|||
# Jackson JSON processor
|
||||
|
||||
Jackson is a high-performance, Free/Open Source JSON processing library.
|
||||
It was originally written by Tatu Saloranta (tatu.saloranta@iki.fi), and has
|
||||
been in development since 2007.
|
||||
It is currently developed by a community of developers, as well as supported
|
||||
commercially by FasterXML.com.
|
||||
|
||||
## Licensing
|
||||
|
||||
Jackson core and extension components may licensed under different licenses.
|
||||
To find the details that apply to this artifact see the accompanying LICENSE file.
|
||||
For more information, including possible other licensing options, contact
|
||||
FasterXML.com (http://fasterxml.com).
|
||||
|
||||
## Credits
|
||||
|
||||
A list of contributors may be found from CREDITS file, which is included
|
||||
in some artifacts (usually source distributions); but is always available
|
||||
from the source code management (SCM) system project uses.
|
|
@ -0,0 +1 @@
|
|||
eb21a035c66ad307e66ec8fce37f5d50fd62d039
|
|
@ -0,0 +1 @@
|
|||
f7b520c458572890807d143670c9b24f4de90897
|
|
@ -0,0 +1,202 @@
|
|||
|
||||
Apache License
|
||||
Version 2.0, January 2004
|
||||
http://www.apache.org/licenses/
|
||||
|
||||
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||
|
||||
1. Definitions.
|
||||
|
||||
"License" shall mean the terms and conditions for use, reproduction,
|
||||
and distribution as defined by Sections 1 through 9 of this document.
|
||||
|
||||
"Licensor" shall mean the copyright owner or entity authorized by
|
||||
the copyright owner that is granting the License.
|
||||
|
||||
"Legal Entity" shall mean the union of the acting entity and all
|
||||
other entities that control, are controlled by, or are under common
|
||||
control with that entity. For the purposes of this definition,
|
||||
"control" means (i) the power, direct or indirect, to cause the
|
||||
direction or management of such entity, whether by contract or
|
||||
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||
|
||||
"You" (or "Your") shall mean an individual or Legal Entity
|
||||
exercising permissions granted by this License.
|
||||
|
||||
"Source" form shall mean the preferred form for making modifications,
|
||||
including but not limited to software source code, documentation
|
||||
source, and configuration files.
|
||||
|
||||
"Object" form shall mean any form resulting from mechanical
|
||||
transformation or translation of a Source form, including but
|
||||
not limited to compiled object code, generated documentation,
|
||||
and conversions to other media types.
|
||||
|
||||
"Work" shall mean the work of authorship, whether in Source or
|
||||
Object form, made available under the License, as indicated by a
|
||||
copyright notice that is included in or attached to the work
|
||||
(an example is provided in the Appendix below).
|
||||
|
||||
"Derivative Works" shall mean any work, whether in Source or Object
|
||||
form, that is based on (or derived from) the Work and for which the
|
||||
editorial revisions, annotations, elaborations, or other modifications
|
||||
represent, as a whole, an original work of authorship. For the purposes
|
||||
of this License, Derivative Works shall not include works that remain
|
||||
separable from, or merely link (or bind by name) to the interfaces of,
|
||||
the Work and Derivative Works thereof.
|
||||
|
||||
"Contribution" shall mean any work of authorship, including
|
||||
the original version of the Work and any modifications or additions
|
||||
to that Work or Derivative Works thereof, that is intentionally
|
||||
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||
or by an individual or Legal Entity authorized to submit on behalf of
|
||||
the copyright owner. For the purposes of this definition, "submitted"
|
||||
means any form of electronic, verbal, or written communication sent
|
||||
to the Licensor or its representatives, including but not limited to
|
||||
communication on electronic mailing lists, source code control systems,
|
||||
and issue tracking systems that are managed by, or on behalf of, the
|
||||
Licensor for the purpose of discussing and improving the Work, but
|
||||
excluding communication that is conspicuously marked or otherwise
|
||||
designated in writing by the copyright owner as "Not a Contribution."
|
||||
|
||||
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||
on behalf of whom a Contribution has been received by Licensor and
|
||||
subsequently incorporated within the Work.
|
||||
|
||||
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
copyright license to reproduce, prepare Derivative Works of,
|
||||
publicly display, publicly perform, sublicense, and distribute the
|
||||
Work and such Derivative Works in Source or Object form.
|
||||
|
||||
3. Grant of Patent License. Subject to the terms and conditions of
|
||||
this License, each Contributor hereby grants to You a perpetual,
|
||||
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||
(except as stated in this section) patent license to make, have made,
|
||||
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||
where such license applies only to those patent claims licensable
|
||||
by such Contributor that are necessarily infringed by their
|
||||
Contribution(s) alone or by combination of their Contribution(s)
|
||||
with the Work to which such Contribution(s) was submitted. If You
|
||||
institute patent litigation against any entity (including a
|
||||
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||
or a Contribution incorporated within the Work constitutes direct
|
||||
or contributory patent infringement, then any patent licenses
|
||||
granted to You under this License for that Work shall terminate
|
||||
as of the date such litigation is filed.
|
||||
|
||||
4. Redistribution. You may reproduce and distribute copies of the
|
||||
Work or Derivative Works thereof in any medium, with or without
|
||||
modifications, and in Source or Object form, provided that You
|
||||
meet the following conditions:
|
||||
|
||||
(a) You must give any other recipients of the Work or
|
||||
Derivative Works a copy of this License; and
|
||||
|
||||
(b) You must cause any modified files to carry prominent notices
|
||||
stating that You changed the files; and
|
||||
|
||||
(c) You must retain, in the Source form of any Derivative Works
|
||||
that You distribute, all copyright, patent, trademark, and
|
||||
attribution notices from the Source form of the Work,
|
||||
excluding those notices that do not pertain to any part of
|
||||
the Derivative Works; and
|
||||
|
||||
(d) If the Work includes a "NOTICE" text file as part of its
|
||||
distribution, then any Derivative Works that You distribute must
|
||||
include a readable copy of the attribution notices contained
|
||||
within such NOTICE file, excluding those notices that do not
|
||||
pertain to any part of the Derivative Works, in at least one
|
||||
of the following places: within a NOTICE text file distributed
|
||||
as part of the Derivative Works; within the Source form or
|
||||
documentation, if provided along with the Derivative Works; or,
|
||||
within a display generated by the Derivative Works, if and
|
||||
wherever such third-party notices normally appear. The contents
|
||||
of the NOTICE file are for informational purposes only and
|
||||
do not modify the License. You may add Your own attribution
|
||||
notices within Derivative Works that You distribute, alongside
|
||||
or as an addendum to the NOTICE text from the Work, provided
|
||||
that such additional attribution notices cannot be construed
|
||||
as modifying the License.
|
||||
|
||||
You may add Your own copyright statement to Your modifications and
|
||||
may provide additional or different license terms and conditions
|
||||
for use, reproduction, or distribution of Your modifications, or
|
||||
for any such Derivative Works as a whole, provided Your use,
|
||||
reproduction, and distribution of the Work otherwise complies with
|
||||
the conditions stated in this License.
|
||||
|
||||
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||
any Contribution intentionally submitted for inclusion in the Work
|
||||
by You to the Licensor shall be under the terms and conditions of
|
||||
this License, without any additional terms or conditions.
|
||||
Notwithstanding the above, nothing herein shall supersede or modify
|
||||
the terms of any separate license agreement you may have executed
|
||||
with Licensor regarding such Contributions.
|
||||
|
||||
6. Trademarks. This License does not grant permission to use the trade
|
||||
names, trademarks, service marks, or product names of the Licensor,
|
||||
except as required for reasonable and customary use in describing the
|
||||
origin of the Work and reproducing the content of the NOTICE file.
|
||||
|
||||
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||
agreed to in writing, Licensor provides the Work (and each
|
||||
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
implied, including, without limitation, any warranties or conditions
|
||||
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||
appropriateness of using or redistributing the Work and assume any
|
||||
risks associated with Your exercise of permissions under this License.
|
||||
|
||||
8. Limitation of Liability. In no event and under no legal theory,
|
||||
whether in tort (including negligence), contract, or otherwise,
|
||||
unless required by applicable law (such as deliberate and grossly
|
||||
negligent acts) or agreed to in writing, shall any Contributor be
|
||||
liable to You for damages, including any direct, indirect, special,
|
||||
incidental, or consequential damages of any character arising as a
|
||||
result of this License or out of the use or inability to use the
|
||||
Work (including but not limited to damages for loss of goodwill,
|
||||
work stoppage, computer failure or malfunction, or any and all
|
||||
other commercial damages or losses), even if such Contributor
|
||||
has been advised of the possibility of such damages.
|
||||
|
||||
9. Accepting Warranty or Additional Liability. While redistributing
|
||||
the Work or Derivative Works thereof, You may choose to offer,
|
||||
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||
or other liability obligations and/or rights consistent with this
|
||||
License. However, in accepting such obligations, You may act only
|
||||
on Your own behalf and on Your sole responsibility, not on behalf
|
||||
of any other Contributor, and only if You agree to indemnify,
|
||||
defend, and hold each Contributor harmless for any liability
|
||||
incurred by, or claims asserted against, such Contributor by reason
|
||||
of your accepting any such warranty or additional liability.
|
||||
|
||||
END OF TERMS AND CONDITIONS
|
||||
|
||||
APPENDIX: How to apply the Apache License to your work.
|
||||
|
||||
To apply the Apache License to your work, attach the following
|
||||
boilerplate notice, with the fields enclosed by brackets "[]"
|
||||
replaced with your own identifying information. (Don't include
|
||||
the brackets!) The text should be enclosed in the appropriate
|
||||
comment syntax for the file format. We also recommend that a
|
||||
file or class name and description of purpose be included on the
|
||||
same "printed page" as the copyright notice for easier
|
||||
identification within third-party archives.
|
||||
|
||||
Copyright [yyyy] [name of copyright owner]
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
|
@ -0,0 +1,5 @@
|
|||
=============================================================================
|
||||
= NOTICE file corresponding to section 4d of the Apache License Version 2.0 =
|
||||
=============================================================================
|
||||
This product includes software developed by
|
||||
Joda.org (http://www.joda.org/).
|
|
@ -7,7 +7,6 @@ package org.elasticsearch.xpack.sql.proto;
|
|||
|
||||
import org.elasticsearch.common.Nullable;
|
||||
import org.elasticsearch.common.ParseField;
|
||||
import org.elasticsearch.common.Strings;
|
||||
import org.elasticsearch.common.xcontent.ConstructingObjectParser;
|
||||
import org.elasticsearch.common.xcontent.ToXContentObject;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
|
@ -74,7 +73,7 @@ public class ColumnInfo implements ToXContentObject {
|
|||
@Override
|
||||
public XContentBuilder toXContent(XContentBuilder builder, Params params) throws IOException {
|
||||
builder.startObject();
|
||||
if (Strings.hasText(table)) {
|
||||
if (table != null && table.isEmpty() == false) {
|
||||
builder.field("table", table);
|
||||
}
|
||||
builder.field("name", name);
|
||||
|
@ -146,6 +145,6 @@ public class ColumnInfo implements ToXContentObject {
|
|||
|
||||
@Override
|
||||
public String toString() {
|
||||
return Strings.toString(this);
|
||||
return ProtoUtils.toString(this);
|
||||
}
|
||||
}
|
|
@ -6,7 +6,6 @@
|
|||
|
||||
package org.elasticsearch.xpack.sql.proto;
|
||||
|
||||
import org.elasticsearch.Build;
|
||||
import org.elasticsearch.common.ParseField;
|
||||
import org.elasticsearch.common.xcontent.ObjectParser;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
|
@ -21,18 +20,16 @@ public class MainResponse {
|
|||
private String version;
|
||||
private String clusterName;
|
||||
private String clusterUuid;
|
||||
// TODO: Add parser for Build
|
||||
private Build build;
|
||||
|
||||
private MainResponse() {
|
||||
}
|
||||
|
||||
public MainResponse(String nodeName, String version, String clusterName, String clusterUuid, Build build) {
|
||||
public MainResponse(String nodeName, String version, String clusterName, String clusterUuid) {
|
||||
this.nodeName = nodeName;
|
||||
this.version = version;
|
||||
this.clusterName = clusterName;
|
||||
this.clusterUuid = clusterUuid;
|
||||
this.build = build;
|
||||
|
||||
}
|
||||
|
||||
public String getNodeName() {
|
||||
|
@ -51,10 +48,6 @@ public class MainResponse {
|
|||
return clusterUuid;
|
||||
}
|
||||
|
||||
public Build getBuild() {
|
||||
return build;
|
||||
}
|
||||
|
||||
private static final ObjectParser<MainResponse, Void> PARSER = new ObjectParser<>(MainResponse.class.getName(), true,
|
||||
MainResponse::new);
|
||||
|
||||
|
@ -65,15 +58,6 @@ public class MainResponse {
|
|||
PARSER.declareString((response, value) -> {
|
||||
}, new ParseField("tagline"));
|
||||
PARSER.declareObject((response, value) -> {
|
||||
final String buildFlavor = (String) value.get("build_flavor");
|
||||
final String buildType = (String) value.get("build_type");
|
||||
response.build =
|
||||
new Build(
|
||||
buildFlavor == null ? Build.Flavor.UNKNOWN : Build.Flavor.fromDisplayName(buildFlavor),
|
||||
buildType == null ? Build.Type.UNKNOWN : Build.Type.fromDisplayName(buildType),
|
||||
(String) value.get("build_hash"),
|
||||
(String) value.get("build_date"),
|
||||
(boolean) value.get("build_snapshot"));
|
||||
response.version = (String) value.get("number");
|
||||
}, (parser, context) -> parser.map(), new ParseField("version"));
|
||||
}
|
||||
|
@ -94,12 +78,11 @@ public class MainResponse {
|
|||
return Objects.equals(nodeName, other.nodeName) &&
|
||||
Objects.equals(version, other.version) &&
|
||||
Objects.equals(clusterUuid, other.clusterUuid) &&
|
||||
Objects.equals(build, other.build) &&
|
||||
Objects.equals(clusterName, other.clusterName);
|
||||
}
|
||||
|
||||
@Override
|
||||
public int hashCode() {
|
||||
return Objects.hash(nodeName, version, clusterUuid, build, clusterName);
|
||||
return Objects.hash(nodeName, version, clusterUuid, clusterName);
|
||||
}
|
||||
}
|
|
@ -0,0 +1,85 @@
|
|||
/*
|
||||
* Copyright Elasticsearch B.V. and/or licensed to Elasticsearch B.V. under one
|
||||
* or more contributor license agreements. Licensed under the Elastic License;
|
||||
* you may not use this file except in compliance with the Elastic License.
|
||||
*/
|
||||
|
||||
package org.elasticsearch.xpack.sql.proto;
|
||||
|
||||
import org.elasticsearch.common.xcontent.ToXContent;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.common.xcontent.json.JsonXContent;
|
||||
|
||||
import java.io.ByteArrayOutputStream;
|
||||
import java.io.IOException;
|
||||
import java.nio.charset.StandardCharsets;
|
||||
import java.util.Locale;
|
||||
|
||||
public final class ProtoUtils {
|
||||
|
||||
private ProtoUtils() {
|
||||
|
||||
}
|
||||
|
||||
/**
|
||||
* Parses a generic value from the XContent stream
|
||||
*/
|
||||
public static Object parseFieldsValue(XContentParser parser) throws IOException {
|
||||
XContentParser.Token token = parser.currentToken();
|
||||
if (token == XContentParser.Token.VALUE_STRING) {
|
||||
//binary values will be parsed back and returned as base64 strings when reading from json and yaml
|
||||
return parser.text();
|
||||
} else if (token == XContentParser.Token.VALUE_NUMBER) {
|
||||
return parser.numberValue();
|
||||
} else if (token == XContentParser.Token.VALUE_BOOLEAN) {
|
||||
return parser.booleanValue();
|
||||
} else if (token == XContentParser.Token.VALUE_NULL) {
|
||||
return null;
|
||||
} else if (token == XContentParser.Token.START_OBJECT) {
|
||||
return parser.mapOrdered();
|
||||
} else if (token == XContentParser.Token.START_ARRAY) {
|
||||
return parser.listOrderedMap();
|
||||
} else {
|
||||
String message = "Failed to parse object: unexpected token [%s] found";
|
||||
throw new IllegalStateException(String.format(Locale.ROOT, message, token));
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns a string representation of the builder (only applicable for text based xcontent).
|
||||
*
|
||||
* @param xContentBuilder builder containing an object to converted to a string
|
||||
*/
|
||||
public static String toString(XContentBuilder xContentBuilder) {
|
||||
byte[] byteArray = ((ByteArrayOutputStream) xContentBuilder.getOutputStream()).toByteArray();
|
||||
return new String(byteArray, StandardCharsets.UTF_8);
|
||||
}
|
||||
|
||||
public static String toString(ToXContent toXContent) {
|
||||
try {
|
||||
XContentBuilder builder = JsonXContent.contentBuilder();
|
||||
if (toXContent.isFragment()) {
|
||||
builder.startObject();
|
||||
}
|
||||
toXContent.toXContent(builder, ToXContent.EMPTY_PARAMS);
|
||||
if (toXContent.isFragment()) {
|
||||
builder.endObject();
|
||||
}
|
||||
builder.close();
|
||||
return toString(builder);
|
||||
} catch (IOException e) {
|
||||
try {
|
||||
XContentBuilder builder = JsonXContent.contentBuilder();
|
||||
builder.startObject();
|
||||
builder.field("error", "error building toString out of XContent: " + e.getMessage());
|
||||
builder.endObject();
|
||||
builder.close();
|
||||
return toString(builder);
|
||||
} catch (IOException e2) {
|
||||
throw new IllegalArgumentException("cannot generate error message for deserialization", e);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
}
|
|
@ -18,7 +18,7 @@ import java.util.Objects;
|
|||
|
||||
import static org.elasticsearch.common.xcontent.ConstructingObjectParser.constructorArg;
|
||||
import static org.elasticsearch.common.xcontent.ConstructingObjectParser.optionalConstructorArg;
|
||||
import static org.elasticsearch.common.xcontent.XContentParserUtils.parseFieldsValue;
|
||||
import static org.elasticsearch.xpack.sql.proto.ProtoUtils.parseFieldsValue;
|
||||
|
||||
/**
|
||||
* Response to perform an sql query for JDBC/CLI client
|
|
@ -11,13 +11,13 @@ import org.elasticsearch.common.xcontent.ObjectParser;
|
|||
import org.elasticsearch.common.xcontent.ToXContentObject;
|
||||
import org.elasticsearch.common.xcontent.XContentBuilder;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.common.xcontent.XContentParserUtils;
|
||||
import org.elasticsearch.xpack.sql.type.DataType;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Objects;
|
||||
|
||||
import static org.elasticsearch.common.xcontent.ConstructingObjectParser.constructorArg;
|
||||
import static org.elasticsearch.xpack.sql.proto.ProtoUtils.parseFieldsValue;
|
||||
|
||||
/**
|
||||
* Represent a strongly typed parameter value
|
||||
|
@ -33,7 +33,7 @@ public class SqlTypedParamValue implements ToXContentObject {
|
|||
private static final ParseField TYPE = new ParseField("type");
|
||||
|
||||
static {
|
||||
PARSER.declareField(constructorArg(), (p, c) -> XContentParserUtils.parseFieldsValue(p), VALUE, ObjectParser.ValueType.VALUE);
|
||||
PARSER.declareField(constructorArg(), (p, c) -> parseFieldsValue(p), VALUE, ObjectParser.ValueType.VALUE);
|
||||
PARSER.declareString(constructorArg(), TYPE);
|
||||
}
|
||||
|
|
@ -0,0 +1,59 @@
|
|||
/*
|
||||
* Copyright Elasticsearch B.V. and/or licensed to Elasticsearch B.V. under one
|
||||
* or more contributor license agreements. Licensed under the Elastic License;
|
||||
* you may not use this file except in compliance with the Elastic License.
|
||||
*/
|
||||
|
||||
package org.elasticsearch.xpack.sql.proto;
|
||||
|
||||
import org.elasticsearch.common.xcontent.DeprecationHandler;
|
||||
import org.elasticsearch.common.xcontent.NamedXContentRegistry;
|
||||
import org.elasticsearch.common.xcontent.XContentParser;
|
||||
import org.elasticsearch.common.xcontent.json.JsonXContent;
|
||||
import org.elasticsearch.test.ESTestCase;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
public class ProtoUtilsTests extends ESTestCase {
|
||||
|
||||
public void testGenericValueParsing() throws IOException {
|
||||
|
||||
String json = ProtoUtils.toString((builder, params) -> {
|
||||
builder.field("int", 42);
|
||||
builder.field("double", 42.5);
|
||||
builder.field("string", "foobar");
|
||||
builder.nullField("null");
|
||||
return builder;
|
||||
});
|
||||
|
||||
XContentParser parser =
|
||||
JsonXContent.jsonXContent.createParser(NamedXContentRegistry.EMPTY, DeprecationHandler.THROW_UNSUPPORTED_OPERATION, json);
|
||||
|
||||
assertEquals(XContentParser.Token.START_OBJECT, parser.nextToken());
|
||||
while (parser.nextToken() != XContentParser.Token.END_OBJECT) {
|
||||
assertEquals(XContentParser.Token.FIELD_NAME, parser.currentToken());
|
||||
String fieldName = parser.currentName();
|
||||
parser.nextToken();
|
||||
Object val = ProtoUtils.parseFieldsValue(parser);
|
||||
switch (fieldName) {
|
||||
case "int":
|
||||
assertEquals(42, val);
|
||||
break;
|
||||
case "double":
|
||||
assertEquals(42.5, val);
|
||||
break;
|
||||
case "string":
|
||||
assertEquals("foobar", val);
|
||||
break;
|
||||
case "null":
|
||||
assertNull(val);
|
||||
break;
|
||||
default:
|
||||
fail("Unexpected value " + fieldName);
|
||||
}
|
||||
}
|
||||
assertNull(parser.nextToken());
|
||||
|
||||
}
|
||||
|
||||
}
|
|
@ -96,6 +96,9 @@ subprojects {
|
|||
|
||||
// CLI testing dependencies
|
||||
testRuntime project(path: xpackModule('sql:sql-cli'), configuration: 'nodeps')
|
||||
testRuntime (xpackProject('plugin:sql:sql-proto')) {
|
||||
transitive = false
|
||||
}
|
||||
testRuntime "org.jline:jline:3.6.0"
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue