Added a new line linter (#2875)
* Added linter to add new line Signed-off-by: Owais Kazi <owaiskazi19@gmail.com> * Fixed new lines Signed-off-by: Owais Kazi <owaiskazi19@gmail.com> * Ignore empty files Signed-off-by: Owais Kazi <owaiskazi19@gmail.com> * Updated DEVELOPER GUIDE Signed-off-by: Owais Kazi <owaiskazi19@gmail.com> * Renamed workflow file Signed-off-by: Owais Kazi <owaiskazi19@gmail.com> * Fixed failing tests Signed-off-by: Owais Kazi <owaiskazi19@gmail.com>
This commit is contained in:
parent
08e4a35839
commit
3c5d997a76
|
@ -12,4 +12,4 @@ ol-7.7
|
|||
sles-12.3 # older version used in Vagrant image
|
||||
sles-12.5
|
||||
sles-15.1
|
||||
sles-15.2
|
||||
sles-15.2
|
||||
|
|
|
@ -0,0 +1,14 @@
|
|||
name: Code Hygiene
|
||||
|
||||
on: [push, pull_request]
|
||||
|
||||
jobs:
|
||||
linelint:
|
||||
runs-on: ubuntu-latest
|
||||
name: Check if all files end in newline
|
||||
steps:
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v2
|
||||
|
||||
- name: Linelint
|
||||
uses: fernandrone/linelint@0.0.4
|
|
@ -12,4 +12,4 @@ jobs:
|
|||
- name: Delete merged branch
|
||||
uses: SvanBoxel/delete-merged-branch@main
|
||||
env:
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||
|
|
|
@ -16,4 +16,4 @@ jobs:
|
|||
args: --accept=200,403,429 --exclude-mail **/*.html **/*.md **/*.txt **/*.json --exclude-file .lychee.excludes
|
||||
fail: true
|
||||
env:
|
||||
GITHUB_TOKEN: ${{secrets.GITHUB_TOKEN}}
|
||||
GITHUB_TOKEN: ${{secrets.GITHUB_TOKEN}}
|
||||
|
|
|
@ -7,4 +7,4 @@ jobs:
|
|||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- uses: gradle/wrapper-validation-action@v1
|
||||
- uses: gradle/wrapper-validation-action@v1
|
||||
|
|
|
@ -0,0 +1,49 @@
|
|||
# 'true' will fix files
|
||||
autofix: true
|
||||
|
||||
ignore:
|
||||
- .git/
|
||||
- .gradle/
|
||||
- .idea/
|
||||
- '*.sha1'
|
||||
- '*.txt'
|
||||
- '.github/CODEOWNERS'
|
||||
- 'buildSrc/src/testKit/opensearch.build/LICENSE'
|
||||
- 'buildSrc/src/testKit/opensearch.build/NOTICE'
|
||||
- 'server/licenses/apache-log4j-extras-DEPENDENCIES'
|
||||
# Empty files
|
||||
- 'doc-tools/missing-doclet/bin/main/org/opensearch/missingdoclet/MissingDoclet.class'
|
||||
- 'buildSrc/src/integTest/resources/org/opensearch/gradle/internal/fake_git/remote/build.gradle'
|
||||
- 'buildSrc/src/integTest/resources/org/opensearch/gradle/internal/fake_git/remote/distribution/archives/oss-darwin-tar/build.gradle'
|
||||
- 'buildSrc/src/integTest/resources/org/opensearch/gradle/internal/fake_git/remote/distribution/bwc/bugfix/build.gradle'
|
||||
- 'buildSrc/src/integTest/resources/org/opensearch/gradle/internal/fake_git/remote/distribution/bwc/minor/build.gradle'
|
||||
- 'buildSrc/src/main/resources/buildSrc.marker'
|
||||
- 'buildSrc/src/testKit/opensearch-build-resources/settings.gradle'
|
||||
- 'buildSrc/src/testKit/opensearch.build/settings.gradle'
|
||||
- 'buildSrc/src/testKit/reaper/settings.gradle'
|
||||
- 'buildSrc/src/testKit/symbolic-link-preserving-tar/settings.gradle'
|
||||
- 'buildSrc/src/testKit/testingConventions/empty_test_task/.gitignore'
|
||||
- 'client/rest-high-level/src/main/resources/META-INF/services/org.opensearch.plugins.spi.NamedXContentProvider'
|
||||
- 'distribution/bwc/bugfix/build.gradle'
|
||||
- 'distribution/bwc/maintenance/build.gradle'
|
||||
- 'distribution/bwc/minor/build.gradle'
|
||||
- 'distribution/bwc/staged/build.gradle'
|
||||
- 'libs/ssl-config/src/test/resources/certs/pem-utils/empty.pem'
|
||||
- 'qa/evil-tests/src/test/resources/org/opensearch/common/logging/does_not_exist/nothing_to_see_here'
|
||||
- 'qa/os/centos-6/build.gradle'
|
||||
- 'qa/os/debian-8/build.gradle'
|
||||
- 'qa/os/oel-6/build.gradle'
|
||||
- 'qa/os/oel-7/build.gradle'
|
||||
- 'qa/os/sles-12/build.gradle'
|
||||
# Test requires no new line for these files
|
||||
- 'server/src/test/resources/org/opensearch/action/bulk/simple-bulk11.json'
|
||||
- 'server/src/test/resources/org/opensearch/action/search/simple-msearch5.json'
|
||||
|
||||
rules:
|
||||
# checks if file ends in a newline character
|
||||
end-of-file:
|
||||
# set to true to enable this rule
|
||||
enable: true
|
||||
|
||||
# if true also checks if file ends in a single newline character
|
||||
single-new-line: true
|
|
@ -48,6 +48,7 @@
|
|||
- [Distributed Framework](#distributed-framework)
|
||||
- [Submitting Changes](#submitting-changes)
|
||||
- [Backports](#backports)
|
||||
- [LineLint](#linelint)
|
||||
|
||||
# Developer Guide
|
||||
|
||||
|
@ -472,3 +473,18 @@ See [CONTRIBUTING](CONTRIBUTING.md).
|
|||
## Backports
|
||||
|
||||
The Github workflow in [`backport.yml`](.github/workflows/backport.yml) creates backport PRs automatically when the original PR with an appropriate label `backport <backport-branch-name>` is merged to main with the backport workflow run successfully on the PR. For example, if a PR on main needs to be backported to `1.x` branch, add a label `backport 1.x` to the PR and make sure the backport workflow runs on the PR along with other checks. Once this PR is merged to main, the workflow will create a backport PR to the `1.x` branch.
|
||||
|
||||
## LineLint
|
||||
A linter in [`code-hygiene.yml`](.github/workflows/code-hygiene.yml) that validates simple newline and whitespace rules in all sorts of files. It can:
|
||||
- Recursively check a directory tree for files that do not end in a newline
|
||||
- Automatically fix these files by adding a newline or trimming extra newlines.
|
||||
|
||||
Rules are defined in `.linelint.yml`.
|
||||
|
||||
Executing the binary will automatically search the local directory tree for linting errors.
|
||||
|
||||
linelint .
|
||||
|
||||
Pass a list of files or directories to limit your search.
|
||||
|
||||
linelint README.md LICENSE
|
||||
|
|
|
@ -45,4 +45,4 @@ Copyright OpenSearch Contributors. See [NOTICE](NOTICE.txt) for details.
|
|||
|
||||
OpenSearch is a registered trademark of Amazon Web Services.
|
||||
|
||||
OpenSearch includes certain Apache-licensed Elasticsearch code from Elasticsearch B.V. and other source code. Elasticsearch B.V. is not the source of that other source code. ELASTICSEARCH is a registered trademark of Elasticsearch B.V.
|
||||
OpenSearch includes certain Apache-licensed Elasticsearch code from Elasticsearch B.V. and other source code. Elasticsearch B.V. is not the source of that other source code. ELASTICSEARCH is a registered trademark of Elasticsearch B.V.
|
||||
|
|
|
@ -1,3 +1,3 @@
|
|||
## Releasing
|
||||
|
||||
This project follows [OpenSearch project branching, labelling, and releasing](https://github.com/opensearch-project/.github/blob/main/RELEASING.md).
|
||||
This project follows [OpenSearch project branching, labelling, and releasing](https://github.com/opensearch-project/.github/blob/main/RELEASING.md).
|
||||
|
|
|
@ -1,3 +1,3 @@
|
|||
## Reporting a Vulnerability
|
||||
|
||||
If you discover a potential security issue in this project we ask that you notify AWS/Amazon Security via our [vulnerability reporting page](http://aws.amazon.com/security/vulnerability-reporting/) or directly via email to aws-security@amazon.com. Please do **not** create a public GitHub issue.
|
||||
If you discover a potential security issue in this project we ask that you notify AWS/Amazon Security via our [vulnerability reporting page](http://aws.amazon.com/security/vulnerability-reporting/) or directly via email to aws-security@amazon.com. Please do **not** create a public GitHub issue.
|
||||
|
|
|
@ -3,4 +3,4 @@ encoding//src/main/java=UTF-8
|
|||
encoding//src/main/resources=UTF-8
|
||||
encoding//src/test/java=UTF-8
|
||||
encoding//src/test/resources=UTF-8
|
||||
encoding/<project>=UTF-8
|
||||
encoding/<project>=UTF-8
|
||||
|
|
|
@ -1 +1 @@
|
|||
7.4.1
|
||||
7.4.1
|
||||
|
|
|
@ -88,6 +88,3 @@ project(':valid_setup_with_base') {
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
|
|
@ -16,4 +16,4 @@ include 'all_classes_in_tasks'
|
|||
include 'not_implementing_base'
|
||||
include 'valid_setup_no_base'
|
||||
include 'valid_setup_with_base'
|
||||
include 'tests_in_main'
|
||||
include 'tests_in_main'
|
||||
|
|
|
@ -9,4 +9,4 @@
|
|||
* GitHub history for details.
|
||||
*/
|
||||
|
||||
include 'sample_jars'
|
||||
include 'sample_jars'
|
||||
|
|
|
@ -136,4 +136,3 @@ if __name__ == "__main__":
|
|||
print('WARNING: no documentation references updates for release %s' % (release_version))
|
||||
|
||||
print('*** Done.')
|
||||
|
||||
|
|
|
@ -28,4 +28,3 @@ done
|
|||
|
||||
# Return non-zero error code if any commits were missing signoff
|
||||
exit $missingSignoff
|
||||
|
||||
|
|
|
@ -7,4 +7,3 @@
|
|||
- is_true: version
|
||||
- is_true: version.number
|
||||
- match: { version.build_type: "docker" }
|
||||
|
||||
|
|
|
@ -123,4 +123,3 @@
|
|||
- match:
|
||||
$body: |
|
||||
/^(\S{5,}\n)+$/
|
||||
|
||||
|
|
|
@ -47,4 +47,3 @@ for VAR_NAME_FILE in OPENSEARCH_PASSWORD_FILE KEYSTORE_PASSWORD_FILE ; do
|
|||
unset "$VAR_NAME_FILE"
|
||||
fi
|
||||
done
|
||||
|
||||
|
|
|
@ -74,4 +74,3 @@ if defined JAVA_OPTS (
|
|||
|
||||
rem check the Java version
|
||||
%JAVA% -cp "%OPENSEARCH_CLASSPATH%" "org.opensearch.tools.java_version_checker.JavaVersionChecker" || exit /b 1
|
||||
|
||||
|
|
|
@ -8,4 +8,3 @@ version '1.0.0-SNAPSHOT'
|
|||
repositories {
|
||||
mavenCentral()
|
||||
}
|
||||
|
||||
|
|
Binary file not shown.
|
@ -40,4 +40,3 @@ dependencies {
|
|||
tasks.named('forbiddenApisMain').configure {
|
||||
replaceSignatureFiles 'jdk-signatures'
|
||||
}
|
||||
|
||||
|
|
|
@ -360,4 +360,4 @@
|
|||
"append": ""
|
||||
}
|
||||
|
||||
]
|
||||
]
|
||||
|
|
|
@ -42,4 +42,3 @@ tasks.named('forbiddenApisMain').configure {
|
|||
// TODO: Need to decide how we want to handle for forbidden signatures with the changes to core
|
||||
replaceSignatureFiles 'jdk-signatures'
|
||||
}
|
||||
|
||||
|
|
|
@ -10,4 +10,3 @@ EXIM_PROTOCOL (P=%{NOTSPACE:protocol})
|
|||
EXIM_MSG_SIZE (S=%{NUMBER:exim_msg_size})
|
||||
EXIM_HEADER_ID (id=%{NOTSPACE:exim_header_id})
|
||||
EXIM_SUBJECT (T=%{QS:exim_subject})
|
||||
|
||||
|
|
|
@ -6,4 +6,3 @@ RT_FLOW1 %{RT_FLOW_EVENT:event}: %{GREEDYDATA:close-reason}: %{IP:src-ip}/%{INT:
|
|||
RT_FLOW2 %{RT_FLOW_EVENT:event}: session created %{IP:src-ip}/%{INT:src-port}->%{IP:dst-ip}/%{INT:dst-port} %{DATA:service} %{IP:nat-src-ip}/%{INT:nat-src-port}->%{IP:nat-dst-ip}/%{INT:nat-dst-port} %{DATA:src-nat-rule-name} %{DATA:dst-nat-rule-name} %{INT:protocol-id} %{DATA:policy-name} %{DATA:from-zone} %{DATA:to-zone} %{INT:session-id} .*
|
||||
|
||||
RT_FLOW3 %{RT_FLOW_EVENT:event}: session denied %{IP:src-ip}/%{INT:src-port}->%{IP:dst-ip}/%{INT:dst-port} %{DATA:service} %{INT:protocol-id}\(\d\) %{DATA:policy-name} %{DATA:from-zone} %{DATA:to-zone} .*
|
||||
|
||||
|
|
|
@ -1,3 +1,2 @@
|
|||
# Default postgresql pg_log format pattern
|
||||
POSTGRESQL %{DATESTAMP:timestamp} %{TZ} %{DATA:user_id} %{GREEDYDATA:connection_id} %{POSINT:pid}
|
||||
|
||||
|
|
|
@ -61,4 +61,3 @@ tasks.test {
|
|||
jvmArgs += ["--add-opens", "java.base/java.security.cert=ALL-UNNAMED"]
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -34,4 +34,4 @@
|
|||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -43,4 +43,4 @@
|
|||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -68,4 +68,3 @@
|
|||
- match: { tokens.1.token: "f" }
|
||||
- match: { tokens.2.token: "g" }
|
||||
- match: { tokens.3.token: "h" }
|
||||
|
||||
|
|
|
@ -119,4 +119,3 @@
|
|||
- match: { indices.analysis.built_in_analyzers.2.name: spanish }
|
||||
- match: { indices.analysis.built_in_analyzers.2.count: 2 }
|
||||
- match: { indices.analysis.built_in_analyzers.2.index_count: 2 }
|
||||
|
||||
|
|
|
@ -76,4 +76,3 @@
|
|||
- match: { tokens.5.token: dude }
|
||||
- match: { tokens.5.position: 4 }
|
||||
- match: { tokens.5.positionLength: null }
|
||||
|
||||
|
|
|
@ -229,4 +229,3 @@ setup:
|
|||
query: bar baz
|
||||
analyzer: lower_graph_syns
|
||||
- match: { hits.total: 1 }
|
||||
|
||||
|
|
|
@ -56,4 +56,3 @@ setup:
|
|||
use_field: text_en
|
||||
max_gaps: 1
|
||||
- match: { hits.total.value: 1 }
|
||||
|
||||
|
|
|
@ -91,4 +91,3 @@ teardown:
|
|||
get:
|
||||
index: test
|
||||
id: 3
|
||||
|
||||
|
|
|
@ -43,4 +43,3 @@ restResources {
|
|||
testClusters.all {
|
||||
extraConfigFile 'ingest-user-agent/test-regexes.yml', file('src/test/test-regexes.yml')
|
||||
}
|
||||
|
||||
|
|
|
@ -1,3 +1,3 @@
|
|||
user_agent_parsers:
|
||||
- regex: '.*'
|
||||
family_replacement: 'Test'
|
||||
family_replacement: 'Test'
|
||||
|
|
|
@ -52,4 +52,3 @@ tasks.named("dependencyLicenses").configure {
|
|||
mapping from: /lucene-.*/, to: 'lucene'
|
||||
mapping from: /asm-.*/, to: 'asm'
|
||||
}
|
||||
|
||||
|
|
|
@ -124,4 +124,4 @@ ID: [_a-zA-Z] [_a-zA-Z0-9]*;
|
|||
mode AFTER_DOT;
|
||||
|
||||
DOTINTEGER: ( '0' | [1-9] [0-9]* ) -> mode(DEFAULT_MODE);
|
||||
DOTID: [_a-zA-Z] [_a-zA-Z0-9]* -> mode(DEFAULT_MODE);
|
||||
DOTID: [_a-zA-Z] [_a-zA-Z0-9]* -> mode(DEFAULT_MODE);
|
||||
|
|
|
@ -139,4 +139,3 @@ setup:
|
|||
- is_false: aggregations.placeholder.buckets.0.str_terms.buckets.1.key_as_string
|
||||
- match: { aggregations.placeholder.buckets.0.str_terms.buckets.1.doc_count: 1 }
|
||||
- match: { aggregations.placeholder.buckets.0.the_bucket_script.value: 2.0 }
|
||||
|
||||
|
|
|
@ -41,4 +41,3 @@ dependencies {
|
|||
testClusters.all {
|
||||
module ':modules:reindex'
|
||||
}
|
||||
|
||||
|
|
|
@ -75,4 +75,3 @@ testClusters.all {
|
|||
"http://snapshot.test*,http://${urlFixture.addressAndPort}"
|
||||
}, PropertyNormalization.IGNORE_VALUE
|
||||
}
|
||||
|
||||
|
|
|
@ -32,4 +32,3 @@ opensearchplugin {
|
|||
description 'Integrates OpenSearch with systemd'
|
||||
classname 'org.opensearch.systemd.SystemdPlugin'
|
||||
}
|
||||
|
||||
|
|
|
@ -18,4 +18,4 @@
|
|||
|
||||
# Apply rule status {200}=RBBI.WORD_LETTER, which is mapped
|
||||
# to <ALPHANUM> token type by DefaultICUTokenizerConfig.
|
||||
.+ {200};
|
||||
.+ {200};
|
||||
|
|
|
@ -46,4 +46,3 @@ restResources {
|
|||
tasks.named("dependencyLicenses").configure {
|
||||
mapping from: /lucene-.*/, to: 'lucene'
|
||||
}
|
||||
|
||||
|
|
|
@ -31,4 +31,3 @@
|
|||
- match: { tokens.1.token: joe }
|
||||
- match: { tokens.2.token: BLKS }
|
||||
- match: { tokens.3.token: bloggs }
|
||||
|
||||
|
|
|
@ -28,4 +28,3 @@
|
|||
|
||||
- length: { tokens: 1 }
|
||||
- match: { tokens.0.token: SPRKLF }
|
||||
|
||||
|
|
|
@ -30,4 +30,3 @@
|
|||
|
||||
- length: { tokens: 1 }
|
||||
- match: { tokens.0.token: Svarts }
|
||||
|
||||
|
|
|
@ -27,4 +27,3 @@
|
|||
|
||||
- length: { tokens: 1 }
|
||||
- match: { tokens.0.token: "645740" }
|
||||
|
||||
|
|
|
@ -47,4 +47,3 @@ restResources {
|
|||
tasks.named("dependencyLicenses").configure {
|
||||
mapping from: /lucene-.*/, to: 'lucene'
|
||||
}
|
||||
|
||||
|
|
|
@ -1 +1 @@
|
|||
europe-west1-b
|
||||
europe-west1-b
|
||||
|
|
|
@ -1 +1 @@
|
|||
metadataserver
|
||||
metadataserver
|
||||
|
|
|
@ -36,4 +36,3 @@ configure(project('painless-whitelist')) {
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -42,4 +42,3 @@ testClusters.all {
|
|||
// Adds a setting in the OpenSearch keystore before running the integration tests
|
||||
keystore 'custom.secured', 'password'
|
||||
}
|
||||
|
||||
|
|
|
@ -2,4 +2,4 @@
|
|||
custom:
|
||||
simple: foo
|
||||
list: [0, 1, 1, 2, 3, 5, 8, 13, 21]
|
||||
filtered: secret
|
||||
filtered: secret
|
||||
|
|
|
@ -56,4 +56,3 @@ javaRestTest {
|
|||
dependsOn exampleFixture
|
||||
nonInputProperties.systemProperty 'external.address', "${-> exampleFixture.addressAndPort}"
|
||||
}
|
||||
|
||||
|
|
|
@ -39,4 +39,3 @@ opensearchplugin {
|
|||
}
|
||||
|
||||
test.enabled = false
|
||||
|
||||
|
|
|
@ -2,4 +2,3 @@
|
|||
= AsciiDoc test
|
||||
|
||||
Here is a test of the asciidoc format.
|
||||
|
||||
|
|
|
@ -12,4 +12,3 @@
|
|||
|
||||
- contains: { 'nodes.$master.plugins': { name: ingest-attachment } }
|
||||
- contains: { 'nodes.$master.ingest.processors': { type: attachment } }
|
||||
|
||||
|
|
|
@ -142,4 +142,3 @@
|
|||
request_cache: false
|
||||
body: { "query" : {"match_phrase" : { "my_field" : {"query": "~MARK0", "analyzer": "whitespace"} } }, "highlight" : { "type" : "annotated", "fields" : { "my_field" : {} } } }
|
||||
- match: {_shards.failed: 0}
|
||||
|
||||
|
|
|
@ -45,4 +45,3 @@
|
|||
- do:
|
||||
snapshot.delete_repository:
|
||||
repository: test_snapshot_repository
|
||||
|
||||
|
|
|
@ -47,4 +47,3 @@
|
|||
- do:
|
||||
snapshot.delete_repository:
|
||||
repository: test_snapshot_repository
|
||||
|
||||
|
|
|
@ -183,4 +183,3 @@ thirdPartyAudit {
|
|||
'io.netty.handler.ssl.util.OpenJdkSelfSignedCertGenerator'
|
||||
)
|
||||
}
|
||||
|
||||
|
|
|
@ -35,4 +35,3 @@ testClusters.javaRestTest {
|
|||
}
|
||||
|
||||
test.enabled = false
|
||||
|
||||
|
|
|
@ -1 +1 @@
|
|||
cmd1 help
|
||||
cmd1 help
|
||||
|
|
|
@ -1 +1 @@
|
|||
tool help
|
||||
tool help
|
||||
|
|
|
@ -220,4 +220,3 @@
|
|||
# When all shards are skipped current logic returns 1 to produce a valid search result
|
||||
- match: { _shards.skipped : 1}
|
||||
- match: { _shards.failed: 0 }
|
||||
|
||||
|
|
|
@ -204,4 +204,3 @@
|
|||
tasks.get:
|
||||
wait_for_completion: true
|
||||
task_id: $task
|
||||
|
||||
|
|
|
@ -111,5 +111,3 @@
|
|||
gte: "2019-02-01T00:00+01:00"
|
||||
lte: "2019-02-01T00:00+01:00"
|
||||
- match: { hits.total: 1 }
|
||||
|
||||
|
||||
|
|
|
@ -133,5 +133,3 @@
|
|||
wait_for_completion: true
|
||||
task_id: $task_id
|
||||
- match: { task.headers.X-Opaque-Id: "Reindexing Again" }
|
||||
|
||||
|
||||
|
|
|
@ -38,4 +38,3 @@
|
|||
time_frame:
|
||||
gte: "2019-02-01T00:00+01:00"
|
||||
lte: "2019-02-01T00:00+01:00"
|
||||
|
||||
|
|
|
@ -112,4 +112,3 @@
|
|||
_id: test_id2
|
||||
pipeline: my_pipeline_1
|
||||
- f1: v2
|
||||
|
||||
|
|
|
@ -102,4 +102,3 @@
|
|||
- match: { error.processor_type: "script" }
|
||||
- match: { error.type: "script_exception" }
|
||||
- match: { error.reason: "compile error" }
|
||||
|
||||
|
|
|
@ -34,4 +34,3 @@
|
|||
id: 1
|
||||
pipeline: "my_timely_pipeline"
|
||||
body: {}
|
||||
|
||||
|
|
|
@ -1 +1 @@
|
|||
ctx.bytes_total = ctx.bytes_in + ctx.bytes_out
|
||||
ctx.bytes_total = ctx.bytes_in + ctx.bytes_out
|
||||
|
|
|
@ -49,4 +49,3 @@ public class SmokeTestPluginsClientYamlTestSuiteIT extends OpenSearchClientYamlS
|
|||
return OpenSearchClientYamlSuiteTestCase.createParameters();
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -411,5 +411,3 @@
|
|||
|
||||
|
||||
Signed-off-by: Abbas Hussain <abbas_10690@yahoo.com>
|
||||
|
||||
|
||||
|
|
|
@ -386,5 +386,3 @@
|
|||
|
||||
|
||||
Signed-off-by: Sooraj Sinha <soosinha@amazon.com>
|
||||
|
||||
|
||||
|
|
|
@ -458,4 +458,3 @@
|
|||
|
||||
|
||||
Signed-off-by: Nicholas Walter Knize <nknize@apache.org>
|
||||
|
||||
|
|
|
@ -72,5 +72,3 @@
|
|||
|
||||
|
||||
Signed-off-by: dblock <dblock@amazon.com>
|
||||
|
||||
|
||||
|
|
|
@ -1295,5 +1295,3 @@
|
|||
[Nick Knize](mailto:nknize@apache.org) - Thu, 4 Nov 2021 14:46:57 -0500
|
||||
|
||||
Signed-off-by: Nicholas Walter Knize <nknize@apache.org>
|
||||
|
||||
|
||||
|
|
|
@ -223,4 +223,3 @@
|
|||
- match: { items.0.index.status: 400 }
|
||||
- match: { items.0.index.error.type: illegal_argument_exception }
|
||||
- match: { items.0.index.error.reason: "no write index is defined for alias [test_index]. The write index may be explicitly disabled using is_write_index=false or the alias points to multiple indices without one being designated as a write index" }
|
||||
|
||||
|
|
|
@ -14,4 +14,3 @@
|
|||
index: test_index
|
||||
|
||||
- match: {count: 2}
|
||||
|
||||
|
|
|
@ -14,4 +14,3 @@
|
|||
index: test_index
|
||||
|
||||
- match: {count: 2}
|
||||
|
||||
|
|
|
@ -68,4 +68,3 @@
|
|||
|
||||
- match: { items.0.update.get._source.foo: garply }
|
||||
- is_false: items.0.update.get._source.bar
|
||||
|
||||
|
|
|
@ -131,4 +131,3 @@
|
|||
- match:
|
||||
$body: |
|
||||
/^(\S{5,}\n)+$/
|
||||
|
||||
|
|
|
@ -11,4 +11,3 @@
|
|||
local: true
|
||||
|
||||
- is_true: tasks
|
||||
|
||||
|
|
|
@ -3,4 +3,3 @@
|
|||
- do:
|
||||
cluster.remote_info: {}
|
||||
- is_true: ''
|
||||
|
||||
|
|
|
@ -104,4 +104,3 @@ teardown:
|
|||
cluster.post_voting_config_exclusions:
|
||||
node_ids: nodeId
|
||||
node_names: nodeName
|
||||
|
||||
|
|
|
@ -39,4 +39,3 @@
|
|||
get:
|
||||
index: test_1
|
||||
id: 1
|
||||
|
||||
|
|
|
@ -30,4 +30,3 @@
|
|||
index: test_1
|
||||
id: 1
|
||||
routing: 5
|
||||
|
||||
|
|
|
@ -14,4 +14,3 @@
|
|||
- match: { _index: test_1 }
|
||||
- match: { _id: '1' }
|
||||
- match: { _source: { foo: "bar" } }
|
||||
|
||||
|
|
|
@ -48,5 +48,3 @@
|
|||
- match: { fields.foo: [bar] }
|
||||
- match: { fields.count: [1] }
|
||||
- match: { _source.foo: bar }
|
||||
|
||||
|
||||
|
|
|
@ -41,4 +41,3 @@
|
|||
get:
|
||||
index: test_1
|
||||
id: 1
|
||||
|
||||
|
|
|
@ -80,4 +80,3 @@
|
|||
id: 1
|
||||
version: 1
|
||||
version_type: external_gte
|
||||
|
||||
|
|
|
@ -40,4 +40,3 @@
|
|||
get:
|
||||
index: test_1
|
||||
id: 1
|
||||
|
||||
|
|
|
@ -221,4 +221,3 @@ setup:
|
|||
catch: param
|
||||
indices.delete_alias:
|
||||
index: "test_index1"
|
||||
|
||||
|
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue