Clean codebase from empty statements (#37822)

* Remove empty statements

There are a couple of instances of undocumented empty statements all across the
code base. While they are mostly harmless, they make the code hard to read and
are potentially error-prone. Removing most of these instances and marking blocks
that look empty by intention as such.

* Change test, slightly more verbose but less confusing
This commit is contained in:
Christoph Büscher 2019-01-25 14:23:02 +01:00 committed by GitHub
parent 49073dd2f6
commit b4b4cd6ebd
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
41 changed files with 97 additions and 93 deletions

View File

@ -37,7 +37,7 @@ public class MultiTermVectorsRequest implements ToXContentObject, Validatable {
* Constructs an empty MultiTermVectorsRequest * Constructs an empty MultiTermVectorsRequest
* After that use {@code add} method to add individual {@code TermVectorsRequest} to it. * After that use {@code add} method to add individual {@code TermVectorsRequest} to it.
*/ */
public MultiTermVectorsRequest() {}; public MultiTermVectorsRequest() {}
/** /**
* Constructs a MultiTermVectorsRequest from the given document ids * Constructs a MultiTermVectorsRequest from the given document ids

View File

@ -38,7 +38,7 @@ public class DetectionRule implements ToXContentObject {
public static final ParseField CONDITIONS_FIELD = new ParseField("conditions"); public static final ParseField CONDITIONS_FIELD = new ParseField("conditions");
public static final ObjectParser<Builder, Void> PARSER = public static final ObjectParser<Builder, Void> PARSER =
new ObjectParser<>(DETECTION_RULE_FIELD.getPreferredName(), true, Builder::new);; new ObjectParser<>(DETECTION_RULE_FIELD.getPreferredName(), true, Builder::new);
static { static {
PARSER.declareStringArray(Builder::setActions, ACTIONS_FIELD); PARSER.declareStringArray(Builder::setActions, ACTIONS_FIELD);

View File

@ -297,7 +297,7 @@ public class ObjectParserTests extends ESTestCase {
enum TestEnum { enum TestEnum {
FOO, BAR FOO, BAR
}; }
public void testParseEnumFromString() throws IOException { public void testParseEnumFromString() throws IOException {
class TestStruct { class TestStruct {

View File

@ -69,7 +69,7 @@ public class KeepTypesFilterFactory extends AbstractTokenFilterFactory {
+ KeepTypesMode.EXCLUDE + "] but was [" + modeString + "]."); + KeepTypesMode.EXCLUDE + "] but was [" + modeString + "].");
} }
} }
}; }
KeepTypesFilterFactory(IndexSettings indexSettings, Environment env, String name, Settings settings) { KeepTypesFilterFactory(IndexSettings indexSettings, Environment env, String name, Settings settings) {
super(indexSettings, name, settings); super(indexSettings, name, settings);

View File

@ -37,7 +37,7 @@ public interface MetricDetail extends ToXContentObject, NamedWriteable {
innerToXContent(builder, params); innerToXContent(builder, params);
builder.endObject(); builder.endObject();
return builder.endObject(); return builder.endObject();
}; }
default String getMetricName() { default String getMetricName() {
return getWriteableName(); return getWriteableName();

View File

@ -60,7 +60,7 @@ public abstract class AbstractAsyncBulkByScrollActionScriptTestCase<
public void execute() { public void execute() {
scriptBody.accept(getCtx()); scriptBody.accept(getCtx());
} }
};; };
when(scriptService.compile(any(), eq(UpdateScript.CONTEXT))).thenReturn(factory); when(scriptService.compile(any(), eq(UpdateScript.CONTEXT))).thenReturn(factory);
AbstractAsyncBulkByScrollAction<Request> action = action(scriptService, request().setScript(mockScript(""))); AbstractAsyncBulkByScrollAction<Request> action = action(scriptService, request().setScript(mockScript("")));
RequestWrapper<?> result = action.buildScriptApplier().apply(AbstractAsyncBulkByScrollAction.wrap(index), doc); RequestWrapper<?> result = action.buildScriptApplier().apply(AbstractAsyncBulkByScrollAction.wrap(index), doc);

View File

@ -20,7 +20,6 @@
package org.elasticsearch.common.settings; package org.elasticsearch.common.settings;
import org.apache.logging.log4j.Level; import org.apache.logging.log4j.Level;
import org.elasticsearch.core.internal.io.IOUtils;
import org.apache.logging.log4j.LogManager; import org.apache.logging.log4j.LogManager;
import org.apache.lucene.util.SetOnce; import org.apache.lucene.util.SetOnce;
import org.elasticsearch.ElasticsearchParseException; import org.elasticsearch.ElasticsearchParseException;
@ -44,6 +43,7 @@ import org.elasticsearch.common.xcontent.XContentFactory;
import org.elasticsearch.common.xcontent.XContentParser; import org.elasticsearch.common.xcontent.XContentParser;
import org.elasticsearch.common.xcontent.XContentParserUtils; import org.elasticsearch.common.xcontent.XContentParserUtils;
import org.elasticsearch.common.xcontent.XContentType; import org.elasticsearch.common.xcontent.XContentType;
import org.elasticsearch.core.internal.io.IOUtils;
import java.io.IOException; import java.io.IOException;
import java.io.InputStream; import java.io.InputStream;
@ -60,12 +60,12 @@ import java.util.HashMap;
import java.util.HashSet; import java.util.HashSet;
import java.util.Iterator; import java.util.Iterator;
import java.util.List; import java.util.List;
import java.util.ListIterator;
import java.util.Map; import java.util.Map;
import java.util.NoSuchElementException; import java.util.NoSuchElementException;
import java.util.Objects; import java.util.Objects;
import java.util.Set; import java.util.Set;
import java.util.TreeMap; import java.util.TreeMap;
import java.util.ListIterator;
import java.util.concurrent.TimeUnit; import java.util.concurrent.TimeUnit;
import java.util.function.Function; import java.util.function.Function;
import java.util.function.Predicate; import java.util.function.Predicate;

View File

@ -20,6 +20,7 @@
package org.elasticsearch.common.util; package org.elasticsearch.common.util;
import com.carrotsearch.hppc.ObjectArrayList; import com.carrotsearch.hppc.ObjectArrayList;
import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.BytesRef;
import org.apache.lucene.util.BytesRefArray; import org.apache.lucene.util.BytesRefArray;
import org.apache.lucene.util.BytesRefBuilder; import org.apache.lucene.util.BytesRefBuilder;
@ -301,8 +302,8 @@ public class CollectionUtils {
public int size() { public int size() {
return in.size(); return in.size();
} }
}
};
public static void sort(final BytesRefArray bytes, final int[] indices) { public static void sort(final BytesRefArray bytes, final int[] indices) {
sort(new BytesRefBuilder(), new BytesRefBuilder(), bytes, indices); sort(new BytesRefBuilder(), new BytesRefBuilder(), bytes, indices);
} }

View File

@ -36,7 +36,7 @@ public class PreBuiltCacheFactory {
* ELASTICSEARCH Exactly one version per elasticsearch version is stored. Useful if you change an analyzer between elasticsearch * ELASTICSEARCH Exactly one version per elasticsearch version is stored. Useful if you change an analyzer between elasticsearch
* releases, when the lucene version does not change * releases, when the lucene version does not change
*/ */
public enum CachingStrategy { ONE, LUCENE, ELASTICSEARCH }; public enum CachingStrategy { ONE, LUCENE, ELASTICSEARCH }
public interface PreBuiltCache<T> { public interface PreBuiltCache<T> {

View File

@ -132,7 +132,7 @@ public final class ScoreScriptUtils {
this.originLat = origin.lat(); this.originLat = origin.lat();
this.originLon = origin.lon(); this.originLon = origin.lon();
this.offset = DistanceUnit.DEFAULT.parse(offsetStr, DistanceUnit.DEFAULT); this.offset = DistanceUnit.DEFAULT.parse(offsetStr, DistanceUnit.DEFAULT);
this.scaling = 0.5 * Math.pow(scale, 2.0) / Math.log(decay);; this.scaling = 0.5 * Math.pow(scale, 2.0) / Math.log(decay);
} }
public double decayGeoGauss(GeoPoint docValue) { public double decayGeoGauss(GeoPoint docValue) {

View File

@ -78,7 +78,7 @@ public class EarlyTerminatingCollector extends FilterCollector {
} }
} }
super.collect(doc); super.collect(doc);
}; }
}; };
} }

View File

@ -93,7 +93,7 @@ public class FieldSortBuilder extends SortBuilder<FieldSortBuilder> {
this.setNestedPath(template.getNestedPath()); this.setNestedPath(template.getNestedPath());
if (template.getNestedSort() != null) { if (template.getNestedSort() != null) {
this.setNestedSort(template.getNestedSort()); this.setNestedSort(template.getNestedSort());
}; }
} }
/** /**

View File

@ -76,7 +76,7 @@ public class RestoreSnapshotRequestTests extends AbstractWireSerializingTestCase
int count = randomInt(3) + 1; int count = randomInt(3) + 1;
for (int i = 0; i < count; ++i) { for (int i = 0; i < count; ++i) {
indexSettings.put(randomAlphaOfLengthBetween(2, 5), randomAlphaOfLengthBetween(2, 5));; indexSettings.put(randomAlphaOfLengthBetween(2, 5), randomAlphaOfLengthBetween(2, 5));
} }
instance.indexSettings(indexSettings); instance.indexSettings(indexSettings);
} }

View File

@ -83,5 +83,5 @@ public class GroupShardsIteratorTests extends ESTestCase {
shardRouting = ShardRoutingHelper.moveToStarted(shardRouting); shardRouting = ShardRoutingHelper.moveToStarted(shardRouting);
} }
return shardRouting; return shardRouting;
}; }
} }

View File

@ -34,8 +34,10 @@ import java.util.ArrayList;
import java.util.Arrays; import java.util.Arrays;
import java.util.Collections; import java.util.Collections;
import java.util.HashMap; import java.util.HashMap;
import java.util.HashSet;
import java.util.List; import java.util.List;
import java.util.Map; import java.util.Map;
import java.util.Set;
import java.util.concurrent.BrokenBarrierException; import java.util.concurrent.BrokenBarrierException;
import java.util.concurrent.CopyOnWriteArrayList; import java.util.concurrent.CopyOnWriteArrayList;
import java.util.concurrent.CountDownLatch; import java.util.concurrent.CountDownLatch;
@ -260,8 +262,10 @@ public class TaskBatcherTests extends TaskExecutorTests {
Map<Integer, TestListener> tasks = new HashMap<>(); Map<Integer, TestListener> tasks = new HashMap<>();
final int numOfTasks = randomInt(10); final int numOfTasks = randomInt(10);
final CountDownLatch latch = new CountDownLatch(numOfTasks); final CountDownLatch latch = new CountDownLatch(numOfTasks);
Set<Integer> usedKeys = new HashSet<>(numOfTasks);
for (int i = 0; i < numOfTasks; i++) { for (int i = 0; i < numOfTasks; i++) {
while (null != tasks.put(randomInt(1024), new TestListener() { int key = randomValueOtherThanMany(k -> usedKeys.contains(k), () -> randomInt(1024));
tasks.put(key, new TestListener() {
@Override @Override
public void processed(String source) { public void processed(String source) {
latch.countDown(); latch.countDown();
@ -271,8 +275,10 @@ public class TaskBatcherTests extends TaskExecutorTests {
public void onFailure(String source, Exception e) { public void onFailure(String source, Exception e) {
fail(ExceptionsHelper.detailedMessage(e)); fail(ExceptionsHelper.detailedMessage(e));
} }
})) ; });
usedKeys.add(key);
} }
assert usedKeys.size() == numOfTasks;
TestExecutor<Integer> executor = taskList -> { TestExecutor<Integer> executor = taskList -> {
assertThat(taskList.size(), equalTo(tasks.size())); assertThat(taskList.size(), equalTo(tasks.size()));

View File

@ -45,7 +45,7 @@ public class GeoDistanceTests extends ESTestCase {
GeoDistance geoDistance = randomFrom(GeoDistance.PLANE, GeoDistance.ARC); GeoDistance geoDistance = randomFrom(GeoDistance.PLANE, GeoDistance.ARC);
try (BytesStreamOutput out = new BytesStreamOutput()) { try (BytesStreamOutput out = new BytesStreamOutput()) {
geoDistance.writeTo(out); geoDistance.writeTo(out);
try (StreamInput in = out.bytes().streamInput()) {; try (StreamInput in = out.bytes().streamInput()) {
GeoDistance copy = GeoDistance.readFromStream(in); GeoDistance copy = GeoDistance.readFromStream(in);
assertEquals(copy.toString() + " vs. " + geoDistance.toString(), copy, geoDistance); assertEquals(copy.toString() + " vs. " + geoDistance.toString(), copy, geoDistance);
} }

View File

@ -19,9 +19,8 @@
package org.elasticsearch.common.geo.builders; package org.elasticsearch.common.geo.builders;
import org.locationtech.jts.geom.Coordinate;
import org.elasticsearch.common.unit.DistanceUnit; import org.elasticsearch.common.unit.DistanceUnit;
import org.locationtech.jts.geom.Coordinate;
import java.io.IOException; import java.io.IOException;
@ -59,7 +58,7 @@ public class CircleBuilderTests extends AbstractShapeBuilderTestCase<CircleBuild
DistanceUnit newRandom = unit; DistanceUnit newRandom = unit;
while (newRandom == unit) { while (newRandom == unit) {
newRandom = randomFrom(DistanceUnit.values()); newRandom = randomFrom(DistanceUnit.values());
}; }
unit = newRandom; unit = newRandom;
} }
return mutation.radius(radius, unit); return mutation.radius(radius, unit);

View File

@ -570,7 +570,7 @@ public class BytesStreamsTests extends ESTestCase {
} }
public void testReadWriteGeoPoint() throws IOException { public void testReadWriteGeoPoint() throws IOException {
try (BytesStreamOutput out = new BytesStreamOutput()) {; try (BytesStreamOutput out = new BytesStreamOutput()) {
GeoPoint geoPoint = new GeoPoint(randomDouble(), randomDouble()); GeoPoint geoPoint = new GeoPoint(randomDouble(), randomDouble());
out.writeGenericValue(geoPoint); out.writeGenericValue(geoPoint);
StreamInput wrap = out.bytes().streamInput(); StreamInput wrap = out.bytes().streamInput();

View File

@ -67,7 +67,7 @@ public class AsyncIOProcessorTests extends ESTestCase {
} catch (Exception ex) { } catch (Exception ex) {
throw new RuntimeException(ex); throw new RuntimeException(ex);
} }
}; }
}; };
thread[i].start(); thread[i].start();
} }
@ -120,7 +120,7 @@ public class AsyncIOProcessorTests extends ESTestCase {
} catch (Exception ex) { } catch (Exception ex) {
throw new RuntimeException(ex); throw new RuntimeException(ex);
} }
}; }
}; };
thread[i].start(); thread[i].start();
} }

View File

@ -392,7 +392,7 @@ public class LiveVersionMapTests extends ESTestCase {
public void testPruneTombstonesWhileLocked() throws InterruptedException, IOException { public void testPruneTombstonesWhileLocked() throws InterruptedException, IOException {
LiveVersionMap map = new LiveVersionMap(); LiveVersionMap map = new LiveVersionMap();
BytesRef uid = uid("1"); BytesRef uid = uid("1");
;
try (Releasable ignore = map.acquireLock(uid)) { try (Releasable ignore = map.acquireLock(uid)) {
map.putDeleteUnderLock(uid, new DeleteVersionValue(0, 0, 0, 0)); map.putDeleteUnderLock(uid, new DeleteVersionValue(0, 0, 0, 0));
map.beforeRefresh(); // refresh otherwise we won't prune since it's tracked by the current map map.beforeRefresh(); // refresh otherwise we won't prune since it's tracked by the current map

View File

@ -51,8 +51,8 @@ import java.util.List;
import java.util.Map; import java.util.Map;
import static java.util.Collections.singletonList; import static java.util.Collections.singletonList;
import static org.apache.lucene.analysis.BaseTokenStreamTestCase.assertTokenStreamContents;
import static java.util.Collections.singletonMap; import static java.util.Collections.singletonMap;
import static org.apache.lucene.analysis.BaseTokenStreamTestCase.assertTokenStreamContents;
import static org.hamcrest.Matchers.containsString; import static org.hamcrest.Matchers.containsString;
import static org.hamcrest.Matchers.equalTo; import static org.hamcrest.Matchers.equalTo;
import static org.hamcrest.Matchers.instanceOf; import static org.hamcrest.Matchers.instanceOf;
@ -81,7 +81,7 @@ public class KeywordFieldMapperTests extends ESSingleNodeTestCase {
}); });
} }
}; }
@Override @Override
protected Collection<Class<? extends Plugin>> getPlugins() { protected Collection<Class<? extends Plugin>> getPlugins() {

View File

@ -861,6 +861,6 @@ public class FunctionScoreTests extends ESTestCase {
@Override @Override
protected int doHashCode() { protected int doHashCode() {
return 0; return 0;
}; }
} }
} }

View File

@ -56,10 +56,10 @@ public class DeleteByQueryRequestTests extends AbstractBulkByScrollRequestTestCa
newIndices[i] = randomSimpleString(random(), 1, 30); newIndices[i] = randomSimpleString(random(), 1, 30);
} }
request.indices(newIndices); request.indices(newIndices);
for (int i = 0; i < numNewIndices; i++) {; for (int i = 0; i < numNewIndices; i++) {
assertEquals(newIndices[i], request.indices()[i]); assertEquals(newIndices[i], request.indices()[i]);
} }
for (int i = 0; i < numNewIndices; i++) {; for (int i = 0; i < numNewIndices; i++) {
assertEquals(newIndices[i], request.getSearchRequest().indices()[i]); assertEquals(newIndices[i], request.getSearchRequest().indices()[i]);
} }
} }

View File

@ -49,10 +49,10 @@ public class UpdateByQueryRequestTests extends AbstractBulkByScrollRequestTestCa
newIndices[i] = randomSimpleString(random(), 1, 30); newIndices[i] = randomSimpleString(random(), 1, 30);
} }
request.indices(newIndices); request.indices(newIndices);
for (int i = 0; i < numNewIndices; i++) {; for (int i = 0; i < numNewIndices; i++) {
assertEquals(newIndices[i], request.indices()[i]); assertEquals(newIndices[i], request.indices()[i]);
} }
for (int i = 0; i < numNewIndices; i++) {; for (int i = 0; i < numNewIndices; i++) {
assertEquals(newIndices[i], request.getSearchRequest().indices()[i]); assertEquals(newIndices[i], request.getSearchRequest().indices()[i]);
} }
} }

View File

@ -28,9 +28,6 @@ import org.elasticsearch.search.DocValueFormat;
import org.elasticsearch.search.aggregations.Aggregation.CommonFields; import org.elasticsearch.search.aggregations.Aggregation.CommonFields;
import org.elasticsearch.search.aggregations.ParsedAggregation; import org.elasticsearch.search.aggregations.ParsedAggregation;
import org.elasticsearch.search.aggregations.metrics.Percentile; import org.elasticsearch.search.aggregations.metrics.Percentile;
import org.elasticsearch.search.aggregations.pipeline.PipelineAggregator;
import org.elasticsearch.search.aggregations.pipeline.InternalPercentilesBucket;
import org.elasticsearch.search.aggregations.pipeline.ParsedPercentilesBucket;
import org.elasticsearch.test.InternalAggregationTestCase; import org.elasticsearch.test.InternalAggregationTestCase;
import java.io.IOException; import java.io.IOException;
@ -192,7 +189,6 @@ public class InternalPercentilesBucketTests extends InternalAggregationTestCase<
String name = instance.getName(); String name = instance.getName();
double[] percents = extractPercents(instance); double[] percents = extractPercents(instance);
double[] percentiles = extractPercentiles(instance); double[] percentiles = extractPercentiles(instance);
;
DocValueFormat formatter = instance.formatter(); DocValueFormat formatter = instance.formatter();
List<PipelineAggregator> pipelineAggregators = instance.pipelineAggregators(); List<PipelineAggregator> pipelineAggregators = instance.pipelineAggregators();
Map<String, Object> metaData = instance.getMetaData(); Map<String, Object> metaData = instance.getMetaData();

View File

@ -152,7 +152,7 @@ public class NestedSortBuilderTests extends ESTestCase {
@Override @Override
protected QueryBuilder doRewrite(org.elasticsearch.index.query.QueryRewriteContext queryShardContext) throws IOException { protected QueryBuilder doRewrite(org.elasticsearch.index.query.QueryRewriteContext queryShardContext) throws IOException {
return new MatchAllQueryBuilder(); return new MatchAllQueryBuilder();
}; }
}; };
// test that filter gets rewritten // test that filter gets rewritten
NestedSortBuilder original = new NestedSortBuilder("path").setFilter(filterThatRewrites); NestedSortBuilder original = new NestedSortBuilder("path").setFilter(filterThatRewrites);

View File

@ -40,7 +40,7 @@ public class EqualsHashCodeTestUtils {
*/ */
public interface CopyFunction<T> { public interface CopyFunction<T> {
T copy(T t) throws IOException; T copy(T t) throws IOException;
}; }
/** /**
* A function that creates a copy of its input argument that is different from its * A function that creates a copy of its input argument that is different from its
@ -48,7 +48,7 @@ public class EqualsHashCodeTestUtils {
*/ */
public interface MutateFunction<T> { public interface MutateFunction<T> {
T mutate(T t) throws IOException; T mutate(T t) throws IOException;
}; }
/** /**
* Perform common equality and hashCode checks on the input object * Perform common equality and hashCode checks on the input object

View File

@ -324,7 +324,7 @@ public final class FrozenEngine extends ReadOnlyEngine {
@Override @Override
public LeafReader wrap(LeafReader reader) { public LeafReader wrap(LeafReader reader) {
return new LazyLeafReader(reader); return new LazyLeafReader(reader);
}; }
}); });
this.delegate = reader; this.delegate = reader;
this.engine = engine; this.engine = engine;

View File

@ -99,7 +99,7 @@ public class DataCounts implements ToXContentObject, Writeable {
p -> TimeUtils.parseTimeField(p, LATEST_EMPTY_BUCKET_TIME.getPreferredName()), LATEST_EMPTY_BUCKET_TIME, ValueType.VALUE); p -> TimeUtils.parseTimeField(p, LATEST_EMPTY_BUCKET_TIME.getPreferredName()), LATEST_EMPTY_BUCKET_TIME, ValueType.VALUE);
PARSER.declareField(ConstructingObjectParser.optionalConstructorArg(), PARSER.declareField(ConstructingObjectParser.optionalConstructorArg(),
p -> TimeUtils.parseTimeField(p, LATEST_SPARSE_BUCKET_TIME.getPreferredName()), LATEST_SPARSE_BUCKET_TIME, ValueType.VALUE); p -> TimeUtils.parseTimeField(p, LATEST_SPARSE_BUCKET_TIME.getPreferredName()), LATEST_SPARSE_BUCKET_TIME, ValueType.VALUE);
PARSER.declareLong((t, u) -> {;}, INPUT_RECORD_COUNT); PARSER.declareLong((t, u) -> {/* intentionally empty */}, INPUT_RECORD_COUNT);
} }
public static String documentId(String jobId) { public static String documentId(String jobId) {

View File

@ -1214,7 +1214,7 @@ public class Cron implements ToXContentFragment {
private static int skipWhiteSpace(int i, String s) { private static int skipWhiteSpace(int i, String s) {
for (; i < s.length() && (s.charAt(i) == ' ' || s.charAt(i) == '\t'); i++) { for (; i < s.length() && (s.charAt(i) == ' ' || s.charAt(i) == '\t'); i++) {
; // intentionally empty
} }
return i; return i;
@ -1222,7 +1222,7 @@ public class Cron implements ToXContentFragment {
private static int findNextWhiteSpace(int i, String s) { private static int findNextWhiteSpace(int i, String s) {
for (; i < s.length() && (s.charAt(i) != ' ' || s.charAt(i) != '\t'); i++) { for (; i < s.length() && (s.charAt(i) != ' ' || s.charAt(i) != '\t'); i++) {
; // intentionally empty
} }
return i; return i;

View File

@ -167,6 +167,7 @@ public class AnalysisLimitsTests extends AbstractSerializingTestCase<AnalysisLim
new AnalysisLimits(1L, 1L); new AnalysisLimits(1L, 1L);
} }
@Override
protected AnalysisLimits mutateInstance(AnalysisLimits instance) throws IOException { protected AnalysisLimits mutateInstance(AnalysisLimits instance) throws IOException {
Long memoryModelLimit = instance.getModelMemoryLimit(); Long memoryModelLimit = instance.getModelMemoryLimit();
Long categorizationExamplesLimit = instance.getCategorizationExamplesLimit(); Long categorizationExamplesLimit = instance.getCategorizationExamplesLimit();
@ -197,5 +198,5 @@ public class AnalysisLimitsTests extends AbstractSerializingTestCase<AnalysisLim
throw new AssertionError("Illegal randomisation branch"); throw new AssertionError("Illegal randomisation branch");
} }
return new AnalysisLimits(memoryModelLimit, categorizationExamplesLimit); return new AnalysisLimits(memoryModelLimit, categorizationExamplesLimit);
}; }
} }

View File

@ -282,6 +282,7 @@ public class DataDescriptionTests extends AbstractSerializingTestCase<DataDescri
return DataDescription.STRICT_PARSER.apply(parser, null).build(); return DataDescription.STRICT_PARSER.apply(parser, null).build();
} }
@Override
protected DataDescription mutateInstance(DataDescription instance) throws java.io.IOException { protected DataDescription mutateInstance(DataDescription instance) throws java.io.IOException {
DataFormat format = instance.getFormat(); DataFormat format = instance.getFormat();
String timeField = instance.getTimeField(); String timeField = instance.getTimeField();
@ -320,5 +321,5 @@ public class DataDescriptionTests extends AbstractSerializingTestCase<DataDescri
throw new AssertionError("Illegal randomisation branch"); throw new AssertionError("Illegal randomisation branch");
} }
return new DataDescription(format, timeField, timeFormat, delimiter, quoteChar); return new DataDescription(format, timeField, timeFormat, delimiter, quoteChar);
}; }
} }

View File

@ -239,7 +239,7 @@ public class DatafeedJobsRestIT extends ESRestTestCase {
+ " \"network_bytes_out\": { \"type\":\"long\"}" + " \"network_bytes_out\": { \"type\":\"long\"}"
+ " }" + " }"
+ " }" + " }"
+ "}");; + "}");
client().performRequest(createIndexRequest); client().performRequest(createIndexRequest);
StringBuilder bulk = new StringBuilder(); StringBuilder bulk = new StringBuilder();

View File

@ -78,7 +78,7 @@ public class TransportForecastJobAction extends TransportJobTaskAction<ForecastJ
ForecastParams params = paramsBuilder.build(); ForecastParams params = paramsBuilder.build();
processManager.forecastJob(task, params, e -> { processManager.forecastJob(task, params, e -> {
if (e == null) { if (e == null) {
; getForecastRequestStats(request.getJobId(), params.getForecastId(), listener); getForecastRequestStats(request.getJobId(), params.getForecastId(), listener);
} else { } else {
listener.onFailure(e); listener.onFailure(e);
} }

View File

@ -11,7 +11,7 @@ import java.util.List;
import java.util.Objects; import java.util.Objects;
public abstract class Normalizable implements ToXContentObject { public abstract class Normalizable implements ToXContentObject {
public enum ChildType {BUCKET_INFLUENCER, RECORD}; public enum ChildType {BUCKET_INFLUENCER, RECORD}
private final String indexName; private final String indexName;
private boolean hadBigNormalizedUpdate; private boolean hadBigNormalizedUpdate;

View File

@ -745,7 +745,7 @@ public class Analyzer extends RuleExecutor<LogicalPlan> {
} }
} }
return ua; return ua;
}; }
} }
// to avoid creating duplicate functions // to avoid creating duplicate functions
@ -920,7 +920,7 @@ public class Analyzer extends RuleExecutor<LogicalPlan> {
} }
return p; return p;
} }
}; }
// //
// Handle aggs in HAVING. To help folding any aggs not found in Aggregation // Handle aggs in HAVING. To help folding any aggs not found in Aggregation

View File

@ -138,7 +138,7 @@ public class AttributeMap<E> {
public String toString() { public String toString() {
return set.toString(); return set.toString();
} }
}; }
private final Map<AttributeWrapper, E> delegate; private final Map<AttributeWrapper, E> delegate;
private Set<Attribute> keySet = null; private Set<Attribute> keySet = null;

View File

@ -57,7 +57,7 @@ public class StringProcessor implements Processor {
int i = n.intValue(); int i = n.intValue();
if (i < 0) { if (i < 0) {
return null; return null;
}; }
char[] spaces = new char[i]; char[] spaces = new char[i];
char whitespace = ' '; char whitespace = ' ';
Arrays.fill(spaces, whitespace); Arrays.fill(spaces, whitespace);

View File

@ -1951,5 +1951,5 @@ public class Optimizer extends RuleExecutor<LogicalPlan> {
enum TransformDirection { enum TransformDirection {
UP, DOWN UP, DOWN
}; }
} }

View File

@ -132,7 +132,7 @@ public class SqlParser {
log.info(format(Locale.ROOT, " %-15s '%s'", log.info(format(Locale.ROOT, " %-15s '%s'",
symbolicName == null ? literalName : symbolicName, symbolicName == null ? literalName : symbolicName,
t.getText())); t.getText()));
}; }
} }
ParserRuleContext tree = parseFunction.apply(parser); ParserRuleContext tree = parseFunction.apply(parser);

View File

@ -35,7 +35,7 @@ public final class Cursors {
private static final NamedWriteableRegistry WRITEABLE_REGISTRY = new NamedWriteableRegistry(getNamedWriteables()); private static final NamedWriteableRegistry WRITEABLE_REGISTRY = new NamedWriteableRegistry(getNamedWriteables());
private Cursors() {}; private Cursors() {}
/** /**
* The {@link NamedWriteable}s required to deserialize {@link Cursor}s. * The {@link NamedWriteable}s required to deserialize {@link Cursor}s.