HADOOP-18022. Add restrict-imports-enforcer-rule for Guava Preconditions and remove remaining usages (#3712)
Reviewed-by: Akira Ajisaka <aajisaka@apache.org> Signed-off-by: Takanobu Asanuma <tasanuma@apache.org>
This commit is contained in:
parent
86705eda82
commit
215388beea
|
@ -24,7 +24,7 @@ import java.util.List;
|
|||
import java.util.concurrent.atomic.AtomicBoolean;
|
||||
import java.util.concurrent.atomic.AtomicInteger;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import com.qcloud.cos.auth.AnonymousCOSCredentials;
|
||||
import com.qcloud.cos.auth.COSCredentials;
|
||||
import com.qcloud.cos.auth.COSCredentialsProvider;
|
||||
|
|
|
@ -19,7 +19,7 @@
|
|||
package org.apache.hadoop.fs.obs;
|
||||
|
||||
import org.apache.hadoop.classification.VisibleForTesting;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.Futures;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListenableFuture;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.util.concurrent.ListeningExecutorService;
|
||||
|
|
|
@ -18,7 +18,7 @@
|
|||
|
||||
package org.apache.hadoop.fs.obs;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import com.obs.services.ObsClient;
|
||||
import com.obs.services.exception.ObsException;
|
||||
import com.obs.services.model.AbortMultipartUploadRequest;
|
||||
|
|
|
@ -19,7 +19,7 @@
|
|||
package org.apache.hadoop.fs.obs;
|
||||
|
||||
import org.apache.hadoop.classification.VisibleForTesting;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FSExceptionMessages;
|
||||
|
|
|
@ -18,7 +18,7 @@
|
|||
|
||||
package org.apache.hadoop.fs.obs;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import com.obs.services.exception.ObsException;
|
||||
|
||||
import java.io.IOException;
|
||||
|
|
|
@ -18,7 +18,7 @@
|
|||
|
||||
package org.apache.hadoop.fs.obs;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import com.obs.services.ObsClient;
|
||||
import com.obs.services.exception.ObsException;
|
||||
import com.obs.services.model.GetObjectRequest;
|
||||
|
|
|
@ -18,7 +18,7 @@
|
|||
|
||||
package org.apache.hadoop.fs.obs;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import com.obs.services.ObsClient;
|
||||
import com.obs.services.exception.ObsException;
|
||||
import com.obs.services.model.AbortMultipartUploadRequest;
|
||||
|
|
|
@ -146,7 +146,7 @@ import org.apache.hadoop.yarn.util.UnitsConversionUtil;
|
|||
import org.apache.hadoop.yarn.util.resource.ResourceUtils;
|
||||
|
||||
import org.apache.hadoop.classification.VisibleForTesting;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
|
|
|
@ -78,7 +78,7 @@ import org.apache.hadoop.yarn.exceptions.YarnRuntimeException;
|
|||
import org.apache.hadoop.yarn.webapp.BadRequestException;
|
||||
import org.apache.hadoop.yarn.webapp.NotFoundException;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import com.google.inject.Inject;
|
||||
|
||||
@Path("/ws/v1/mapreduce")
|
||||
|
|
|
@ -28,7 +28,7 @@ import java.util.Iterator;
|
|||
import java.util.Map;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Joiner;
|
||||
import static org.apache.hadoop.thirdparty.com.google.common.base.Preconditions.*;
|
||||
import static org.apache.hadoop.util.Preconditions.*;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.collect.AbstractIterator;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.collect.Iterators;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.collect.Maps;
|
||||
|
|
|
@ -18,7 +18,7 @@
|
|||
|
||||
package org.apache.hadoop.mapreduce.counters;
|
||||
|
||||
import static org.apache.hadoop.thirdparty.com.google.common.base.Preconditions.checkNotNull;
|
||||
import static org.apache.hadoop.util.Preconditions.checkNotNull;
|
||||
|
||||
import java.io.DataInput;
|
||||
import java.io.DataOutput;
|
||||
|
|
|
@ -37,7 +37,7 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext;
|
|||
import org.apache.hadoop.mapreduce.TaskAttemptID;
|
||||
|
||||
import org.apache.hadoop.classification.VisibleForTesting;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import org.apache.hadoop.util.DurationInfo;
|
||||
import org.apache.hadoop.util.Progressable;
|
||||
import org.slf4j.Logger;
|
||||
|
|
|
@ -21,7 +21,7 @@ package org.apache.hadoop.mapreduce.lib.output;
|
|||
import java.io.IOException;
|
||||
import java.text.NumberFormat;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.classification.InterfaceStability;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
|
|
|
@ -22,7 +22,7 @@ import java.io.IOException;
|
|||
import java.lang.reflect.Constructor;
|
||||
import java.lang.reflect.InvocationTargetException;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
|
|
|
@ -20,7 +20,7 @@ package org.apache.hadoop.mapreduce.lib.output;
|
|||
|
||||
import java.io.IOException;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
|
|
|
@ -23,7 +23,7 @@ import java.nio.ByteBuffer;
|
|||
import org.apache.hadoop.classification.InterfaceAudience;
|
||||
import org.apache.hadoop.mapred.nativetask.NativeDataTarget;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
|
||||
/**
|
||||
* DataOutputStream implementation which buffers data in a fixed-size
|
||||
|
|
|
@ -19,7 +19,7 @@ package org.apache.hadoop.mapred.nativetask.testutil;
|
|||
|
||||
import java.util.Random;
|
||||
|
||||
import org.apache.hadoop.thirdparty.com.google.common.base.Preconditions;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.primitives.Ints;
|
||||
import org.apache.hadoop.thirdparty.com.google.common.primitives.Longs;
|
||||
|
||||
|
|
|
@ -22,8 +22,7 @@ import org.apache.hadoop.fs.BlockLocation;
|
|||
import org.apache.hadoop.fs.EtagSource;
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.LocatedFileStatus;
|
||||
|
||||
import static org.apache.hadoop.thirdparty.com.google.common.base.Preconditions.checkNotNull;
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
|
||||
/**
|
||||
* {@link LocatedFileStatus} extended to also carry an ETag.
|
||||
|
@ -38,7 +37,7 @@ public class AbfsLocatedFileStatus extends LocatedFileStatus implements EtagSour
|
|||
private final String etag;
|
||||
|
||||
public AbfsLocatedFileStatus(FileStatus status, BlockLocation[] locations) {
|
||||
super(checkNotNull(status), locations);
|
||||
super(Preconditions.checkNotNull(status), locations);
|
||||
if (status instanceof EtagSource) {
|
||||
this.etag = ((EtagSource) status).getEtag();
|
||||
} else {
|
||||
|
|
|
@ -24,8 +24,6 @@ import org.apache.hadoop.classification.InterfaceStability.Unstable;
|
|||
import org.apache.hadoop.yarn.api.records.ApplicationId;
|
||||
import org.apache.hadoop.yarn.proto.YarnProtos.ApplicationIdProto;
|
||||
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
|
||||
@Private
|
||||
@Unstable
|
||||
public class ApplicationIdPBImpl extends ApplicationId {
|
||||
|
@ -46,24 +44,32 @@ public class ApplicationIdPBImpl extends ApplicationId {
|
|||
|
||||
@Override
|
||||
public int getId() {
|
||||
Preconditions.checkNotNull(proto);
|
||||
if (proto == null) {
|
||||
throw new NullPointerException("The argument object is NULL");
|
||||
}
|
||||
return proto.getId();
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void setId(int id) {
|
||||
Preconditions.checkNotNull(builder);
|
||||
if (builder == null) {
|
||||
throw new NullPointerException("The argument object is NULL");
|
||||
}
|
||||
builder.setId(id);
|
||||
}
|
||||
@Override
|
||||
public long getClusterTimestamp() {
|
||||
Preconditions.checkNotNull(proto);
|
||||
if (proto == null) {
|
||||
throw new NullPointerException("The argument object is NULL");
|
||||
}
|
||||
return proto.getClusterTimestamp();
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void setClusterTimestamp(long clusterTimestamp) {
|
||||
Preconditions.checkNotNull(builder);
|
||||
if (builder == null) {
|
||||
throw new NullPointerException("The argument object is NULL");
|
||||
}
|
||||
builder.setClusterTimestamp((clusterTimestamp));
|
||||
}
|
||||
|
||||
|
|
|
@ -58,8 +58,6 @@ import org.apache.hadoop.yarn.server.timelineservice.storage.common.RowKeyPrefix
|
|||
import org.apache.hadoop.yarn.server.timelineservice.storage.common.TimelineStorageUtils;
|
||||
import org.apache.hadoop.yarn.webapp.BadRequestException;
|
||||
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
|
||||
/**
|
||||
* Timeline entity reader for application entities that are stored in the
|
||||
* application table.
|
||||
|
@ -336,21 +334,29 @@ class ApplicationEntityReader extends GenericEntityReader {
|
|||
|
||||
@Override
|
||||
protected void validateParams() {
|
||||
Preconditions.checkNotNull(getContext(), "context shouldn't be null");
|
||||
Preconditions.checkNotNull(
|
||||
getDataToRetrieve(), "data to retrieve shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getClusterId(),
|
||||
"clusterId shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getEntityType(),
|
||||
"entityType shouldn't be null");
|
||||
if (getContext() == null) {
|
||||
throw new NullPointerException("context shouldn't be null");
|
||||
}
|
||||
if (getDataToRetrieve() == null) {
|
||||
throw new NullPointerException("data to retrieve shouldn't be null");
|
||||
}
|
||||
if (getContext().getClusterId() == null) {
|
||||
throw new NullPointerException("clusterId shouldn't be null");
|
||||
}
|
||||
if (getContext().getEntityType() == null) {
|
||||
throw new NullPointerException("entityType shouldn't be null");
|
||||
}
|
||||
if (isSingleEntityRead()) {
|
||||
Preconditions.checkNotNull(getContext().getAppId(),
|
||||
"appId shouldn't be null");
|
||||
if (getContext().getAppId() == null) {
|
||||
throw new NullPointerException("appId shouldn't be null");
|
||||
}
|
||||
} else {
|
||||
Preconditions.checkNotNull(getContext().getUserId(),
|
||||
"userId shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getFlowName(),
|
||||
"flowName shouldn't be null");
|
||||
if (getContext().getUserId() == null) {
|
||||
throw new NullPointerException("userId shouldn't be null");
|
||||
}
|
||||
if (getContext().getFlowName() == null) {
|
||||
throw new NullPointerException("flowName shouldn't be null");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -17,7 +17,6 @@
|
|||
*/
|
||||
package org.apache.hadoop.yarn.server.timelineservice.storage.reader;
|
||||
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hbase.client.Connection;
|
||||
import org.apache.hadoop.hbase.client.Result;
|
||||
|
@ -115,11 +114,15 @@ public final class EntityTypeReader extends AbstractTimelineStorageReader {
|
|||
|
||||
@Override
|
||||
protected void validateParams() {
|
||||
Preconditions.checkNotNull(getContext(), "context shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getClusterId(),
|
||||
"clusterId shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getAppId(),
|
||||
"appId shouldn't be null");
|
||||
if (getContext() == null) {
|
||||
throw new NullPointerException("context shouldn't be null");
|
||||
}
|
||||
if (getContext().getClusterId() == null) {
|
||||
throw new NullPointerException("clusterId shouldn't be null");
|
||||
}
|
||||
if (getContext().getAppId() == null) {
|
||||
throw new NullPointerException("appId shouldn't be null");
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -46,8 +46,6 @@ import org.apache.hadoop.yarn.server.timelineservice.storage.flow.FlowActivityRo
|
|||
import org.apache.hadoop.yarn.server.timelineservice.storage.flow.FlowActivityTableRW;
|
||||
import org.apache.hadoop.yarn.webapp.BadRequestException;
|
||||
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
|
||||
/**
|
||||
* Timeline entity reader for flow activity entities that are stored in the
|
||||
* flow activity table.
|
||||
|
@ -82,8 +80,10 @@ class FlowActivityEntityReader extends TimelineEntityReader {
|
|||
|
||||
@Override
|
||||
protected void validateParams() {
|
||||
Preconditions.checkNotNull(getContext().getClusterId(),
|
||||
"clusterId shouldn't be null");
|
||||
String clusterId = getContext().getClusterId();
|
||||
if (clusterId == null) {
|
||||
throw new NullPointerException("clusterId shouldn't be null");
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
|
|
|
@ -57,8 +57,6 @@ import org.apache.hadoop.yarn.server.timelineservice.storage.flow.FlowRunRowKeyP
|
|||
import org.apache.hadoop.yarn.server.timelineservice.storage.flow.FlowRunTableRW;
|
||||
import org.apache.hadoop.yarn.webapp.BadRequestException;
|
||||
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
|
||||
/**
|
||||
* Timeline entity reader for flow run entities that are stored in the flow run
|
||||
* table.
|
||||
|
@ -86,18 +84,25 @@ class FlowRunEntityReader extends TimelineEntityReader {
|
|||
|
||||
@Override
|
||||
protected void validateParams() {
|
||||
Preconditions.checkNotNull(getContext(), "context shouldn't be null");
|
||||
Preconditions.checkNotNull(getDataToRetrieve(),
|
||||
"data to retrieve shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getClusterId(),
|
||||
"clusterId shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getUserId(),
|
||||
"userId shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getFlowName(),
|
||||
"flowName shouldn't be null");
|
||||
if (getContext() == null) {
|
||||
throw new NullPointerException("context shouldn't be null");
|
||||
}
|
||||
if (getDataToRetrieve() == null) {
|
||||
throw new NullPointerException("data to retrieve shouldn't be null");
|
||||
}
|
||||
if (getContext().getClusterId() == null) {
|
||||
throw new NullPointerException("clusterId shouldn't be null");
|
||||
}
|
||||
if (getContext().getUserId() == null) {
|
||||
throw new NullPointerException("userId shouldn't be null");
|
||||
}
|
||||
if (getContext().getFlowName() == null) {
|
||||
throw new NullPointerException("flowName shouldn't be null");
|
||||
}
|
||||
if (isSingleEntityRead()) {
|
||||
Preconditions.checkNotNull(getContext().getFlowRunId(),
|
||||
"flowRunId shouldn't be null");
|
||||
if (getContext().getFlowRunId() == null) {
|
||||
throw new NullPointerException("flowRunId shouldn't be null");
|
||||
}
|
||||
}
|
||||
EnumSet<Field> fieldsToRetrieve = getDataToRetrieve().getFieldsToRetrieve();
|
||||
if (!isSingleEntityRead() && fieldsToRetrieve != null) {
|
||||
|
|
|
@ -63,8 +63,6 @@ import org.apache.hadoop.yarn.server.timelineservice.storage.entity.EntityRowKey
|
|||
import org.apache.hadoop.yarn.server.timelineservice.storage.entity.EntityTableRW;
|
||||
import org.apache.hadoop.yarn.webapp.BadRequestException;
|
||||
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
|
||||
/**
|
||||
* Timeline entity reader for generic entities that are stored in the entity
|
||||
* table.
|
||||
|
@ -406,18 +404,25 @@ class GenericEntityReader extends TimelineEntityReader {
|
|||
|
||||
@Override
|
||||
protected void validateParams() {
|
||||
Preconditions.checkNotNull(getContext(), "context shouldn't be null");
|
||||
Preconditions.checkNotNull(getDataToRetrieve(),
|
||||
"data to retrieve shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getClusterId(),
|
||||
"clusterId shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getAppId(),
|
||||
"appId shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getEntityType(),
|
||||
"entityType shouldn't be null");
|
||||
if (getContext() == null) {
|
||||
throw new NullPointerException("context shouldn't be null");
|
||||
}
|
||||
if (getDataToRetrieve() == null) {
|
||||
throw new NullPointerException("data to retrieve shouldn't be null");
|
||||
}
|
||||
if (getContext().getClusterId() == null) {
|
||||
throw new NullPointerException("clusterId shouldn't be null");
|
||||
}
|
||||
if (getContext().getAppId() == null) {
|
||||
throw new NullPointerException("appId shouldn't be null");
|
||||
}
|
||||
if (getContext().getEntityType() == null) {
|
||||
throw new NullPointerException("entityType shouldn't be null");
|
||||
}
|
||||
if (isSingleEntityRead()) {
|
||||
Preconditions.checkNotNull(getContext().getEntityId(),
|
||||
"entityId shouldn't be null");
|
||||
if (getContext().getEntityId() == null) {
|
||||
throw new NullPointerException("entityId shouldn't be null");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -56,8 +56,6 @@ import org.apache.hadoop.yarn.server.timelineservice.storage.subapplication.SubA
|
|||
import org.apache.hadoop.yarn.server.timelineservice.storage.subapplication.SubApplicationTableRW;
|
||||
import org.apache.hadoop.yarn.webapp.BadRequestException;
|
||||
|
||||
import org.apache.hadoop.util.Preconditions;
|
||||
|
||||
class SubApplicationEntityReader extends GenericEntityReader {
|
||||
private static final SubApplicationTableRW SUB_APPLICATION_TABLE =
|
||||
new SubApplicationTableRW();
|
||||
|
@ -308,15 +306,21 @@ class SubApplicationEntityReader extends GenericEntityReader {
|
|||
|
||||
@Override
|
||||
protected void validateParams() {
|
||||
Preconditions.checkNotNull(getContext(), "context shouldn't be null");
|
||||
Preconditions.checkNotNull(getDataToRetrieve(),
|
||||
"data to retrieve shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getClusterId(),
|
||||
"clusterId shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getDoAsUser(),
|
||||
"DoAsUser shouldn't be null");
|
||||
Preconditions.checkNotNull(getContext().getEntityType(),
|
||||
"entityType shouldn't be null");
|
||||
if (getContext() == null) {
|
||||
throw new NullPointerException("context shouldn't be null");
|
||||
}
|
||||
if (getDataToRetrieve() == null) {
|
||||
throw new NullPointerException("data to retrieve shouldn't be null");
|
||||
}
|
||||
if (getContext().getClusterId() == null) {
|
||||
throw new NullPointerException("clusterId shouldn't be null");
|
||||
}
|
||||
if (getContext().getDoAsUser() == null) {
|
||||
throw new NullPointerException("DoAsUser shouldn't be null");
|
||||
}
|
||||
if (getContext().getEntityType() == null) {
|
||||
throw new NullPointerException("entityType shouldn't be null");
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
|
|
8
pom.xml
8
pom.xml
|
@ -254,6 +254,14 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 https://maven.apache.org/x
|
|||
<bannedImport>org.apache.hadoop.thirdparty.com.google.common.collect.ImmutableListMultimap</bannedImport>
|
||||
</bannedImports>
|
||||
</restrictImports>
|
||||
<restrictImports implementation="de.skuzzle.enforcer.restrictimports.rule.RestrictImports">
|
||||
<includeTestCode>true</includeTestCode>
|
||||
<reason>Use hadoop-common provided Preconditions rather than Guava provided</reason>
|
||||
<bannedImports>
|
||||
<bannedImport>org.apache.hadoop.thirdparty.com.google.common.base.Preconditions</bannedImport>
|
||||
<bannedImport>static org.apache.hadoop.thirdparty.com.google.common.base.Preconditions.**</bannedImport>
|
||||
</bannedImports>
|
||||
</restrictImports>
|
||||
</rules>
|
||||
</configuration>
|
||||
</execution>
|
||||
|
|
Loading…
Reference in New Issue