mirror of https://github.com/apache/lucene.git
SOLR-14910: Use in-line tags for logger declarations in Gradle ValidateLogCalls that are non-standard, change //logok to //nowarn
This commit is contained in:
parent
7e04e4d0ca
commit
f6c4f8a755
|
@ -67,7 +67,7 @@ class ValidateLogCallsTask extends DefaultTask {
|
||||||
boolean violation = false
|
boolean violation = false
|
||||||
|
|
||||||
// If the line has been explicitly OK'd, then it's OK!
|
// If the line has been explicitly OK'd, then it's OK!
|
||||||
if (line.replaceAll("\\s", "").toLowerCase().contains("//logok")) {
|
if (line.replaceAll("\\s", "").toLowerCase().contains("//nowarn")) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// Strip all of the comments, things in quotes and the like.
|
// Strip all of the comments, things in quotes and the like.
|
||||||
|
@ -133,7 +133,7 @@ class ValidateLogCallsTask extends DefaultTask {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Always report toString(). Note, this over-reports some constructs
|
// Always report toString(). Note, this over-reports some constructs
|
||||||
// but just add //logOK if it's really OK.
|
// but just add //nowarn if it's really OK.
|
||||||
if (violation == false) {
|
if (violation == false) {
|
||||||
if (line.contains("toString(") == true && prevLineNotIf) {
|
if (line.contains("toString(") == true && prevLineNotIf) {
|
||||||
cause = "Line contains toString"
|
cause = "Line contains toString"
|
||||||
|
@ -151,27 +151,13 @@ class ValidateLogCallsTask extends DefaultTask {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Require all our logger definitions lower case "log", except a couple of special ones.
|
// Require all our logger definitions lower case "log", except if they have //nowarn
|
||||||
def checkLogName(File file, String line) {
|
def checkLogName(File file, String line) {
|
||||||
// It's many times faster to do check this way than use a regex
|
// It's many times faster to do check this way than use a regex
|
||||||
if (line.contains("static ") && line.contains("getLogger") && line.contains(" log ") == false) {
|
if (line.contains("static ") && line.contains("getLogger") && line.contains(" log ") == false) {
|
||||||
String name = file.name
|
if (line.replaceAll("\\s", "").toLowerCase().contains("//nowarn")) {
|
||||||
if (name.equals("LoggerFactory.java")) {
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
if (name.equals("SolrCore.java") && (line.contains("requestLog") || line.contains("slowLog"))) {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if (name.equals("StartupLoggingUtils.java") && line.contains("getLoggerImplStr")) {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
// Apparently the Hadoop code expectes upper-case LOG, so...
|
|
||||||
|
|
||||||
if ((name.equals("HttpServer2.java") || name.equals("BlockPoolSlice.java") || name.equals("FileUtil.java"))
|
|
||||||
&& line.contains(" LOG ")) {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
reportViolation("Change the logger name to lower-case 'log' in " + file.name + " " + line + " project" + project)
|
reportViolation("Change the logger name to lower-case 'log' in " + file.name + " " + line + " project" + project)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -60,7 +60,7 @@ NOTES:
|
||||||
simple concatenation. This last is something of a style check.
|
simple concatenation. This last is something of a style check.
|
||||||
|
|
||||||
- You can get into some pretty convolued consructs trying to pass some of these
|
- You can get into some pretty convolued consructs trying to pass some of these
|
||||||
checks. Adding //logok, with or without spaces will cause the line to pass
|
checks. Adding //nowarn, with or without spaces will cause the line to pass
|
||||||
no matter what. Please use this hack sparingly and be conscientious about
|
no matter what. Please use this hack sparingly and be conscientious about
|
||||||
surrounding with 'if (log.is*Enabled)'.
|
surrounding with 'if (log.is*Enabled)'.
|
||||||
|
|
||||||
|
|
|
@ -67,7 +67,7 @@ public abstract class StrategyTestCase extends SpatialTestCase {
|
||||||
protected boolean storeShape = true;
|
protected boolean storeShape = true;
|
||||||
|
|
||||||
protected void executeQueries(SpatialMatchConcern concern, String... testQueryFile) throws IOException {
|
protected void executeQueries(SpatialMatchConcern concern, String... testQueryFile) throws IOException {
|
||||||
log.info("testing queried for strategy "+strategy); // logOk
|
log.info("testing queried for strategy "+strategy); // nowarn
|
||||||
for( String path : testQueryFile ) {
|
for( String path : testQueryFile ) {
|
||||||
Iterator<SpatialTestQuery> testQueryIterator = getTestQueries(path, ctx);
|
Iterator<SpatialTestQuery> testQueryIterator = getTestQueries(path, ctx);
|
||||||
runTestQueries(testQueryIterator, concern);
|
runTestQueries(testQueryIterator, concern);
|
||||||
|
|
|
@ -67,7 +67,7 @@ public class HeatmapFacetCounterTest extends StrategyTestCase {
|
||||||
|
|
||||||
@After
|
@After
|
||||||
public void after() {
|
public void after() {
|
||||||
log.info("Validated " + cellsValidated + " cells, " + cellValidatedNonZero + " non-zero"); // logOK
|
log.info("Validated " + cellsValidated + " cells, " + cellValidatedNonZero + " non-zero"); // nowarn
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
|
|
|
@ -87,7 +87,7 @@ public class RandomSpatialOpFuzzyPrefixTreeTest extends StrategyTestCase {
|
||||||
((PrefixTreeStrategy) strategy).setPointsOnly(true);
|
((PrefixTreeStrategy) strategy).setPointsOnly(true);
|
||||||
}
|
}
|
||||||
|
|
||||||
log.info("Strategy: " + strategy.toString()); // logOk
|
log.info("Strategy: " + strategy.toString()); // nowarn
|
||||||
}
|
}
|
||||||
|
|
||||||
private void setupCtx2D(SpatialContext ctx) {
|
private void setupCtx2D(SpatialContext ctx) {
|
||||||
|
|
|
@ -294,6 +294,9 @@ Other Changes
|
||||||
* SOLR-12987: Deprecated plugins/features are now logged once and with log category org.apache.solr.DEPRECATED
|
* SOLR-12987: Deprecated plugins/features are now logged once and with log category org.apache.solr.DEPRECATED
|
||||||
(David Smiley)
|
(David Smiley)
|
||||||
|
|
||||||
|
* SOLR-14910: Use in-line tags for logger declarations in Gradle ValidateLogCalls that are non-standard,
|
||||||
|
change //logok to //nowarn (Erick Erickson)
|
||||||
|
|
||||||
================== 8.6.2 ==================
|
================== 8.6.2 ==================
|
||||||
|
|
||||||
Consult the LUCENE_CHANGES.txt file for additional, low level, changes in this release.
|
Consult the LUCENE_CHANGES.txt file for additional, low level, changes in this release.
|
||||||
|
|
|
@ -163,7 +163,7 @@ public class CarrotClusteringEngine extends SearchClusteringEngine {
|
||||||
if (attributeXmls.length > 0) {
|
if (attributeXmls.length > 0) {
|
||||||
if (attributeXmls.length > 1) {
|
if (attributeXmls.length > 1) {
|
||||||
log.warn("More than one attribute file found, first one will be used: {}"
|
log.warn("More than one attribute file found, first one will be used: {}"
|
||||||
, Arrays.toString(attributeXmls)); // logOk
|
, Arrays.toString(attributeXmls)); // nowarn
|
||||||
}
|
}
|
||||||
|
|
||||||
withContextClassLoader(core.getResourceLoader().getClassLoader(), () -> {
|
withContextClassLoader(core.getResourceLoader().getClassLoader(), () -> {
|
||||||
|
|
|
@ -230,7 +230,7 @@ public class ExtractingDocumentLoader extends ContentStreamLoader {
|
||||||
} catch (TikaException e) {
|
} catch (TikaException e) {
|
||||||
if(ignoreTikaException)
|
if(ignoreTikaException)
|
||||||
log.warn(new StringBuilder("skip extracting text due to ").append(e.getLocalizedMessage())
|
log.warn(new StringBuilder("skip extracting text due to ").append(e.getLocalizedMessage())
|
||||||
.append(". metadata=").append(metadata.toString()).toString()); // logOk
|
.append(". metadata=").append(metadata.toString()).toString()); // nowarn
|
||||||
else
|
else
|
||||||
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, e);
|
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, e);
|
||||||
}
|
}
|
||||||
|
|
|
@ -93,7 +93,7 @@ public class SchedulerMetricsCollector implements Closeable {
|
||||||
try {
|
try {
|
||||||
metricSamples.addAll(future.get());
|
metricSamples.addAll(future.get());
|
||||||
} catch (ExecutionException e) {
|
} catch (ExecutionException e) {
|
||||||
log.error("Error occurred during metrics collection", e.getCause());//logok
|
log.error("Error occurred during metrics collection", e.getCause());//nowarn
|
||||||
// continue any ways; do not fail
|
// continue any ways; do not fail
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -119,11 +119,11 @@ public class ActiveReplicaWatcher implements CollectionStateWatcher {
|
||||||
log.debug("-- onStateChanged@{}: replicaIds={}, solrCoreNames={} {}\ncollectionState {}"
|
log.debug("-- onStateChanged@{}: replicaIds={}, solrCoreNames={} {}\ncollectionState {}"
|
||||||
, Long.toHexString(hashCode()), replicaIds, solrCoreNames
|
, Long.toHexString(hashCode()), replicaIds, solrCoreNames
|
||||||
, (latch != null ? "\nlatch count=" + latch.getCount() : "")
|
, (latch != null ? "\nlatch count=" + latch.getCount() : "")
|
||||||
, collectionState); // logOk
|
, collectionState); // nowarn
|
||||||
}
|
}
|
||||||
if (collectionState == null) { // collection has been deleted - don't wait
|
if (collectionState == null) { // collection has been deleted - don't wait
|
||||||
if (log.isDebugEnabled()) {
|
if (log.isDebugEnabled()) {
|
||||||
log.debug("-- collection deleted, decrementing latch by {} ", replicaIds.size() + solrCoreNames.size()); // logOk
|
log.debug("-- collection deleted, decrementing latch by {} ", replicaIds.size() + solrCoreNames.size()); // nowarn
|
||||||
}
|
}
|
||||||
if (latch != null) {
|
if (latch != null) {
|
||||||
for (int i = 0; i < replicaIds.size() + solrCoreNames.size(); i++) {
|
for (int i = 0; i < replicaIds.size() + solrCoreNames.size(); i++) {
|
||||||
|
|
|
@ -93,7 +93,7 @@ public class CloudUtil {
|
||||||
}
|
}
|
||||||
log.error("{}",
|
log.error("{}",
|
||||||
new SolrException(ErrorCode.SERVER_ERROR, "Will not load SolrCore " + desc.getName()
|
new SolrException(ErrorCode.SERVER_ERROR, "Will not load SolrCore " + desc.getName()
|
||||||
+ " because it has been replaced due to failover.")); // logOk
|
+ " because it has been replaced due to failover.")); // nowarn
|
||||||
throw new SolrException(ErrorCode.SERVER_ERROR,
|
throw new SolrException(ErrorCode.SERVER_ERROR,
|
||||||
"Will not load SolrCore " + desc.getName()
|
"Will not load SolrCore " + desc.getName()
|
||||||
+ " because it has been replaced due to failover.");
|
+ " because it has been replaced due to failover.");
|
||||||
|
|
|
@ -632,9 +632,9 @@ public class OverseerTaskProcessor implements Runnable, Closeable {
|
||||||
private void printTrackingMaps() {
|
private void printTrackingMaps() {
|
||||||
if (log.isDebugEnabled()) {
|
if (log.isDebugEnabled()) {
|
||||||
log.debug("RunningTasks: {}", runningTasks);
|
log.debug("RunningTasks: {}", runningTasks);
|
||||||
log.debug("BlockedTasks: {}", blockedTasks.keySet()); // logOk
|
log.debug("BlockedTasks: {}", blockedTasks.keySet()); // nowarn
|
||||||
log.debug("CompletedTasks: {}", completedTasks.keySet()); // logOk
|
log.debug("CompletedTasks: {}", completedTasks.keySet()); // nowarn
|
||||||
log.debug("RunningZKTasks: {}", runningZKTasks); // logOk
|
log.debug("RunningZKTasks: {}", runningZKTasks); // nowarn
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -189,8 +189,8 @@ public final class SolrCore implements SolrInfoBean, Closeable {
|
||||||
public static final String version = "1.0";
|
public static final String version = "1.0";
|
||||||
|
|
||||||
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
private static final Logger requestLog = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass().getName() + ".Request");
|
private static final Logger requestLog = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass().getName() + ".Request"); //nowarn
|
||||||
private static final Logger slowLog = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass().getName() + ".SlowRequest");
|
private static final Logger slowLog = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass().getName() + ".SlowRequest"); //nowarn
|
||||||
|
|
||||||
private String name;
|
private String name;
|
||||||
private String logid; // used to show what name is set
|
private String logid; // used to show what name is set
|
||||||
|
|
|
@ -437,7 +437,7 @@ public class IndexFetcher {
|
||||||
|
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("Follower's generation: {}", commit.getGeneration());
|
log.info("Follower's generation: {}", commit.getGeneration());
|
||||||
log.info("Follower's version: {}", IndexDeletionPolicyWrapper.getCommitTimestamp(commit)); // logOK
|
log.info("Follower's version: {}", IndexDeletionPolicyWrapper.getCommitTimestamp(commit)); // nowarn
|
||||||
}
|
}
|
||||||
|
|
||||||
if (latestVersion == 0L) {
|
if (latestVersion == 0L) {
|
||||||
|
@ -1249,7 +1249,7 @@ public class IndexFetcher {
|
||||||
try {
|
try {
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("From dir files: {}", Arrays.asList(tmpIdxDir.listAll()));
|
log.info("From dir files: {}", Arrays.asList(tmpIdxDir.listAll()));
|
||||||
log.info("To dir files: {}", Arrays.asList(indexDir.listAll())); //logOk
|
log.info("To dir files: {}", Arrays.asList(indexDir.listAll())); //nowarn
|
||||||
}
|
}
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new RuntimeException(e);
|
throw new RuntimeException(e);
|
||||||
|
|
|
@ -1606,7 +1606,7 @@ public class ReplicationHandler extends RequestHandlerBase implements SolrCoreAw
|
||||||
}
|
}
|
||||||
fos.write(buf, 0, read);
|
fos.write(buf, 0, read);
|
||||||
fos.flush();
|
fos.flush();
|
||||||
log.debug("Wrote {} bytes for file {}", offset + read, fileName); // logOK
|
log.debug("Wrote {} bytes for file {}", offset + read, fileName); // nowarn
|
||||||
|
|
||||||
//Pause if necessary
|
//Pause if necessary
|
||||||
maxBytesBeforePause += read;
|
maxBytesBeforePause += read;
|
||||||
|
|
|
@ -1349,7 +1349,7 @@ public class CollectionsHandler extends RequestHandlerBase implements Permission
|
||||||
success = true;
|
success = true;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
log.warn("Force leader attempt {}. Waiting 5 secs for an active leader. State of the slice: {}", (i + 1), slice); //logok
|
log.warn("Force leader attempt {}. Waiting 5 secs for an active leader. State of the slice: {}", (i + 1), slice); //nowarn
|
||||||
}
|
}
|
||||||
|
|
||||||
if (success) {
|
if (success) {
|
||||||
|
|
|
@ -135,7 +135,7 @@ class PrepRecoveryOp implements CoreAdminHandler.CoreAdminOp {
|
||||||
", live=" + live + ", checkLive=" + checkLive + ", currentState=" + state
|
", live=" + live + ", checkLive=" + checkLive + ", currentState=" + state
|
||||||
+ ", localState=" + localState + ", nodeName=" + nodeName +
|
+ ", localState=" + localState + ", nodeName=" + nodeName +
|
||||||
", coreNodeName=" + coreNodeName + ", onlyIfActiveCheckResult=" + onlyIfActiveCheckResult
|
", coreNodeName=" + coreNodeName + ", onlyIfActiveCheckResult=" + onlyIfActiveCheckResult
|
||||||
+ ", nodeProps: " + replica); //LOGOK
|
+ ", nodeProps: " + replica); //nowarn
|
||||||
}
|
}
|
||||||
if (!onlyIfActiveCheckResult && replica != null && (state == waitForState || leaderDoesNotNeedRecovery)) {
|
if (!onlyIfActiveCheckResult && replica != null && (state == waitForState || leaderDoesNotNeedRecovery)) {
|
||||||
if (checkLive == null) {
|
if (checkLive == null) {
|
||||||
|
|
|
@ -233,7 +233,7 @@ public class HadoopAuthPlugin extends AuthenticationPlugin {
|
||||||
log.info("----------HTTP Request---------");
|
log.info("----------HTTP Request---------");
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("{} : {}", request.getMethod(), request.getRequestURI());
|
log.info("{} : {}", request.getMethod(), request.getRequestURI());
|
||||||
log.info("Query : {}", request.getQueryString()); // logOk
|
log.info("Query : {}", request.getQueryString()); // nowarn
|
||||||
}
|
}
|
||||||
log.info("Headers :");
|
log.info("Headers :");
|
||||||
Enumeration<String> headers = request.getHeaderNames();
|
Enumeration<String> headers = request.getHeaderNames();
|
||||||
|
|
|
@ -493,7 +493,7 @@ public class HttpSolrCall {
|
||||||
}
|
}
|
||||||
if (statusCode == AuthorizationResponse.FORBIDDEN.statusCode) {
|
if (statusCode == AuthorizationResponse.FORBIDDEN.statusCode) {
|
||||||
if (log.isDebugEnabled()) {
|
if (log.isDebugEnabled()) {
|
||||||
log.debug("UNAUTHORIZED auth header {} context : {}, msg: {}", req.getHeader("Authorization"), context, authResponse.getMessage()); // logOk
|
log.debug("UNAUTHORIZED auth header {} context : {}, msg: {}", req.getHeader("Authorization"), context, authResponse.getMessage()); // nowarn
|
||||||
}
|
}
|
||||||
sendError(statusCode,
|
sendError(statusCode,
|
||||||
"Unauthorized request, Response code: " + statusCode);
|
"Unauthorized request, Response code: " + statusCode);
|
||||||
|
@ -503,7 +503,7 @@ public class HttpSolrCall {
|
||||||
return RETURN;
|
return RETURN;
|
||||||
}
|
}
|
||||||
if (!(statusCode == HttpStatus.SC_ACCEPTED) && !(statusCode == HttpStatus.SC_OK)) {
|
if (!(statusCode == HttpStatus.SC_ACCEPTED) && !(statusCode == HttpStatus.SC_OK)) {
|
||||||
log.warn("ERROR {} during authentication: {}", statusCode, authResponse.getMessage()); // logOk
|
log.warn("ERROR {} during authentication: {}", statusCode, authResponse.getMessage()); // nowarn
|
||||||
sendError(statusCode,
|
sendError(statusCode,
|
||||||
"ERROR during authorization, Response code: " + statusCode);
|
"ERROR during authorization, Response code: " + statusCode);
|
||||||
if (shouldAudit(EventType.ERROR)) {
|
if (shouldAudit(EventType.ERROR)) {
|
||||||
|
|
|
@ -184,7 +184,7 @@ public class SpellCheckCollator {
|
||||||
collations.add(collation);
|
collations.add(collation);
|
||||||
}
|
}
|
||||||
if (log.isDebugEnabled()) {
|
if (log.isDebugEnabled()) {
|
||||||
log.debug("Collation: {} {}", collationQueryStr, (verifyCandidateWithQuery ? (" will return " + hits + " hits.") : "")); // logOk
|
log.debug("Collation: {} {}", collationQueryStr, (verifyCandidateWithQuery ? (" will return " + hits + " hits.") : "")); // nowarn
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return collations;
|
return collations;
|
||||||
|
|
|
@ -312,7 +312,7 @@ public class SolrIndexSplitter {
|
||||||
for (int segmentNumber = 0; segmentNumber<leaves.size(); segmentNumber++) {
|
for (int segmentNumber = 0; segmentNumber<leaves.size(); segmentNumber++) {
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("SolrIndexSplitter: partition # {} partitionCount={} {} segment #={} segmentCount={}", partitionNumber, numPieces
|
log.info("SolrIndexSplitter: partition # {} partitionCount={} {} segment #={} segmentCount={}", partitionNumber, numPieces
|
||||||
, (ranges != null ? " range=" + ranges.get(partitionNumber) : ""), segmentNumber, leaves.size()); // logOk
|
, (ranges != null ? " range=" + ranges.get(partitionNumber) : ""), segmentNumber, leaves.size()); // nowarn
|
||||||
}
|
}
|
||||||
CodecReader subReader = SlowCodecReaderWrapper.wrap(leaves.get(segmentNumber).reader());
|
CodecReader subReader = SlowCodecReaderWrapper.wrap(leaves.get(segmentNumber).reader());
|
||||||
iw.addIndexes(new LiveDocsReader(subReader, segmentDocSets.get(segmentNumber)[partitionNumber]));
|
iw.addIndexes(new LiveDocsReader(subReader, segmentDocSets.get(segmentNumber)[partitionNumber]));
|
||||||
|
|
|
@ -53,7 +53,7 @@ public final class StartupLoggingUtils {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public static String getLoggerImplStr() {
|
public static String getLoggerImplStr() { //nowarn
|
||||||
return binder.getLoggerFactoryClassStr();
|
return binder.getLoggerFactoryClassStr();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -76,7 +76,8 @@ import org.slf4j.LoggerFactory;
|
||||||
public class FileUtil {
|
public class FileUtil {
|
||||||
public static final Object SOLR_HACK_FOR_CLASS_VERIFICATION = new Object();
|
public static final Object SOLR_HACK_FOR_CLASS_VERIFICATION = new Object();
|
||||||
|
|
||||||
private static final Logger LOG = LoggerFactory.getLogger(FileUtil.class);
|
// Apparently the Hadoop code expectes upper-case LOG, so...
|
||||||
|
private static final Logger LOG = LoggerFactory.getLogger(FileUtil.class); //nowarn
|
||||||
|
|
||||||
/* The error code is defined in winutils to indicate insufficient
|
/* The error code is defined in winutils to indicate insufficient
|
||||||
* privilege to create symbolic links. This value need to keep in
|
* privilege to create symbolic links. This value need to keep in
|
||||||
|
|
|
@ -81,7 +81,8 @@ import com.google.common.annotations.VisibleForTesting;
|
||||||
public class BlockPoolSlice {
|
public class BlockPoolSlice {
|
||||||
public static final Object SOLR_HACK_FOR_CLASS_VERIFICATION = new Object();
|
public static final Object SOLR_HACK_FOR_CLASS_VERIFICATION = new Object();
|
||||||
|
|
||||||
static final Logger LOG = LoggerFactory.getLogger(BlockPoolSlice.class);
|
// Apparently the Hadoop code expectes upper-case LOG, so...
|
||||||
|
static final Logger LOG = LoggerFactory.getLogger(BlockPoolSlice.class); //nowarn
|
||||||
|
|
||||||
private final String bpid;
|
private final String bpid;
|
||||||
private final FsVolumeImpl volume; // volume to which this BlockPool belongs to
|
private final FsVolumeImpl volume; // volume to which this BlockPool belongs to
|
||||||
|
|
|
@ -118,7 +118,8 @@ import org.slf4j.LoggerFactory;
|
||||||
public final class HttpServer2 implements FilterContainer {
|
public final class HttpServer2 implements FilterContainer {
|
||||||
public static final Object SOLR_HACK_FOR_CLASS_VERIFICATION = new Object();
|
public static final Object SOLR_HACK_FOR_CLASS_VERIFICATION = new Object();
|
||||||
|
|
||||||
public static final Logger LOG = LoggerFactory.getLogger(HttpServer2.class);
|
// Apparently the Hadoop code expectes upper-case LOG, so...
|
||||||
|
public static final Logger LOG = LoggerFactory.getLogger(HttpServer2.class); //nowarn
|
||||||
|
|
||||||
public static final String HTTP_SCHEME = "http";
|
public static final String HTTP_SCHEME = "http";
|
||||||
public static final String HTTPS_SCHEME = "https";
|
public static final String HTTPS_SCHEME = "https";
|
||||||
|
|
|
@ -206,7 +206,7 @@ public class ChaosMonkeySafeLeaderWithPullReplicasTest extends AbstractFullDistr
|
||||||
|
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("control docs:{}\n\n", controlClient.query(new SolrQuery("*:*")).getResults().getNumFound());
|
log.info("control docs:{}\n\n", controlClient.query(new SolrQuery("*:*")).getResults().getNumFound());
|
||||||
log.info("collection state: {}", printClusterStateInfo(DEFAULT_COLLECTION)); // logOk
|
log.info("collection state: {}", printClusterStateInfo(DEFAULT_COLLECTION)); // nowarn
|
||||||
}
|
}
|
||||||
|
|
||||||
waitForReplicationFromReplicas(DEFAULT_COLLECTION, cloudClient.getZkStateReader(), new TimeOut(30, TimeUnit.SECONDS, TimeSource.NANO_TIME));
|
waitForReplicationFromReplicas(DEFAULT_COLLECTION, cloudClient.getZkStateReader(), new TimeOut(30, TimeUnit.SECONDS, TimeSource.NANO_TIME));
|
||||||
|
|
|
@ -99,7 +99,7 @@ public class OverseerRolesTest extends SolrCloudTestCase {
|
||||||
private void logOverseerState() throws KeeperException, InterruptedException {
|
private void logOverseerState() throws KeeperException, InterruptedException {
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("Overseer: {}", getLeaderNode(zkClient()));
|
log.info("Overseer: {}", getLeaderNode(zkClient()));
|
||||||
log.info("Election queue: {}", getSortedElectionNodes(zkClient(), "/overseer_elect/election")); // logOk
|
log.info("Election queue: {}", getSortedElectionNodes(zkClient(), "/overseer_elect/election")); // nowarn
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1171,14 +1171,14 @@ public class OverseerTest extends SolrTestCaseJ4 {
|
||||||
Snapshot snapshot = timer.getSnapshot();
|
Snapshot snapshot = timer.getSnapshot();
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("\t avgRequestsPerSecond: {}", timer.getMeanRate());
|
log.info("\t avgRequestsPerSecond: {}", timer.getMeanRate());
|
||||||
log.info("\t 5minRateRequestsPerSecond: {}", timer.getFiveMinuteRate()); // logOk
|
log.info("\t 5minRateRequestsPerSecond: {}", timer.getFiveMinuteRate()); // nowarn
|
||||||
log.info("\t 15minRateRequestsPerSecond: {}", timer.getFifteenMinuteRate()); // logOk
|
log.info("\t 15minRateRequestsPerSecond: {}", timer.getFifteenMinuteRate()); // nowarn
|
||||||
log.info("\t avgTimePerRequest: {}", nsToMs(snapshot.getMean())); // logOk
|
log.info("\t avgTimePerRequest: {}", nsToMs(snapshot.getMean())); // nowarn
|
||||||
log.info("\t medianRequestTime: {}", nsToMs(snapshot.getMedian())); // logOk
|
log.info("\t medianRequestTime: {}", nsToMs(snapshot.getMedian())); // nowarn
|
||||||
log.info("\t 75thPcRequestTime: {}", nsToMs(snapshot.get75thPercentile())); // logOk
|
log.info("\t 75thPcRequestTime: {}", nsToMs(snapshot.get75thPercentile())); // nowarn
|
||||||
log.info("\t 95thPcRequestTime: {}", nsToMs(snapshot.get95thPercentile())); // logOk
|
log.info("\t 95thPcRequestTime: {}", nsToMs(snapshot.get95thPercentile())); // nowarn
|
||||||
log.info("\t 99thPcRequestTime: {}", nsToMs(snapshot.get99thPercentile())); // logOk
|
log.info("\t 99thPcRequestTime: {}", nsToMs(snapshot.get99thPercentile())); // nowarn
|
||||||
log.info("\t 999thPcRequestTime: {}", nsToMs(snapshot.get999thPercentile())); // logOk
|
log.info("\t 999thPcRequestTime: {}", nsToMs(snapshot.get999thPercentile())); // nowarn
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -83,7 +83,7 @@ public class RollingRestartTest extends AbstractFullDistribZkTestBase {
|
||||||
boolean sawLiveDesignate = false;
|
boolean sawLiveDesignate = false;
|
||||||
int numRestarts = 1 + random().nextInt(TEST_NIGHTLY ? 12 : 2);
|
int numRestarts = 1 + random().nextInt(TEST_NIGHTLY ? 12 : 2);
|
||||||
for (int i = 0; i < numRestarts; i++) {
|
for (int i = 0; i < numRestarts; i++) {
|
||||||
log.info("Rolling restart #{}", i + 1); // logOk
|
log.info("Rolling restart #{}", i + 1); // nowarn
|
||||||
for (CloudJettyRunner cloudJetty : designateJettys) {
|
for (CloudJettyRunner cloudJetty : designateJettys) {
|
||||||
log.info("Restarting {}", cloudJetty);
|
log.info("Restarting {}", cloudJetty);
|
||||||
chaosMonkey.stopJetty(cloudJetty);
|
chaosMonkey.stopJetty(cloudJetty);
|
||||||
|
|
|
@ -752,7 +752,7 @@ public class TestReplicationHandler extends SolrTestCaseJ4 {
|
||||||
break;
|
break;
|
||||||
} catch (NumberFormatException | AssertionError notYet) {
|
} catch (NumberFormatException | AssertionError notYet) {
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("{}th attempt failure on {} details are {}", retries + 1, notYet, followerDetails); // logOk
|
log.info("{}th attempt failure on {} details are {}", retries + 1, notYet, followerDetails); // nowarn
|
||||||
}
|
}
|
||||||
if (retries>9) {
|
if (retries>9) {
|
||||||
log.error("giving up: ", notYet);
|
log.error("giving up: ", notYet);
|
||||||
|
|
|
@ -265,7 +265,7 @@ public class BasicAuthIntegrationTest extends SolrCloudAuthTestCase {
|
||||||
assertTrue(obj.containsKey("memory"));
|
assertTrue(obj.containsKey("memory"));
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
log.error("RunExampleTool failed due to: {}; stdout from tool prior to failure: {}"
|
log.error("RunExampleTool failed due to: {}; stdout from tool prior to failure: {}"
|
||||||
, e, baos.toString(StandardCharsets.UTF_8.name())); // logOk
|
, e, baos.toString(StandardCharsets.UTF_8.name())); // nowarn
|
||||||
}
|
}
|
||||||
|
|
||||||
SolrParams params = new MapSolrParams(Collections.singletonMap("q", "*:*"));
|
SolrParams params = new MapSolrParams(Collections.singletonMap("q", "*:*"));
|
||||||
|
|
|
@ -943,7 +943,7 @@ public class TestInPlaceUpdatesDistrib extends AbstractFullDistribZkTestBase {
|
||||||
|
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("Non leader 0: {}", ((HttpSolrClient) NONLEADERS.get(0)).getBaseURL());
|
log.info("Non leader 0: {}", ((HttpSolrClient) NONLEADERS.get(0)).getBaseURL());
|
||||||
log.info("Non leader 1: {}", ((HttpSolrClient) NONLEADERS.get(1)).getBaseURL()); // logOk
|
log.info("Non leader 1: {}", ((HttpSolrClient) NONLEADERS.get(1)).getBaseURL()); // nowarn
|
||||||
}
|
}
|
||||||
|
|
||||||
SolrDocument doc0 = NONLEADERS.get(0).getById(String.valueOf(0), params("distrib", "false"));
|
SolrDocument doc0 = NONLEADERS.get(0).getById(String.valueOf(0), params("distrib", "false"));
|
||||||
|
@ -1034,7 +1034,7 @@ public class TestInPlaceUpdatesDistrib extends AbstractFullDistribZkTestBase {
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("Testing client (Fetch missing test): {}", ((HttpSolrClient) client).getBaseURL());
|
log.info("Testing client (Fetch missing test): {}", ((HttpSolrClient) client).getBaseURL());
|
||||||
log.info("Version at {} is: {}"
|
log.info("Version at {} is: {}"
|
||||||
, ((HttpSolrClient) client).getBaseURL(), getReplicaValue(client, 1, "_version_")); // logOk
|
, ((HttpSolrClient) client).getBaseURL(), getReplicaValue(client, 1, "_version_")); // nowarn
|
||||||
}
|
}
|
||||||
assertReplicaValue(client, 1, "inplace_updatable_float", (newinplace_updatable_float + 2.0f),
|
assertReplicaValue(client, 1, "inplace_updatable_float", (newinplace_updatable_float + 2.0f),
|
||||||
"inplace_updatable_float didn't match for replica at client: " + ((HttpSolrClient) client).getBaseURL());
|
"inplace_updatable_float didn't match for replica at client: " + ((HttpSolrClient) client).getBaseURL());
|
||||||
|
@ -1344,7 +1344,7 @@ public class TestInPlaceUpdatesDistrib extends AbstractFullDistribZkTestBase {
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("Testing client (testDBQUsingUpdatedFieldFromDroppedUpdate): {}", ((HttpSolrClient) client).getBaseURL());
|
log.info("Testing client (testDBQUsingUpdatedFieldFromDroppedUpdate): {}", ((HttpSolrClient) client).getBaseURL());
|
||||||
log.info("Version at {} is: {}", ((HttpSolrClient) client).getBaseURL(),
|
log.info("Version at {} is: {}", ((HttpSolrClient) client).getBaseURL(),
|
||||||
getReplicaValue(client, 1, "_version_")); // logOk
|
getReplicaValue(client, 1, "_version_")); // nowarn
|
||||||
}
|
}
|
||||||
assertNull(client.getById("1", params("distrib", "false")));
|
assertNull(client.getById("1", params("distrib", "false")));
|
||||||
}
|
}
|
||||||
|
|
|
@ -76,7 +76,7 @@ public class CategoryRoutedAliasUpdateProcessorTest extends RoutedAliasUpdatePro
|
||||||
//log this to help debug potential causes of problems
|
//log this to help debug potential causes of problems
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("SolrClient: {}", solrClient);
|
log.info("SolrClient: {}", solrClient);
|
||||||
log.info("ClusterStateProvider {}", solrClient.getClusterStateProvider()); // logOk
|
log.info("ClusterStateProvider {}", solrClient.getClusterStateProvider()); // nowarn
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -74,7 +74,7 @@ public class DimensionalRoutedAliasUpdateProcessorTest extends RoutedAliasUpdate
|
||||||
//log this to help debug potential causes of problems
|
//log this to help debug potential causes of problems
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("SolrClient: {}", solrClient);
|
log.info("SolrClient: {}", solrClient);
|
||||||
log.info("ClusterStateProvider {}", solrClient.getClusterStateProvider()); // logOk
|
log.info("ClusterStateProvider {}", solrClient.getClusterStateProvider()); // nowarn
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -91,7 +91,7 @@ public class TimeRoutedAliasUpdateProcessorTest extends RoutedAliasUpdateProcess
|
||||||
//log this to help debug potential causes of problems
|
//log this to help debug potential causes of problems
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("SolrClient: {}", solrClient);
|
log.info("SolrClient: {}", solrClient);
|
||||||
log.info("ClusterStateProvider {}", solrClient.getClusterStateProvider()); // logOk
|
log.info("ClusterStateProvider {}", solrClient.getClusterStateProvider()); // nowarn
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -359,7 +359,7 @@ public class TestSolrCLIRunExample extends SolrTestCaseJ4 {
|
||||||
assertEquals("it should be ok "+tool+" "+Arrays.toString(toolArgs),0, status);
|
assertEquals("it should be ok "+tool+" "+Arrays.toString(toolArgs),0, status);
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
log.error("RunExampleTool failed due to: {}; stdout from tool prior to failure: {}"
|
log.error("RunExampleTool failed due to: {}; stdout from tool prior to failure: {}"
|
||||||
, e , baos.toString(StandardCharsets.UTF_8.name())); // logOk
|
, e , baos.toString(StandardCharsets.UTF_8.name())); // nowarn
|
||||||
throw e;
|
throw e;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -154,10 +154,10 @@ public class NodesSysPropsCacher implements SolrCloseable {
|
||||||
Thread.sleep(backOffTime);
|
Thread.sleep(backOffTime);
|
||||||
} catch (InterruptedException e1) {
|
} catch (InterruptedException e1) {
|
||||||
Thread.currentThread().interrupt();
|
Thread.currentThread().interrupt();
|
||||||
log.info("Exception on caching node:{} system.properties:{}, retry {}/{}", node, tags, i+1, NUM_RETRY, e); // logOk
|
log.info("Exception on caching node:{} system.properties:{}, retry {}/{}", node, tags, i+1, NUM_RETRY, e); // nowarn
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
log.info("Exception on caching node:{} system.properties:{}, retry {}/{}", node, tags, i+1, NUM_RETRY, e); // logOk
|
log.info("Exception on caching node:{} system.properties:{}, retry {}/{}", node, tags, i+1, NUM_RETRY, e); // nowarn
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -800,7 +800,7 @@ public class Utils {
|
||||||
int statusCode = rsp.getStatusLine().getStatusCode();
|
int statusCode = rsp.getStatusLine().getStatusCode();
|
||||||
if (statusCode != 200) {
|
if (statusCode != 200) {
|
||||||
try {
|
try {
|
||||||
log.error("Failed a request to: {}, status: {}, body: {}", url, rsp.getStatusLine(), EntityUtils.toString(rsp.getEntity(), StandardCharsets.UTF_8)); // logOk
|
log.error("Failed a request to: {}, status: {}, body: {}", url, rsp.getStatusLine(), EntityUtils.toString(rsp.getEntity(), StandardCharsets.UTF_8)); // nowarn
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
log.error("could not print error", e);
|
log.error("could not print error", e);
|
||||||
}
|
}
|
||||||
|
|
|
@ -431,7 +431,7 @@ public abstract class AbstractFullDistribZkTestBase extends AbstractDistribZkTes
|
||||||
if (useTlogReplicas()) {
|
if (useTlogReplicas()) {
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("create jetty {} in directory {} of type {} in shard {}"
|
log.info("create jetty {} in directory {} of type {} in shard {}"
|
||||||
, i, jettyDir, Replica.Type.TLOG, ((currentI % sliceCount) + 1)); // logOk
|
, i, jettyDir, Replica.Type.TLOG, ((currentI % sliceCount) + 1)); // nowarn
|
||||||
}
|
}
|
||||||
customThreadPool.submit(() -> {
|
customThreadPool.submit(() -> {
|
||||||
try {
|
try {
|
||||||
|
@ -463,7 +463,7 @@ public abstract class AbstractFullDistribZkTestBase extends AbstractDistribZkTes
|
||||||
} else {
|
} else {
|
||||||
if (log.isInfoEnabled()) {
|
if (log.isInfoEnabled()) {
|
||||||
log.info("create jetty {} in directory {} of type {} for shard{}"
|
log.info("create jetty {} in directory {} of type {} for shard{}"
|
||||||
, i, jettyDir, Replica.Type.NRT, ((currentI % sliceCount) + 1)); // logOk
|
, i, jettyDir, Replica.Type.NRT, ((currentI % sliceCount) + 1)); // nowarn
|
||||||
}
|
}
|
||||||
|
|
||||||
customThreadPool.submit(() -> {
|
customThreadPool.submit(() -> {
|
||||||
|
@ -492,7 +492,7 @@ public abstract class AbstractFullDistribZkTestBase extends AbstractDistribZkTes
|
||||||
addedReplicas++;
|
addedReplicas++;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
log.info("create jetty {} in directory {} of type {} for shard{}", i, jettyDir, Replica.Type.PULL, ((currentI % sliceCount) + 1)); // logOk
|
log.info("create jetty {} in directory {} of type {} for shard{}", i, jettyDir, Replica.Type.PULL, ((currentI % sliceCount) + 1)); // nowarn
|
||||||
customThreadPool.submit(() -> {
|
customThreadPool.submit(() -> {
|
||||||
try {
|
try {
|
||||||
JettySolrRunner j = createJetty(jettyDir, useJettyDataDir ? getDataDir(testDir + "/jetty"
|
JettySolrRunner j = createJetty(jettyDir, useJettyDataDir ? getDataDir(testDir + "/jetty"
|
||||||
|
|
Loading…
Reference in New Issue