mirror of https://github.com/apache/lucene.git
SOLR-12558: solr/core (private) logger renames
This commit is contained in:
parent
abd6b07ea9
commit
9262ed7e56
|
@ -44,7 +44,7 @@ import org.slf4j.LoggerFactory;
|
||||||
*/
|
*/
|
||||||
public abstract class ConfigSetService {
|
public abstract class ConfigSetService {
|
||||||
|
|
||||||
private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
public static ConfigSetService createConfigSetService(NodeConfig nodeConfig, SolrResourceLoader loader, ZkController zkController) {
|
public static ConfigSetService createConfigSetService(NodeConfig nodeConfig, SolrResourceLoader loader, ZkController zkController) {
|
||||||
if (zkController != null)
|
if (zkController != null)
|
||||||
return new CloudConfigSetService(loader, zkController);
|
return new CloudConfigSetService(loader, zkController);
|
||||||
|
@ -228,15 +228,15 @@ public abstract class ConfigSetService {
|
||||||
try {
|
try {
|
||||||
String cachedName = cacheName(schemaFile);
|
String cachedName = cacheName(schemaFile);
|
||||||
return schemaCache.get(cachedName, () -> {
|
return schemaCache.get(cachedName, () -> {
|
||||||
logger.info("Creating new index schema for core {}", cd.getName());
|
log.info("Creating new index schema for core {}", cd.getName());
|
||||||
return IndexSchemaFactory.buildIndexSchema(cd.getSchemaName(), solrConfig);
|
return IndexSchemaFactory.buildIndexSchema(cd.getSchemaName(), solrConfig);
|
||||||
});
|
});
|
||||||
} catch (ExecutionException e) {
|
} catch (ExecutionException e) {
|
||||||
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR,
|
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR,
|
||||||
"Error creating index schema for core " + cd.getName(), e);
|
"Error creating index schema for core " + cd.getName(), e);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
logger.warn("Couldn't get last modified time for schema file {}: {}", schemaFile, e.getMessage());
|
log.warn("Couldn't get last modified time for schema file {}: {}", schemaFile, e.getMessage());
|
||||||
logger.warn("Will not use schema cache");
|
log.warn("Will not use schema cache");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return IndexSchemaFactory.buildIndexSchema(cd.getSchemaName(), solrConfig);
|
return IndexSchemaFactory.buildIndexSchema(cd.getSchemaName(), solrConfig);
|
||||||
|
|
|
@ -50,13 +50,13 @@ public class CorePropertiesLocator implements CoresLocator {
|
||||||
|
|
||||||
public static final String PROPERTIES_FILENAME = "core.properties";
|
public static final String PROPERTIES_FILENAME = "core.properties";
|
||||||
|
|
||||||
private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
|
|
||||||
private final Path rootDirectory;
|
private final Path rootDirectory;
|
||||||
|
|
||||||
public CorePropertiesLocator(Path coreDiscoveryRoot) {
|
public CorePropertiesLocator(Path coreDiscoveryRoot) {
|
||||||
this.rootDirectory = coreDiscoveryRoot;
|
this.rootDirectory = coreDiscoveryRoot;
|
||||||
logger.debug("Config-defined core root directory: {}", this.rootDirectory);
|
log.debug("Config-defined core root directory: {}", this.rootDirectory);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -92,7 +92,7 @@ public class CorePropertiesLocator implements CoresLocator {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
catch (IOException e) {
|
catch (IOException e) {
|
||||||
logger.error("Couldn't persist core properties to {}: {}", propfile, e.getMessage());
|
log.error("Couldn't persist core properties to {}: {}", propfile, e.getMessage());
|
||||||
throw new SolrException(SolrException.ErrorCode.BAD_REQUEST,
|
throw new SolrException(SolrException.ErrorCode.BAD_REQUEST,
|
||||||
"Couldn't persist core properties to " + propfile.toAbsolutePath().toString() + " : " + e.getMessage());
|
"Couldn't persist core properties to " + propfile.toAbsolutePath().toString() + " : " + e.getMessage());
|
||||||
}
|
}
|
||||||
|
@ -109,7 +109,7 @@ public class CorePropertiesLocator implements CoresLocator {
|
||||||
try {
|
try {
|
||||||
Files.deleteIfExists(propfile);
|
Files.deleteIfExists(propfile);
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
logger.warn("Couldn't delete core properties file {}: {}", propfile, e.getMessage());
|
log.warn("Couldn't delete core properties file {}: {}", propfile, e.getMessage());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -132,7 +132,7 @@ public class CorePropertiesLocator implements CoresLocator {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public List<CoreDescriptor> discover(final CoreContainer cc) {
|
public List<CoreDescriptor> discover(final CoreContainer cc) {
|
||||||
logger.debug("Looking for core definitions underneath {}", rootDirectory);
|
log.debug("Looking for core definitions underneath {}", rootDirectory);
|
||||||
final List<CoreDescriptor> cds = Lists.newArrayList();
|
final List<CoreDescriptor> cds = Lists.newArrayList();
|
||||||
try {
|
try {
|
||||||
Set<FileVisitOption> options = new HashSet<>();
|
Set<FileVisitOption> options = new HashSet<>();
|
||||||
|
@ -144,7 +144,7 @@ public class CorePropertiesLocator implements CoresLocator {
|
||||||
if (file.getFileName().toString().equals(PROPERTIES_FILENAME)) {
|
if (file.getFileName().toString().equals(PROPERTIES_FILENAME)) {
|
||||||
CoreDescriptor cd = buildCoreDescriptor(file, cc);
|
CoreDescriptor cd = buildCoreDescriptor(file, cc);
|
||||||
if (cd != null) {
|
if (cd != null) {
|
||||||
logger.debug("Found core {} in {}", cd.getName(), cd.getInstanceDir());
|
log.debug("Found core {} in {}", cd.getName(), cd.getInstanceDir());
|
||||||
cds.add(cd);
|
cds.add(cd);
|
||||||
}
|
}
|
||||||
return FileVisitResult.SKIP_SIBLINGS;
|
return FileVisitResult.SKIP_SIBLINGS;
|
||||||
|
@ -157,19 +157,19 @@ public class CorePropertiesLocator implements CoresLocator {
|
||||||
// if we get an error on the root, then fail the whole thing
|
// if we get an error on the root, then fail the whole thing
|
||||||
// otherwise, log a warning and continue to try and load other cores
|
// otherwise, log a warning and continue to try and load other cores
|
||||||
if (file.equals(rootDirectory)) {
|
if (file.equals(rootDirectory)) {
|
||||||
logger.error("Error reading core root directory {}: {}", file, exc);
|
log.error("Error reading core root directory {}: {}", file, exc);
|
||||||
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, "Error reading core root directory");
|
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, "Error reading core root directory");
|
||||||
}
|
}
|
||||||
logger.warn("Error visiting {}: {}", file, exc);
|
log.warn("Error visiting {}: {}", file, exc);
|
||||||
return FileVisitResult.CONTINUE;
|
return FileVisitResult.CONTINUE;
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, "Couldn't walk file tree under " + this.rootDirectory, e);
|
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, "Couldn't walk file tree under " + this.rootDirectory, e);
|
||||||
}
|
}
|
||||||
logger.info("Found {} core definitions underneath {}", cds.size(), rootDirectory);
|
log.info("Found {} core definitions underneath {}", cds.size(), rootDirectory);
|
||||||
if (cds.size() > 0) {
|
if (cds.size() > 0) {
|
||||||
logger.info("Cores are: {}", cds.stream().map(CoreDescriptor::getName).collect(Collectors.toList()));
|
log.info("Cores are: {}", cds.stream().map(CoreDescriptor::getName).collect(Collectors.toList()));
|
||||||
}
|
}
|
||||||
return cds;
|
return cds;
|
||||||
}
|
}
|
||||||
|
@ -190,7 +190,7 @@ public class CorePropertiesLocator implements CoresLocator {
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
catch (IOException e) {
|
catch (IOException e) {
|
||||||
logger.error("Couldn't load core descriptor from {}:{}", propertiesFile, e.toString());
|
log.error("Couldn't load core descriptor from {}:{}", propertiesFile, e.toString());
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -54,7 +54,7 @@ import org.slf4j.LoggerFactory;
|
||||||
public class GraphHandler extends RequestHandlerBase implements SolrCoreAware, PermissionNameProvider {
|
public class GraphHandler extends RequestHandlerBase implements SolrCoreAware, PermissionNameProvider {
|
||||||
|
|
||||||
private StreamFactory streamFactory = new DefaultStreamFactory();
|
private StreamFactory streamFactory = new DefaultStreamFactory();
|
||||||
private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
private String coreName;
|
private String coreName;
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -110,7 +110,7 @@ public class GraphHandler extends RequestHandlerBase implements SolrCoreAware, P
|
||||||
tupleStream = this.streamFactory.constructStream(params.get("expr"));
|
tupleStream = this.streamFactory.constructStream(params.get("expr"));
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
//Catch exceptions that occur while the stream is being created. This will include streaming expression parse rules.
|
//Catch exceptions that occur while the stream is being created. This will include streaming expression parse rules.
|
||||||
SolrException.log(logger, e);
|
SolrException.log(log, e);
|
||||||
Map requestContext = req.getContext();
|
Map requestContext = req.getContext();
|
||||||
requestContext.put("stream", new DummyErrorStream(e));
|
requestContext.put("stream", new DummyErrorStream(e));
|
||||||
return;
|
return;
|
||||||
|
|
|
@ -49,7 +49,7 @@ import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
public class SQLHandler extends RequestHandlerBase implements SolrCoreAware, PermissionNameProvider {
|
public class SQLHandler extends RequestHandlerBase implements SolrCoreAware, PermissionNameProvider {
|
||||||
|
|
||||||
private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
|
|
||||||
private static String defaultZkhost = null;
|
private static String defaultZkhost = null;
|
||||||
private static String defaultWorkerCollection = null;
|
private static String defaultWorkerCollection = null;
|
||||||
|
@ -124,7 +124,7 @@ public class SQLHandler extends RequestHandlerBase implements SolrCoreAware, Per
|
||||||
if(tupleStream != null) {
|
if(tupleStream != null) {
|
||||||
tupleStream.close();
|
tupleStream.close();
|
||||||
}
|
}
|
||||||
SolrException.log(logger, e);
|
SolrException.log(log, e);
|
||||||
rsp.add("result-set", new StreamHandler.DummyErrorStream(e));
|
rsp.add("result-set", new StreamHandler.DummyErrorStream(e));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -66,7 +66,7 @@ public class StreamHandler extends RequestHandlerBase implements SolrCoreAware,
|
||||||
static SolrClientCache clientCache = new SolrClientCache();
|
static SolrClientCache clientCache = new SolrClientCache();
|
||||||
static ModelCache modelCache = null;
|
static ModelCache modelCache = null;
|
||||||
private StreamFactory streamFactory = new DefaultStreamFactory();
|
private StreamFactory streamFactory = new DefaultStreamFactory();
|
||||||
private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
private String coreName;
|
private String coreName;
|
||||||
private Map<String,DaemonStream> daemons = Collections.synchronizedMap(new HashMap());
|
private Map<String,DaemonStream> daemons = Collections.synchronizedMap(new HashMap());
|
||||||
|
|
||||||
|
@ -157,7 +157,7 @@ public class StreamHandler extends RequestHandlerBase implements SolrCoreAware,
|
||||||
} catch (Exception e) {
|
} catch (Exception e) {
|
||||||
// Catch exceptions that occur while the stream is being created. This will include streaming expression parse
|
// Catch exceptions that occur while the stream is being created. This will include streaming expression parse
|
||||||
// rules.
|
// rules.
|
||||||
SolrException.log(logger, e);
|
SolrException.log(log, e);
|
||||||
rsp.add("result-set", new DummyErrorStream(e));
|
rsp.add("result-set", new DummyErrorStream(e));
|
||||||
|
|
||||||
return;
|
return;
|
||||||
|
|
|
@ -30,7 +30,7 @@ import java.util.Map;
|
||||||
|
|
||||||
/** Enumerator that reads from a Solr collection. */
|
/** Enumerator that reads from a Solr collection. */
|
||||||
class SolrEnumerator implements Enumerator<Object> {
|
class SolrEnumerator implements Enumerator<Object> {
|
||||||
private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
|
|
||||||
private final TupleStream tupleStream;
|
private final TupleStream tupleStream;
|
||||||
private final List<Map.Entry<String, Class>> fields;
|
private final List<Map.Entry<String, Class>> fields;
|
||||||
|
@ -126,7 +126,7 @@ class SolrEnumerator implements Enumerator<Object> {
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
logger.error("IOException", e);
|
log.error("IOException", e);
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -77,7 +77,7 @@ import org.slf4j.LoggerFactory;
|
||||||
*/
|
*/
|
||||||
public class Grouping {
|
public class Grouping {
|
||||||
|
|
||||||
private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
|
|
||||||
private final SolrIndexSearcher searcher;
|
private final SolrIndexSearcher searcher;
|
||||||
private final QueryResult qr;
|
private final QueryResult qr;
|
||||||
|
@ -384,8 +384,8 @@ public class Grouping {
|
||||||
cachedCollector.replay(secondPhaseCollectors);
|
cachedCollector.replay(secondPhaseCollectors);
|
||||||
} else {
|
} else {
|
||||||
signalCacheWarning = true;
|
signalCacheWarning = true;
|
||||||
logger.warn(String.format(Locale.ROOT, "The grouping cache is active, but not used because it exceeded the max cache limit of %d percent", maxDocsPercentageToCache));
|
log.warn(String.format(Locale.ROOT, "The grouping cache is active, but not used because it exceeded the max cache limit of %d percent", maxDocsPercentageToCache));
|
||||||
logger.warn("Please increase cache size or disable group caching.");
|
log.warn("Please increase cache size or disable group caching.");
|
||||||
searchWithTimeLimiter(luceneFilter, secondPhaseCollectors);
|
searchWithTimeLimiter(luceneFilter, secondPhaseCollectors);
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
|
@ -447,7 +447,7 @@ public class Grouping {
|
||||||
}
|
}
|
||||||
searcher.search(q, collector);
|
searcher.search(q, collector);
|
||||||
} catch (TimeLimitingCollector.TimeExceededException | ExitableDirectoryReader.ExitingReaderException x) {
|
} catch (TimeLimitingCollector.TimeExceededException | ExitableDirectoryReader.ExitingReaderException x) {
|
||||||
logger.warn( "Query: " + query + "; " + x.getMessage() );
|
log.warn( "Query: " + query + "; " + x.getMessage() );
|
||||||
qr.setPartialResults(true);
|
qr.setPartialResults(true);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -115,7 +115,7 @@ public class CommandHandler {
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
|
|
||||||
private final QueryCommand queryCommand;
|
private final QueryCommand queryCommand;
|
||||||
private final List<Command> commands;
|
private final List<Command> commands;
|
||||||
|
@ -243,7 +243,7 @@ public class CommandHandler {
|
||||||
searcher.search(query, collector);
|
searcher.search(query, collector);
|
||||||
} catch (TimeLimitingCollector.TimeExceededException | ExitableDirectoryReader.ExitingReaderException x) {
|
} catch (TimeLimitingCollector.TimeExceededException | ExitableDirectoryReader.ExitingReaderException x) {
|
||||||
partialResults = true;
|
partialResults = true;
|
||||||
logger.warn( "Query: " + query + "; " + x.getMessage() );
|
log.warn( "Query: " + query + "; " + x.getMessage() );
|
||||||
}
|
}
|
||||||
|
|
||||||
if (includeHitCount) {
|
if (includeHitCount) {
|
||||||
|
|
|
@ -45,7 +45,7 @@ public class HdfsLocalityReporter implements SolrInfoBean, SolrMetricProducer {
|
||||||
public static final String LOCALITY_BLOCKS_LOCAL = "locality.blocks.local";
|
public static final String LOCALITY_BLOCKS_LOCAL = "locality.blocks.local";
|
||||||
public static final String LOCALITY_BLOCKS_RATIO = "locality.blocks.ratio";
|
public static final String LOCALITY_BLOCKS_RATIO = "locality.blocks.ratio";
|
||||||
|
|
||||||
private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||||
|
|
||||||
private String hostname;
|
private String hostname;
|
||||||
private final ConcurrentMap<HdfsDirectory,ConcurrentMap<FileStatus,BlockLocation[]>> cache;
|
private final ConcurrentMap<HdfsDirectory,ConcurrentMap<FileStatus,BlockLocation[]>> cache;
|
||||||
|
@ -129,7 +129,7 @@ public class HdfsLocalityReporter implements SolrInfoBean, SolrMetricProducer {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} catch (IOException e) {
|
} catch (IOException e) {
|
||||||
logger.warn("Could not retrieve locality information for {} due to exception: {}",
|
log.warn("Could not retrieve locality information for {} due to exception: {}",
|
||||||
hdfsDirectory.getHdfsDirPath(), e);
|
hdfsDirectory.getHdfsDirPath(), e);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -160,7 +160,7 @@ public class HdfsLocalityReporter implements SolrInfoBean, SolrMetricProducer {
|
||||||
* The directory to keep metrics on.
|
* The directory to keep metrics on.
|
||||||
*/
|
*/
|
||||||
public void registerDirectory(HdfsDirectory dir) {
|
public void registerDirectory(HdfsDirectory dir) {
|
||||||
logger.info("Registering direcotry {} for locality metrics.", dir.getHdfsDirPath().toString());
|
log.info("Registering direcotry {} for locality metrics.", dir.getHdfsDirPath().toString());
|
||||||
cache.put(dir, new ConcurrentHashMap<FileStatus, BlockLocation[]>());
|
cache.put(dir, new ConcurrentHashMap<FileStatus, BlockLocation[]>());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -181,7 +181,7 @@ public class HdfsLocalityReporter implements SolrInfoBean, SolrMetricProducer {
|
||||||
FileStatus[] statuses = fs.listStatus(dir.getHdfsDirPath());
|
FileStatus[] statuses = fs.listStatus(dir.getHdfsDirPath());
|
||||||
List<FileStatus> statusList = Arrays.asList(statuses);
|
List<FileStatus> statusList = Arrays.asList(statuses);
|
||||||
|
|
||||||
logger.debug("Updating locality information for: {}", statusList);
|
log.debug("Updating locality information for: {}", statusList);
|
||||||
|
|
||||||
// Keep only the files that still exist
|
// Keep only the files that still exist
|
||||||
cachedStatuses.retainAll(statusList);
|
cachedStatuses.retainAll(statusList);
|
||||||
|
|
Loading…
Reference in New Issue