mirror of https://github.com/apache/archiva.git
MRM-708 - Migrate from Plexus Logging to Slf4J
* Finishing up AbstractLogEnabled conversions to Slf4J. git-svn-id: https://svn.apache.org/repos/asf/maven/archiva/trunk@629704 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
45fca34ec2
commit
5f27f502ff
|
@ -39,7 +39,7 @@ import org.slf4j.LoggerFactory;
|
|||
*/
|
||||
public class Checksums
|
||||
{
|
||||
private static final Logger LOG = LoggerFactory.getLogger(Checksums.class);
|
||||
private Logger log = LoggerFactory.getLogger(Checksums.class);
|
||||
|
||||
/**
|
||||
* @plexus.requirement role-hint="sha1"
|
||||
|
@ -66,7 +66,7 @@ public class Checksums
|
|||
// Both files missing is a failure.
|
||||
if ( !sha1File.exists() && !md5File.exists() )
|
||||
{
|
||||
LOG.error( "File " + file.getPath() + " has no checksum files (sha1 or md5)." );
|
||||
log.error( "File " + file.getPath() + " has no checksum files (sha1 or md5)." );
|
||||
checksPass = false;
|
||||
}
|
||||
|
||||
|
@ -75,7 +75,7 @@ public class Checksums
|
|||
// Bad sha1 checksum is a failure.
|
||||
if ( !validateChecksum( sha1File, "sha1" ) )
|
||||
{
|
||||
LOG.warn( "SHA1 is incorrect for " + file.getPath() );
|
||||
log.warn( "SHA1 is incorrect for " + file.getPath() );
|
||||
checksPass = false;
|
||||
}
|
||||
}
|
||||
|
@ -85,7 +85,7 @@ public class Checksums
|
|||
// Bad md5 checksum is a failure.
|
||||
if ( !validateChecksum( md5File, "md5" ) )
|
||||
{
|
||||
LOG.warn( "MD5 is incorrect for " + file.getPath() );
|
||||
log.warn( "MD5 is incorrect for " + file.getPath() );
|
||||
checksPass = false;
|
||||
}
|
||||
}
|
||||
|
@ -139,12 +139,12 @@ public class Checksums
|
|||
}
|
||||
catch ( DigesterException e )
|
||||
{
|
||||
LOG.warn( "Unable to create " + digester.getFilenameExtension() + " file: " + e.getMessage(), e );
|
||||
log.warn( "Unable to create " + digester.getFilenameExtension() + " file: " + e.getMessage(), e );
|
||||
return false;
|
||||
}
|
||||
catch ( IOException e )
|
||||
{
|
||||
LOG.warn( "Unable to create " + digester.getFilenameExtension() + " file: " + e.getMessage(), e );
|
||||
log.warn( "Unable to create " + digester.getFilenameExtension() + " file: " + e.getMessage(), e );
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
@ -169,28 +169,28 @@ public class Checksums
|
|||
{
|
||||
if ( checksumFile.isValidChecksum( hashFile ) )
|
||||
{
|
||||
LOG.debug( "Valid checksum: " + hashFile.getPath() );
|
||||
log.debug( "Valid checksum: " + hashFile.getPath() );
|
||||
return true;
|
||||
}
|
||||
else
|
||||
{
|
||||
LOG.debug( "Not valid checksum: " + hashFile.getPath() );
|
||||
log.debug( "Not valid checksum: " + hashFile.getPath() );
|
||||
return createChecksum( localFile, digester );
|
||||
}
|
||||
}
|
||||
catch ( FileNotFoundException e )
|
||||
{
|
||||
LOG.warn( "Unable to find " + ext + " file: " + hashFile.getAbsolutePath(), e );
|
||||
log.warn( "Unable to find " + ext + " file: " + hashFile.getAbsolutePath(), e );
|
||||
return false;
|
||||
}
|
||||
catch ( DigesterException e )
|
||||
{
|
||||
LOG.warn( "Unable to process " + ext + " file: " + hashFile.getAbsolutePath(), e );
|
||||
log.warn( "Unable to process " + ext + " file: " + hashFile.getAbsolutePath(), e );
|
||||
return false;
|
||||
}
|
||||
catch ( IOException e )
|
||||
{
|
||||
LOG.warn( "Unable to process " + ext + " file: " + hashFile.getAbsolutePath(), e );
|
||||
log.warn( "Unable to process " + ext + " file: " + hashFile.getAbsolutePath(), e );
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
@ -213,27 +213,27 @@ public class Checksums
|
|||
boolean validity = checksumFile.isValidChecksum( hashFile );
|
||||
if ( validity )
|
||||
{
|
||||
LOG.debug( "Valid checksum: " + hashFile.getPath() );
|
||||
log.debug( "Valid checksum: " + hashFile.getPath() );
|
||||
}
|
||||
else
|
||||
{
|
||||
LOG.debug( "Not valid checksum: " + hashFile.getPath() );
|
||||
log.debug( "Not valid checksum: " + hashFile.getPath() );
|
||||
}
|
||||
return validity;
|
||||
}
|
||||
catch ( FileNotFoundException e )
|
||||
{
|
||||
LOG.warn( "Unable to find " + type + " file: " + hashFile.getAbsolutePath(), e );
|
||||
log.warn( "Unable to find " + type + " file: " + hashFile.getAbsolutePath(), e );
|
||||
return false;
|
||||
}
|
||||
catch ( DigesterException e )
|
||||
{
|
||||
LOG.warn( "Unable to process " + type + " file: " + hashFile.getAbsolutePath(), e );
|
||||
log.warn( "Unable to process " + type + " file: " + hashFile.getAbsolutePath(), e );
|
||||
return false;
|
||||
}
|
||||
catch ( IOException e )
|
||||
{
|
||||
LOG.warn( "Unable to process " + type + " file: " + hashFile.getAbsolutePath(), e );
|
||||
log.warn( "Unable to process " + type + " file: " + hashFile.getAbsolutePath(), e );
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -36,12 +36,13 @@ import org.codehaus.plexus.evaluator.DefaultExpressionEvaluator;
|
|||
import org.codehaus.plexus.evaluator.EvaluatorException;
|
||||
import org.codehaus.plexus.evaluator.ExpressionEvaluator;
|
||||
import org.codehaus.plexus.evaluator.sources.SystemPropertyExpressionSource;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.codehaus.plexus.registry.Registry;
|
||||
import org.codehaus.plexus.registry.RegistryException;
|
||||
import org.codehaus.plexus.registry.RegistryListener;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
|
@ -83,9 +84,10 @@ import java.util.Map.Entry;
|
|||
* @plexus.component role="org.apache.maven.archiva.configuration.ArchivaConfiguration"
|
||||
*/
|
||||
public class DefaultArchivaConfiguration
|
||||
extends AbstractLogEnabled
|
||||
implements ArchivaConfiguration, RegistryListener, Initializable
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger(DefaultArchivaConfiguration.class);
|
||||
|
||||
/**
|
||||
* Plexus registry to read the configuration from.
|
||||
*
|
||||
|
@ -287,7 +289,7 @@ public class DefaultArchivaConfiguration
|
|||
else
|
||||
{
|
||||
// Policy key doesn't exist. Don't add it to golden version.
|
||||
getLogger().warn( "Policy [" + policyId + "] does not exist." );
|
||||
log.warn( "Policy [" + policyId + "] does not exist." );
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -323,13 +325,13 @@ public class DefaultArchivaConfiguration
|
|||
{
|
||||
if ( MapUtils.isEmpty( prePolicies ) )
|
||||
{
|
||||
getLogger().error( "No PreDownloadPolicies found!" );
|
||||
log.error( "No PreDownloadPolicies found!" );
|
||||
return null;
|
||||
}
|
||||
|
||||
if ( MapUtils.isEmpty( postPolicies ) )
|
||||
{
|
||||
getLogger().error( "No PostDownloadPolicies found!" );
|
||||
log.error( "No PostDownloadPolicies found!" );
|
||||
return null;
|
||||
}
|
||||
|
||||
|
@ -354,13 +356,13 @@ public class DefaultArchivaConfiguration
|
|||
{
|
||||
if ( MapUtils.isEmpty( prePolicies ) )
|
||||
{
|
||||
getLogger().error( "No PreDownloadPolicies found!" );
|
||||
log.error( "No PreDownloadPolicies found!" );
|
||||
return false;
|
||||
}
|
||||
|
||||
if ( MapUtils.isEmpty( postPolicies ) )
|
||||
{
|
||||
getLogger().error( "No PostDownloadPolicies found!" );
|
||||
log.error( "No PostDownloadPolicies found!" );
|
||||
return false;
|
||||
}
|
||||
|
||||
|
@ -505,7 +507,7 @@ public class DefaultArchivaConfiguration
|
|||
}
|
||||
catch ( IOException e )
|
||||
{
|
||||
getLogger().error( "Unable to create " + filetype + " file: " + e.getMessage(), e );
|
||||
log.error( "Unable to create " + filetype + " file: " + e.getMessage(), e );
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
@ -521,7 +523,7 @@ public class DefaultArchivaConfiguration
|
|||
}
|
||||
catch ( Throwable t )
|
||||
{
|
||||
getLogger().warn( "Unable to notify of saved configuration event.", t );
|
||||
log.warn( "Unable to notify of saved configuration event.", t );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,17 +19,6 @@ package org.apache.maven.archiva.configuration;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.commons.collections.Predicate;
|
||||
import org.apache.commons.configuration.CombinedConfiguration;
|
||||
import org.apache.maven.archiva.configuration.functors.FiletypeSelectionPredicate;
|
||||
import org.apache.maven.archiva.configuration.io.registry.ConfigurationRegistryReader;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.codehaus.plexus.registry.RegistryException;
|
||||
import org.codehaus.plexus.registry.commons.CommonsConfigurationRegistry;
|
||||
|
||||
import java.lang.reflect.Field;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
|
@ -37,6 +26,19 @@ import java.util.HashMap;
|
|||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.commons.collections.Predicate;
|
||||
import org.apache.commons.configuration.CombinedConfiguration;
|
||||
import org.apache.maven.archiva.common.utils.Slf4JPlexusLogger;
|
||||
import org.apache.maven.archiva.configuration.functors.FiletypeSelectionPredicate;
|
||||
import org.apache.maven.archiva.configuration.io.registry.ConfigurationRegistryReader;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.codehaus.plexus.registry.RegistryException;
|
||||
import org.codehaus.plexus.registry.commons.CommonsConfigurationRegistry;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* FileTypes
|
||||
*
|
||||
|
@ -46,9 +48,10 @@ import java.util.Map;
|
|||
* @plexus.component role="org.apache.maven.archiva.configuration.FileTypes"
|
||||
*/
|
||||
public class FileTypes
|
||||
extends AbstractLogEnabled
|
||||
implements Initializable
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger(FileTypes.class);
|
||||
|
||||
public static final String ARTIFACTS = "artifacts";
|
||||
|
||||
public static final String AUTO_REMOVE = "auto-remove";
|
||||
|
@ -122,7 +125,7 @@ public class FileTypes
|
|||
Field fld = commonsRegistry.getClass().getDeclaredField( "configuration" );
|
||||
fld.setAccessible( true );
|
||||
fld.set( commonsRegistry, new CombinedConfiguration() );
|
||||
commonsRegistry.enableLogging( getLogger() );
|
||||
commonsRegistry.enableLogging( new Slf4JPlexusLogger( FileTypes.class ) );
|
||||
commonsRegistry.addConfigurationFromResource( "org/apache/maven/archiva/configuration/default-archiva.xml" );
|
||||
|
||||
// Read configuration as it was intended.
|
||||
|
|
|
@ -22,7 +22,8 @@ package org.apache.maven.archiva.indexer.functors;
|
|||
import org.apache.commons.collections.Predicate;
|
||||
import org.apache.maven.archiva.indexer.RepositoryContentIndex;
|
||||
import org.apache.maven.archiva.indexer.RepositoryIndexException;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* Test the {@link RepositoryContentIndex} object for the existance of an index.
|
||||
|
@ -35,9 +36,10 @@ import org.codehaus.plexus.logging.AbstractLogEnabled;
|
|||
* role-hint="index-exists"
|
||||
*/
|
||||
public class IndexExistsPredicate
|
||||
extends AbstractLogEnabled
|
||||
implements Predicate
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( IndexExistsPredicate.class );
|
||||
|
||||
public boolean evaluate( Object object )
|
||||
{
|
||||
boolean satisfies = false;
|
||||
|
@ -51,7 +53,7 @@ public class IndexExistsPredicate
|
|||
}
|
||||
catch ( RepositoryIndexException e )
|
||||
{
|
||||
getLogger().info(
|
||||
log.info(
|
||||
"Repository Content Index [" + index.getId() + "] for repository ["
|
||||
+ index.getRepository().getId() + "] does not exist yet in ["
|
||||
+ index.getIndexDirectory().getAbsolutePath() + "]." );
|
||||
|
|
|
@ -22,7 +22,8 @@ package org.apache.maven.archiva.indexer.functors;
|
|||
import org.apache.commons.collections.Transformer;
|
||||
import org.apache.maven.archiva.indexer.RepositoryIndexSearchException;
|
||||
import org.apache.maven.archiva.indexer.lucene.LuceneRepositoryContentIndex;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* SearchableTransformer
|
||||
|
@ -33,9 +34,10 @@ import org.codehaus.plexus.logging.AbstractLogEnabled;
|
|||
* @plexus.component role="org.apache.commons.collections.Transformer" role-hint="searchable"
|
||||
*/
|
||||
public class SearchableTransformer
|
||||
extends AbstractLogEnabled
|
||||
implements Transformer
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( SearchableTransformer.class );
|
||||
|
||||
public Object transform( Object input )
|
||||
{
|
||||
if ( input instanceof LuceneRepositoryContentIndex )
|
||||
|
@ -47,7 +49,7 @@ public class SearchableTransformer
|
|||
}
|
||||
catch ( RepositoryIndexSearchException e )
|
||||
{
|
||||
getLogger().warn("Unable to get searchable for index:" + e.getMessage(), e);
|
||||
log.warn("Unable to get searchable for index:" + e.getMessage(), e);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -19,9 +19,10 @@ package org.apache.maven.archiva.indexer.search;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.commons.collections.Predicate;
|
||||
import org.apache.commons.collections.Transformer;
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.lucene.document.Document;
|
||||
import org.apache.lucene.queryParser.MultiFieldQueryParser;
|
||||
import org.apache.lucene.queryParser.ParseException;
|
||||
|
@ -38,21 +39,17 @@ import org.apache.maven.archiva.indexer.RepositoryIndexException;
|
|||
import org.apache.maven.archiva.indexer.RepositoryIndexSearchException;
|
||||
import org.apache.maven.archiva.indexer.bytecode.BytecodeHandlers;
|
||||
import org.apache.maven.archiva.indexer.filecontent.FileContentHandlers;
|
||||
import org.apache.maven.archiva.indexer.functors.UserAllowedToSearchRepositoryPredicate;
|
||||
import org.apache.maven.archiva.indexer.hashcodes.HashcodesHandlers;
|
||||
import org.apache.maven.archiva.indexer.hashcodes.HashcodesKeys;
|
||||
import org.apache.maven.archiva.indexer.lucene.LuceneEntryConverter;
|
||||
import org.apache.maven.archiva.indexer.lucene.LuceneQuery;
|
||||
import org.apache.maven.archiva.indexer.lucene.LuceneRepositoryContentRecord;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.codehaus.plexus.registry.Registry;
|
||||
import org.codehaus.plexus.registry.RegistryListener;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* DefaultCrossRepositorySearch
|
||||
|
@ -62,9 +59,10 @@ import java.util.List;
|
|||
* @plexus.component role="org.apache.maven.archiva.indexer.search.CrossRepositorySearch" role-hint="default"
|
||||
*/
|
||||
public class DefaultCrossRepositorySearch
|
||||
extends AbstractLogEnabled
|
||||
implements CrossRepositorySearch, RegistryListener, Initializable
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( DefaultCrossRepositorySearch.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement role-hint="lucene"
|
||||
*/
|
||||
|
@ -93,7 +91,7 @@ public class DefaultCrossRepositorySearch
|
|||
}
|
||||
catch ( ParseException e )
|
||||
{
|
||||
getLogger().warn( "Unable to parse query [" + checksum + "]: " + e.getMessage(), e );
|
||||
log.warn( "Unable to parse query [" + checksum + "]: " + e.getMessage(), e );
|
||||
}
|
||||
|
||||
// empty results.
|
||||
|
@ -115,7 +113,7 @@ public class DefaultCrossRepositorySearch
|
|||
}
|
||||
catch ( ParseException e )
|
||||
{
|
||||
getLogger().warn( "Unable to parse query [" + term + "]: " + e.getMessage(), e );
|
||||
log.warn( "Unable to parse query [" + term + "]: " + e.getMessage(), e );
|
||||
}
|
||||
|
||||
// empty results.
|
||||
|
@ -137,7 +135,7 @@ public class DefaultCrossRepositorySearch
|
|||
}
|
||||
catch ( ParseException e )
|
||||
{
|
||||
getLogger().warn( "Unable to parse query [" + term + "]: " + e.getMessage(), e );
|
||||
log.warn( "Unable to parse query [" + term + "]: " + e.getMessage(), e );
|
||||
}
|
||||
|
||||
// empty results.
|
||||
|
@ -212,7 +210,7 @@ public class DefaultCrossRepositorySearch
|
|||
}
|
||||
catch ( java.text.ParseException e )
|
||||
{
|
||||
getLogger().warn( "Unable to parse document into record: " + e.getMessage(), e );
|
||||
log.warn( "Unable to parse document into record: " + e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -220,7 +218,7 @@ public class DefaultCrossRepositorySearch
|
|||
}
|
||||
catch ( IOException e )
|
||||
{
|
||||
getLogger().error( "Unable to setup multi-search: " + e.getMessage(), e );
|
||||
log.error( "Unable to setup multi-search: " + e.getMessage(), e );
|
||||
}
|
||||
finally
|
||||
{
|
||||
|
@ -233,7 +231,7 @@ public class DefaultCrossRepositorySearch
|
|||
}
|
||||
catch ( IOException ie )
|
||||
{
|
||||
getLogger().error( "Unable to close index searcher: " + ie.getMessage(), ie );
|
||||
log.error( "Unable to close index searcher: " + ie.getMessage(), ie );
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -251,7 +249,7 @@ public class DefaultCrossRepositorySearch
|
|||
}
|
||||
catch ( RepositoryIndexSearchException e )
|
||||
{
|
||||
getLogger().warn( "Unable to get searchable for index [" + contentIndex.getId() + "] :"
|
||||
log.warn( "Unable to get searchable for index [" + contentIndex.getId() + "] :"
|
||||
+ e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
|
@ -329,7 +327,7 @@ public class DefaultCrossRepositorySearch
|
|||
}
|
||||
catch ( RepositoryIndexException e )
|
||||
{
|
||||
getLogger().info(
|
||||
log.info(
|
||||
"Repository Content Index [" + index.getId() + "] for repository ["
|
||||
+ index.getRepository().getId() + "] does not exist yet in ["
|
||||
+ index.getIndexDirectory().getAbsolutePath() + "]." );
|
||||
|
|
|
@ -19,16 +19,17 @@ package org.apache.maven.archiva.policies;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.common.utils.VersionUtil;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.io.File;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Calendar;
|
||||
import java.util.List;
|
||||
import java.util.Properties;
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.common.utils.VersionUtil;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* AbstractUpdatePolicy
|
||||
*
|
||||
|
@ -36,9 +37,10 @@ import java.util.Properties;
|
|||
* @version $Id$
|
||||
*/
|
||||
public abstract class AbstractUpdatePolicy
|
||||
extends AbstractLogEnabled
|
||||
implements PreDownloadPolicy
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( AbstractUpdatePolicy.class );
|
||||
|
||||
/**
|
||||
* The ALWAYS policy setting means that the artifact is always uipdated from the remote repo.
|
||||
*/
|
||||
|
@ -127,20 +129,20 @@ public abstract class AbstractUpdatePolicy
|
|||
if ( ALWAYS.equals( policySetting ) )
|
||||
{
|
||||
// Skip means ok to update.
|
||||
getLogger().debug( "OK to update, " + getUpdateMode() + " policy set to ALWAYS." );
|
||||
log.debug( "OK to update, " + getUpdateMode() + " policy set to ALWAYS." );
|
||||
return;
|
||||
}
|
||||
|
||||
// Test for mismatches.
|
||||
if ( !isSnapshotVersion && isSnapshotPolicy() )
|
||||
{
|
||||
getLogger().debug( "OK to update, snapshot policy does not apply for non-snapshot versions." );
|
||||
log.debug( "OK to update, snapshot policy does not apply for non-snapshot versions." );
|
||||
return;
|
||||
}
|
||||
|
||||
if ( isSnapshotVersion && !isSnapshotPolicy() )
|
||||
{
|
||||
getLogger().debug( "OK to update, release policy does not apply for snapshot versions." );
|
||||
log.debug( "OK to update, release policy does not apply for snapshot versions." );
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -153,7 +155,7 @@ public abstract class AbstractUpdatePolicy
|
|||
if ( !localFile.exists() )
|
||||
{
|
||||
// No file means it's ok.
|
||||
getLogger().debug( "OK to update " + getUpdateMode() + ", local file does not exist." );
|
||||
log.debug( "OK to update " + getUpdateMode() + ", local file does not exist." );
|
||||
return;
|
||||
}
|
||||
|
||||
|
|
|
@ -19,15 +19,16 @@ package org.apache.maven.archiva.policies;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.policies.urlcache.UrlFailureCache;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.io.File;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Properties;
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.policies.urlcache.UrlFailureCache;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* {@link PreDownloadPolicy} to check if the requested url has failed before.
|
||||
*
|
||||
|
@ -37,9 +38,10 @@ import java.util.Properties;
|
|||
* role-hint="cache-failures"
|
||||
*/
|
||||
public class CachedFailuresPolicy
|
||||
extends AbstractLogEnabled
|
||||
implements PreDownloadPolicy
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( CachedFailuresPolicy.class );
|
||||
|
||||
/**
|
||||
* The NO policy setting means that the the existence of old failures is <strong>not</strong> checked.
|
||||
* All resource requests are allowed thru to the remote repo.
|
||||
|
@ -78,7 +80,7 @@ public class CachedFailuresPolicy
|
|||
if ( NO.equals( policySetting ) )
|
||||
{
|
||||
// Skip.
|
||||
getLogger().debug( "OK to fetch, check-failures policy set to NO." );
|
||||
log.debug( "OK to fetch, check-failures policy set to NO." );
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -92,7 +94,7 @@ public class CachedFailuresPolicy
|
|||
}
|
||||
}
|
||||
|
||||
getLogger().debug( "OK to fetch, check-failures detected no issues." );
|
||||
log.debug( "OK to fetch, check-failures detected no issues." );
|
||||
}
|
||||
|
||||
public String getDefaultOption()
|
||||
|
|
|
@ -19,15 +19,16 @@ package org.apache.maven.archiva.policies;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.common.utils.Checksums;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.io.File;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Properties;
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.common.utils.Checksums;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* ChecksumPolicy - a policy applied after the download to see if the file has been downloaded
|
||||
* successfully and completely (or not).
|
||||
|
@ -39,9 +40,10 @@ import java.util.Properties;
|
|||
* role-hint="checksum"
|
||||
*/
|
||||
public class ChecksumPolicy
|
||||
extends AbstractLogEnabled
|
||||
implements PostDownloadPolicy
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( ChecksumPolicy.class );
|
||||
|
||||
/**
|
||||
* The IGNORE policy indicates that if the checksum policy is ignored, and
|
||||
* the state of, contents of, or validity of the checksum files are not
|
||||
|
@ -90,7 +92,7 @@ public class ChecksumPolicy
|
|||
if ( IGNORE.equals( policySetting ) )
|
||||
{
|
||||
// Ignore.
|
||||
getLogger().debug( "Checksum policy set to IGNORE." );
|
||||
log.debug( "Checksum policy set to IGNORE." );
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -131,7 +133,7 @@ public class ChecksumPolicy
|
|||
{
|
||||
if( checksums.update( localFile ) )
|
||||
{
|
||||
getLogger().debug( "Checksum policy set to FIX, checksum files have been updated." );
|
||||
log.debug( "Checksum policy set to FIX, checksum files have been updated." );
|
||||
return;
|
||||
}
|
||||
else
|
||||
|
|
|
@ -19,9 +19,18 @@ package org.apache.maven.archiva.proxy;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Properties;
|
||||
import java.util.Map.Entry;
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.commons.io.FileUtils;
|
||||
import org.apache.commons.lang.ArrayUtils;
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
|
||||
import org.apache.maven.archiva.configuration.ConfigurationNames;
|
||||
|
@ -56,24 +65,13 @@ import org.apache.maven.wagon.authentication.AuthenticationException;
|
|||
import org.apache.maven.wagon.authentication.AuthenticationInfo;
|
||||
import org.apache.maven.wagon.proxy.ProxyInfo;
|
||||
import org.apache.maven.wagon.repository.Repository;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.codehaus.plexus.registry.Registry;
|
||||
import org.codehaus.plexus.registry.RegistryListener;
|
||||
import org.codehaus.plexus.util.SelectorUtils;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.net.URLClassLoader;
|
||||
import java.net.URL;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
import java.util.Properties;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* DefaultRepositoryProxyConnectors
|
||||
|
@ -83,9 +81,10 @@ import java.util.Properties;
|
|||
* @plexus.component role-hint="default"
|
||||
*/
|
||||
public class DefaultRepositoryProxyConnectors
|
||||
extends AbstractLogEnabled
|
||||
implements RepositoryProxyConnectors, RegistryListener, Initializable
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( DefaultRepositoryProxyConnectors.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement
|
||||
*/
|
||||
|
@ -160,29 +159,29 @@ public class DefaultRepositoryProxyConnectors
|
|||
|
||||
if ( fileExists( downloadedFile ) )
|
||||
{
|
||||
getLogger().debug( "Successfully transferred: " + downloadedFile.getAbsolutePath() );
|
||||
log.debug( "Successfully transferred: " + downloadedFile.getAbsolutePath() );
|
||||
return downloadedFile;
|
||||
}
|
||||
}
|
||||
catch ( NotFoundException e )
|
||||
{
|
||||
getLogger().debug( "Artifact " + Keys.toKey( artifact ) + " not found on repository \""
|
||||
log.debug( "Artifact " + Keys.toKey( artifact ) + " not found on repository \""
|
||||
+ targetRepository.getRepository().getId() + "\"." );
|
||||
}
|
||||
catch ( NotModifiedException e )
|
||||
{
|
||||
getLogger().debug( "Artifact " + Keys.toKey( artifact ) + " not updated on repository \""
|
||||
log.debug( "Artifact " + Keys.toKey( artifact ) + " not updated on repository \""
|
||||
+ targetRepository.getRepository().getId() + "\"." );
|
||||
}
|
||||
catch ( ProxyException e )
|
||||
{
|
||||
getLogger().warn( "Transfer error from repository \"" + targetRepository.getRepository().getId() +
|
||||
log.warn( "Transfer error from repository \"" + targetRepository.getRepository().getId() +
|
||||
"\" for artifact " + Keys.toKey( artifact ) + ", continuing to next repository. Error message: " +
|
||||
e.getMessage() );
|
||||
getLogger().debug( "Full stack trace", e );
|
||||
log.debug( "Full stack trace", e );
|
||||
}
|
||||
}
|
||||
getLogger().debug( "Exhausted all target repositories, artifact " + Keys.toKey( artifact ) + " not found." );
|
||||
log.debug( "Exhausted all target repositories, artifact " + Keys.toKey( artifact ) + " not found." );
|
||||
|
||||
return null;
|
||||
}
|
||||
|
@ -221,22 +220,22 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
catch ( NotFoundException e )
|
||||
{
|
||||
getLogger().debug( "Versioned Metadata " + Keys.toKey( metadata )
|
||||
log.debug( "Versioned Metadata " + Keys.toKey( metadata )
|
||||
+ " not found on remote repository \""
|
||||
+ targetRepository.getRepository().getId() + "\"." );
|
||||
}
|
||||
catch ( NotModifiedException e )
|
||||
{
|
||||
getLogger().debug( "Versioned Metadata " + Keys.toKey( metadata )
|
||||
log.debug( "Versioned Metadata " + Keys.toKey( metadata )
|
||||
+ " not updated on remote repository \""
|
||||
+ targetRepository.getRepository().getId() + "\"." );
|
||||
}
|
||||
catch ( ProxyException e )
|
||||
{
|
||||
getLogger().warn( "Transfer error from repository \"" + targetRepository.getRepository().getId() +
|
||||
log.warn( "Transfer error from repository \"" + targetRepository.getRepository().getId() +
|
||||
"\" for versioned Metadata " + Keys.toKey( metadata ) +
|
||||
", continuing to next repository. Error message: " + e.getMessage() );
|
||||
getLogger().debug( "Full stack trace", e );
|
||||
log.debug( "Full stack trace", e );
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -253,24 +252,24 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
catch ( LayoutException e )
|
||||
{
|
||||
getLogger().warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage() );
|
||||
log.warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage() );
|
||||
// TODO: add into repository report?
|
||||
}
|
||||
catch ( RepositoryMetadataException e )
|
||||
{
|
||||
getLogger()
|
||||
log
|
||||
.warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage(), e );
|
||||
// TODO: add into repository report?
|
||||
}
|
||||
catch ( IOException e )
|
||||
{
|
||||
getLogger()
|
||||
log
|
||||
.warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage(), e );
|
||||
// TODO: add into repository report?
|
||||
}
|
||||
catch ( ContentNotFoundException e )
|
||||
{
|
||||
getLogger()
|
||||
log
|
||||
.warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage(), e );
|
||||
// TODO: add into repository report?
|
||||
}
|
||||
|
@ -339,21 +338,21 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
catch ( NotFoundException e )
|
||||
{
|
||||
getLogger().debug( "Project Metadata " + Keys.toKey( metadata ) + " not found on remote repository \""
|
||||
log.debug( "Project Metadata " + Keys.toKey( metadata ) + " not found on remote repository \""
|
||||
+ targetRepository.getRepository().getId() + "\"." );
|
||||
}
|
||||
catch ( NotModifiedException e )
|
||||
{
|
||||
getLogger().debug( "Project Metadata " + Keys.toKey( metadata )
|
||||
log.debug( "Project Metadata " + Keys.toKey( metadata )
|
||||
+ " not updated on remote repository \""
|
||||
+ targetRepository.getRepository().getId() + "\"." );
|
||||
}
|
||||
catch ( ProxyException e )
|
||||
{
|
||||
getLogger().warn( "Transfer error from repository \"" + targetRepository.getRepository().getId() +
|
||||
log.warn( "Transfer error from repository \"" + targetRepository.getRepository().getId() +
|
||||
"\" for project metadata " + Keys.toKey( metadata ) +
|
||||
", continuing to next repository. Error message: " + e.getMessage() );
|
||||
getLogger().debug( "Full stack trace", e );
|
||||
log.debug( "Full stack trace", e );
|
||||
}
|
||||
|
||||
}
|
||||
|
@ -371,24 +370,24 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
catch ( LayoutException e )
|
||||
{
|
||||
getLogger().warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage() );
|
||||
log.warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage() );
|
||||
// TODO: add into repository report?
|
||||
}
|
||||
catch ( RepositoryMetadataException e )
|
||||
{
|
||||
getLogger()
|
||||
log
|
||||
.warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage(), e );
|
||||
// TODO: add into repository report?
|
||||
}
|
||||
catch ( IOException e )
|
||||
{
|
||||
getLogger()
|
||||
log
|
||||
.warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage(), e );
|
||||
// TODO: add into repository report?
|
||||
}
|
||||
catch ( ContentNotFoundException e )
|
||||
{
|
||||
getLogger()
|
||||
log
|
||||
.warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage(), e );
|
||||
// TODO: add into repository report?
|
||||
}
|
||||
|
@ -495,7 +494,7 @@ public class DefaultRepositoryProxyConnectors
|
|||
// Path must belong to whitelist.
|
||||
if ( !matchesPattern( remotePath, connector.getWhitelist() ) )
|
||||
{
|
||||
getLogger().debug( "Path [" + remotePath +
|
||||
log.debug( "Path [" + remotePath +
|
||||
"] is not part of defined whitelist (skipping transfer from repository [" +
|
||||
remoteRepository.getRepository().getName() + "])." );
|
||||
return null;
|
||||
|
@ -505,7 +504,7 @@ public class DefaultRepositoryProxyConnectors
|
|||
// Is target path part of blacklist?
|
||||
if ( matchesPattern( remotePath, connector.getBlacklist() ) )
|
||||
{
|
||||
getLogger().debug( "Path [" + remotePath + "] is part of blacklist (skipping transfer from repository [" +
|
||||
log.debug( "Path [" + remotePath + "] is part of blacklist (skipping transfer from repository [" +
|
||||
remoteRepository.getRepository().getName() + "])." );
|
||||
return null;
|
||||
}
|
||||
|
@ -520,11 +519,11 @@ public class DefaultRepositoryProxyConnectors
|
|||
String emsg = "Transfer not attempted on " + url + " : " + e.getMessage();
|
||||
if ( fileExists( localFile ) )
|
||||
{
|
||||
getLogger().info( emsg + ": using already present local file." );
|
||||
log.info( emsg + ": using already present local file." );
|
||||
return localFile;
|
||||
}
|
||||
|
||||
getLogger().info( emsg );
|
||||
log.info( emsg );
|
||||
return null;
|
||||
}
|
||||
|
||||
|
@ -573,7 +572,7 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
catch ( ConnectionException e )
|
||||
{
|
||||
getLogger().warn( "Unable to disconnect wagon.", e );
|
||||
log.warn( "Unable to disconnect wagon.", e );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -585,7 +584,7 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
catch ( PolicyViolationException e )
|
||||
{
|
||||
getLogger().info( "Transfer invalidated from " + url + " : " + e.getMessage() );
|
||||
log.info( "Transfer invalidated from " + url + " : " + e.getMessage() );
|
||||
if ( fileExists( localFile ) )
|
||||
{
|
||||
return localFile;
|
||||
|
@ -629,22 +628,22 @@ public class DefaultRepositoryProxyConnectors
|
|||
{
|
||||
File hashFile = new File( localFile.getAbsolutePath() + type );
|
||||
transferSimpleFile( wagon, remoteRepository, remotePath + type, hashFile );
|
||||
getLogger().debug( "Checksum" + type + " Downloaded: " + hashFile );
|
||||
log.debug( "Checksum" + type + " Downloaded: " + hashFile );
|
||||
}
|
||||
catch ( NotFoundException e )
|
||||
{
|
||||
getLogger().debug( "Transfer failed, checksum not found: " + url );
|
||||
log.debug( "Transfer failed, checksum not found: " + url );
|
||||
// Consume it, do not pass this on.
|
||||
}
|
||||
catch ( NotModifiedException e )
|
||||
{
|
||||
getLogger().debug( "Transfer skipped, checksum not modified: " + url );
|
||||
log.debug( "Transfer skipped, checksum not modified: " + url );
|
||||
// Consume it, do not pass this on.
|
||||
}
|
||||
catch ( ProxyException e )
|
||||
{
|
||||
urlFailureCache.cacheFailure( url + type );
|
||||
getLogger().warn( "Transfer failed on checksum: " + url + " : " + e.getMessage(), e );
|
||||
log.warn( "Transfer failed on checksum: " + url + " : " + e.getMessage(), e );
|
||||
// Critical issue, pass it on.
|
||||
throw e;
|
||||
}
|
||||
|
@ -678,7 +677,7 @@ public class DefaultRepositoryProxyConnectors
|
|||
|
||||
if ( !localFile.exists() )
|
||||
{
|
||||
getLogger().debug( "Retrieving " + remotePath + " from " + remoteRepository.getRepository().getName() );
|
||||
log.debug( "Retrieving " + remotePath + " from " + remoteRepository.getRepository().getName() );
|
||||
wagon.get( remotePath, temp );
|
||||
success = true;
|
||||
|
||||
|
@ -688,11 +687,11 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
|
||||
// You wouldn't get here on failure, a WagonException would have been thrown.
|
||||
getLogger().debug( "Downloaded successfully." );
|
||||
log.debug( "Downloaded successfully." );
|
||||
}
|
||||
else
|
||||
{
|
||||
getLogger().debug( "Retrieving " + remotePath + " from " + remoteRepository.getRepository().getName()
|
||||
log.debug( "Retrieving " + remotePath + " from " + remoteRepository.getRepository().getName()
|
||||
+ " if updated" );
|
||||
success = wagon.getIfNewer( remotePath, temp, localFile.lastModified() );
|
||||
if ( !success )
|
||||
|
@ -703,7 +702,7 @@ public class DefaultRepositoryProxyConnectors
|
|||
|
||||
if ( temp.exists() )
|
||||
{
|
||||
getLogger().debug( "Downloaded successfully." );
|
||||
log.debug( "Downloaded successfully." );
|
||||
moveTempToTarget( temp, localFile );
|
||||
}
|
||||
}
|
||||
|
@ -748,14 +747,14 @@ public class DefaultRepositoryProxyConnectors
|
|||
String defaultSetting = policy.getDefaultOption();
|
||||
String setting = StringUtils.defaultString( (String) settings.get( key ), defaultSetting );
|
||||
|
||||
getLogger().debug( "Applying [" + key + "] policy with [" + setting + "]" );
|
||||
log.debug( "Applying [" + key + "] policy with [" + setting + "]" );
|
||||
try
|
||||
{
|
||||
policy.applyPolicy( setting, request, localFile );
|
||||
}
|
||||
catch ( PolicyConfigurationException e )
|
||||
{
|
||||
getLogger().error( e.getMessage(), e );
|
||||
log.error( e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -778,7 +777,7 @@ public class DefaultRepositoryProxyConnectors
|
|||
|
||||
if ( !temp.renameTo( target ) )
|
||||
{
|
||||
getLogger().warn( "Unable to rename tmp file to its final name... resorting to copy command." );
|
||||
log.warn( "Unable to rename tmp file to its final name... resorting to copy command." );
|
||||
|
||||
try
|
||||
{
|
||||
|
@ -821,7 +820,7 @@ public class DefaultRepositoryProxyConnectors
|
|||
|
||||
if ( StringUtils.isNotBlank( username ) && StringUtils.isNotBlank( password ) )
|
||||
{
|
||||
getLogger().debug( "Using username " + username + " to connect to remote repository "
|
||||
log.debug( "Using username " + username + " to connect to remote repository "
|
||||
+ remoteRepository.getURL() );
|
||||
authInfo = new AuthenticationInfo();
|
||||
authInfo.setUserName( username );
|
||||
|
@ -829,7 +828,7 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
else
|
||||
{
|
||||
getLogger().debug( "No authentication for remote repository needed" );
|
||||
log.debug( "No authentication for remote repository needed" );
|
||||
}
|
||||
|
||||
//Convert seconds to milliseconds
|
||||
|
@ -851,14 +850,14 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
catch ( ConnectionException e )
|
||||
{
|
||||
getLogger().warn(
|
||||
log.warn(
|
||||
"Could not connect to " + remoteRepository.getRepository().getName() + ": "
|
||||
+ e.getMessage() );
|
||||
connected = false;
|
||||
}
|
||||
catch ( AuthenticationException e )
|
||||
{
|
||||
getLogger().warn(
|
||||
log.warn(
|
||||
"Could not connect to " + remoteRepository.getRepository().getName() + ": "
|
||||
+ e.getMessage() );
|
||||
connected = false;
|
||||
|
@ -925,6 +924,16 @@ public class DefaultRepositoryProxyConnectors
|
|||
{
|
||||
/* do nothing */
|
||||
}
|
||||
|
||||
private void logProcess( String managedRepoId, String resource, String event )
|
||||
{
|
||||
|
||||
}
|
||||
|
||||
private void logRejection( String managedRepoId, String remoteRepoId, String resource, String reason )
|
||||
{
|
||||
|
||||
}
|
||||
|
||||
private void initConnectorsAndNetworkProxies()
|
||||
{
|
||||
|
@ -988,11 +997,11 @@ public class DefaultRepositoryProxyConnectors
|
|||
}
|
||||
catch ( RepositoryNotFoundException e )
|
||||
{
|
||||
getLogger().warn( "Unable to use proxy connector: " + e.getMessage(), e );
|
||||
log.warn( "Unable to use proxy connector: " + e.getMessage(), e );
|
||||
}
|
||||
catch ( RepositoryException e )
|
||||
{
|
||||
getLogger().warn( "Unable to use proxy connector: " + e.getMessage(), e );
|
||||
log.warn( "Unable to use proxy connector: " + e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -19,6 +19,10 @@ package org.apache.maven.archiva.proxy;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.commons.io.FileUtils;
|
||||
import org.apache.maven.wagon.ConnectionException;
|
||||
import org.apache.maven.wagon.ResourceDoesNotExistException;
|
||||
|
@ -31,11 +35,8 @@ import org.apache.maven.wagon.events.SessionListener;
|
|||
import org.apache.maven.wagon.events.TransferListener;
|
||||
import org.apache.maven.wagon.proxy.ProxyInfo;
|
||||
import org.apache.maven.wagon.repository.Repository;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.util.List;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* A dummy wagon implementation
|
||||
|
@ -43,9 +44,10 @@ import java.util.List;
|
|||
* @author <a href="mailto:brett@apache.org">Brett Porter</a>
|
||||
*/
|
||||
public class WagonDelegate
|
||||
extends AbstractLogEnabled
|
||||
implements Wagon
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( WagonDelegate.class );
|
||||
|
||||
private Wagon delegate;
|
||||
|
||||
private String contentToGet;
|
||||
|
@ -53,7 +55,7 @@ public class WagonDelegate
|
|||
public void get( String resourceName, File destination )
|
||||
throws TransferFailedException, ResourceDoesNotExistException, AuthorizationException
|
||||
{
|
||||
getLogger().debug( ".get(" + resourceName + ", " + destination + ")" );
|
||||
log.debug( ".get(" + resourceName + ", " + destination + ")" );
|
||||
delegate.get( resourceName, destination );
|
||||
create( destination );
|
||||
}
|
||||
|
@ -61,7 +63,7 @@ public class WagonDelegate
|
|||
public boolean getIfNewer( String resourceName, File destination, long timestamp )
|
||||
throws TransferFailedException, ResourceDoesNotExistException, AuthorizationException
|
||||
{
|
||||
getLogger().info( ".getIfNewer(" + resourceName + ", " + destination + ", " + timestamp + ")" );
|
||||
log.info( ".getIfNewer(" + resourceName + ", " + destination + ", " + timestamp + ")" );
|
||||
|
||||
boolean result = delegate.getIfNewer( resourceName, destination, timestamp );
|
||||
createIfMissing( destination );
|
||||
|
|
|
@ -19,7 +19,8 @@ package org.apache.maven.archiva.repository.audit;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.apache.log4j.Logger;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* AuditLog - Audit Log.
|
||||
|
@ -33,7 +34,7 @@ import org.apache.log4j.Logger;
|
|||
public class AuditLog
|
||||
implements AuditListener
|
||||
{
|
||||
public static final Logger logger = Logger.getLogger( "org.apache.archiva.AuditLog" );
|
||||
public static final Logger logger = LoggerFactory.getLogger( "org.apache.archiva.AuditLog" );
|
||||
|
||||
private static final char DELIM = ' ';
|
||||
|
||||
|
|
|
@ -19,6 +19,8 @@ package org.apache.maven.archiva.repository.project;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
|
||||
import org.apache.maven.archiva.configuration.ConfigurationNames;
|
||||
|
@ -29,13 +31,12 @@ import org.apache.maven.archiva.repository.RepositoryException;
|
|||
import org.apache.maven.archiva.repository.project.resolvers.ManagedRepositoryProjectResolver;
|
||||
import org.apache.maven.archiva.repository.project.resolvers.NopProjectResolver;
|
||||
import org.apache.maven.archiva.repository.project.resolvers.ProjectModelResolverStack;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.codehaus.plexus.registry.Registry;
|
||||
import org.codehaus.plexus.registry.RegistryListener;
|
||||
|
||||
import java.util.List;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* Factory for ProjectModelResolver objects
|
||||
|
@ -45,9 +46,10 @@ import java.util.List;
|
|||
* @plexus.component role="org.apache.maven.archiva.repository.project.ProjectModelResolverFactory"
|
||||
*/
|
||||
public class ProjectModelResolverFactory
|
||||
extends AbstractLogEnabled
|
||||
implements RegistryListener, Initializable
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( ProjectModelResolverFactory.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement
|
||||
*/
|
||||
|
@ -128,7 +130,7 @@ public class ProjectModelResolverFactory
|
|||
}
|
||||
catch ( RepositoryException e )
|
||||
{
|
||||
getLogger().warn( e.getMessage(), e );
|
||||
log.warn( e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -19,6 +19,10 @@ package org.apache.maven.archiva.repository.scanner;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.io.File;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.maven.archiva.configuration.FileTypes;
|
||||
import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration;
|
||||
|
@ -26,12 +30,9 @@ import org.apache.maven.archiva.consumers.InvalidRepositoryContentConsumer;
|
|||
import org.apache.maven.archiva.consumers.KnownRepositoryContentConsumer;
|
||||
import org.apache.maven.archiva.consumers.RepositoryContentConsumer;
|
||||
import org.apache.maven.archiva.repository.RepositoryException;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.util.DirectoryWalker;
|
||||
|
||||
import java.io.File;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* DefaultRepositoryScanner
|
||||
|
@ -41,9 +42,10 @@ import java.util.List;
|
|||
* @plexus.component role="org.apache.maven.archiva.repository.scanner.RepositoryScanner"
|
||||
*/
|
||||
public class DefaultRepositoryScanner
|
||||
extends AbstractLogEnabled
|
||||
implements RepositoryScanner
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( DefaultRepositoryScanner.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement
|
||||
*/
|
||||
|
@ -112,7 +114,7 @@ public class DefaultRepositoryScanner
|
|||
|
||||
// Setup the Scan Instance
|
||||
RepositoryScannerInstance scannerInstance = new RepositoryScannerInstance( repository, knownContentConsumers,
|
||||
invalidContentConsumers, getLogger(), changesSince );
|
||||
invalidContentConsumers, changesSince );
|
||||
|
||||
dirWalker.addDirectoryWalkListener( scannerInstance );
|
||||
|
||||
|
|
|
@ -19,6 +19,9 @@ package org.apache.maven.archiva.repository.scanner;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.io.File;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.commons.collections.Closure;
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.commons.collections.functors.IfClosure;
|
||||
|
@ -30,12 +33,9 @@ import org.apache.maven.archiva.consumers.KnownRepositoryContentConsumer;
|
|||
import org.apache.maven.archiva.repository.scanner.functors.ConsumerProcessFileClosure;
|
||||
import org.apache.maven.archiva.repository.scanner.functors.ConsumerWantsFilePredicate;
|
||||
import org.apache.maven.archiva.repository.scanner.functors.TriggerBeginScanClosure;
|
||||
import org.codehaus.plexus.logging.Logger;
|
||||
import org.codehaus.plexus.util.DirectoryWalkListener;
|
||||
|
||||
import java.io.File;
|
||||
import java.util.List;
|
||||
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
/**
|
||||
* RepositoryScannerInstance
|
||||
*
|
||||
|
@ -45,6 +45,8 @@ import java.util.List;
|
|||
public class RepositoryScannerInstance
|
||||
implements DirectoryWalkListener
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( RepositoryScannerInstance.class );
|
||||
|
||||
/**
|
||||
* Consumers that process known content.
|
||||
*/
|
||||
|
@ -65,24 +67,21 @@ public class RepositoryScannerInstance
|
|||
|
||||
private ConsumerWantsFilePredicate consumerWantsFile;
|
||||
|
||||
private Logger logger;
|
||||
|
||||
public RepositoryScannerInstance( ManagedRepositoryConfiguration repository,
|
||||
List<KnownRepositoryContentConsumer> knownConsumerList,
|
||||
List<InvalidRepositoryContentConsumer> invalidConsumerList, Logger logger )
|
||||
List<InvalidRepositoryContentConsumer> invalidConsumerList )
|
||||
{
|
||||
this.repository = repository;
|
||||
this.knownConsumers = knownConsumerList;
|
||||
this.invalidConsumers = invalidConsumerList;
|
||||
this.logger = logger;
|
||||
|
||||
this.consumerProcessFile = new ConsumerProcessFileClosure( logger );
|
||||
this.consumerProcessFile = new ConsumerProcessFileClosure();
|
||||
this.consumerWantsFile = new ConsumerWantsFilePredicate();
|
||||
|
||||
stats = new RepositoryScanStatistics();
|
||||
stats.setRepositoryId( repository.getId() );
|
||||
|
||||
Closure triggerBeginScan = new TriggerBeginScanClosure( repository, logger );
|
||||
Closure triggerBeginScan = new TriggerBeginScanClosure( repository );
|
||||
|
||||
CollectionUtils.forAllDo( knownConsumerList, triggerBeginScan );
|
||||
CollectionUtils.forAllDo( invalidConsumerList, triggerBeginScan );
|
||||
|
@ -95,10 +94,9 @@ public class RepositoryScannerInstance
|
|||
|
||||
public RepositoryScannerInstance( ManagedRepositoryConfiguration repository,
|
||||
List<KnownRepositoryContentConsumer> knownContentConsumers,
|
||||
List<InvalidRepositoryContentConsumer> invalidContentConsumers, Logger logger,
|
||||
long changesSince )
|
||||
List<InvalidRepositoryContentConsumer> invalidContentConsumers, long changesSince )
|
||||
{
|
||||
this( repository, knownContentConsumers, invalidContentConsumers, logger );
|
||||
this( repository, knownContentConsumers, invalidContentConsumers );
|
||||
|
||||
consumerWantsFile.setChangesSince( changesSince );
|
||||
|
||||
|
@ -112,13 +110,13 @@ public class RepositoryScannerInstance
|
|||
|
||||
public void directoryWalkStarting( File basedir )
|
||||
{
|
||||
logger.info( "Walk Started: [" + this.repository.getId() + "] " + this.repository.getLocation() );
|
||||
log.info( "Walk Started: [" + this.repository.getId() + "] " + this.repository.getLocation() );
|
||||
stats.triggerStart();
|
||||
}
|
||||
|
||||
public void directoryWalkStep( int percentage, File file )
|
||||
{
|
||||
logger.debug( "Walk Step: " + percentage + ", " + file );
|
||||
log.debug( "Walk Step: " + percentage + ", " + file );
|
||||
|
||||
stats.increaseFileCount();
|
||||
|
||||
|
@ -146,7 +144,7 @@ public class RepositoryScannerInstance
|
|||
|
||||
public void directoryWalkFinished()
|
||||
{
|
||||
logger.info( "Walk Finished: [" + this.repository.getId() + "] " + this.repository.getLocation() );
|
||||
log.info( "Walk Finished: [" + this.repository.getId() + "] " + this.repository.getLocation() );
|
||||
stats.triggerFinished();
|
||||
}
|
||||
|
||||
|
@ -155,6 +153,6 @@ public class RepositoryScannerInstance
|
|||
*/
|
||||
public void debug( String message )
|
||||
{
|
||||
logger.debug( "Repository Scanner: " + message );
|
||||
log.debug( "Repository Scanner: " + message );
|
||||
}
|
||||
}
|
||||
|
|
|
@ -23,7 +23,8 @@ import org.apache.commons.collections.Closure;
|
|||
import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration;
|
||||
import org.apache.maven.archiva.consumers.ConsumerException;
|
||||
import org.apache.maven.archiva.consumers.RepositoryContentConsumer;
|
||||
import org.codehaus.plexus.logging.Logger;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* TriggerBeginScanClosure
|
||||
|
@ -34,14 +35,13 @@ import org.codehaus.plexus.logging.Logger;
|
|||
public class TriggerBeginScanClosure
|
||||
implements Closure
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( TriggerBeginScanClosure.class );
|
||||
|
||||
private ManagedRepositoryConfiguration repository;
|
||||
|
||||
private Logger logger;
|
||||
|
||||
public TriggerBeginScanClosure( ManagedRepositoryConfiguration repository, Logger logger )
|
||||
public TriggerBeginScanClosure( ManagedRepositoryConfiguration repository )
|
||||
{
|
||||
this.repository = repository;
|
||||
this.logger = logger;
|
||||
}
|
||||
|
||||
public void execute( Object input )
|
||||
|
@ -56,7 +56,7 @@ public class TriggerBeginScanClosure
|
|||
}
|
||||
catch ( ConsumerException e )
|
||||
{
|
||||
logger.warn( "Consumer [" + consumer.getId() + "] cannot begin: " + e.getMessage(), e );
|
||||
log.warn( "Consumer [" + consumer.getId() + "] cannot begin: " + e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -21,7 +21,6 @@ package org.apache.maven.archiva.transaction;
|
|||
|
||||
import org.codehaus.plexus.digest.Digester;
|
||||
import org.codehaus.plexus.digest.DigesterException;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.util.FileUtils;
|
||||
import org.codehaus.plexus.util.IOUtil;
|
||||
|
||||
|
@ -43,7 +42,6 @@ import java.util.Map;
|
|||
* @version $Id$
|
||||
*/
|
||||
public abstract class AbstractTransactionEvent
|
||||
extends AbstractLogEnabled
|
||||
implements TransactionEvent
|
||||
{
|
||||
private Map backups = new HashMap();
|
||||
|
|
|
@ -19,6 +19,12 @@ package org.apache.maven.archiva.database.browsing;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.util.Collections;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.commons.collections.PredicateUtils;
|
||||
import org.apache.commons.collections.functors.NotPredicate;
|
||||
|
@ -35,13 +41,8 @@ import org.apache.maven.archiva.database.updater.DatabaseUpdater;
|
|||
import org.apache.maven.archiva.model.ArchivaArtifact;
|
||||
import org.apache.maven.archiva.model.ArchivaProjectModel;
|
||||
import org.apache.maven.archiva.model.Keys;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.util.Collections;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* DefaultRepositoryBrowsing
|
||||
|
@ -51,9 +52,10 @@ import java.util.Map.Entry;
|
|||
* @plexus.component role="org.apache.maven.archiva.database.browsing.RepositoryBrowsing"
|
||||
*/
|
||||
public class DefaultRepositoryBrowsing
|
||||
extends AbstractLogEnabled
|
||||
implements RepositoryBrowsing
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( DefaultRepositoryBrowsing.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement role-hint="jdo"
|
||||
*/
|
||||
|
@ -201,7 +203,7 @@ public class DefaultRepositoryBrowsing
|
|||
{
|
||||
Map<String, String> snapshots = new HashMap<String, String>();
|
||||
|
||||
getLogger().info( "Processing snapshots." );
|
||||
log.info( "Processing snapshots." );
|
||||
|
||||
for ( String version : versions )
|
||||
{
|
||||
|
@ -253,7 +255,7 @@ public class DefaultRepositoryBrowsing
|
|||
{
|
||||
if ( VersionUtil.getBaseVersion( uniqueVersion ).equals( version ) )
|
||||
{
|
||||
getLogger().info( "Retrieving artifact with version " + uniqueVersion );
|
||||
log.info( "Retrieving artifact with version " + uniqueVersion );
|
||||
pomArtifact = dao.getArtifactDAO().getArtifact( groupId, artifactId, uniqueVersion, null, "pom" );
|
||||
|
||||
return pomArtifact;
|
||||
|
|
|
@ -28,7 +28,6 @@ import org.apache.maven.archiva.database.SimpleConstraint;
|
|||
import org.apache.maven.archiva.database.constraints.AbstractSimpleConstraint;
|
||||
import org.apache.maven.archiva.model.CompoundKey;
|
||||
import org.codehaus.plexus.jdo.JdoFactory;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
|
||||
|
@ -60,7 +59,6 @@ import javax.jdo.spi.PersistenceCapable;
|
|||
* @plexus.component role="org.apache.maven.archiva.database.jdo.JdoAccess" role-hint="archiva"
|
||||
*/
|
||||
public class JdoAccess
|
||||
extends AbstractLogEnabled
|
||||
implements Initializable, InstanceLifecycleListener, StoreLifecycleListener
|
||||
{
|
||||
/**
|
||||
|
|
|
@ -25,7 +25,6 @@ import org.apache.maven.archiva.database.ProjectModelDAO;
|
|||
import org.apache.maven.archiva.database.RepositoryContentStatisticsDAO;
|
||||
import org.apache.maven.archiva.database.RepositoryProblemDAO;
|
||||
import org.apache.maven.archiva.database.SimpleConstraint;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.io.Serializable;
|
||||
import java.util.List;
|
||||
|
@ -39,7 +38,6 @@ import java.util.List;
|
|||
* @plexus.component role-hint="jdo"
|
||||
*/
|
||||
public class JdoArchivaDAO
|
||||
extends AbstractLogEnabled
|
||||
implements ArchivaDAO
|
||||
{
|
||||
/**
|
||||
|
|
|
@ -26,7 +26,6 @@ import org.apache.maven.archiva.database.ObjectNotFoundException;
|
|||
import org.apache.maven.archiva.model.ArchivaArtifact;
|
||||
import org.apache.maven.archiva.model.ArchivaArtifactModel;
|
||||
import org.apache.maven.archiva.model.jpox.ArchivaArtifactModelKey;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Iterator;
|
||||
|
@ -41,7 +40,6 @@ import java.util.List;
|
|||
* @plexus.component role-hint="jdo"
|
||||
*/
|
||||
public class JdoArtifactDAO
|
||||
extends AbstractLogEnabled
|
||||
implements ArtifactDAO
|
||||
{
|
||||
/**
|
||||
|
|
|
@ -19,6 +19,8 @@ package org.apache.maven.archiva.database.project;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.maven.archiva.database.ArchivaDAO;
|
||||
import org.apache.maven.archiva.database.ArchivaDatabaseException;
|
||||
import org.apache.maven.archiva.database.ObjectNotFoundException;
|
||||
|
@ -28,9 +30,8 @@ import org.apache.maven.archiva.repository.project.ProjectModelException;
|
|||
import org.apache.maven.archiva.repository.project.ProjectModelResolver;
|
||||
import org.apache.maven.archiva.repository.project.resolvers.FilesystemBasedResolver;
|
||||
import org.apache.maven.archiva.repository.project.resolvers.ProjectModelResolutionListener;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.util.List;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* Just in Time save of project models to the database, implemented as a listener
|
||||
|
@ -44,9 +45,10 @@ import java.util.List;
|
|||
* role-hint="model-to-db"
|
||||
*/
|
||||
public class ProjectModelToDatabaseListener
|
||||
extends AbstractLogEnabled
|
||||
implements ProjectModelResolutionListener
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( ProjectModelToDatabaseListener.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement role-hint="jdo"
|
||||
*/
|
||||
|
@ -147,7 +149,7 @@ public class ProjectModelToDatabaseListener
|
|||
}
|
||||
catch ( ProjectModelException e )
|
||||
{
|
||||
getLogger().warn( e.getMessage(), e );
|
||||
log.warn( e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,6 +19,10 @@ package org.apache.maven.archiva.database.updater;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.util.Date;
|
||||
import java.util.Iterator;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.commons.collections.IteratorUtils;
|
||||
import org.apache.commons.collections.Predicate;
|
||||
|
@ -29,11 +33,8 @@ import org.apache.maven.archiva.database.ArchivaDatabaseException;
|
|||
import org.apache.maven.archiva.database.constraints.ArtifactsProcessedConstraint;
|
||||
import org.apache.maven.archiva.model.ArchivaArtifact;
|
||||
import org.apache.maven.archiva.model.functors.UnprocessedArtifactPredicate;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.util.Date;
|
||||
import java.util.Iterator;
|
||||
import java.util.List;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* JdoDatabaseUpdater
|
||||
|
@ -45,9 +46,10 @@ import java.util.List;
|
|||
* role-hint="jdo"
|
||||
*/
|
||||
public class JdoDatabaseUpdater
|
||||
extends AbstractLogEnabled
|
||||
implements DatabaseUpdater
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( JdoDatabaseUpdater.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement role-hint="jdo"
|
||||
*/
|
||||
|
@ -144,7 +146,7 @@ public class JdoDatabaseUpdater
|
|||
|
||||
if ( CollectionUtils.isEmpty( consumers ) )
|
||||
{
|
||||
getLogger().warn( "There are no selected consumers for unprocessed artifacts." );
|
||||
log.warn( "There are no selected consumers for unprocessed artifacts." );
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -162,7 +164,7 @@ public class JdoDatabaseUpdater
|
|||
|
||||
if ( CollectionUtils.isEmpty( consumers ) )
|
||||
{
|
||||
getLogger().warn( "There are no selected consumers for artifact cleanup." );
|
||||
log.warn( "There are no selected consumers for artifact cleanup." );
|
||||
return;
|
||||
}
|
||||
|
||||
|
|
|
@ -23,7 +23,8 @@ import org.apache.commons.collections.Closure;
|
|||
import org.apache.maven.archiva.consumers.ArchivaArtifactConsumer;
|
||||
import org.apache.maven.archiva.consumers.ConsumerException;
|
||||
import org.apache.maven.archiva.model.ArchivaArtifact;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* ProcessArchivaArtifactClosure
|
||||
|
@ -36,9 +37,10 @@ import org.codehaus.plexus.logging.AbstractLogEnabled;
|
|||
* instantiation-strategy="per-lookup"
|
||||
*/
|
||||
class ProcessArchivaArtifactClosure
|
||||
extends AbstractLogEnabled
|
||||
implements Closure
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( ProcessArchivaArtifactClosure.class );
|
||||
|
||||
private ArchivaArtifact artifact;
|
||||
|
||||
public void execute( Object input )
|
||||
|
@ -53,9 +55,7 @@ class ProcessArchivaArtifactClosure
|
|||
}
|
||||
catch ( ConsumerException e )
|
||||
{
|
||||
getLogger().warn(
|
||||
"Unable to process artifact [" + artifact + "] with consumer [" + consumer.getId()
|
||||
+ "]" );
|
||||
log.warn( "Unable to process artifact [" + artifact + "] with consumer [" + consumer.getId() + "]" );
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -19,8 +19,6 @@ package org.apache.maven.archiva.reporting;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import java.util.Map;
|
||||
|
||||
/**
|
||||
|
@ -33,7 +31,6 @@ import java.util.Map;
|
|||
* role-hint="default"
|
||||
*/
|
||||
public class DefaultReportingManager
|
||||
extends AbstractLogEnabled
|
||||
implements ReportingManager
|
||||
{
|
||||
/**
|
||||
|
|
|
@ -19,6 +19,9 @@ package org.apache.maven.archiva.scheduled;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.text.ParseException;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.maven.archiva.common.ArchivaException;
|
||||
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
|
||||
|
@ -27,7 +30,6 @@ import org.apache.maven.archiva.scheduled.tasks.ArchivaTask;
|
|||
import org.apache.maven.archiva.scheduled.tasks.DatabaseTask;
|
||||
import org.apache.maven.archiva.scheduled.tasks.RepositoryTask;
|
||||
import org.apache.maven.archiva.scheduled.tasks.RepositoryTaskSelectionPredicate;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Startable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.StartingException;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.StoppingException;
|
||||
|
@ -43,9 +45,8 @@ import org.quartz.CronTrigger;
|
|||
import org.quartz.JobDataMap;
|
||||
import org.quartz.JobDetail;
|
||||
import org.quartz.SchedulerException;
|
||||
|
||||
import java.text.ParseException;
|
||||
import java.util.List;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* Default implementation of a scheduling component for archiva.
|
||||
|
@ -55,9 +56,10 @@ import java.util.List;
|
|||
* @plexus.component role="org.apache.maven.archiva.scheduled.ArchivaTaskScheduler" role-hint="default"
|
||||
*/
|
||||
public class DefaultArchivaTaskScheduler
|
||||
extends AbstractLogEnabled
|
||||
implements ArchivaTaskScheduler, Startable, RegistryListener
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( DefaultArchivaTaskScheduler.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement
|
||||
*/
|
||||
|
@ -134,7 +136,7 @@ public class DefaultArchivaTaskScheduler
|
|||
{
|
||||
if ( repoConfig.getRefreshCronExpression() == null )
|
||||
{
|
||||
getLogger().warn( "Skipping job, no cron expression for " + repoConfig.getId() );
|
||||
log.warn( "Skipping job, no cron expression for " + repoConfig.getId() );
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -144,7 +146,7 @@ public class DefaultArchivaTaskScheduler
|
|||
CronExpressionValidator cronValidator = new CronExpressionValidator();
|
||||
if ( !cronValidator.validate( cronString ) )
|
||||
{
|
||||
getLogger().warn( "Cron expression [" + cronString + "] for repository [" + repoConfig.getId() +
|
||||
log.warn( "Cron expression [" + cronString + "] for repository [" + repoConfig.getId() +
|
||||
"] is invalid. Defaulting to hourly." );
|
||||
cronString = CRON_HOURLY;
|
||||
}
|
||||
|
@ -168,7 +170,7 @@ public class DefaultArchivaTaskScheduler
|
|||
}
|
||||
catch ( ParseException e )
|
||||
{
|
||||
getLogger().error(
|
||||
log.error(
|
||||
"ParseException in repository scanning cron expression, disabling repository scanning for '" +
|
||||
repoConfig.getId() + "': " + e.getMessage() );
|
||||
}
|
||||
|
@ -190,7 +192,7 @@ public class DefaultArchivaTaskScheduler
|
|||
CronExpressionValidator cronValidator = new CronExpressionValidator();
|
||||
if ( !cronValidator.validate( cronString ) )
|
||||
{
|
||||
getLogger().warn(
|
||||
log.warn(
|
||||
"Cron expression [" + cronString + "] for database update is invalid. Defaulting to hourly." );
|
||||
cronString = CRON_HOURLY;
|
||||
}
|
||||
|
@ -203,7 +205,7 @@ public class DefaultArchivaTaskScheduler
|
|||
}
|
||||
catch ( ParseException e )
|
||||
{
|
||||
getLogger().error(
|
||||
log.error(
|
||||
"ParseException in database scanning cron expression, disabling database scanning: " + e.getMessage() );
|
||||
}
|
||||
|
||||
|
@ -235,7 +237,7 @@ public class DefaultArchivaTaskScheduler
|
|||
// cronExpression comes from the database scanning section
|
||||
if ( "cronExpression".equals( propertyName ) )
|
||||
{
|
||||
getLogger().debug( "Restarting the database scheduled task after property change: " + propertyName );
|
||||
log.debug( "Restarting the database scheduled task after property change: " + propertyName );
|
||||
|
||||
try
|
||||
{
|
||||
|
@ -245,7 +247,7 @@ public class DefaultArchivaTaskScheduler
|
|||
}
|
||||
catch ( SchedulerException e )
|
||||
{
|
||||
getLogger().error( "Error restarting the database scanning job after property change." );
|
||||
log.error( "Error restarting the database scanning job after property change." );
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -269,7 +271,7 @@ public class DefaultArchivaTaskScheduler
|
|||
}
|
||||
catch ( SchedulerException e )
|
||||
{
|
||||
getLogger().error( "error restarting job: " + REPOSITORY_JOB + ":" + repoConfig.getId() );
|
||||
log.error( "error restarting job: " + REPOSITORY_JOB + ":" + repoConfig.getId() );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -22,12 +22,13 @@ package org.apache.maven.archiva.scheduled.executors;
|
|||
import org.apache.maven.archiva.database.ArchivaDatabaseException;
|
||||
import org.apache.maven.archiva.database.updater.DatabaseUpdater;
|
||||
import org.apache.maven.archiva.scheduled.tasks.DatabaseTask;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.codehaus.plexus.taskqueue.Task;
|
||||
import org.codehaus.plexus.taskqueue.execution.TaskExecutionException;
|
||||
import org.codehaus.plexus.taskqueue.execution.TaskExecutor;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* ArchivaDatabaseTaskExecutor
|
||||
|
@ -40,9 +41,10 @@ import org.codehaus.plexus.taskqueue.execution.TaskExecutor;
|
|||
* role-hint="database-update"
|
||||
*/
|
||||
public class ArchivaDatabaseUpdateTaskExecutor
|
||||
extends AbstractLogEnabled
|
||||
implements TaskExecutor, Initializable
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( ArchivaDatabaseUpdateTaskExecutor.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement role-hint="jdo"
|
||||
*/
|
||||
|
@ -51,7 +53,7 @@ public class ArchivaDatabaseUpdateTaskExecutor
|
|||
public void initialize()
|
||||
throws InitializationException
|
||||
{
|
||||
getLogger().info( "Initialized " + this.getClass().getName() );
|
||||
log.info( "Initialized " + this.getClass().getName() );
|
||||
}
|
||||
|
||||
public void executeTask( Task task )
|
||||
|
@ -59,12 +61,12 @@ public class ArchivaDatabaseUpdateTaskExecutor
|
|||
{
|
||||
DatabaseTask dbtask = (DatabaseTask) task;
|
||||
|
||||
getLogger().info( "Executing task from queue with job name: " + dbtask.getName() );
|
||||
log.info( "Executing task from queue with job name: " + dbtask.getName() );
|
||||
long time = System.currentTimeMillis();
|
||||
|
||||
try
|
||||
{
|
||||
getLogger().info( "Task: Updating unprocessed artifacts" );
|
||||
log.info( "Task: Updating unprocessed artifacts" );
|
||||
databaseUpdater.updateAllUnprocessed();
|
||||
}
|
||||
catch ( ArchivaDatabaseException e )
|
||||
|
@ -74,7 +76,7 @@ public class ArchivaDatabaseUpdateTaskExecutor
|
|||
|
||||
try
|
||||
{
|
||||
getLogger().info( "Task: Updating processed artifacts" );
|
||||
log.info( "Task: Updating processed artifacts" );
|
||||
databaseUpdater.updateAllProcessed();
|
||||
}
|
||||
catch ( ArchivaDatabaseException e )
|
||||
|
@ -84,6 +86,6 @@ public class ArchivaDatabaseUpdateTaskExecutor
|
|||
|
||||
time = System.currentTimeMillis() - time;
|
||||
|
||||
getLogger().info( "Finished database task in " + time + "ms." );
|
||||
log.info( "Finished database task in " + time + "ms." );
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,26 +19,26 @@ package org.apache.maven.archiva.scheduled.executors;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
|
||||
import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration;
|
||||
import org.apache.maven.archiva.database.ArchivaDAO;
|
||||
import org.apache.maven.archiva.database.ArchivaDatabaseException;
|
||||
import org.apache.maven.archiva.database.constraints.MostRecentRepositoryScanStatistics;
|
||||
import org.apache.maven.archiva.model.RepositoryContentStatistics;
|
||||
import org.apache.maven.archiva.repository.RepositoryException;
|
||||
import org.apache.maven.archiva.repository.scanner.RepositoryScanStatistics;
|
||||
import org.apache.maven.archiva.repository.scanner.RepositoryScanner;
|
||||
import org.apache.maven.archiva.scheduled.tasks.RepositoryTask;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.codehaus.plexus.taskqueue.Task;
|
||||
import org.codehaus.plexus.taskqueue.execution.TaskExecutionException;
|
||||
import org.codehaus.plexus.taskqueue.execution.TaskExecutor;
|
||||
|
||||
import java.util.List;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* ArchivaRepositoryScanningTaskExecutor
|
||||
|
@ -51,9 +51,10 @@ import java.util.List;
|
|||
* role-hint="repository-scanning"
|
||||
*/
|
||||
public class ArchivaRepositoryScanningTaskExecutor
|
||||
extends AbstractLogEnabled
|
||||
implements TaskExecutor, Initializable
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( ArchivaRepositoryScanningTaskExecutor.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement role-hint="jdo"
|
||||
*/
|
||||
|
@ -74,7 +75,7 @@ public class ArchivaRepositoryScanningTaskExecutor
|
|||
public void initialize()
|
||||
throws InitializationException
|
||||
{
|
||||
getLogger().info( "Initialized " + this.getClass().getName() );
|
||||
log.info( "Initialized " + this.getClass().getName() );
|
||||
}
|
||||
|
||||
public void executeTask( Task task )
|
||||
|
@ -87,7 +88,7 @@ public class ArchivaRepositoryScanningTaskExecutor
|
|||
throw new TaskExecutionException("Unable to execute RepositoryTask with blank repository Id.");
|
||||
}
|
||||
|
||||
getLogger().info( "Executing task from queue with job name: " + repoTask.getName() );
|
||||
log.info( "Executing task from queue with job name: " + repoTask.getName() );
|
||||
|
||||
try
|
||||
{
|
||||
|
@ -105,7 +106,7 @@ public class ArchivaRepositoryScanningTaskExecutor
|
|||
|
||||
RepositoryScanStatistics stats = repoScanner.scan( arepo, sinceWhen );
|
||||
|
||||
getLogger().info( "Finished repository task: " + stats.toDump( arepo ) );
|
||||
log.info( "Finished repository task: " + stats.toDump( arepo ) );
|
||||
|
||||
// I hate jpox and modello
|
||||
RepositoryContentStatistics dbstats = new RepositoryContentStatistics();
|
||||
|
|
|
@ -19,11 +19,12 @@ package org.apache.maven.archiva.security;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import java.util.List;
|
||||
|
||||
import org.codehaus.plexus.redback.rbac.RBACManager;
|
||||
import org.codehaus.plexus.redback.system.check.EnvironmentCheck;
|
||||
|
||||
import java.util.List;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* ArchivaStandardRolesCheck tests for the existance of expected / standard roles and permissions.
|
||||
|
@ -35,9 +36,10 @@ import java.util.List;
|
|||
* role-hint="required-roles"
|
||||
*/
|
||||
public class ArchivaStandardRolesCheck
|
||||
extends AbstractLogEnabled
|
||||
implements EnvironmentCheck
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( ArchivaStandardRolesCheck.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement role-hint="cached"
|
||||
*/
|
||||
|
@ -61,7 +63,7 @@ public class ArchivaStandardRolesCheck
|
|||
ArchivaRoleConstants.REGISTERED_USER_ROLE,
|
||||
ArchivaRoleConstants.USER_ADMINISTRATOR_ROLE };
|
||||
|
||||
getLogger().info( "Checking the existance of required roles." );
|
||||
log.info( "Checking the existance of required roles." );
|
||||
|
||||
for ( String roleName : expectedRoles )
|
||||
{
|
||||
|
@ -85,7 +87,7 @@ public class ArchivaStandardRolesCheck
|
|||
ArchivaRoleConstants.OPERATION_REPOSITORY_ACCESS,
|
||||
"archiva-guest" };
|
||||
|
||||
getLogger().info( "Checking the existance of required operations." );
|
||||
log.info( "Checking the existance of required operations." );
|
||||
|
||||
for ( String operation : expectedOperations )
|
||||
{
|
||||
|
|
|
@ -19,23 +19,24 @@ package org.apache.maven.archiva.security;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.maven.archiva.common.ArchivaException;
|
||||
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
|
||||
import org.apache.maven.archiva.configuration.ConfigurationNames;
|
||||
import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.redback.rbac.RBACManager;
|
||||
import org.codehaus.plexus.redback.rbac.RbacManagerException;
|
||||
import org.codehaus.plexus.redback.rbac.UserAssignment;
|
||||
import org.codehaus.plexus.redback.system.check.EnvironmentCheck;
|
||||
import org.codehaus.plexus.registry.Registry;
|
||||
import org.codehaus.plexus.registry.RegistryListener;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* SecurityStartup
|
||||
|
@ -46,9 +47,10 @@ import java.util.Map.Entry;
|
|||
* @plexus.component role="org.apache.maven.archiva.security.SecurityStartup"
|
||||
*/
|
||||
public class SecurityStartup
|
||||
extends AbstractLogEnabled
|
||||
implements RegistryListener
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( SecurityStartup.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement
|
||||
*/
|
||||
|
@ -104,7 +106,7 @@ public class SecurityStartup
|
|||
}
|
||||
catch ( RbacManagerException e )
|
||||
{
|
||||
getLogger().warn(
|
||||
log.warn(
|
||||
"Unable to add role [" + ArchivaRoleConstants.toRepositoryObserverRoleName( repoId )
|
||||
+ "] to " + principal + " user.", e );
|
||||
}
|
||||
|
@ -129,7 +131,7 @@ public class SecurityStartup
|
|||
}
|
||||
catch ( ArchivaSecurityException e )
|
||||
{
|
||||
getLogger().warn( e.getMessage(), e );
|
||||
log.warn( e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -145,7 +147,7 @@ public class SecurityStartup
|
|||
}
|
||||
catch ( ArchivaSecurityException e )
|
||||
{
|
||||
getLogger().warn( e.getMessage(), e );
|
||||
log.warn( e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -164,7 +166,7 @@ public class SecurityStartup
|
|||
for ( Entry<String, EnvironmentCheck> entry : checkers.entrySet() )
|
||||
{
|
||||
EnvironmentCheck check = entry.getValue();
|
||||
getLogger().info( "Running Environment Check: " + entry.getKey() );
|
||||
log.info( "Running Environment Check: " + entry.getKey() );
|
||||
check.validateEnvironment( violations );
|
||||
}
|
||||
|
||||
|
@ -183,7 +185,7 @@ public class SecurityStartup
|
|||
|
||||
msg.append( "\n" );
|
||||
msg.append( "======================================================================" );
|
||||
getLogger().fatalError( msg.toString() );
|
||||
log.error( msg.toString() );
|
||||
|
||||
throw new ArchivaException( "Unable to initialize Redback Security Environment, [" + violations.size()
|
||||
+ "] violation(s) encountered, See log for details." );
|
||||
|
|
|
@ -23,7 +23,6 @@ import com.opensymphony.webwork.ServletActionContext;
|
|||
import com.opensymphony.xwork.ActionInvocation;
|
||||
import com.opensymphony.xwork.interceptor.Interceptor;
|
||||
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
import javax.servlet.ServletContext;
|
||||
|
||||
|
@ -35,7 +34,6 @@ import javax.servlet.ServletContext;
|
|||
* role-hint="configurationInterceptor"
|
||||
*/
|
||||
public class ConfigurationInterceptor
|
||||
extends AbstractLogEnabled
|
||||
implements Interceptor
|
||||
{
|
||||
/**
|
||||
|
|
|
@ -21,9 +21,10 @@ package org.apache.maven.archiva.web.startup;
|
|||
|
||||
import org.apache.maven.archiva.common.ArchivaException;
|
||||
import org.apache.maven.archiva.scheduled.ArchivaTaskScheduler;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* ArchivaStartup - the startup of all archiva features in a deterministic order.
|
||||
|
@ -36,7 +37,6 @@ import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationExce
|
|||
* role-hint="default"
|
||||
*/
|
||||
public class ArchivaStartup
|
||||
extends AbstractLogEnabled
|
||||
implements Initializable
|
||||
{
|
||||
/**
|
||||
|
@ -57,7 +57,7 @@ public class ArchivaStartup
|
|||
public void initialize()
|
||||
throws InitializationException
|
||||
{
|
||||
Banner.display( getLogger(), ArchivaVersion.determineVersion( this.getClass().getClassLoader() ) );
|
||||
Banner.display( ArchivaVersion.determineVersion( this.getClass().getClassLoader() ) );
|
||||
|
||||
try
|
||||
{
|
||||
|
|
|
@ -19,14 +19,13 @@ package org.apache.maven.archiva.web.startup;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.logging.Logger;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
|
||||
import java.util.regex.Matcher;
|
||||
import java.util.regex.Pattern;
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* Banner
|
||||
*
|
||||
|
@ -38,7 +37,6 @@ import java.util.regex.Pattern;
|
|||
* role-hint="default"
|
||||
*/
|
||||
public class Banner
|
||||
extends AbstractLogEnabled
|
||||
{
|
||||
public static String encode( String raw )
|
||||
{
|
||||
|
@ -224,15 +222,15 @@ public class Banner
|
|||
return injectVersion( decode( encodedBanner ), version );
|
||||
}
|
||||
|
||||
public static void display( Logger logger, String version )
|
||||
public static void display( String version )
|
||||
{
|
||||
String banner = getBanner( version );
|
||||
logger.info( StringUtils.repeat( "_", 25 ) + "\n" + banner );
|
||||
LoggerFactory.getLogger( Banner.class ).info( StringUtils.repeat( "_", 25 ) + "\n" + banner );
|
||||
}
|
||||
|
||||
public void initialize()
|
||||
throws InitializationException
|
||||
{
|
||||
Banner.display( getLogger(), ArchivaVersion.determineVersion( this.getClass().getClassLoader() ) );
|
||||
Banner.display( ArchivaVersion.determineVersion( this.getClass().getClassLoader() ) );
|
||||
}
|
||||
}
|
||||
|
|
|
@ -23,7 +23,6 @@ import org.apache.maven.archiva.common.ArchivaException;
|
|||
import org.apache.maven.archiva.database.project.ProjectModelToDatabaseListener;
|
||||
import org.apache.maven.archiva.repository.project.ProjectModelResolver;
|
||||
import org.apache.maven.archiva.repository.project.ProjectModelResolverFactory;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
|
||||
/**
|
||||
* ResolverFactoryInit - Initialize the Resolver Factory, and hook it up to
|
||||
|
@ -37,7 +36,6 @@ import org.codehaus.plexus.logging.AbstractLogEnabled;
|
|||
* role-hint="default"
|
||||
*/
|
||||
public class ResolverFactoryInit
|
||||
extends AbstractLogEnabled
|
||||
{
|
||||
/**
|
||||
* @plexus.requirement role-hint="database"
|
||||
|
|
|
@ -19,13 +19,17 @@ package org.apache.maven.archiva.web.startup;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
|
||||
import org.apache.commons.collections.CollectionUtils;
|
||||
import org.apache.maven.archiva.common.ArchivaException;
|
||||
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
|
||||
import org.apache.maven.archiva.configuration.ConfigurationNames;
|
||||
import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration;
|
||||
import org.apache.maven.archiva.security.ArchivaRoleConstants;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.redback.rbac.RBACManager;
|
||||
import org.codehaus.plexus.redback.rbac.RbacManagerException;
|
||||
import org.codehaus.plexus.redback.rbac.UserAssignment;
|
||||
|
@ -34,11 +38,8 @@ import org.codehaus.plexus.redback.role.RoleManagerException;
|
|||
import org.codehaus.plexus.redback.system.check.EnvironmentCheck;
|
||||
import org.codehaus.plexus.registry.Registry;
|
||||
import org.codehaus.plexus.registry.RegistryListener;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Map.Entry;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* ConfigurationSynchronization
|
||||
|
@ -50,9 +51,10 @@ import java.util.Map.Entry;
|
|||
* role-hint="default"
|
||||
*/
|
||||
public class SecuritySynchronization
|
||||
extends AbstractLogEnabled
|
||||
implements RegistryListener
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( SecuritySynchronization.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement role-hint="default"
|
||||
*/
|
||||
|
@ -112,7 +114,7 @@ public class SecuritySynchronization
|
|||
catch ( RoleManagerException e )
|
||||
{
|
||||
// Log error.
|
||||
getLogger().error( "Unable to create roles for configured repositories: " + e.getMessage(), e );
|
||||
log.error( "Unable to create roles for configured repositories: " + e.getMessage(), e );
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -145,7 +147,7 @@ public class SecuritySynchronization
|
|||
for ( Entry<String, EnvironmentCheck> entry : checkers.entrySet() )
|
||||
{
|
||||
EnvironmentCheck check = entry.getValue();
|
||||
getLogger().info( "Running Environment Check: " + entry.getKey() );
|
||||
log.info( "Running Environment Check: " + entry.getKey() );
|
||||
check.validateEnvironment( violations );
|
||||
}
|
||||
|
||||
|
@ -164,7 +166,7 @@ public class SecuritySynchronization
|
|||
|
||||
msg.append( "\n" );
|
||||
msg.append( "======================================================================" );
|
||||
getLogger().fatalError( msg.toString() );
|
||||
log.error( msg.toString() );
|
||||
|
||||
throw new ArchivaException( "Unable to initialize Redback Security Environment, [" + violations.size()
|
||||
+ "] violation(s) encountered, See log for details." );
|
||||
|
@ -198,7 +200,7 @@ public class SecuritySynchronization
|
|||
}
|
||||
catch ( RbacManagerException e )
|
||||
{
|
||||
getLogger().warn( "Unable to add role [" + ArchivaRoleConstants.toRepositoryObserverRoleName( repoId )
|
||||
log.warn( "Unable to add role [" + ArchivaRoleConstants.toRepositoryObserverRoleName( repoId )
|
||||
+ "] to " + principal + " user.", e );
|
||||
}
|
||||
}
|
||||
|
|
|
@ -19,15 +19,17 @@ package org.apache.maven.archiva.web.tags;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import org.apache.commons.lang.StringEscapeUtils;
|
||||
import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration;
|
||||
import org.apache.maven.archiva.web.util.ContextUtils;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import java.io.IOException;
|
||||
|
||||
import javax.servlet.jsp.JspException;
|
||||
import javax.servlet.jsp.JspWriter;
|
||||
import javax.servlet.jsp.PageContext;
|
||||
import java.io.IOException;
|
||||
|
||||
import org.apache.commons.lang.StringEscapeUtils;
|
||||
import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration;
|
||||
import org.apache.maven.archiva.web.util.ContextUtils;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* CopyPasteSnippet
|
||||
|
@ -37,8 +39,9 @@ import java.io.IOException;
|
|||
* @plexus.component role="org.apache.maven.archiva.web.tags.CopyPasteSnippet"
|
||||
*/
|
||||
public class CopyPasteSnippet
|
||||
extends AbstractLogEnabled
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( CopyPasteSnippet.class );
|
||||
|
||||
public static final String PRE = "pre";
|
||||
|
||||
public static final String TOGGLE = "toggle";
|
||||
|
@ -53,7 +56,7 @@ public class CopyPasteSnippet
|
|||
if ( o == null )
|
||||
{
|
||||
buf.append( "Error generating snippet." );
|
||||
getLogger().error( "Unable to generate snippet for null object." );
|
||||
log.error( "Unable to generate snippet for null object." );
|
||||
}
|
||||
else if ( o instanceof ManagedRepositoryConfiguration )
|
||||
{
|
||||
|
|
|
@ -19,6 +19,12 @@ package org.apache.maven.archiva.web.tags;
|
|||
* under the License.
|
||||
*/
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Stack;
|
||||
|
||||
import javax.servlet.jsp.PageContext;
|
||||
|
||||
import org.apache.commons.lang.StringUtils;
|
||||
import org.apache.maven.archiva.common.ArchivaException;
|
||||
import org.apache.maven.archiva.dependency.DependencyGraphFactory;
|
||||
|
@ -34,15 +40,10 @@ import org.apache.maven.archiva.model.ArtifactReference;
|
|||
import org.apache.maven.archiva.model.DependencyScope;
|
||||
import org.apache.maven.archiva.model.Keys;
|
||||
import org.apache.maven.archiva.model.VersionedReference;
|
||||
import org.codehaus.plexus.logging.AbstractLogEnabled;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
|
||||
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
import java.util.Stack;
|
||||
|
||||
import javax.servlet.jsp.PageContext;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
/**
|
||||
* DependencyTree
|
||||
|
@ -53,9 +54,10 @@ import javax.servlet.jsp.PageContext;
|
|||
* @plexus.component role="org.apache.maven.archiva.web.tags.DependencyTree"
|
||||
*/
|
||||
public class DependencyTree
|
||||
extends AbstractLogEnabled
|
||||
implements Initializable
|
||||
{
|
||||
private Logger log = LoggerFactory.getLogger( DependencyTree.class );
|
||||
|
||||
/**
|
||||
* @plexus.requirement
|
||||
* role="org.apache.maven.archiva.dependency.graph.DependencyGraphBuilder"
|
||||
|
@ -121,7 +123,7 @@ public class DependencyTree
|
|||
{
|
||||
String emsg = "Error generating dependency tree [" + Keys.toKey( groupId, artifactId, modelVersion )
|
||||
+ "]: groupId is blank.";
|
||||
getLogger().error( emsg );
|
||||
log.error( emsg );
|
||||
throw new ArchivaException( emsg );
|
||||
}
|
||||
|
||||
|
@ -129,7 +131,7 @@ public class DependencyTree
|
|||
{
|
||||
String emsg = "Error generating dependency tree [" + Keys.toKey( groupId, artifactId, modelVersion )
|
||||
+ "]: artifactId is blank.";
|
||||
getLogger().error( emsg );
|
||||
log.error( emsg );
|
||||
throw new ArchivaException( emsg );
|
||||
}
|
||||
|
||||
|
@ -137,7 +139,7 @@ public class DependencyTree
|
|||
{
|
||||
String emsg = "Error generating dependency tree [" + Keys.toKey( groupId, artifactId, modelVersion )
|
||||
+ "]: version is blank.";
|
||||
getLogger().error( emsg );
|
||||
log.error( emsg );
|
||||
throw new ArchivaException( emsg );
|
||||
}
|
||||
|
||||
|
@ -249,7 +251,7 @@ public class DependencyTree
|
|||
catch ( GraphTaskException e )
|
||||
{
|
||||
String emsg = "Unable to generate graph for [" + Keys.toKey( projectRef ) + "] : " + e.getMessage();
|
||||
getLogger().warn( emsg, e );
|
||||
log.warn( emsg, e );
|
||||
throw new ArchivaException( emsg, e );
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue