[MRM-1442] track time spent in each consumer during a scan, to help diagnose poor scanning performance

Merged from: r1042689


git-svn-id: https://svn.apache.org/repos/asf/archiva/trunk@1042695 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Brett Porter 2010-12-06 16:03:02 +00:00
parent 9565cefd5b
commit 98d9fea131
4 changed files with 130 additions and 15 deletions

View File

@ -21,6 +21,7 @@ package org.apache.archiva.repository.scanner;
import java.text.SimpleDateFormat;
import java.util.List;
import java.util.Map;
import org.apache.commons.collections.CollectionUtils;
import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration;
@ -70,6 +71,10 @@ public class RepositoryScanStatistics
*/
private long totalSize = 0;
private Map<String, Long> consumerCounts;
private Map<String, Long> consumerTimings;
public void triggerStart()
{
startTimestamp = System.currentTimeMillis();
@ -127,6 +132,17 @@ public class RepositoryScanStatistics
for ( String id : knownConsumers )
{
buf.append( "\n " ).append( id );
if ( consumerTimings.containsKey( id ) )
{
long time = consumerTimings.get( id );
buf.append( " (Total: " ).append( time ).append( "ms" );
if ( consumerCounts.containsKey( id ) )
{
long total = consumerCounts.get( id );
buf.append( "; Avg.: " + ( time / total ) + "; Count: " + total );
}
buf.append( ")" );
}
}
}
else
@ -141,6 +157,17 @@ public class RepositoryScanStatistics
for ( String id : invalidConsumers )
{
buf.append( "\n " ).append( id );
if ( consumerTimings.containsKey( id ) )
{
long time = consumerTimings.get( id );
buf.append( " (Total: " ).append( time ).append( "ms" );
if ( consumerCounts.containsKey( id ) )
{
long total = consumerCounts.get( id );
buf.append( "; Avg.: " + ( time / total ) + "ms; Count: " + total );
}
buf.append( ")" );
}
}
}
else
@ -205,4 +232,14 @@ public class RepositoryScanStatistics
{
return totalSize;
}
public void setConsumerCounts( Map<String, Long> consumerCounts )
{
this.consumerCounts = consumerCounts;
}
public void setConsumerTimings( Map<String, Long> consumerTimings )
{
this.consumerTimings = consumerTimings;
}
}

View File

@ -21,7 +21,9 @@ package org.apache.archiva.repository.scanner;
import java.io.File;
import java.util.Date;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import org.apache.archiva.repository.scanner.functors.ConsumerProcessFileClosure;
import org.apache.archiva.repository.scanner.functors.TriggerBeginScanClosure;
@ -68,6 +70,10 @@ public class RepositoryScannerInstance
private ConsumerWantsFilePredicate consumerWantsFile;
private Map<String, Long> consumerTimings;
private Map<String, Long> consumerCounts;
public RepositoryScannerInstance( ManagedRepositoryConfiguration repository,
List<KnownRepositoryContentConsumer> knownConsumerList,
List<InvalidRepositoryContentConsumer> invalidConsumerList )
@ -76,7 +82,14 @@ public class RepositoryScannerInstance
this.knownConsumers = knownConsumerList;
this.invalidConsumers = invalidConsumerList;
consumerTimings = new HashMap<String,Long>();
consumerCounts = new HashMap<String,Long>();
this.consumerProcessFile = new ConsumerProcessFileClosure();
consumerProcessFile.setExecuteOnEntireRepo( true );
consumerProcessFile.setConsumerTimings( consumerTimings );
consumerProcessFile.setConsumerCounts( consumerCounts );
this.consumerWantsFile = new ConsumerWantsFilePredicate();
stats = new RepositoryScanStatistics();
@ -109,6 +122,16 @@ public class RepositoryScannerInstance
return stats;
}
public Map<String, Long> getConsumerTimings()
{
return consumerTimings;
}
public Map<String, Long> getConsumerCounts()
{
return consumerCounts;
}
public void directoryWalkStarting( File basedir )
{
log.info( "Walk Started: [" + this.repository.getId() + "] " + this.repository.getLocation() );
@ -123,20 +146,19 @@ public class RepositoryScannerInstance
// consume files regardless - the predicate will check the timestamp
BaseFile basefile = new BaseFile( repository.getLocation(), file );
// Timestamp finished points to the last successful scan, not this current one.
if ( file.lastModified() >= changesSince )
{
stats.increaseNewFileCount();
stats.increaseNewFileCount();
}
consumerProcessFile.setBasefile( basefile );
consumerProcessFile.setExecuteOnEntireRepo( true );
consumerWantsFile.setBasefile( basefile );
Closure processIfWanted = IfClosure.getInstance( consumerWantsFile, consumerProcessFile );
CollectionUtils.forAllDo( this.knownConsumers, processIfWanted );
if ( consumerWantsFile.getWantedFileCount() <= 0 )
{
// Nothing known processed this file. It is invalid!
@ -150,6 +172,9 @@ public class RepositoryScannerInstance
CollectionUtils.forAllDo( knownConsumers, scanCompletedClosure );
CollectionUtils.forAllDo( invalidConsumers, scanCompletedClosure );
stats.setConsumerTimings( consumerTimings );
stats.setConsumerCounts( consumerCounts );
log.info( "Walk Finished: [" + this.repository.getId() + "] " + this.repository.getLocation() );
stats.triggerFinished();
}

View File

@ -25,6 +25,9 @@ import org.apache.maven.archiva.consumers.RepositoryContentConsumer;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import java.util.HashMap;
import java.util.Map;
/**
* ConsumerProcessFileClosure
*
@ -39,24 +42,43 @@ public class ConsumerProcessFileClosure
private boolean executeOnEntireRepo;
private Map<String,Long> consumerTimings;
private Map<String,Long> consumerCounts;
public void execute( Object input )
{
if ( input instanceof RepositoryContentConsumer )
{
RepositoryContentConsumer consumer = (RepositoryContentConsumer) input;
String id = consumer.getId();
try
{
log.debug( "Sending to consumer: " + consumer.getId() );
log.debug( "Sending to consumer: " + id );
long startTime = System.currentTimeMillis();
consumer.processFile( basefile.getRelativePath(), executeOnEntireRepo );
long endTime = System.currentTimeMillis();
if ( consumerTimings != null )
{
Long value = consumerTimings.get( id );
consumerTimings.put( id, ( value != null ? value : 0 ) + endTime - startTime );
}
if ( consumerCounts != null )
{
Long value = consumerCounts.get( id );
consumerCounts.put( id, ( value != null ? value : 0 ) + 1 );
}
}
catch ( Exception e )
{
/* Intentionally Catch all exceptions.
* So that the discoverer processing can continue.
*/
log.error( "Consumer [" + consumer.getId() + "] had an error when processing file ["
log.error( "Consumer [" + id + "] had an error when processing file ["
+ basefile.getAbsolutePath() + "]: " + e.getMessage(), e );
}
}
@ -82,6 +104,16 @@ public class ConsumerProcessFileClosure
this.executeOnEntireRepo = executeOnEntireRepo;
}
public void setConsumerTimings( Map<String, Long> consumerTimings )
{
this.consumerTimings = consumerTimings;
}
public void setConsumerCounts( Map<String, Long> consumerCounts )
{
this.consumerCounts = consumerCounts;
}
public Logger getLogger()
{
return log;

View File

@ -49,7 +49,7 @@
<c:set var="queue" value="${queueEntry.value.queueSnapshot}"/>
<tr>
<td>${queueEntry.key}</td>
<td>${fn:length(queue)}</td>
<td align="right">${fn:length(queue)}</td>
</tr>
</c:forEach>
</table>
@ -67,8 +67,29 @@
<c:forEach var="scan" items="${scanner.inProgressScans}">
<tr>
<td>${scan.repository.name} (${scan.repository.id})</td>
<td>${scan.stats.totalFileCount}</td>
<td>${scan.stats.newFileCount}</td>
<td align="right">${scan.stats.totalFileCount}</td>
<td align="right">${scan.stats.newFileCount}</td>
</tr>
<tr>
<td colspan="3">
<table>
<tr>
<th>Name</th>
<th>Total</th>
<th>Average</th>
<th>Invocations</th>
</tr>
<c:forEach var="entry" items="${scan.consumerTimings}">
<tr>
<c:set var="total" value="${scan.consumerCounts[entry.key]}"/>
<td>${entry.key}</td>
<td align="right">${entry.value}ms</td>
<td align="right"><fmt:formatNumber value="${entry.value / total}" pattern="#"/>ms</td>
<td align="right">${total}</td>
</tr>
</c:forEach>
</table>
</td>
</tr>
</c:forEach>
</table>
@ -92,10 +113,10 @@
<c:forEach var="cacheEntry" items="${caches}">
<tr>
<td>${cacheEntry.key}</td>
<td>${cacheEntry.value.statistics.size}</td>
<td>${cacheEntry.value.statistics.cacheHits}</td>
<td>${cacheEntry.value.statistics.cacheMiss}</td>
<td><fmt:formatNumber value="${cacheEntry.value.statistics.cacheHitRate}" pattern="#%"/></td>
<td align="right">${cacheEntry.value.statistics.size}</td>
<td align="right">${cacheEntry.value.statistics.cacheHits}</td>
<td align="right">${cacheEntry.value.statistics.cacheMiss}</td>
<td align="right"><fmt:formatNumber value="${cacheEntry.value.statistics.cacheHitRate}" pattern="#%"/></td>
<td><a href="javascript:alert('Not yet implemented')">Flush</a></td>
</tr>
</c:forEach>