import org.apache.maven.archiva.database.constraints.UniqueGroupIdConstraint;
import org.apache.maven.archiva.model.ArchivaArtifact;
import org.apache.maven.archiva.model.RepositoryContentStatistics;
+import org.apache.maven.archiva.repository.events.RepositoryListener;
import org.apache.maven.archiva.scheduled.tasks.RepositoryTask;
import org.codehaus.plexus.personality.plexus.lifecycle.phase.Initializable;
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
import org.slf4j.LoggerFactory;
/**
- * ArchivaRepositoryScanningTaskExecutor
+ * ArchivaRepositoryScanningTaskExecutor
*
* @version $Id$
- *
+ *
* @plexus.component
* role="org.codehaus.plexus.taskqueue.execution.TaskExecutor"
* role-hint="repository-scanning"
implements TaskExecutor, Initializable
{
private Logger log = LoggerFactory.getLogger( ArchivaRepositoryScanningTaskExecutor.class );
-
+
/**
* TODO: just for stats, remove this and use the main stats module
- *
+ *
* @plexus.requirement role-hint="jdo"
*/
private ArchivaDAO dao;
-
+
/**
* @plexus.requirement
*/
/**
* The repository scanner component.
- *
+ *
* @plexus.requirement
*/
private RepositoryScanner repoScanner;
-
+
+ /**
+ * @plexus.requirement role="org.apache.maven.archiva.repository.events.RepositoryListener"
+ */
+ private List<RepositoryListener> repositoryListeners;
+
/**
* @plexus.requirement
*/
private RepositoryContentConsumers consumers;
-
+
private Task task;
public void initialize()
public void executeTask( Task task )
throws TaskExecutionException
{
+
+ // TODO: replace this whole class with the prescribed content scanning service/action
+ // - scan repository for artifacts that do not have corresponding metadata or have been updated and
+ // send events for each
+ // - scan metadata for artifacts that have been removed and send events for each
+ // - scan metadata for missing plugin data
+ // - store information so that it can restart upon failure (publish event on the server recovery
+ // queue, remove it on successful completion)
+
this.task = task;
-
+
RepositoryTask repoTask = (RepositoryTask) task;
-
+
if ( StringUtils.isBlank( repoTask.getRepositoryId() ) )
{
throw new TaskExecutionException("Unable to execute RepositoryTask with blank repository Id.");
}
ManagedRepositoryConfiguration arepo = archivaConfiguration.getConfiguration().findManagedRepositoryById( repoTask.getRepositoryId() );
-
+
// execute consumers on resource file if set
if( repoTask.getResourceFile() != null )
{
else
{
log.info( "Executing task from queue with job name: " + repoTask );
-
+
// otherwise, execute consumers on whole repository
try
- {
+ {
if ( arepo == null )
{
throw new TaskExecutionException( "Unable to execute RepositoryTask with invalid repository id: " + repoTask.getRepositoryId() );
}
-
+
long sinceWhen = RepositoryScanner.FRESH_SCAN;
-
+
List<RepositoryContentStatistics> results = (List<RepositoryContentStatistics>) dao.query( new MostRecentRepositoryScanStatistics( arepo.getId() ) );
-
+
if ( CollectionUtils.isNotEmpty( results ) )
{
RepositoryContentStatistics lastStats = results.get( 0 );
sinceWhen = lastStats.getWhenGathered().getTime() + lastStats.getDuration();
}
}
-
+
RepositoryScanStatistics stats = repoScanner.scan( arepo, sinceWhen );
-
- log.info( "Finished repository task: " + stats.toDump( arepo ) );
-
+
+ log.info( "Finished first scan: " + stats.toDump( arepo ) );
+
RepositoryContentStatistics dbstats = constructRepositoryStatistics( arepo, sinceWhen, results, stats );
-
- dao.getRepositoryContentStatisticsDAO().saveRepositoryContentStatistics( dbstats );
-
+
+ dao.getRepositoryContentStatisticsDAO().saveRepositoryContentStatistics( dbstats );
+
+// log.info( "Scanning for removed repository content" );
+
+ // FIXME: remove hardcoding
+// MetadataRepository metadataRepository =
+// new FileMetadataRepository( new File( arepo.getLocation(), ".metadata" ) );
+
+// metadataRepository.findAllProjects();
+ // FIXME: do something
+
+ log.info( "Finished repository task: " + repoTask );
+
this.task = null;
}
catch ( RepositoryScannerException e )
private RepositoryContentStatistics constructRepositoryStatistics( ManagedRepositoryConfiguration arepo,
long sinceWhen,
List<RepositoryContentStatistics> results,
- RepositoryScanStatistics stats )
+ RepositoryScanStatistics stats )
{
// I hate jpox and modello <-- and so do I
RepositoryContentStatistics dbstats = new RepositoryContentStatistics();
dbstats.setRepositoryId( stats.getRepositoryId() );
dbstats.setTotalFileCount( stats.getTotalFileCount() );
dbstats.setWhenGathered( stats.getWhenGathered() );
-
+
// total artifact count
try
{
- List<ArchivaArtifact> artifacts = dao.getArtifactDAO().queryArtifacts(
- new ArtifactsByRepositoryConstraint( arepo.getId(), stats.getWhenGathered(), "groupId", true ) );
+ List<ArchivaArtifact> artifacts = dao.getArtifactDAO().queryArtifacts(
+ new ArtifactsByRepositoryConstraint( arepo.getId(), stats.getWhenGathered(), "groupId", true ) );
dbstats.setTotalArtifactCount( artifacts.size() );
}
catch ( ObjectNotFoundException oe )
log.error( "Object not found in the database : " + oe.getMessage() );
}
catch ( ArchivaDatabaseException ae )
- {
+ {
log.error( "Error occurred while querying artifacts for artifact count : " + ae.getMessage() );
}
-
+
// total repo size
long size = FileUtils.sizeOfDirectory( new File( arepo.getLocation() ) );
dbstats.setTotalSize( size );
-
+
// total unique groups
List<String> repos = new ArrayList<String>();
- repos.add( arepo.getId() );
-
+ repos.add( arepo.getId() );
+
List<String> groupIds = (List<String>) dao.query( new UniqueGroupIdConstraint( repos ) );
dbstats.setTotalGroupCount( groupIds.size() );
-
+
List<Object[]> artifactIds = (List<Object[]>) dao.query( new UniqueArtifactIdConstraint( arepo.getId(), true ) );
dbstats.setTotalProjectCount( artifactIds.size() );
-
+
return dbstats;
- }
-
+ }
+
public Task getCurrentTaskInExecution()
{
return task;
+++ /dev/null
-Stage 1: remove use of database and index from core consumers (move implementation into respective database and index modules)
-
-Done!
-
-Stage 2: separate model from JPOX annotated classes, centralising JPOX use in database
-
-Done!
-
-Stage 3: add a basic repository querying API for base artifact information and retrieval of metadata
-
-* RSS, browse
-* consider repository-api refactorings
-* replace archiva-model with separated repository APIs
-* at this point, should be able to have functional Archiva without a database
-* note that metadata need not be stored with the artifacts themselves, but will be by default
-
-Stage 4: incorporation of event API
-
-* used to centralise arrival, removal, etc of files/artifacts in the repository
-* errors should be events as well to avoid exceptions in the logs and instead meaningful handling/reporting
-* could also be used for configuration events
-* consider hooking the audit log to this as well
-
-Stage 5: isolate scanning code
-
-* Repository should operate without scanning code, it should push events if enabled
-* better assessment of its progress, performance
-* removal of database / repository scanning duality - all operations are driven by the event bus
-* move some database operations to a housekeeping scheduled task (same for index), make scheduled tasks a listable item based on available plugins
-