* under the License.
*/
-import org.codehaus.plexus.logging.AbstractLogEnabled;
-
import java.util.HashSet;
import java.util.Iterator;
import java.util.Set;
* @version $Id$
*/
public abstract class AbstractMonitoredConsumer
- extends AbstractLogEnabled
implements BaseConsumer
{
private Set<ConsumerMonitor> monitors = new HashSet<ConsumerMonitor>();
* under the License.
*/
+import java.io.File;
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.List;
+
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
import org.apache.maven.archiva.configuration.ConfigurationNames;
import org.apache.maven.archiva.configuration.FileTypes;
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
import org.codehaus.plexus.registry.Registry;
import org.codehaus.plexus.registry.RegistryListener;
-
-import java.io.File;
-import java.io.IOException;
-import java.util.ArrayList;
-import java.util.List;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* MetadataUpdaterConsumer will create and update the metadata present within the repository.
extends AbstractMonitoredConsumer
implements KnownRepositoryContentConsumer, RegistryListener, Initializable
{
+ private Logger log = LoggerFactory.getLogger( MetadataUpdaterConsumer.class );
+
/**
* @plexus.configuration default-value="metadata-updater"
*/
if ( projectMetadata.exists() && ( projectMetadata.lastModified() >= this.scanStartTimestamp ) )
{
// This metadata is up to date. skip it.
- getLogger().debug( "Skipping uptodate metadata: " + this.metadataTools.toPath( projectRef ) );
+ log.debug( "Skipping uptodate metadata: " + this.metadataTools.toPath( projectRef ) );
return;
}
metadataTools.updateMetadata( this.repository, projectRef );
- getLogger().debug( "Updated metadata: " + this.metadataTools.toPath( projectRef ) );
+ log.debug( "Updated metadata: " + this.metadataTools.toPath( projectRef ) );
}
catch ( LayoutException e )
{
if ( projectMetadata.exists() && ( projectMetadata.lastModified() >= this.scanStartTimestamp ) )
{
// This metadata is up to date. skip it.
- getLogger().debug( "Skipping uptodate metadata: " + this.metadataTools.toPath( versionRef ) );
+ log.debug( "Skipping uptodate metadata: " + this.metadataTools.toPath( versionRef ) );
return;
}
metadataTools.updateMetadata( this.repository, versionRef );
- getLogger().debug( "Updated metadata: " + this.metadataTools.toPath( versionRef ) );
+ log.debug( "Updated metadata: " + this.metadataTools.toPath( versionRef ) );
}
catch ( LayoutException e )
{
* under the License.
*/
+import java.io.File;
+import java.util.ArrayList;
+import java.util.List;
+
import org.apache.commons.lang.StringUtils;
import org.apache.maven.archiva.common.utils.VersionUtil;
import org.apache.maven.archiva.consumers.AbstractMonitoredConsumer;
import org.apache.maven.archiva.repository.project.ProjectModelFilter;
import org.apache.maven.archiva.repository.project.ProjectModelReader;
import org.apache.maven.archiva.repository.project.filters.EffectiveProjectModelFilter;
-
-import java.io.File;
-import java.util.ArrayList;
-import java.util.List;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* ProjectModelToDatabaseConsumer
extends AbstractMonitoredConsumer
implements DatabaseUnprocessedArtifactConsumer
{
+ private Logger log = LoggerFactory.getLogger( ProjectModelToDatabaseConsumer.class );
+
/**
* @plexus.configuration default-value="update-db-project"
*/
if ( isValidModel( model, repo, artifact ) )
{
- getLogger().debug( "Adding project model to database - " + Keys.toKey( model ) );
+ log.debug( "Adding project model to database - " + Keys.toKey( model ) );
dao.getProjectModelDAO().saveProjectModel( model );
}
else
{
- getLogger().warn(
+ log.warn(
"Invalid or corrupt pom. Project model not added to database - " + Keys.toKey( model ) );
}
}
catch ( ProjectModelException e )
{
- getLogger().warn( "Unable to read project model " + artifactFile + " : " + e.getMessage(), e );
+ log.warn( "Unable to read project model " + artifactFile + " : " + e.getMessage(), e );
addProblem( artifact, "Unable to read project model " + artifactFile + " : " + e.getMessage() );
}
catch ( ArchivaDatabaseException e )
{
- getLogger().warn( "Unable to save project model " + artifactFile + " to the database : " + e.getMessage(),
+ log.warn( "Unable to save project model " + artifactFile + " to the database : " + e.getMessage(),
e );
}
catch ( Throwable t )
{
// Catch the other errors in the process to allow the rest of the process to complete.
- getLogger().error( "Unable to process model " + artifactFile + " due to : " + t.getClass().getName() +
+ log.error( "Unable to process model " + artifactFile + " due to : " + t.getClass().getName() +
" : " + t.getMessage(), t );
}
}
emsg.append( "]: The model artifactId [" ).append( model.getArtifactId() );
emsg.append( "] does not match the artifactId portion of the filename: " ).append( artifact.getArtifactId() );
- getLogger().warn(emsg.toString() );
+ log.warn(emsg.toString() );
addProblem( artifact, emsg.toString() );
return false;
emsg.append( "]; The model version [" ).append( model.getVersion() );
emsg.append( "] does not match the version portion of the filename: " ).append( artifact.getVersion() );
- getLogger().warn(emsg.toString() );
+ log.warn(emsg.toString() );
addProblem( artifact, emsg.toString() );
return false;
catch ( ArchivaDatabaseException e )
{
String emsg = "Unable to save problem with artifact location to DB: " + e.getMessage();
- getLogger().warn( emsg, e );
+ log.warn( emsg, e );
throw new ConsumerException( emsg, e );
}
}
* under the License.
*/
+import java.util.HashMap;
+import java.util.Iterator;
+import java.util.List;
+import java.util.Map;
+
import org.apache.commons.lang.StringUtils;
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
import org.apache.maven.archiva.configuration.ConfigurationNames;
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
import org.codehaus.plexus.registry.Registry;
import org.codehaus.plexus.registry.RegistryListener;
-
-import java.util.HashMap;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* IndexArtifactConsumer
extends AbstractMonitoredConsumer
implements DatabaseUnprocessedArtifactConsumer, RegistryListener, Initializable
{
+ private Logger log = LoggerFactory.getLogger( IndexArtifactConsumer.class );
+
private static final String INDEX_ERROR = "indexing_error";
/**
}
catch ( RepositoryException e )
{
- getLogger().error( "Unable to load repository content object: " + e.getMessage(), e );
+ log.error( "Unable to load repository content object: " + e.getMessage(), e );
}
}
}
* under the License.
*/
+import java.util.ArrayList;
+import java.util.List;
+
import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration;
import org.apache.maven.archiva.consumers.AbstractMonitoredConsumer;
import org.apache.maven.archiva.consumers.ConsumerException;
import org.apache.maven.artifact.Artifact;
import org.apache.maven.artifact.factory.ArtifactFactory;
import org.apache.maven.artifact.repository.ArtifactRepository;
-
-import java.util.ArrayList;
-import java.util.List;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* LegacyConverterArtifactConsumer - convert artifacts as they are found
extends AbstractMonitoredConsumer
implements KnownRepositoryContentConsumer
{
+ private Logger log = LoggerFactory.getLogger( LegacyConverterArtifactConsumer.class );
+
/**
* @plexus.requirement role-hint="legacy-to-default"
*/
}
catch ( LayoutException e )
{
- getLogger().warn( "Unable to convert artifact: " + path + " : " + e.getMessage(), e );
+ log.warn( "Unable to convert artifact: " + path + " : " + e.getMessage(), e );
}
catch ( ArtifactConversionException e )
{
- getLogger().warn( "Unable to convert artifact: " + path + " : " + e.getMessage(), e );
+ log.warn( "Unable to convert artifact: " + path + " : " + e.getMessage(), e );
}
}
* under the License.
*/
+import java.io.File;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
import org.apache.commons.collections.Closure;
import org.apache.commons.collections.CollectionUtils;
import org.apache.commons.collections.functors.IfClosure;
import org.apache.maven.archiva.repository.scanner.functors.ConsumerProcessFileClosure;
import org.apache.maven.archiva.repository.scanner.functors.ConsumerWantsFilePredicate;
import org.apache.maven.archiva.repository.scanner.functors.TriggerBeginScanClosure;
-import org.codehaus.plexus.logging.AbstractLogEnabled;
-
-import java.io.File;
-import java.util.ArrayList;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* RepositoryContentConsumerUtil
* @plexus.component role="org.apache.maven.archiva.repository.scanner.RepositoryContentConsumers"
*/
public class RepositoryContentConsumers
- extends AbstractLogEnabled
{
+ private Logger log = LoggerFactory.getLogger( RepositoryContentConsumers.class );
+
/**
* @plexus.requirement
*/
// Run the repository consumers
try
{
- Closure triggerBeginScan = new TriggerBeginScanClosure( repository, getLogger() );
+ Closure triggerBeginScan = new TriggerBeginScanClosure( repository );
List<KnownRepositoryContentConsumer> selectedKnownConsumers = getSelectedKnownConsumers();
List<InvalidRepositoryContentConsumer> selectedInvalidConsumers = getSelectedInvalidConsumers();
BaseFile baseFile = new BaseFile( repository.getLocation(), localFile );
ConsumerWantsFilePredicate predicate = new ConsumerWantsFilePredicate();
predicate.setBasefile( baseFile );
- ConsumerProcessFileClosure closure = new ConsumerProcessFileClosure( getLogger() );
+ ConsumerProcessFileClosure closure = new ConsumerProcessFileClosure();
closure.setBasefile( baseFile );
predicate.setCaseSensitive( false );
Closure processIfWanted = IfClosure.getInstance( predicate, closure );
import org.apache.commons.collections.Closure;
import org.apache.maven.archiva.common.utils.BaseFile;
import org.apache.maven.archiva.consumers.RepositoryContentConsumer;
-import org.codehaus.plexus.logging.Logger;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* ConsumerProcessFileClosure
public class ConsumerProcessFileClosure
implements Closure
{
+ private Logger log = LoggerFactory.getLogger( ConsumerProcessFileClosure.class );
+
private BaseFile basefile;
- private Logger logger;
-
- public ConsumerProcessFileClosure( Logger logger )
- {
- // Lame. I know, but seeing as plexus doesn't like to cleanup after
- // application loaded/lookup'd components, this is the best I can do.
- this.logger = logger;
- }
-
public void execute( Object input )
{
if ( input instanceof RepositoryContentConsumer )
try
{
- logger.debug( "Sending to consumer: " + consumer.getId() );
+ log.debug( "Sending to consumer: " + consumer.getId() );
consumer.processFile( basefile.getRelativePath() );
}
/* Intentionally Catch all exceptions.
* So that the discoverer processing can continue.
*/
- logger.error( "Consumer [" + consumer.getId() + "] had an error when processing file ["
+ log.error( "Consumer [" + consumer.getId() + "] had an error when processing file ["
+ basefile.getAbsolutePath() + "]: " + e.getMessage(), e );
}
}
public Logger getLogger()
{
- return logger;
+ return log;
}
public void setLogger( Logger logger )
{
- this.logger = logger;
+ this.log = logger;
}
}
* under the License.
*/
+import java.util.ArrayList;
+import java.util.List;
+
import org.apache.maven.archiva.consumers.AbstractMonitoredConsumer;
import org.apache.maven.archiva.consumers.ConsumerException;
import org.apache.maven.archiva.consumers.DatabaseUnprocessedArtifactConsumer;
import org.apache.maven.archiva.model.ArchivaArtifact;
-
-import java.util.ArrayList;
-import java.util.List;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* TestDatabaseUnprocessedConsumer
extends AbstractMonitoredConsumer
implements DatabaseUnprocessedArtifactConsumer
{
+ private Logger log = LoggerFactory.getLogger( TestDatabaseUnprocessedConsumer.class );
+
private int countBegin = 0;
private int countComplete = 0;
public void processArchivaArtifact( ArchivaArtifact artifact )
throws ConsumerException
{
- getLogger().info( "Processing Artifact: " + artifact );
+ log.info( "Processing Artifact: " + artifact );
countProcessed++;
}
* under the License.
*/
+import java.util.ArrayList;
+import java.util.List;
+
import org.apache.commons.collections.CollectionUtils;
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
import org.apache.maven.archiva.configuration.ConfigurationNames;
import org.codehaus.plexus.personality.plexus.lifecycle.phase.InitializationException;
import org.codehaus.plexus.registry.Registry;
import org.codehaus.plexus.registry.RegistryListener;
-
-import java.util.ArrayList;
-import java.util.List;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* Search the database of known SHA1 Checksums for potential duplicate artifacts.
extends AbstractMonitoredConsumer
implements ArchivaArtifactConsumer, RegistryListener, Initializable
{
+ private Logger log = LoggerFactory.getLogger( DuplicateArtifactsConsumer.class );
+
/**
* @plexus.configuration default-value="duplicate-artifacts"
*/
}
catch ( ObjectNotFoundException e )
{
- getLogger().debug( "No duplicates for artifact: " + artifact );
+ log.debug( "No duplicates for artifact: " + artifact );
return;
}
catch ( ArchivaDatabaseException e )
{
- getLogger().warn( "Unable to query DB for potential duplicates with : " + artifact );
+ log.warn( "Unable to query DB for potential duplicates with : " + artifact );
return;
}
if ( results.size() <= 1 )
{
// No duplicates detected.
- getLogger().debug( "Found no duplicate artifact results on: " + artifact );
+ log.debug( "Found no duplicate artifact results on: " + artifact );
return;
}
try
{
- getLogger().debug( "Found duplicate artifact: " + problem );
+ log.debug( "Found duplicate artifact: " + problem );
dao.getRepositoryProblemDAO().saveRepositoryProblem( problem );
}
catch ( ArchivaDatabaseException e )
{
String emsg = "Unable to save problem with duplicate artifact to DB: " + e.getMessage();
- getLogger().warn( emsg, e );
+ log.warn( emsg, e );
throw new ConsumerException( emsg, e );
}
}
}
catch ( RepositoryException e )
{
- getLogger().warn( "Unable to calculate path for artifact: " + artifact );
+ log.warn( "Unable to calculate path for artifact: " + artifact );
return "";
}
}
* under the License.
*/
+import java.io.File;
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.Enumeration;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+import java.util.jar.JarEntry;
+import java.util.jar.JarFile;
+
import org.apache.commons.lang.StringUtils;
import org.apache.maven.archiva.configuration.ArchivaConfiguration;
import org.apache.maven.archiva.configuration.ConfigurationNames;
import org.codehaus.plexus.registry.Registry;
import org.codehaus.plexus.registry.RegistryListener;
import org.codehaus.plexus.util.SelectorUtils;
-
-import java.io.File;
-import java.io.IOException;
-import java.util.ArrayList;
-import java.util.Enumeration;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-import java.util.jar.JarEntry;
-import java.util.jar.JarFile;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* Validate the location of the artifact based on the values indicated
extends AbstractMonitoredConsumer
implements ArchivaArtifactConsumer, RegistryListener, Initializable
{
+ private Logger log = LoggerFactory.getLogger( LocationArtifactsConsumer.class );
+
/**
* @plexus.configuration default-value="duplicate-artifacts"
*/
catch ( ArchivaDatabaseException e )
{
String emsg = "Unable to save problem with artifact location to DB: " + e.getMessage();
- getLogger().warn( emsg, e );
+ log.warn( emsg, e );
throw new ConsumerException( emsg, e );
}
}
}
catch ( RepositoryException e )
{
- getLogger().warn( "Unable to calculate path for artifact: " + artifact );
+ log.warn( "Unable to calculate path for artifact: " + artifact );
return "";
}
}
* under the License.
*/
+import java.util.ArrayList;
+import java.util.List;
+
import org.apache.maven.archiva.consumers.AbstractMonitoredConsumer;
import org.apache.maven.archiva.consumers.ConsumerException;
import org.apache.maven.archiva.consumers.DatabaseUnprocessedArtifactConsumer;
import org.apache.maven.archiva.model.ArchivaArtifact;
-
-import java.util.ArrayList;
-import java.util.List;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
/**
* TestDatabaseUnprocessedConsumer
extends AbstractMonitoredConsumer
implements DatabaseUnprocessedArtifactConsumer
{
+ private Logger log = LoggerFactory.getLogger( TestDatabaseUnprocessedConsumer.class );
+
private int countBegin = 0;
private int countComplete = 0;
public void processArchivaArtifact( ArchivaArtifact artifact )
throws ConsumerException
{
- getLogger().info( "Processing Artifact: " + artifact );
+ log.info( "Processing Artifact: " + artifact );
countProcessed++;
}