git-svn-id: https://svn.apache.org/repos/asf/archiva/trunk@1546890 13f79535-47bb-0310-9956-ffa450edef68tags/archiva-2.0.0-RC1
@@ -175,7 +175,7 @@ public class ChecksummedFile | |||
} | |||
catch ( IOException e ) | |||
{ | |||
log.warn( "Unable to update checksum:" + e.getMessage() ); | |||
log.warn( "Unable to update checksum:{}", e.getMessage() ); | |||
return false; | |||
} | |||
@@ -200,7 +200,7 @@ public class ChecksummedFile | |||
} | |||
catch ( IOException e ) | |||
{ | |||
log.warn( "Unable to read / parse checksum: " + e.getMessage() ); | |||
log.warn( "Unable to read / parse checksum: {}", e.getMessage() ); | |||
return false; | |||
} | |||
@@ -336,7 +336,7 @@ public class DefaultArchivaConfiguration | |||
else | |||
{ | |||
// Policy key doesn't exist. Don't add it to golden version. | |||
log.warn( "Policy [" + policyId + "] does not exist." ); | |||
log.warn( "Policy [{}] does not exist.", policyId ); | |||
} | |||
} | |||
@@ -568,7 +568,7 @@ public class ArchivaConfigurationTest | |||
assertEquals( "check managed repositories", "default", repository.getLayout() ); | |||
assertTrue( "check managed repositories", repository.isScanned() ); | |||
log.info( "knowContentConsumers " + configuration.getRepositoryScanning().getKnownContentConsumers() ); | |||
log.info( "knowContentConsumers {}", configuration.getRepositoryScanning().getKnownContentConsumers() ); | |||
assertFalse( | |||
configuration.getRepositoryScanning().getKnownContentConsumers().contains( "update-db-artifact" ) ); |
@@ -206,15 +206,15 @@ public class ArchivaMetadataCreationConsumer | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( | |||
"Error occurred persisting metadata for artifact:{} (repository:{}); message: {}" + e.getMessage(), | |||
new Object[]{ path, repoId, e.getMessage() }, e ); | |||
"Error occurred persisting metadata for artifact:{} (repository:{}); message: {}" , | |||
path, repoId, e.getMessage(), e ); | |||
repositorySession.revert(); | |||
} | |||
catch ( RepositoryStorageRuntimeException e ) | |||
{ | |||
log.warn( | |||
"Error occurred persisting metadata for artifact:{} (repository:{}); message: {}" + e.getMessage(), | |||
new Object[]{ path, repoId, e.getMessage() }, e ); | |||
"Error occurred persisting metadata for artifact:{} (repository:{}); message: {}", | |||
path, repoId, e.getMessage(), e ); | |||
repositorySession.revert(); | |||
} | |||
finally |
@@ -134,11 +134,11 @@ public class LegacyConverterArtifactConsumer | |||
} | |||
catch ( LayoutException e ) | |||
{ | |||
log.warn( "Unable to convert artifact: " + path + " : " + e.getMessage(), e ); | |||
log.warn( "Unable to convert artifact: {} : {}",path , e.getMessage(), e ); | |||
} | |||
catch ( ArtifactConversionException e ) | |||
{ | |||
log.warn( "Unable to convert artifact: " + path + " : " + e.getMessage(), e ); | |||
log.warn( "Unable to convert artifact: {} : {}",path , e.getMessage(), e ); | |||
} | |||
} | |||
@@ -828,7 +828,7 @@ public class MavenRepositorySearchTest | |||
archivaConfigControl.verify(); | |||
log.info( "groupIds: " + groupIds ); | |||
log.info( "groupIds: {}", groupIds ); | |||
assertEquals( 3, groupIds.size() ); | |||
assertTrue( groupIds.contains( "com" ) ); |
@@ -110,7 +110,7 @@ public class DefaultWagonFactory | |||
} | |||
catch ( Exception e ) | |||
{ | |||
logger.warn( "fail to configure User-Agent: " + e.getMessage(), e ); | |||
logger.warn( "fail to configure User-Agent: {}", e.getMessage(), e ); | |||
} | |||
} | |||
} |
@@ -268,11 +268,11 @@ public class DefaultRepositoryProxyConnectors | |||
} | |||
catch ( RepositoryNotFoundException e ) | |||
{ | |||
log.warn( "Unable to use proxy connector: " + e.getMessage(), e ); | |||
log.warn( "Unable to use proxy connector: {}", e.getMessage(), e ); | |||
} | |||
catch ( RepositoryException e ) | |||
{ | |||
log.warn( "Unable to use proxy connector: " + e.getMessage(), e ); | |||
log.warn( "Unable to use proxy connector: {}", e.getMessage(), e ); | |||
} | |||
@@ -447,8 +447,8 @@ public class DefaultRepositoryProxyConnectors | |||
catch ( ProxyException e ) | |||
{ | |||
log.warn( | |||
"Transfer error from repository \"" + targetRepository.getRepository().getId() + "\" for resource " | |||
+ path + ", continuing to next repository. Error message: {}", e.getMessage() ); | |||
"Transfer error from repository {} for resource {}, continuing to next repository. Error message: {}", | |||
targetRepository.getRepository().getId(), path, e.getMessage() ); | |||
log.debug( MarkerFactory.getDetachedMarker( "transfer.error" ), | |||
"Transfer error from repository \"" + targetRepository.getRepository().getId() | |||
+ "\" for resource " + path + ", continuing to next repository. Error message: {}", | |||
@@ -457,8 +457,7 @@ public class DefaultRepositoryProxyConnectors | |||
catch ( RepositoryAdminException e ) | |||
{ | |||
log.debug( MarkerFactory.getDetachedMarker( "transfer.error" ), | |||
"Transfer error from repository \"" + targetRepository.getRepository().getId() | |||
+ "\" for resource " + path + ", continuing to next repository. Error message: {}", | |||
"Transfer error from repository {} for resource {}, continuing to next repository. Error message: {}",targetRepository.getRepository().getId(), path, | |||
e.getMessage(), e ); | |||
log.debug( MarkerFactory.getDetachedMarker( "transfer.error" ), "Full stack trace", e ); | |||
} | |||
@@ -517,16 +516,14 @@ public class DefaultRepositoryProxyConnectors | |||
} | |||
catch ( ProxyException e ) | |||
{ | |||
log.warn( "Transfer error from repository \"" + targetRepository.getRepository().getId() | |||
+ "\" for versioned Metadata " + logicalPath | |||
+ ", continuing to next repository. Error message: " + e.getMessage() ); | |||
log.warn( "Transfer error from repository {} for versioned Metadata {}, continuing to next repository. Error message: {}", | |||
targetRepository.getRepository().getId(), logicalPath, e.getMessage() ); | |||
log.debug( "Full stack trace", e ); | |||
} | |||
catch ( RepositoryAdminException e ) | |||
{ | |||
log.warn( "Transfer error from repository \"" + targetRepository.getRepository().getId() | |||
+ "\" for versioned Metadata " + logicalPath | |||
+ ", continuing to next repository. Error message: " + e.getMessage() ); | |||
log.warn( "Transfer error from repository {} for versioned Metadata {}, continuing to next repository. Error message: {}", | |||
targetRepository.getRepository().getId(), logicalPath, e.getMessage() ); | |||
log.debug( "Full stack trace", e ); | |||
} | |||
} | |||
@@ -544,7 +541,7 @@ public class DefaultRepositoryProxyConnectors | |||
} | |||
catch ( RepositoryMetadataException e ) | |||
{ | |||
log.warn( "Unable to update metadata " + localFile.getAbsolutePath() + ": " + e.getMessage(), e ); | |||
log.warn( "Unable to update metadata {}:{}", localFile.getAbsolutePath(), e.getMessage(), e ); | |||
} | |||
} | |||
@@ -928,7 +925,7 @@ public class DefaultRepositoryProxyConnectors | |||
catch ( ProxyException e ) | |||
{ | |||
urlFailureCache.cacheFailure( url ); | |||
log.warn( "Transfer failed on checksum: " + url + " : " + e.getMessage(), e ); | |||
log.warn( "Transfer failed on checksum: {} : {}",url ,e.getMessage(), e ); | |||
// Critical issue, pass it on. | |||
throw e; | |||
} | |||
@@ -1088,8 +1085,8 @@ public class DefaultRepositoryProxyConnectors | |||
} | |||
log.warn( | |||
"Transfer error from repository \"" + content.getRepository().getId() + "\" for artifact " + Keys.toKey( | |||
artifact ) + ", continuing to next repository. Error message: " + exception.getMessage() ); | |||
"Transfer error from repository {} for artifact {} , continuing to next repository. Error message: {}", | |||
content.getRepository().getId(), Keys.toKey( artifact), exception.getMessage() ); | |||
log.debug( "Full stack trace", exception ); | |||
} | |||
@@ -1214,12 +1211,12 @@ public class DefaultRepositoryProxyConnectors | |||
} | |||
catch ( ConnectionException e ) | |||
{ | |||
log.warn( "Could not connect to " + remoteRepository.getRepository().getName() + ": " + e.getMessage() ); | |||
log.warn( "Could not connect to {}: {}", remoteRepository.getRepository().getName(), e.getMessage() ); | |||
connected = false; | |||
} | |||
catch ( AuthenticationException e ) | |||
{ | |||
log.warn( "Could not connect to " + remoteRepository.getRepository().getName() + ": " + e.getMessage() ); | |||
log.warn( "Could not connect to {}: {}", remoteRepository.getRepository().getName(), e.getMessage() ); | |||
connected = false; | |||
} | |||
@@ -213,7 +213,7 @@ public abstract class AbstractProxyTestCase | |||
CacheManager.getInstance().clearAll(); | |||
log.info( "\n.\\ " + name + "() \\._________________________________________\n" ); | |||
log.info( "\n.\\ {}() \\._________________________________________\n", name ); | |||
} | |||
@After |
@@ -65,7 +65,7 @@ public class WagonDelegate | |||
public boolean getIfNewer( String resourceName, File destination, long timestamp ) | |||
throws TransferFailedException, ResourceDoesNotExistException, AuthorizationException | |||
{ | |||
log.info( ".getIfNewer(" + resourceName + ", " + destination + ", " + timestamp + ")" ); | |||
log.info( ".getIfNewer({}, {}, {})", resourceName, destination, timestamp ); | |||
boolean result = delegate.getIfNewer( resourceName, destination, timestamp ); | |||
createIfMissing( destination ); |
@@ -389,7 +389,7 @@ public class DefaultManagedRepositoryAdmin | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
//throw new RepositoryAdminException( e.getMessage(), e ); | |||
log.warn( "skip error during removing repository from MetadatRepository:" + e.getMessage(), e ); | |||
log.warn( "skip error during removing repository from MetadatRepository:{}", e.getMessage(), e ); | |||
} | |||
finally | |||
{ |
@@ -48,7 +48,7 @@ public class ManagedRepositoryAdminTest | |||
List<ManagedRepository> repos = managedRepositoryAdmin.getManagedRepositories(); | |||
assertNotNull( repos ); | |||
assertTrue( repos.size() > 0 ); | |||
log.info( "repos " + repos ); | |||
log.info( "repos {}", repos ); | |||
// check default internal | |||
ManagedRepository internal = findManagedRepoById( repos, "internal" ); |
@@ -39,7 +39,7 @@ public class RemoteRepositoryAdminTest | |||
List<RemoteRepository> remoteRepositories = remoteRepositoryAdmin.getRemoteRepositories(); | |||
assertNotNull( remoteRepositories ); | |||
assertTrue( remoteRepositories.size() > 0 ); | |||
log.info( "remote " + remoteRepositories ); | |||
log.info( "remote {}", remoteRepositories ); | |||
} | |||
@Test |
@@ -376,7 +376,7 @@ public class MetadataTools | |||
{ | |||
// TODO: [monitor] consider a monitor for this event. | |||
// TODO: consider a read-redo on monitor return code? | |||
log.warn( "Unable to read metadata: " + metadataFile.getAbsolutePath(), e ); | |||
log.warn( "Unable to read metadata: {}", metadataFile.getAbsolutePath(), e ); | |||
return null; | |||
} | |||
} | |||
@@ -401,7 +401,7 @@ public class MetadataTools | |||
{ | |||
// TODO: [monitor] consider a monitor for this event. | |||
// TODO: consider a read-redo on monitor return code? | |||
log.warn( "Unable to read metadata: " + metadataFile.getAbsolutePath(), e ); | |||
log.warn( "Unable to read metadata: {}", metadataFile.getAbsolutePath(), e ); | |||
return null; | |||
} | |||
} | |||
@@ -426,7 +426,7 @@ public class MetadataTools | |||
{ | |||
// TODO: [monitor] consider a monitor for this event. | |||
// TODO: consider a read-redo on monitor return code? | |||
log.warn( "Unable to read metadata: " + metadataFile.getAbsolutePath(), e ); | |||
log.warn( "Unable to read metadata: {}", metadataFile.getAbsolutePath(), e ); | |||
return null; | |||
} | |||
} |
@@ -73,7 +73,7 @@ public class TriggerBeginScanClosure | |||
} | |||
catch ( ConsumerException e ) | |||
{ | |||
log.warn( "Consumer [" + consumer.getId() + "] cannot begin: " + e.getMessage(), e ); | |||
log.warn( "Consumer [{}] cannot begin: {}",consumer.getId(), e.getMessage(), e ); | |||
} | |||
} | |||
} |
@@ -205,7 +205,7 @@ public class DefaultDownloadRemoteIndexScheduler | |||
} | |||
catch ( IllegalArgumentException e ) | |||
{ | |||
log.warn( "Unable to schedule remote index download: " + e.getLocalizedMessage() ); | |||
log.warn( "Unable to schedule remote index download: {}", e.getLocalizedMessage() ); | |||
} | |||
if ( remoteRepository.isDownloadRemoteIndexOnStartup() ) |
@@ -229,7 +229,7 @@ public class DownloadRemoteIndexTask | |||
deleteDirectoryQuiet( tempIndexDirectory ); | |||
this.runningRemoteDownloadIds.remove( this.remoteRepository.getId() ); | |||
} | |||
log.info( "end download remote index for remote repository " + this.remoteRepository.getId() ); | |||
log.info( "end download remote index for remote repository {}", this.remoteRepository.getId() ); | |||
} | |||
private void deleteDirectoryQuiet( File f ) |
@@ -88,7 +88,7 @@ public class DownloadRemoteIndexTaskTest | |||
this.server.start(); | |||
Connector connector = this.server.getConnectors()[0]; | |||
this.port = connector.getLocalPort(); | |||
log.info( "start server on port " + this.port ); | |||
log.info( "start server on port {}", this.port ); | |||
nexusIndexer = plexusSisuBridge.lookup( NexusIndexer.class ); | |||
} | |||
@@ -140,7 +140,7 @@ public class DownloadRemoteIndexTaskTest | |||
FlatSearchResponse response = nexusIndexer.searchFlat( rq ); | |||
log.info( "returned hit count:" + response.getReturnedHitsCount() ); | |||
log.info( "returned hit count:{}", response.getReturnedHitsCount() ); | |||
assertEquals( 8, response.getReturnedHitsCount() ); | |||
} | |||
@@ -281,8 +281,8 @@ public class DefaultBrowseService | |||
catch ( MetadataResolutionException e ) | |||
{ | |||
log.warn( | |||
"Skipping invalid metadata while compiling shared model for " + groupId + ":" + artifactId | |||
+ " in repo " + repoId + ": " + e.getMessage() ); | |||
"Skipping invalid metadata while compiling shared model for {}:{} in repo {}: {}", | |||
groupId, artifactId, repoId, e.getMessage() ); | |||
} | |||
} | |||
} |
@@ -165,7 +165,7 @@ public class DefaultManagedRepositoriesService | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Error retrieving repository statistics: " + e.getMessage(), e ); | |||
log.warn( "Error retrieving repository statistics: {}", e.getMessage(), e ); | |||
} | |||
if ( stats != null ) | |||
{ |
@@ -82,7 +82,7 @@ public class DefaultReportRepositoriesService | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to retrieve stats, assuming is empty: " + e.getMessage(), e ); | |||
log.warn( "Unable to retrieve stats, assuming is empty: {}", e.getMessage(), e ); | |||
} | |||
} | |||
@@ -109,7 +109,7 @@ public class DefaultReportRepositoriesService | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to retrieve stats, assuming is empty: " + e.getMessage(), e ); | |||
log.warn( "Unable to retrieve stats, assuming is empty: {}", e.getMessage(), e ); | |||
} | |||
if ( stats == null || stats.isEmpty() ) | |||
{ |
@@ -160,7 +160,7 @@ public class ManagedRepositoriesServiceTest | |||
assertNotNull( archivaRepositoryStatistics ); | |||
log.info( "archivaRepositoryStatistics:" + archivaRepositoryStatistics.toString() ); | |||
log.info( "archivaRepositoryStatistics: {}", archivaRepositoryStatistics.toString() ); | |||
assertEquals( 92, archivaRepositoryStatistics.getNewFileCount() ); | |||
assertEquals( 92, archivaRepositoryStatistics.getTotalFileCount() ); |
@@ -130,7 +130,7 @@ public class SearchServiceTest | |||
for ( Artifact artifact : artifacts ) | |||
{ | |||
log.info( "url:" + artifact.getUrl() ); | |||
log.info( "url: {}", artifact.getUrl() ); | |||
String version = artifact.getVersion(); | |||
assertTrue( artifact.getUrl().contains( version ) ); | |||
@@ -231,7 +231,7 @@ public class SearchServiceTest | |||
for ( Artifact artifact : artifacts ) | |||
{ | |||
log.info( "url:" + artifact.getUrl() ); | |||
log.info( "url: {}", artifact.getUrl() ); | |||
String version = artifact.getVersion(); | |||
assertEquals( "http://localhost:" + port | |||
+ "/repository/test-repo/org/apache/karaf/features/org.apache.karaf.features.core/" | |||
@@ -269,7 +269,7 @@ public class SearchServiceTest | |||
assertTrue( " not 1 results for Bundle ExportPackage org.apache.karaf.features.command.completers but " | |||
+ artifacts.size() + ":" + artifacts, artifacts.size() == 1 ); | |||
log.info( "artifact url " + artifacts.get( 0 ).getUrl() ); | |||
log.info( "artifact url {}", artifacts.get( 0 ).getUrl() ); | |||
deleteTestRepo( testRepoId ); | |||
} | |||
@@ -333,7 +333,7 @@ public class SearchServiceTest | |||
assertTrue( " not 0 results for Bundle ExportPackage org.apache.karaf.features.command.completers but " | |||
+ artifacts.size() + ":" + artifacts, artifacts.size() == 1 ); | |||
log.info( "artifact url " + artifacts.get( 0 ).getUrl() ); | |||
log.info( "artifact url {}", artifacts.get( 0 ).getUrl() ); | |||
deleteTestRepo( testRepoId ); | |||
} | |||
@@ -354,7 +354,7 @@ public class SearchServiceTest | |||
SearchService searchService = getSearchService( authorizationHeader ); | |||
Collection<String> groupIds = searchService.getAllGroupIds( Arrays.asList( testRepoId ) ).getGroupIds(); | |||
log.info( "groupIds " + groupIds ); | |||
log.info( "groupIds {}", groupIds ); | |||
assertFalse( groupIds.isEmpty() ); | |||
assertTrue( groupIds.contains( "commons-cli" ) ); | |||
assertTrue( groupIds.contains( "org.apache.felix" ) ); | |||
@@ -384,7 +384,7 @@ public class SearchServiceTest | |||
new SearchRequest( "org.foo", "studio-all-update-site", null, null, null, Arrays.asList( "test-repo" ) ); | |||
List<Artifact> artifacts = searchService.searchArtifacts( searchRequest ); | |||
log.info( "artifacts:" + artifacts ); | |||
log.info( "artifacts: {}", artifacts ); | |||
assertEquals( 1, artifacts.size() ); | |||
deleteTestRepo( testRepoId ); | |||
} |
@@ -77,8 +77,8 @@ public class ArchivaServletAuthenticator | |||
{ | |||
if ( authzResult.getException() != null ) | |||
{ | |||
log.info( "Authorization Denied [ip=" + request.getRemoteAddr() + ",permission=" + permission + ",repo=" | |||
+ repositoryId + "] : " + authzResult.getException().getMessage() ); | |||
log.info( "Authorization Denied [ip={},permission={},repo={}] : {}", request.getRemoteAddr(), | |||
permission, repositoryId, authzResult.getException().getMessage() ); | |||
throw new UnauthorizedException( "Access denied for repository " + repositoryId ); | |||
} |
@@ -263,7 +263,7 @@ public class RssFeedServlet | |||
} | |||
catch ( DecoderException ie ) | |||
{ | |||
log.warn( "Error decoding username and password.", ie.getMessage() ); | |||
log.warn( "Error decoding username and password: {}", ie.getMessage() ); | |||
} | |||
if ( usernamePassword == null || usernamePassword.trim().equals( "" ) ) |
@@ -134,7 +134,7 @@ public class ArchivaLockedAdminEnvironmentCheck | |||
} | |||
catch ( RbacManagerException e ) | |||
{ | |||
log.warn( "Exception when checking for locked admin user: " + e.getMessage(), e ); | |||
log.warn( "Exception when checking for locked admin user: {}", e.getMessage(), e ); | |||
} | |||
checked = true; |
@@ -224,6 +224,6 @@ public class Banner | |||
public static void display( String version ) | |||
{ | |||
String banner = getBanner( version ); | |||
LoggerFactory.getLogger( Banner.class ).info( StringUtils.repeat( "_", 25 ) + eol + banner ); | |||
LoggerFactory.getLogger( Banner.class ).info( "{} {}, {}" , StringUtils.repeat( "_", 25 ), eol, banner ); | |||
} | |||
} |
@@ -188,7 +188,7 @@ public class SecuritySynchronization | |||
EnvironmentCheck check = entry.getValue(); | |||
List<String> v = new ArrayList<String>(); | |||
check.validateEnvironment( v ); | |||
log.info( "Environment Check: " + entry.getKey() + " -> " + v.size() + " violation(s)" ); | |||
log.info( "Environment Check: {} -> {} violation(s)", entry.getKey(), v.size() ); | |||
for ( String s : v ) | |||
{ | |||
violations.add( "[" + entry.getKey() + "] " + s ); | |||
@@ -247,8 +247,7 @@ public class SecuritySynchronization | |||
} | |||
catch ( RbacManagerException e ) | |||
{ | |||
log.warn( "Unable to add role [" + ArchivaRoleConstants.toRepositoryObserverRoleName( repoId ) + "] to " | |||
+ principal + " user.", e ); | |||
log.warn( "Unable to add role [{}] to {} user.", ArchivaRoleConstants.toRepositoryObserverRoleName( repoId ), principal, e ); | |||
} | |||
} | |||
} |
@@ -128,7 +128,7 @@ public abstract class AbstractDownloadTest | |||
this.server.start(); | |||
Connector connector = this.server.getConnectors()[0]; | |||
this.port = connector.getLocalPort(); | |||
log.info( "start server on port " + this.port ); | |||
log.info( "start server on port {}", this.port ); | |||
User user = new User(); | |||
user.setEmail( "toto@toto.fr" ); |
@@ -197,7 +197,7 @@ public class DownloadArtifactsTest | |||
} | |||
catch ( Throwable e ) | |||
{ | |||
log.info( "fail to get zipEntries " + e.getMessage(), e ); | |||
log.info( "fail to get zipEntries {}", e.getMessage(), e ); | |||
} | |||
return Collections.emptyList(); | |||
} |
@@ -629,7 +629,7 @@ public class ArchivaDavResourceFactory | |||
} | |||
catch ( LayoutException e ) | |||
{ | |||
log.warn( "Artifact path '" + resourcePath + "' is invalid." ); | |||
log.warn( "Artifact path '{}' is invalid." ,resourcePath ); | |||
} | |||
} | |||
@@ -150,7 +150,7 @@ public abstract class AbstractRepositoryServletProxiedTestCase | |||
int port = repo.server.getConnectors()[0].getLocalPort(); | |||
repo.url = "http://localhost:" + port + repo.context; | |||
log.info( "Remote HTTP Server started on " + repo.url ); | |||
log.info( "Remote HTTP Server started on {}", repo.url ); | |||
repo.config = createRemoteRepository( repo.id, "Testable [" + repo.id + "] Remote Repo", repo.url ); | |||
@@ -126,7 +126,7 @@ public class DefaultMetadataResolver | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to persist resolved information: " + e.getMessage(), e ); | |||
log.warn( "Unable to persist resolved information: {}", e.getMessage(), e ); | |||
} | |||
session.markDirty(); | |||
@@ -193,7 +193,7 @@ public class DefaultMetadataResolver | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to persist resolved information: " + e.getMessage(), e ); | |||
log.warn( "Unable to persist resolved information: {}", e.getMessage(), e ); | |||
} | |||
} | |||
session.markDirty(); | |||
@@ -233,7 +233,7 @@ public class DefaultMetadataResolver | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to persist resolved information: " + e.getMessage(), e ); | |||
log.warn( "Unable to persist resolved information: {}", e.getMessage(), e ); | |||
} | |||
} | |||
session.markDirty(); | |||
@@ -276,7 +276,7 @@ public class DefaultMetadataResolver | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to persist resolved information: " + e.getMessage(), e ); | |||
log.warn( "Unable to persist resolved information: {}", e.getMessage(), e ); | |||
} | |||
} | |||
} | |||
@@ -327,13 +327,13 @@ public class DefaultMetadataResolver | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to persist resolved information: " + e.getMessage(), e ); | |||
log.warn( "Unable to persist resolved information: {}", e.getMessage(), e ); | |||
} | |||
catch ( RepositoryStorageMetadataInvalidException e ) | |||
{ | |||
log.warn( | |||
"Not update project in metadata repository due to an error resolving it from storage: " | |||
+ e.getMessage() ); | |||
"Not update project in metadata repository due to an error resolving it from storage: {}", | |||
e.getMessage() ); | |||
for ( RepositoryListener listener : listeners ) | |||
{ | |||
@@ -391,7 +391,7 @@ public class DefaultMetadataResolver | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to persist resolved information: " + e.getMessage(), e ); | |||
log.warn( "Unable to persist resolved information: {}", e.getMessage(), e ); | |||
} | |||
} | |||
session.markDirty(); |
@@ -64,7 +64,7 @@ public class MetadataAuditListener | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to write audit event to repository: " + e.getMessage(), e ); | |||
log.warn( "Unable to write audit event to repository: {}", e.getMessage(), e ); | |||
} | |||
finally | |||
{ |
@@ -211,7 +211,7 @@ public class Maven2RepositoryStorage | |||
catch ( XMLException e ) | |||
{ | |||
// unable to parse metadata - log it, and continue with the version as the original SNAPSHOT version | |||
log.warn( "Invalid metadata: " + metadataFile + " - " + e.getMessage() ); | |||
log.warn( "Invalid metadata: {} - {}", metadataFile, e.getMessage() ); | |||
} | |||
} | |||
@@ -290,8 +290,8 @@ public class Maven2RepositoryStorage | |||
if ( ( problem.getException() instanceof FileNotFoundException && e.getModelId() != null && | |||
!e.getModelId().equals( problem.getModelId() ) ) ) | |||
{ | |||
log.warn( "The artifact's parent POM file '" + file + "' cannot be resolved. " + | |||
"Using defaults for project version metadata.." ); | |||
log.warn( "The artifact's parent POM file '{}' cannot be resolved. " + | |||
"Using defaults for project version metadata..", file ); | |||
ProjectVersionMetadata metadata = new ProjectVersionMetadata(); | |||
metadata.setId( readMetadataRequest.getProjectVersion() ); |
@@ -142,13 +142,13 @@ public class RepositoryModelResolver | |||
{ | |||
log.info( | |||
"An exception was caught while attempting to retrieve model '{}' from remote repository '{}'.Reason:{}", | |||
new Object[]{ model.getAbsolutePath(), remoteRepository.getId(), e.getMessage() } ); | |||
model.getAbsolutePath(), remoteRepository.getId(), e.getMessage() ); | |||
} | |||
catch ( Exception e ) | |||
{ | |||
log.warn( | |||
"An exception was caught while attempting to retrieve model '{}' from remote repository '{}'.Reason:{}", | |||
new Object[]{ model.getAbsolutePath(), remoteRepository.getId(), e.getMessage() } ); | |||
model.getAbsolutePath(), remoteRepository.getId(), e.getMessage() ); | |||
continue; | |||
} |
@@ -51,7 +51,7 @@ public class RepositoryProblemEventListener | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to remove metadata facet as part of delete event: " + e.getMessage(), e ); | |||
log.warn( "Unable to remove metadata facet as part of delete event: {}", e.getMessage(), e ); | |||
} | |||
} | |||
@@ -69,8 +69,8 @@ public class RepositoryProblemEventListener | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to remove repository problem facets for the version being corrected in the repository: " | |||
+ e.getMessage(), e ); | |||
log.warn( "Unable to remove repository problem facets for the version being corrected in the repository: {}", | |||
e.getMessage(), e ); | |||
} | |||
} | |||
@@ -92,7 +92,7 @@ public class RepositoryProblemEventListener | |||
} | |||
catch ( MetadataRepositoryException e ) | |||
{ | |||
log.warn( "Unable to add repository problem facets for the version being removed: " + e.getMessage(), e ); | |||
log.warn( "Unable to add repository problem facets for the version being removed: {}", e.getMessage(), e ); | |||
} | |||
} | |||
@@ -178,7 +178,7 @@ public class DuplicateArtifactsConsumer | |||
} | |||
catch ( Exception e ) | |||
{ | |||
log.warn( "Not reporting problem for invalid artifact in checksum check: " + e.getMessage() ); | |||
log.warn( "Not reporting problem for invalid artifact in checksum check: {}", e.getMessage() ); | |||
return; | |||
} | |||