diff options
author | Brett Porter <brett@apache.org> | 2014-06-25 14:22:18 +1000 |
---|---|---|
committer | Olivier Lamy <olamy@apache.org> | 2014-06-25 16:21:43 +1000 |
commit | c648adb3061202572de0fa575150e48ce4c486de (patch) | |
tree | c8611c9dd673c505d3f2bff57ee985ae4c6a1bae | |
parent | d93e36e62d6fd85eaac59682e8b9ad839004b8f1 (diff) | |
download | archiva-c648adb3061202572de0fa575150e48ce4c486de.tar.gz archiva-c648adb3061202572de0fa575150e48ce4c486de.zip |
[MRM-1796] support artifacts larger than 4G
-rw-r--r-- | archiva-modules/archiva-web/archiva-webdav/src/main/java/org/apache/maven/archiva/webdav/ArchivaDavResource.java | 706 |
1 files changed, 706 insertions, 0 deletions
diff --git a/archiva-modules/archiva-web/archiva-webdav/src/main/java/org/apache/maven/archiva/webdav/ArchivaDavResource.java b/archiva-modules/archiva-web/archiva-webdav/src/main/java/org/apache/maven/archiva/webdav/ArchivaDavResource.java new file mode 100644 index 000000000..f359c9c77 --- /dev/null +++ b/archiva-modules/archiva-web/archiva-webdav/src/main/java/org/apache/maven/archiva/webdav/ArchivaDavResource.java @@ -0,0 +1,706 @@ +package org.apache.maven.archiva.webdav; + +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +import java.io.File; +import java.io.FileInputStream; +import java.io.FileOutputStream; +import java.io.IOException; +import java.util.ArrayList; +import java.util.Calendar; +import java.util.List; + +import javax.servlet.http.HttpServletResponse; + +import org.apache.commons.io.FileUtils; +import org.apache.commons.io.IOUtils; +import org.apache.jackrabbit.util.Text; +import org.apache.jackrabbit.webdav.DavConstants; +import org.apache.jackrabbit.webdav.DavException; +import org.apache.jackrabbit.webdav.DavResource; +import org.apache.jackrabbit.webdav.DavResourceFactory; +import org.apache.jackrabbit.webdav.DavResourceIterator; +import org.apache.jackrabbit.webdav.DavResourceIteratorImpl; +import org.apache.jackrabbit.webdav.DavResourceLocator; +import org.apache.jackrabbit.webdav.DavServletResponse; +import org.apache.jackrabbit.webdav.DavSession; +import org.apache.jackrabbit.webdav.MultiStatusResponse; +import org.apache.jackrabbit.webdav.io.InputContext; +import org.apache.jackrabbit.webdav.io.OutputContext; +import org.apache.jackrabbit.webdav.lock.ActiveLock; +import org.apache.jackrabbit.webdav.lock.LockInfo; +import org.apache.jackrabbit.webdav.lock.LockManager; +import org.apache.jackrabbit.webdav.lock.Scope; +import org.apache.jackrabbit.webdav.lock.Type; +import org.apache.jackrabbit.webdav.property.DavProperty; +import org.apache.jackrabbit.webdav.property.DavPropertyName; +import org.apache.jackrabbit.webdav.property.DavPropertyNameSet; +import org.apache.jackrabbit.webdav.property.DavPropertySet; +import org.apache.jackrabbit.webdav.property.DefaultDavProperty; +import org.apache.jackrabbit.webdav.property.ResourceType; +import org.apache.maven.archiva.configuration.ManagedRepositoryConfiguration; +import org.apache.maven.archiva.database.ArchivaAuditLogsDao; +import org.apache.maven.archiva.model.ArchivaAuditLogs; +import org.apache.maven.archiva.repository.audit.AuditEvent; +import org.apache.maven.archiva.repository.audit.AuditListener; +import org.apache.maven.archiva.scheduled.ArchivaTaskScheduler; +import org.apache.maven.archiva.scheduled.tasks.RepositoryTask; +import org.apache.maven.archiva.scheduled.tasks.TaskCreator; +import org.apache.maven.archiva.webdav.util.IndexWriter; +import org.apache.maven.archiva.webdav.util.MimeTypes; +import org.codehaus.plexus.taskqueue.TaskQueueException; +import org.joda.time.DateTime; +import org.joda.time.format.DateTimeFormatter; +import org.joda.time.format.ISODateTimeFormat; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +/** + */ +public class ArchivaDavResource + implements DavResource +{ + public static final String HIDDEN_PATH_PREFIX = "."; + + private final ArchivaDavResourceLocator locator; + + private final DavResourceFactory factory; + + private final File localResource; + + private final String logicalResource; + + private DavPropertySet properties = null; + + private LockManager lockManager; + + private final DavSession session; + + private String remoteAddr; + + private final ManagedRepositoryConfiguration repository; + + private final MimeTypes mimeTypes; + + private List<AuditListener> auditListeners; + + private String principal; + + public static final String COMPLIANCE_CLASS = "1, 2"; + + private ArchivaTaskScheduler scheduler; + + private Logger log = LoggerFactory.getLogger( ArchivaDavResource.class ); + + private ArchivaAuditLogsDao auditLogsDao; + + public ArchivaDavResource( String localResource, String logicalResource, ManagedRepositoryConfiguration repository, + DavSession session, ArchivaDavResourceLocator locator, DavResourceFactory factory, + MimeTypes mimeTypes, List<AuditListener> auditListeners, + ArchivaTaskScheduler scheduler, ArchivaAuditLogsDao auditLogsDao ) + { + this.localResource = new File( localResource ); + this.logicalResource = logicalResource; + this.locator = locator; + this.factory = factory; + this.session = session; + + // TODO: push into locator as well as moving any references out of the resource factory + this.repository = repository; + + // TODO: these should be pushed into the repository layer, along with the physical file operations in this class + this.mimeTypes = mimeTypes; + this.auditListeners = auditListeners; + this.scheduler = scheduler; + this.auditLogsDao = auditLogsDao; + } + + public ArchivaDavResource( String localResource, String logicalResource, ManagedRepositoryConfiguration repository, + String remoteAddr, String principal, DavSession session, ArchivaDavResourceLocator locator, + DavResourceFactory factory, MimeTypes mimeTypes, List<AuditListener> auditListeners, + ArchivaTaskScheduler scheduler, ArchivaAuditLogsDao auditLogsDao ) + { + this( localResource, logicalResource, repository, session, locator, factory, mimeTypes, auditListeners, + scheduler, auditLogsDao ); + + this.remoteAddr = remoteAddr; + this.principal = principal; + } + + public String getComplianceClass() + { + return COMPLIANCE_CLASS; + } + + public String getSupportedMethods() + { + return METHODS; + } + + public boolean exists() + { + return localResource.exists(); + } + + public boolean isCollection() + { + return localResource.isDirectory(); + } + + public String getDisplayName() + { + String resPath = getResourcePath(); + return ( resPath != null ) ? Text.getName( resPath ) : resPath; + } + + public DavResourceLocator getLocator() + { + return locator; + } + + public File getLocalResource() + { + return localResource; + } + + public String getResourcePath() + { + return locator.getResourcePath(); + } + + public String getHref() + { + return locator.getHref( isCollection() ); + } + + public long getModificationTime() + { + return localResource.lastModified(); + } + + public void spool( OutputContext outputContext ) + throws IOException + { + if ( !isCollection() ) + { + outputContext.setProperty( DavConstants.HEADER_CONTENT_LENGTH, Long.toString( localResource.length() ) ); + outputContext.setContentType( mimeTypes.getMimeType( localResource.getName() ) ); + } + + if ( !isCollection() && outputContext.hasStream() ) + { + FileInputStream is = null; + try + { + // Write content to stream + is = new FileInputStream( localResource ); + IOUtils.copy( is, outputContext.getOutputStream() ); + } + finally + { + IOUtils.closeQuietly( is ); + } + } + else if ( outputContext.hasStream() ) + { + IndexWriter writer = new IndexWriter( this, localResource, logicalResource ); + writer.write( outputContext ); + } + } + + public DavPropertyName[] getPropertyNames() + { + return getProperties().getPropertyNames(); + } + + public DavProperty getProperty( DavPropertyName name ) + { + return getProperties().get( name ); + } + + public DavPropertySet getProperties() + { + return initProperties(); + } + + public void setProperty( DavProperty property ) + throws DavException + { + } + + public void removeProperty( DavPropertyName propertyName ) + throws DavException + { + } + + public MultiStatusResponse alterProperties( DavPropertySet setProperties, DavPropertyNameSet removePropertyNames ) + throws DavException + { + return null; + } + + @SuppressWarnings( "unchecked" ) + public MultiStatusResponse alterProperties( List changeList ) + throws DavException + { + return null; + } + + public DavResource getCollection() + { + DavResource parent = null; + if ( getResourcePath() != null && !getResourcePath().equals( "/" ) ) + { + String parentPath = Text.getRelativeParent( getResourcePath(), 1 ); + if ( parentPath.equals( "" ) ) + { + parentPath = "/"; + } + DavResourceLocator parentloc = locator.getFactory().createResourceLocator( locator.getPrefix(), + parentPath ); + try + { + parent = factory.createResource( parentloc, session ); + } + catch ( DavException e ) + { + // should not occur + } + } + return parent; + } + + public void addMember( DavResource resource, InputContext inputContext ) + throws DavException + { + File localFile = new File( localResource, resource.getDisplayName() ); + boolean exists = localFile.exists(); + + if ( isCollection() && inputContext.hasStream() ) // New File + { + FileOutputStream stream = null; + try + { + stream = new FileOutputStream( localFile ); + IOUtils.copy( inputContext.getInputStream(), stream ); + } + catch ( IOException e ) + { + throw new DavException( HttpServletResponse.SC_INTERNAL_SERVER_ERROR, e ); + } + finally + { + IOUtils.closeQuietly( stream ); + } + + // TODO: a bad deployment shouldn't delete an existing file - do we need to write to a temporary location first? + long expectedContentLength = Long.parseLong( inputContext.getProperty( DavConstants.HEADER_CONTENT_LENGTH ) ); + long actualContentLength = localFile.length(); + // length of -1 is given for a chunked request or unknown length, in which case we accept what was uploaded + if ( expectedContentLength >= 0 && expectedContentLength != actualContentLength ) + { + String msg = + "Content Header length was " + expectedContentLength + " but was " + actualContentLength; + log.debug( "Upload failed: " + msg ); + + FileUtils.deleteQuietly( localFile ); + throw new DavException( HttpServletResponse.SC_BAD_REQUEST, msg ); + } + + queueRepositoryTask( localFile ); + + log.debug( + "File '" + resource.getDisplayName() + ( exists ? "' modified " : "' created " ) + "(current user '" + + this.principal + "')" ); + + triggerAuditEvent( resource, exists ? AuditEvent.MODIFY_FILE : AuditEvent.CREATE_FILE ); + } + else if ( !inputContext.hasStream() && isCollection() ) // New directory + { + localFile.mkdir(); + + log.debug( "Directory '" + resource.getDisplayName() + "' (current user '" + this.principal + "')" ); + + triggerAuditEvent( resource, AuditEvent.CREATE_DIR ); + } + else + { + String msg = "Could not write member " + resource.getResourcePath() + " at " + getResourcePath() + + " as this is not a DAV collection"; + log.debug( msg ); + throw new DavException( HttpServletResponse.SC_BAD_REQUEST, msg ); + } + } + + public DavResourceIterator getMembers() + { + List<DavResource> list = new ArrayList<DavResource>(); + if ( exists() && isCollection() ) + { + for ( String item : localResource.list() ) + { + try + { + if ( !item.startsWith( HIDDEN_PATH_PREFIX ) ) + { + String path = locator.getResourcePath() + '/' + item; + DavResourceLocator resourceLocator = locator.getFactory().createResourceLocator( + locator.getPrefix(), path ); + DavResource resource = factory.createResource( resourceLocator, session ); + + if ( resource != null ) + { + list.add( resource ); + } + log.debug( "Resource '" + item + "' retrieved by '" + this.principal + "'" ); + } + } + catch ( DavException e ) + { + // Should not occur + } + } + } + return new DavResourceIteratorImpl( list ); + } + + public void removeMember( DavResource member ) + throws DavException + { + File resource = checkDavResourceIsArchivaDavResource( member ).getLocalResource(); + + if ( resource.exists() ) + { + try + { + if ( resource.isDirectory() ) + { + FileUtils.deleteDirectory( resource ); + + triggerAuditEvent( member, AuditEvent.REMOVE_DIR ); + } + else + { + if ( !resource.delete() ) + { + throw new IOException( "Could not remove file" ); + } + + triggerAuditEvent( member, AuditEvent.REMOVE_FILE ); + } + log.debug( ( resource.isDirectory() ? "Directory '" : "File '" ) + member.getDisplayName() + + "' removed (current user '" + this.principal + "')" ); + } + catch ( IOException e ) + { + throw new DavException( HttpServletResponse.SC_INTERNAL_SERVER_ERROR ); + } + } + else + { + throw new DavException( HttpServletResponse.SC_NOT_FOUND ); + } + } + + private void triggerAuditEvent( DavResource member, String event ) + throws DavException + { + String path = logicalResource + "/" + member.getDisplayName(); + + ArchivaDavResource resource = checkDavResourceIsArchivaDavResource( member ); + triggerAuditEvent( resource.remoteAddr, resource.principal, locator.getRepositoryId(), path, + event ); + } + + public void move( DavResource destination ) + throws DavException + { + if ( !exists() ) + { + throw new DavException( HttpServletResponse.SC_NOT_FOUND, "Resource to copy does not exist." ); + } + + try + { + ArchivaDavResource resource = checkDavResourceIsArchivaDavResource( destination ); + if ( isCollection() ) + { + FileUtils.moveDirectory( getLocalResource(), resource.getLocalResource() ); + + triggerAuditEvent( remoteAddr, principal, locator.getRepositoryId(), logicalResource, AuditEvent.MOVE_DIRECTORY ); + } + else + { + FileUtils.moveFile( getLocalResource(), resource.getLocalResource() ); + + triggerAuditEvent( remoteAddr, principal, locator.getRepositoryId(), logicalResource, AuditEvent.MOVE_FILE ); + } + + log.debug( ( isCollection() ? "Directory '" : "File '" ) + getLocalResource().getName() + "' moved to '" + + destination + "' (current user '" + this.principal + "')" ); + } + catch ( IOException e ) + { + throw new DavException( HttpServletResponse.SC_INTERNAL_SERVER_ERROR, e ); + } + } + + public void copy( DavResource destination, boolean shallow ) + throws DavException + { + if ( !exists() ) + { + throw new DavException( HttpServletResponse.SC_NOT_FOUND, "Resource to copy does not exist." ); + } + + if ( shallow && isCollection() ) + { + throw new DavException( DavServletResponse.SC_FORBIDDEN, "Unable to perform shallow copy for collection" ); + } + + try + { + ArchivaDavResource resource = checkDavResourceIsArchivaDavResource( destination ); + if ( isCollection() ) + { + FileUtils.copyDirectory( getLocalResource(), resource.getLocalResource() ); + + triggerAuditEvent( remoteAddr, principal, locator.getRepositoryId(), logicalResource, AuditEvent.COPY_DIRECTORY ); + } + else + { + FileUtils.copyFile( getLocalResource(), resource.getLocalResource() ); + + triggerAuditEvent( remoteAddr, principal, locator.getRepositoryId(), logicalResource, AuditEvent.COPY_FILE ); + } + log.debug( ( isCollection() ? "Directory '" : "File '" ) + getLocalResource().getName() + "' copied to '" + + destination + "' (current user '" + this.principal + "')" ); + } + catch ( IOException e ) + { + throw new DavException( HttpServletResponse.SC_INTERNAL_SERVER_ERROR, e ); + } + } + + public boolean isLockable( Type type, Scope scope ) + { + return Type.WRITE.equals( type ) && Scope.EXCLUSIVE.equals( scope ); + } + + public boolean hasLock( Type type, Scope scope ) + { + return getLock( type, scope ) != null; + } + + public ActiveLock getLock( Type type, Scope scope ) + { + ActiveLock lock = null; + if ( exists() && Type.WRITE.equals( type ) && Scope.EXCLUSIVE.equals( scope ) ) + { + lock = lockManager.getLock( type, scope, this ); + } + return lock; + } + + public ActiveLock[] getLocks() + { + ActiveLock writeLock = getLock( Type.WRITE, Scope.EXCLUSIVE ); + return ( writeLock != null ) ? new ActiveLock[]{writeLock} : new ActiveLock[0]; + } + + public ActiveLock lock( LockInfo lockInfo ) + throws DavException + { + ActiveLock lock = null; + if ( isLockable( lockInfo.getType(), lockInfo.getScope() ) ) + { + lock = lockManager.createLock( lockInfo, this ); + } + else + { + throw new DavException( DavServletResponse.SC_PRECONDITION_FAILED, "Unsupported lock type or scope." ); + } + return lock; + } + + public ActiveLock refreshLock( LockInfo lockInfo, String lockToken ) + throws DavException + { + if ( !exists() ) + { + throw new DavException( DavServletResponse.SC_NOT_FOUND ); + } + ActiveLock lock = getLock( lockInfo.getType(), lockInfo.getScope() ); + if ( lock == null ) + { + throw new DavException( DavServletResponse.SC_PRECONDITION_FAILED, + "No lock with the given type/scope present on resource " + getResourcePath() ); + } + + lock = lockManager.refreshLock( lockInfo, lockToken, this ); + + return lock; + } + + public void unlock( String lockToken ) + throws DavException + { + ActiveLock lock = getLock( Type.WRITE, Scope.EXCLUSIVE ); + if ( lock == null ) + { + throw new DavException( HttpServletResponse.SC_PRECONDITION_FAILED ); + } + else if ( lock.isLockedByToken( lockToken ) ) + { + lockManager.releaseLock( lockToken, this ); + } + else + { + throw new DavException( DavServletResponse.SC_LOCKED ); + } + } + + public void addLockManager( LockManager lockManager ) + { + this.lockManager = lockManager; + } + + public DavResourceFactory getFactory() + { + return factory; + } + + public DavSession getSession() + { + return session; + } + + /** + * Fill the set of properties + */ + protected DavPropertySet initProperties() + { + if ( !exists() ) + { + properties = new DavPropertySet(); + } + + if ( properties != null ) + { + return properties; + } + + DavPropertySet properties = new DavPropertySet(); + + // set (or reset) fundamental properties + if ( getDisplayName() != null ) + { + properties.add( new DefaultDavProperty( DavPropertyName.DISPLAYNAME, getDisplayName() ) ); + } + if ( isCollection() ) + { + properties.add( new ResourceType( ResourceType.COLLECTION ) ); + // Windows XP support + properties.add( new DefaultDavProperty( DavPropertyName.ISCOLLECTION, "1" ) ); + } + else + { + properties.add( new ResourceType( ResourceType.DEFAULT_RESOURCE ) ); + + // Windows XP support + properties.add( new DefaultDavProperty( DavPropertyName.ISCOLLECTION, "0" ) ); + } + + // Need to get the ISO8601 date for properties + DateTime dt = new DateTime( localResource.lastModified() ); + DateTimeFormatter fmt = ISODateTimeFormat.dateTime(); + String modifiedDate = fmt.print( dt ); + + properties.add( new DefaultDavProperty( DavPropertyName.GETLASTMODIFIED, modifiedDate ) ); + + properties.add( new DefaultDavProperty( DavPropertyName.CREATIONDATE, modifiedDate ) ); + + properties.add( new DefaultDavProperty( DavPropertyName.GETCONTENTLENGTH, localResource.length() ) ); + + this.properties = properties; + + return properties; + } + + private ArchivaDavResource checkDavResourceIsArchivaDavResource( DavResource resource ) + throws DavException + { + if ( !( resource instanceof ArchivaDavResource ) ) + { + throw new DavException( HttpServletResponse.SC_INTERNAL_SERVER_ERROR, + "DavResource is not instance of ArchivaDavResource" ); + } + return (ArchivaDavResource) resource; + } + + private void triggerAuditEvent( String remoteIP, String principal, String repositoryId, String resource, + String action ) + { + AuditEvent event = new AuditEvent( repositoryId, principal, resource, action ); + event.setRemoteIP( remoteIP ); + + for ( AuditListener listener : auditListeners ) + { + listener.auditEvent( event ); + } + + // identify as artifact deployment/upload + if( action.equals( AuditEvent.CREATE_FILE ) ) + { + action = AuditEvent.UPLOAD_FILE; + } + + String user = principal; + if( principal == null ) + { + user = "guest"; + } + + ArchivaAuditLogs auditLogs = new ArchivaAuditLogs(); + auditLogs.setArtifact( resource ); + auditLogs.setEvent( action ); + auditLogs.setEventDate( Calendar.getInstance().getTime() ); + auditLogs.setRepositoryId( repositoryId ); + auditLogs.setUsername( user ); + + auditLogsDao.saveAuditLogs( auditLogs ); + } + + private void queueRepositoryTask( File localFile ) + { + RepositoryTask task = TaskCreator.createRepositoryTask( repository.getId(), localFile, false, true ); + + try + { + scheduler.queueRepositoryTask( task ); + } + catch ( TaskQueueException e ) + { + log.error( + "Unable to queue repository task to execute consumers on resource file ['" + localFile.getName() + + "']." ); + } + } +} |