2014-12-22 23:24:45 +01:00
// Package s3 provides a storagedriver.StorageDriver implementation to
// store blobs in Amazon S3 cloud storage.
//
2016-01-22 03:17:53 +01:00
// This package leverages the official aws client library for interfacing with
// S3.
2014-12-22 23:24:45 +01:00
//
2016-01-22 03:17:53 +01:00
// Because S3 is a key, value store the Stat call does not support last modification
2014-12-22 23:24:45 +01:00
// time for directories (directories are an abstraction for key, value stores)
//
2016-01-22 03:17:53 +01:00
// Keep in mind that S3 guarantees only read-after-write consistency for new
// objects, but no read-after-update or list-after-write consistency.
2014-10-25 01:37:25 +02:00
package s3
import (
"bytes"
2014-10-29 02:15:40 +01:00
"fmt"
2014-10-25 01:37:25 +02:00
"io"
2014-12-19 18:16:51 +01:00
"io/ioutil"
2014-10-25 01:37:25 +02:00
"net/http"
2015-04-22 23:31:34 +02:00
"reflect"
2014-10-25 01:37:25 +02:00
"strconv"
2014-12-19 18:16:51 +01:00
"strings"
2015-04-23 00:07:18 +02:00
"sync"
2014-12-19 18:16:51 +01:00
"time"
2014-10-25 01:37:25 +02:00
2015-04-24 05:07:32 +02:00
"github.com/Sirupsen/logrus"
2016-01-22 03:17:53 +01:00
"github.com/aws/aws-sdk-go/aws"
"github.com/aws/aws-sdk-go/aws/awserr"
"github.com/aws/aws-sdk-go/aws/credentials"
"github.com/aws/aws-sdk-go/aws/credentials/ec2rolecreds"
"github.com/aws/aws-sdk-go/aws/ec2metadata"
"github.com/aws/aws-sdk-go/aws/request"
"github.com/aws/aws-sdk-go/aws/session"
"github.com/aws/aws-sdk-go/service/s3"
2015-04-28 00:58:58 +02:00
"github.com/docker/distribution/context"
2016-01-21 01:40:58 +01:00
"github.com/docker/distribution/registry/client/transport"
2015-02-11 03:14:23 +01:00
storagedriver "github.com/docker/distribution/registry/storage/driver"
"github.com/docker/distribution/registry/storage/driver/base"
"github.com/docker/distribution/registry/storage/driver/factory"
2014-10-25 01:37:25 +02:00
)
2016-01-22 03:17:53 +01:00
const driverName = "s3aws"
2014-10-29 02:15:40 +01:00
2014-10-29 20:14:19 +01:00
// minChunkSize defines the minimum multipart upload chunk size
// S3 API requires multipart upload chunks to be at least 5MB
2015-01-24 01:46:43 +01:00
const minChunkSize = 5 << 20
const defaultChunkSize = 2 * minChunkSize
2014-10-25 01:37:25 +02:00
2014-12-19 18:16:51 +01:00
// listMax is the largest amount of objects you can request from S3 in a list call
const listMax = 1000
2014-10-25 01:37:25 +02:00
2016-01-22 03:17:53 +01:00
// validRegions maps known s3 region identifiers to region descriptors
var validRegions = map [ string ] struct { } { }
2015-01-07 11:18:42 +01:00
//DriverParameters A struct that encapsulates all of the driver parameters after all values have been set
type DriverParameters struct {
AccessKey string
SecretKey string
Bucket string
2016-01-22 03:17:53 +01:00
Region string
2015-01-07 11:18:42 +01:00
Encrypt bool
Secure bool
2015-01-24 01:46:43 +01:00
ChunkSize int64
2015-01-07 11:18:42 +01:00
RootDirectory string
2016-01-22 03:17:53 +01:00
StorageClass string
2016-01-21 01:40:58 +01:00
UserAgent string
2015-01-07 11:18:42 +01:00
}
2014-10-29 02:15:40 +01:00
func init ( ) {
2016-01-22 03:17:53 +01:00
for _ , region := range [ ] string {
"us-east-1" ,
"us-west-1" ,
"us-west-2" ,
"eu-west-1" ,
"eu-central-1" ,
"ap-southeast-1" ,
"ap-southeast-2" ,
"ap-northeast-1" ,
"ap-northeast-2" ,
"sa-east-1" ,
} {
validRegions [ region ] = struct { } { }
}
// Register this as the default s3 driver in addition to s3aws
factory . Register ( "s3" , & s3DriverFactory { } )
2014-11-18 00:44:07 +01:00
factory . Register ( driverName , & s3DriverFactory { } )
2014-10-29 02:15:40 +01:00
}
2014-10-29 20:14:19 +01:00
// s3DriverFactory implements the factory.StorageDriverFactory interface
2014-10-29 02:15:40 +01:00
type s3DriverFactory struct { }
2014-12-19 18:16:51 +01:00
func ( factory * s3DriverFactory ) Create ( parameters map [ string ] interface { } ) ( storagedriver . StorageDriver , error ) {
2014-10-29 02:15:40 +01:00
return FromParameters ( parameters )
}
2015-02-04 01:54:52 +01:00
type driver struct {
2014-12-19 18:16:51 +01:00
S3 * s3 . S3
2016-01-22 03:17:53 +01:00
Bucket string
2015-01-24 01:46:43 +01:00
ChunkSize int64
2014-12-19 18:16:51 +01:00
Encrypt bool
2015-01-24 01:46:43 +01:00
RootDirectory string
2016-01-22 03:17:53 +01:00
StorageClass string
2015-04-23 00:07:18 +02:00
pool sync . Pool // pool []byte buffers used for WriteStream
zeros [ ] byte // shared, zero-valued buffer used for WriteStream
2014-10-25 01:37:25 +02:00
}
2015-02-04 01:54:52 +01:00
type baseEmbed struct {
base . Base
}
// Driver is a storagedriver.StorageDriver implementation backed by Amazon S3
// Objects are stored at absolute keys in the provided bucket.
type Driver struct {
baseEmbed
}
2014-11-18 00:44:07 +01:00
// FromParameters constructs a new Driver with a given parameters map
2014-10-29 02:15:40 +01:00
// Required parameters:
// - accesskey
// - secretkey
// - region
// - bucket
// - encrypt
2014-12-19 18:16:51 +01:00
func FromParameters ( parameters map [ string ] interface { } ) ( * Driver , error ) {
2014-12-22 23:24:45 +01:00
// Providing no values for these is valid in case the user is authenticating
// with an IAM on an ec2 instance (in which case the instance credentials will
// be summoned when GetAuth is called)
2016-01-22 03:17:53 +01:00
accessKey , ok := parameters [ "accesskey" ]
if ! ok {
2015-01-24 00:50:55 +01:00
accessKey = ""
}
2016-01-22 03:17:53 +01:00
secretKey , ok := parameters [ "secretkey" ]
if ! ok {
2015-01-24 00:50:55 +01:00
secretKey = ""
}
2014-10-29 02:15:40 +01:00
2016-01-22 03:17:53 +01:00
regionName , ok := parameters [ "region" ]
if ! ok || fmt . Sprint ( regionName ) == "" {
2014-10-29 02:15:40 +01:00
return nil , fmt . Errorf ( "No region parameter provided" )
}
2016-01-22 03:17:53 +01:00
region := fmt . Sprint ( regionName )
_ , ok = validRegions [ region ]
if ! ok {
2014-11-13 02:19:19 +01:00
return nil , fmt . Errorf ( "Invalid region provided: %v" , region )
2014-10-29 02:15:40 +01:00
}
2016-01-22 03:17:53 +01:00
bucket , ok := parameters [ "bucket" ]
if ! ok || fmt . Sprint ( bucket ) == "" {
2014-10-29 02:15:40 +01:00
return nil , fmt . Errorf ( "No bucket parameter provided" )
}
2015-01-07 10:51:29 +01:00
encryptBool := false
2016-01-22 03:17:53 +01:00
encrypt , ok := parameters [ "encrypt" ]
if ok {
encryptBool , ok = encrypt . ( bool )
if ! ok {
2014-12-30 18:31:12 +01:00
return nil , fmt . Errorf ( "The encrypt parameter should be a boolean" )
}
2014-10-29 02:15:40 +01:00
}
2014-12-19 18:16:51 +01:00
2015-01-07 10:45:31 +01:00
secureBool := true
2016-01-22 03:17:53 +01:00
secure , ok := parameters [ "secure" ]
if ok {
secureBool , ok = secure . ( bool )
if ! ok {
2014-12-29 21:29:54 +01:00
return nil , fmt . Errorf ( "The secure parameter should be a boolean" )
}
2015-01-07 10:45:31 +01:00
}
2015-01-24 01:46:43 +01:00
chunkSize := int64 ( defaultChunkSize )
2016-01-22 03:17:53 +01:00
chunkSizeParam , ok := parameters [ "chunksize" ]
if ok {
switch v := chunkSizeParam . ( type ) {
case string :
vv , err := strconv . ParseInt ( v , 0 , 64 )
if err != nil {
return nil , fmt . Errorf ( "chunksize parameter must be an integer, %v invalid" , chunkSizeParam )
}
chunkSize = vv
case int64 :
chunkSize = v
case int , uint , int32 , uint32 , uint64 :
chunkSize = reflect . ValueOf ( v ) . Convert ( reflect . TypeOf ( chunkSize ) ) . Int ( )
default :
return nil , fmt . Errorf ( "invalid valud for chunksize: %#v" , chunkSizeParam )
2015-04-22 23:31:34 +02:00
}
2016-01-22 03:17:53 +01:00
if chunkSize < minChunkSize {
return nil , fmt . Errorf ( "The chunksize %#v parameter should be a number that is larger than or equal to %d" , chunkSize , minChunkSize )
}
2015-01-24 01:46:43 +01:00
}
2016-01-22 03:17:53 +01:00
rootDirectory , ok := parameters [ "rootdirectory" ]
if ! ok {
2014-12-30 18:31:12 +01:00
rootDirectory = ""
2014-12-19 18:16:51 +01:00
}
2016-01-22 03:17:53 +01:00
storageClass := s3 . StorageClassStandard
storageClassParam , ok := parameters [ "storageclass" ]
if ok {
2016-01-29 00:48:49 +01:00
storageClassString , ok := storageClassParam . ( string )
if ! ok {
2016-01-22 03:17:53 +01:00
return nil , fmt . Errorf ( "The storageclass parameter must be one of %v, %v invalid" , [ ] string { s3 . StorageClassStandard , s3 . StorageClassReducedRedundancy } , storageClassParam )
2016-01-29 00:48:49 +01:00
}
// All valid storage class parameters are UPPERCASE, so be a bit more flexible here
2016-01-22 03:17:53 +01:00
storageClassString = strings . ToUpper ( storageClassString )
if storageClassString != s3 . StorageClassStandard && storageClassString != s3 . StorageClassReducedRedundancy {
return nil , fmt . Errorf ( "The storageclass parameter must be one of %v, %v invalid" , [ ] string { s3 . StorageClassStandard , s3 . StorageClassReducedRedundancy } , storageClassParam )
2016-01-29 00:48:49 +01:00
}
2016-01-22 03:17:53 +01:00
storageClass = storageClassString
2016-01-29 00:48:49 +01:00
}
2016-01-22 03:17:53 +01:00
userAgent , ok := parameters [ "useragent" ]
if ! ok {
2016-01-21 01:40:58 +01:00
userAgent = ""
}
2015-01-07 11:18:42 +01:00
params := DriverParameters {
fmt . Sprint ( accessKey ) ,
fmt . Sprint ( secretKey ) ,
fmt . Sprint ( bucket ) ,
region ,
encryptBool ,
secureBool ,
2015-01-24 01:46:43 +01:00
chunkSize ,
2015-01-07 11:18:42 +01:00
fmt . Sprint ( rootDirectory ) ,
2016-01-29 00:48:49 +01:00
storageClass ,
2016-01-21 01:40:58 +01:00
fmt . Sprint ( userAgent ) ,
2015-01-07 11:18:42 +01:00
}
return New ( params )
2014-10-29 02:15:40 +01:00
}
2014-11-18 00:44:07 +01:00
// New constructs a new Driver with the given AWS credentials, region, encryption flag, and
2014-10-29 20:14:19 +01:00
// bucketName
2015-01-07 11:18:42 +01:00
func New ( params DriverParameters ) ( * Driver , error ) {
2016-01-22 03:17:53 +01:00
awsConfig := aws . NewConfig ( )
creds := credentials . NewChainCredentials ( [ ] credentials . Provider {
& credentials . StaticProvider {
Value : credentials . Value {
AccessKeyID : params . AccessKey ,
SecretAccessKey : params . SecretKey ,
} ,
} ,
& credentials . EnvProvider { } ,
& credentials . SharedCredentialsProvider { } ,
& ec2rolecreds . EC2RoleProvider { Client : ec2metadata . New ( session . New ( ) ) } ,
} )
2014-12-29 21:29:54 +01:00
2016-01-22 03:17:53 +01:00
awsConfig . WithCredentials ( creds )
awsConfig . WithRegion ( params . Region )
awsConfig . WithDisableSSL ( ! params . Secure )
// awsConfig.WithMaxRetries(10)
2016-01-21 01:40:58 +01:00
if params . UserAgent != "" {
2016-01-22 03:17:53 +01:00
awsConfig . WithHTTPClient ( & http . Client {
Transport : transport . NewTransport ( http . DefaultTransport , transport . NewHeaderRequestModifier ( http . Header { http . CanonicalHeaderKey ( "User-Agent" ) : [ ] string { params . UserAgent } } ) ) ,
} )
2016-01-21 01:40:58 +01:00
}
2014-10-25 01:37:25 +02:00
2016-01-22 03:17:53 +01:00
s3obj := s3 . New ( session . New ( awsConfig ) )
2016-01-21 01:40:58 +01:00
2014-12-21 07:48:42 +01:00
// TODO Currently multipart uploads have no timestamps, so this would be unwise
// if you initiated a new s3driver while another one is running on the same bucket.
// multis, _, err := bucket.ListMulti("", "")
// if err != nil {
// return nil, err
// }
// for _, multi := range multis {
// err := multi.Abort()
// //TODO appropriate to do this error checking?
// if err != nil {
// return nil, err
// }
// }
2014-12-19 18:16:51 +01:00
2015-02-04 01:54:52 +01:00
d := & driver {
2015-01-24 01:46:43 +01:00
S3 : s3obj ,
2016-01-22 03:17:53 +01:00
Bucket : params . Bucket ,
2015-01-24 01:46:43 +01:00
ChunkSize : params . ChunkSize ,
Encrypt : params . Encrypt ,
2015-02-04 01:54:52 +01:00
RootDirectory : params . RootDirectory ,
2016-01-29 00:48:49 +01:00
StorageClass : params . StorageClass ,
2015-04-23 00:07:18 +02:00
zeros : make ( [ ] byte , params . ChunkSize ) ,
}
d . pool . New = func ( ) interface { } {
return make ( [ ] byte , d . ChunkSize )
2015-02-04 01:54:52 +01:00
}
return & Driver {
baseEmbed : baseEmbed {
Base : base . Base {
StorageDriver : d ,
} ,
} ,
} , nil
2014-10-25 01:37:25 +02:00
}
2014-10-29 20:14:19 +01:00
// Implement the storagedriver.StorageDriver interface
2015-04-23 02:30:01 +02:00
func ( d * driver ) Name ( ) string {
return driverName
}
2014-11-18 00:44:07 +01:00
// GetContent retrieves the content stored at "path" as a []byte.
2015-04-28 00:58:58 +02:00
func ( d * driver ) GetContent ( ctx context . Context , path string ) ( [ ] byte , error ) {
2016-01-22 03:17:53 +01:00
reader , err := d . ReadStream ( ctx , path , 0 )
2014-11-19 02:41:48 +01:00
if err != nil {
2016-01-22 03:17:53 +01:00
return nil , err
2014-11-19 02:41:48 +01:00
}
2016-01-22 03:17:53 +01:00
return ioutil . ReadAll ( reader )
2014-10-25 01:37:25 +02:00
}
2014-11-18 00:44:07 +01:00
// PutContent stores the []byte content at a location designated by "path".
2015-04-28 00:58:58 +02:00
func ( d * driver ) PutContent ( ctx context . Context , path string , contents [ ] byte ) error {
2016-01-22 03:17:53 +01:00
_ , err := d . S3 . PutObject ( & s3 . PutObjectInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
ContentType : d . getContentType ( ) ,
ACL : d . getACL ( ) ,
ServerSideEncryption : d . getEncryptionMode ( ) ,
StorageClass : d . getStorageClass ( ) ,
Body : bytes . NewReader ( contents ) ,
} )
return parseError ( path , err )
2014-10-25 01:37:25 +02:00
}
2014-11-18 00:44:07 +01:00
// ReadStream retrieves an io.ReadCloser for the content stored at "path" with a
// given byte offset.
2015-04-28 00:58:58 +02:00
func ( d * driver ) ReadStream ( ctx context . Context , path string , offset int64 ) ( io . ReadCloser , error ) {
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . GetObject ( & s3 . GetObjectInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
Range : aws . String ( "bytes=" + strconv . FormatInt ( offset , 10 ) + "-" ) ,
} )
2014-10-25 01:37:25 +02:00
2014-11-19 02:41:48 +01:00
if err != nil {
2016-01-22 03:17:53 +01:00
if s3Err , ok := err . ( awserr . Error ) ; ok && s3Err . Code ( ) == "InvalidRange" {
2014-12-19 18:16:51 +01:00
return ioutil . NopCloser ( bytes . NewReader ( nil ) ) , nil
}
return nil , parseError ( path , err )
2014-10-25 01:37:25 +02:00
}
2014-11-19 02:41:48 +01:00
return resp . Body , nil
2014-10-25 01:37:25 +02:00
}
2014-12-23 09:54:01 +01:00
// WriteStream stores the contents of the provided io.Reader at a
2014-12-22 23:24:45 +01:00
// location designated by the given path. The driver will know it has
// received the full contents when the reader returns io.EOF. The number
2014-12-23 09:54:01 +01:00
// of successfully READ bytes will be returned, even if an error is
2014-12-22 23:24:45 +01:00
// returned. May be used to resume writing a stream by providing a nonzero
// offset. Offsets past the current size will write from the position
// beyond the end of the file.
2015-04-28 00:58:58 +02:00
func ( d * driver ) WriteStream ( ctx context . Context , path string , offset int64 , reader io . Reader ) ( totalRead int64 , err error ) {
2016-01-22 03:17:53 +01:00
var partNumber int64 = 1
2014-12-19 18:20:07 +01:00
bytesRead := 0
2015-01-27 02:51:59 +01:00
var putErrChan chan error
2016-01-22 03:17:53 +01:00
parts := [ ] * s3 . CompletedPart { }
2015-04-28 23:06:24 +02:00
done := make ( chan struct { } ) // stopgap to free up waiting goroutines
2014-12-19 18:16:51 +01:00
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . CreateMultipartUpload ( & s3 . CreateMultipartUploadInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
ContentType : d . getContentType ( ) ,
ACL : d . getACL ( ) ,
ServerSideEncryption : d . getEncryptionMode ( ) ,
StorageClass : d . getStorageClass ( ) ,
} )
2014-10-25 01:37:25 +02:00
if err != nil {
2014-12-19 18:16:51 +01:00
return 0 , err
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
uploadID := resp . UploadId
2015-04-23 00:07:18 +02:00
buf := d . getbuf ( )
2014-12-19 18:16:51 +01:00
// We never want to leave a dangling multipart upload, our only consistent state is
// when there is a whole object at path. This is in order to remain consistent with
// the stat call.
//
// Note that if the machine dies before executing the defer, we will be left with a dangling
// multipart upload, which will eventually be cleaned up, but we will lose all of the progress
// made prior to the machine crashing.
defer func ( ) {
2015-01-27 02:51:59 +01:00
if putErrChan != nil {
if putErr := <- putErrChan ; putErr != nil {
err = putErr
}
}
2014-12-19 18:16:51 +01:00
if len ( parts ) > 0 {
2016-01-22 03:17:53 +01:00
_ , err := d . S3 . CompleteMultipartUpload ( & s3 . CompleteMultipartUploadInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
UploadId : uploadID ,
MultipartUpload : & s3 . CompletedMultipartUpload {
Parts : parts ,
} ,
} )
if err != nil {
// TODO (brianbland): log errors here
d . S3 . AbortMultipartUpload ( & s3 . AbortMultipartUploadInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
UploadId : uploadID ,
} )
2014-12-19 18:16:51 +01:00
}
}
2015-04-23 00:07:18 +02:00
d . putbuf ( buf ) // needs to be here to pick up new buf value
2015-04-28 23:06:24 +02:00
close ( done ) // free up any waiting goroutines
2014-12-19 18:16:51 +01:00
} ( )
2014-10-25 01:37:25 +02:00
2014-12-19 18:20:07 +01:00
// Fills from 0 to total from current
fromSmallCurrent := func ( total int64 ) error {
2015-04-28 00:58:58 +02:00
current , err := d . ReadStream ( ctx , path , 0 )
2014-12-19 18:16:51 +01:00
if err != nil {
2014-12-19 18:20:07 +01:00
return err
2014-12-19 18:16:51 +01:00
}
2014-12-19 18:20:07 +01:00
bytesRead = 0
for int64 ( bytesRead ) < total {
//The loop should very rarely enter a second iteration
2014-12-20 09:32:48 +01:00
nn , err := current . Read ( buf [ bytesRead : total ] )
2014-12-19 23:18:27 +01:00
bytesRead += nn
if err != nil {
2014-12-20 09:32:48 +01:00
if err != io . EOF {
return err
}
break
2014-12-19 18:16:51 +01:00
}
2014-12-19 18:20:07 +01:00
}
return nil
}
// Fills from parameter to chunkSize from reader
fromReader := func ( from int64 ) error {
bytesRead = 0
2015-01-24 01:46:43 +01:00
for from + int64 ( bytesRead ) < d . ChunkSize {
2014-12-20 09:32:48 +01:00
nn , err := reader . Read ( buf [ from + int64 ( bytesRead ) : ] )
2014-12-19 18:20:07 +01:00
totalRead += int64 ( nn )
bytesRead += nn
2014-12-19 23:18:27 +01:00
if err != nil {
2014-12-20 09:32:48 +01:00
if err != io . EOF {
2014-12-19 23:18:27 +01:00
return err
}
break
2014-12-19 18:20:07 +01:00
}
}
2015-01-27 02:51:59 +01:00
if putErrChan == nil {
putErrChan = make ( chan error )
} else {
if putErr := <- putErrChan ; putErr != nil {
putErrChan = nil
return putErr
2014-12-20 09:32:48 +01:00
}
2014-12-19 18:20:07 +01:00
}
2015-01-27 02:51:59 +01:00
go func ( bytesRead int , from int64 , buf [ ] byte ) {
2015-04-23 00:07:18 +02:00
defer d . putbuf ( buf ) // this buffer gets dropped after this call
2015-04-24 05:07:32 +02:00
// DRAGONS(stevvooe): There are few things one might want to know
// about this section. First, the putErrChan is expecting an error
// and a nil or just a nil to come through the channel. This is
// covered by the silly defer below. The other aspect is the s3
// retry backoff to deal with RequestTimeout errors. Even though
// the underlying s3 library should handle it, it doesn't seem to
// be part of the shouldRetry function (see AdRoll/goamz/s3).
defer func ( ) {
2015-04-28 23:06:24 +02:00
select {
case putErrChan <- nil : // for some reason, we do this no matter what.
case <- done :
return // ensure we don't leak the goroutine
}
2015-04-24 05:07:32 +02:00
} ( )
if bytesRead <= 0 {
return
}
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . UploadPart ( & s3 . UploadPartInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
PartNumber : aws . Int64 ( partNumber ) ,
UploadId : uploadID ,
Body : bytes . NewReader ( buf [ 0 : int64 ( bytesRead ) + from ] ) ,
} )
2015-04-24 05:07:32 +02:00
if err != nil {
logrus . Errorf ( "error putting part, aborting: %v" , err )
2015-04-28 23:06:24 +02:00
select {
case putErrChan <- err :
case <- done :
return // don't leak the goroutine
}
2015-04-24 05:07:32 +02:00
}
// parts and partNumber are safe, because this function is the
// only one modifying them and we force it to be executed
// serially.
2016-01-22 03:17:53 +01:00
parts = append ( parts , & s3 . CompletedPart {
ETag : resp . ETag ,
PartNumber : aws . Int64 ( partNumber ) ,
} )
2015-04-24 05:07:32 +02:00
partNumber ++
2015-01-27 02:51:59 +01:00
} ( bytesRead , from , buf )
2015-04-23 00:07:18 +02:00
buf = d . getbuf ( ) // use a new buffer for the next call
2014-12-19 18:20:07 +01:00
return nil
}
if offset > 0 {
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . HeadObject ( & s3 . HeadObjectInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
} )
2014-12-19 18:20:07 +01:00
if err != nil {
2016-01-22 03:17:53 +01:00
if s3Err , ok := err . ( awserr . Error ) ; ! ok || s3Err . Code ( ) != "NoSuchKey" {
2014-12-19 18:16:51 +01:00
return 0 , err
}
2014-12-19 18:20:07 +01:00
}
2014-12-19 18:16:51 +01:00
2014-12-19 18:20:07 +01:00
currentLength := int64 ( 0 )
2016-01-22 03:17:53 +01:00
if err == nil && resp . ContentLength != nil {
currentLength = * resp . ContentLength
2014-12-19 18:20:07 +01:00
}
2014-12-19 18:16:51 +01:00
2014-12-19 18:20:07 +01:00
if currentLength >= offset {
2015-01-24 01:46:43 +01:00
if offset < d . ChunkSize {
2014-12-19 18:20:07 +01:00
// chunkSize > currentLength >= offset
if err = fromSmallCurrent ( offset ) ; err != nil {
return totalRead , err
}
if err = fromReader ( offset ) ; err != nil {
return totalRead , err
}
2014-12-20 09:32:48 +01:00
2015-01-24 01:46:43 +01:00
if totalRead + offset < d . ChunkSize {
2014-12-20 09:32:48 +01:00
return totalRead , nil
}
2014-12-19 18:20:07 +01:00
} else {
// currentLength >= offset >= chunkSize
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . UploadPartCopy ( & s3 . UploadPartCopyInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
PartNumber : aws . Int64 ( partNumber ) ,
UploadId : uploadID ,
CopySource : aws . String ( d . Bucket + "/" + d . s3Path ( path ) ) ,
CopySourceRange : aws . String ( "bytes=0-" + strconv . FormatInt ( offset - 1 , 10 ) ) ,
} )
2014-12-19 18:20:07 +01:00
if err != nil {
return 0 , err
}
2014-12-19 18:16:51 +01:00
2016-01-22 03:17:53 +01:00
parts = append ( parts , & s3 . CompletedPart {
ETag : resp . CopyPartResult . ETag ,
PartNumber : aws . Int64 ( partNumber ) ,
} )
2014-12-20 09:32:48 +01:00
partNumber ++
2014-12-19 18:16:51 +01:00
}
} else {
2014-12-19 18:20:07 +01:00
// Fills between parameters with 0s but only when to - from <= chunkSize
fromZeroFillSmall := func ( from , to int64 ) error {
bytesRead = 0
for from + int64 ( bytesRead ) < to {
2015-04-23 00:07:18 +02:00
nn , err := bytes . NewReader ( d . zeros ) . Read ( buf [ from + int64 ( bytesRead ) : to ] )
2014-12-19 18:20:07 +01:00
bytesRead += nn
2014-12-19 23:18:27 +01:00
if err != nil {
2014-12-19 18:20:07 +01:00
return err
}
}
return nil
2014-12-19 18:16:51 +01:00
}
2014-12-19 18:20:07 +01:00
// Fills between parameters with 0s, making new parts
fromZeroFillLarge := func ( from , to int64 ) error {
bytesRead64 := int64 ( 0 )
2015-01-24 01:46:43 +01:00
for to - ( from + bytesRead64 ) >= d . ChunkSize {
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . UploadPart ( & s3 . UploadPartInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
PartNumber : aws . Int64 ( partNumber ) ,
UploadId : uploadID ,
Body : bytes . NewReader ( d . zeros ) ,
} )
2014-12-19 18:20:07 +01:00
if err != nil {
return err
}
2015-01-24 01:46:43 +01:00
bytesRead64 += d . ChunkSize
2014-12-19 18:20:07 +01:00
2016-01-22 03:17:53 +01:00
parts = append ( parts , & s3 . CompletedPart {
ETag : resp . ETag ,
PartNumber : aws . Int64 ( partNumber ) ,
} )
2014-12-19 18:20:07 +01:00
partNumber ++
}
2015-01-24 01:46:43 +01:00
return fromZeroFillSmall ( 0 , ( to - from ) % d . ChunkSize )
2014-12-19 18:20:07 +01:00
}
// currentLength < offset
2015-01-24 01:46:43 +01:00
if currentLength < d . ChunkSize {
if offset < d . ChunkSize {
2014-12-19 18:20:07 +01:00
// chunkSize > offset > currentLength
if err = fromSmallCurrent ( currentLength ) ; err != nil {
return totalRead , err
}
if err = fromZeroFillSmall ( currentLength , offset ) ; err != nil {
return totalRead , err
}
if err = fromReader ( offset ) ; err != nil {
return totalRead , err
}
2015-01-24 01:46:43 +01:00
if totalRead + offset < d . ChunkSize {
2014-12-19 18:20:07 +01:00
return totalRead , nil
}
} else {
// offset >= chunkSize > currentLength
if err = fromSmallCurrent ( currentLength ) ; err != nil {
return totalRead , err
}
2015-01-24 01:46:43 +01:00
if err = fromZeroFillSmall ( currentLength , d . ChunkSize ) ; err != nil {
2014-12-19 18:20:07 +01:00
return totalRead , err
}
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . UploadPart ( & s3 . UploadPartInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
PartNumber : aws . Int64 ( partNumber ) ,
UploadId : uploadID ,
Body : bytes . NewReader ( buf ) ,
} )
2014-12-19 18:20:07 +01:00
if err != nil {
return totalRead , err
}
2016-01-22 03:17:53 +01:00
parts = append ( parts , & s3 . CompletedPart {
ETag : resp . ETag ,
PartNumber : aws . Int64 ( partNumber ) ,
} )
2014-12-19 18:20:07 +01:00
partNumber ++
//Zero fill from chunkSize up to offset, then some reader
2015-01-24 01:46:43 +01:00
if err = fromZeroFillLarge ( d . ChunkSize , offset ) ; err != nil {
2014-12-19 18:20:07 +01:00
return totalRead , err
}
2015-01-24 01:46:43 +01:00
if err = fromReader ( offset % d . ChunkSize ) ; err != nil {
2014-12-19 18:20:07 +01:00
return totalRead , err
}
2015-01-24 01:46:43 +01:00
if totalRead + ( offset % d . ChunkSize ) < d . ChunkSize {
2014-12-19 18:20:07 +01:00
return totalRead , nil
}
}
} else {
// offset > currentLength >= chunkSize
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . UploadPartCopy ( & s3 . UploadPartCopyInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
PartNumber : aws . Int64 ( partNumber ) ,
UploadId : uploadID ,
CopySource : aws . String ( d . Bucket + "/" + d . s3Path ( path ) ) ,
} )
2014-12-19 18:20:07 +01:00
if err != nil {
return 0 , err
}
2016-01-22 03:17:53 +01:00
parts = append ( parts , & s3 . CompletedPart {
ETag : resp . CopyPartResult . ETag ,
PartNumber : aws . Int64 ( partNumber ) ,
} )
2014-12-19 18:20:07 +01:00
partNumber ++
//Zero fill from currentLength up to offset, then some reader
if err = fromZeroFillLarge ( currentLength , offset ) ; err != nil {
return totalRead , err
}
2015-01-24 01:46:43 +01:00
if err = fromReader ( ( offset - currentLength ) % d . ChunkSize ) ; err != nil {
2014-12-19 18:20:07 +01:00
return totalRead , err
}
2015-01-24 01:46:43 +01:00
if totalRead + ( ( offset - currentLength ) % d . ChunkSize ) < d . ChunkSize {
2014-12-19 18:20:07 +01:00
return totalRead , nil
}
}
2014-12-19 18:16:51 +01:00
}
2014-10-25 01:37:25 +02:00
}
for {
2014-12-20 09:32:48 +01:00
if err = fromReader ( 0 ) ; err != nil {
2014-12-19 18:16:51 +01:00
return totalRead , err
}
2014-10-25 01:37:25 +02:00
2015-01-24 01:46:43 +01:00
if int64 ( bytesRead ) < d . ChunkSize {
2014-12-19 18:16:51 +01:00
break
2014-10-25 01:37:25 +02:00
}
}
2014-12-19 18:16:51 +01:00
return totalRead , nil
2014-10-25 01:37:25 +02:00
}
2014-12-19 18:16:51 +01:00
// Stat retrieves the FileInfo for the given path, including the current size
// in bytes and the creation time.
2015-04-28 00:58:58 +02:00
func ( d * driver ) Stat ( ctx context . Context , path string ) ( storagedriver . FileInfo , error ) {
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . ListObjects ( & s3 . ListObjectsInput {
Bucket : aws . String ( d . Bucket ) ,
Prefix : aws . String ( d . s3Path ( path ) ) ,
MaxKeys : aws . Int64 ( 1 ) ,
} )
2014-10-25 01:37:25 +02:00
if err != nil {
2014-12-19 18:16:51 +01:00
return nil , err
2014-10-25 01:37:25 +02:00
}
2014-12-19 18:16:51 +01:00
fi := storagedriver . FileInfoFields {
Path : path ,
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
if len ( resp . Contents ) == 1 {
if * resp . Contents [ 0 ] . Key != d . s3Path ( path ) {
2014-12-19 18:16:51 +01:00
fi . IsDir = true
} else {
fi . IsDir = false
2016-01-22 03:17:53 +01:00
fi . Size = * resp . Contents [ 0 ] . Size
fi . ModTime = * resp . Contents [ 0 ] . LastModified
2014-12-19 18:16:51 +01:00
}
2016-01-22 03:17:53 +01:00
} else if len ( resp . CommonPrefixes ) == 1 {
2014-12-19 18:16:51 +01:00
fi . IsDir = true
} else {
return nil , storagedriver . PathNotFoundError { Path : path }
}
return storagedriver . FileInfoInternal { FileInfoFields : fi } , nil
2014-10-25 01:37:25 +02:00
}
2014-12-19 18:16:51 +01:00
// List returns a list of the objects that are direct descendants of the given path.
2015-12-08 20:02:40 +01:00
func ( d * driver ) List ( ctx context . Context , opath string ) ( [ ] string , error ) {
path := opath
2014-12-19 18:16:51 +01:00
if path != "/" && path [ len ( path ) - 1 ] != '/' {
2014-11-07 21:58:48 +01:00
path = path + "/"
2014-11-04 01:20:38 +01:00
}
2015-02-20 01:31:34 +01:00
// This is to cover for the cases when the rootDirectory of the driver is either "" or "/".
// In those cases, there is no root prefix to replace and we must actually add a "/" to all
// results in order to keep them as valid paths as recognized by storagedriver.PathRegexp
prefix := ""
if d . s3Path ( "" ) == "" {
prefix = "/"
}
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . ListObjects ( & s3 . ListObjectsInput {
Bucket : aws . String ( d . Bucket ) ,
Prefix : aws . String ( d . s3Path ( path ) ) ,
Delimiter : aws . String ( "/" ) ,
MaxKeys : aws . Int64 ( listMax ) ,
} )
2014-10-25 01:37:25 +02:00
if err != nil {
2015-12-08 20:02:40 +01:00
return nil , parseError ( opath , err )
2015-11-24 23:23:12 +01:00
}
2014-10-25 01:37:25 +02:00
files := [ ] string { }
directories := [ ] string { }
2014-11-04 01:20:38 +01:00
for {
2016-01-22 03:17:53 +01:00
for _ , key := range resp . Contents {
files = append ( files , strings . Replace ( * key . Key , d . s3Path ( "" ) , prefix , 1 ) )
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
for _ , commonPrefix := range resp . CommonPrefixes {
commonPrefix := * commonPrefix . Prefix
2015-02-20 01:31:34 +01:00
directories = append ( directories , strings . Replace ( commonPrefix [ 0 : len ( commonPrefix ) - 1 ] , d . s3Path ( "" ) , prefix , 1 ) )
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
if * resp . IsTruncated {
resp , err = d . S3 . ListObjects ( & s3 . ListObjectsInput {
Bucket : aws . String ( d . Bucket ) ,
Prefix : aws . String ( d . s3Path ( path ) ) ,
Delimiter : aws . String ( "/" ) ,
MaxKeys : aws . Int64 ( listMax ) ,
Marker : resp . NextMarker ,
} )
2014-11-04 01:20:38 +01:00
if err != nil {
return nil , err
}
2014-10-25 01:37:25 +02:00
} else {
2014-11-04 01:20:38 +01:00
break
2014-10-25 01:37:25 +02:00
}
}
2015-12-08 20:02:40 +01:00
if opath != "/" {
if len ( files ) == 0 && len ( directories ) == 0 {
// Treat empty response as missing directory, since we don't actually
// have directories in s3.
return nil , storagedriver . PathNotFoundError { Path : opath }
}
}
2014-10-25 01:37:25 +02:00
return append ( files , directories ... ) , nil
}
2014-11-18 00:44:07 +01:00
// Move moves an object stored at sourcePath to destPath, removing the original
// object.
2015-04-28 00:58:58 +02:00
func ( d * driver ) Move ( ctx context . Context , sourcePath string , destPath string ) error {
2014-10-25 01:37:25 +02:00
/* This is terrible, but aws doesn't have an actual move. */
2016-01-22 03:17:53 +01:00
_ , err := d . S3 . CopyObject ( & s3 . CopyObjectInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( destPath ) ) ,
ContentType : d . getContentType ( ) ,
ACL : d . getACL ( ) ,
ServerSideEncryption : d . getEncryptionMode ( ) ,
StorageClass : d . getStorageClass ( ) ,
CopySource : aws . String ( d . Bucket + "/" + d . s3Path ( sourcePath ) ) ,
} )
2014-10-25 01:37:25 +02:00
if err != nil {
2014-12-19 18:16:51 +01:00
return parseError ( sourcePath , err )
2014-10-25 01:37:25 +02:00
}
2015-04-28 00:58:58 +02:00
return d . Delete ( ctx , sourcePath )
2014-10-25 01:37:25 +02:00
}
2014-11-18 00:44:07 +01:00
// Delete recursively deletes all objects stored at "path" and its subpaths.
2015-04-28 00:58:58 +02:00
func ( d * driver ) Delete ( ctx context . Context , path string ) error {
2016-01-22 03:17:53 +01:00
resp , err := d . S3 . ListObjects ( & s3 . ListObjectsInput {
Bucket : aws . String ( d . Bucket ) ,
Prefix : aws . String ( d . s3Path ( path ) ) ,
} )
if err != nil || len ( resp . Contents ) == 0 {
2014-11-13 02:19:19 +01:00
return storagedriver . PathNotFoundError { Path : path }
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
s3Objects := make ( [ ] * s3 . ObjectIdentifier , 0 , listMax )
2014-10-25 01:37:25 +02:00
2016-01-22 03:17:53 +01:00
for len ( resp . Contents ) > 0 {
for _ , key := range resp . Contents {
s3Objects = append ( s3Objects , & s3 . ObjectIdentifier {
Key : key . Key ,
} )
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
_ , err := d . S3 . DeleteObjects ( & s3 . DeleteObjectsInput {
Bucket : aws . String ( d . Bucket ) ,
Delete : & s3 . Delete {
Objects : s3Objects ,
Quiet : aws . Bool ( false ) ,
} ,
} )
2014-10-25 01:37:25 +02:00
if err != nil {
return nil
}
2016-01-22 03:17:53 +01:00
resp , err = d . S3 . ListObjects ( & s3 . ListObjectsInput {
Bucket : aws . String ( d . Bucket ) ,
Prefix : aws . String ( d . s3Path ( path ) ) ,
} )
2014-10-25 01:37:25 +02:00
if err != nil {
return err
}
}
return nil
}
2015-01-07 17:31:38 +01:00
// URLFor returns a URL which may be used to retrieve the content stored at the given path.
// May return an UnsupportedMethodErr in certain StorageDriver implementations.
2015-04-28 00:58:58 +02:00
func ( d * driver ) URLFor ( ctx context . Context , path string , options map [ string ] interface { } ) ( string , error ) {
2015-01-14 20:31:11 +01:00
methodString := "GET"
method , ok := options [ "method" ]
if ok {
methodString , ok = method . ( string )
if ! ok || ( methodString != "GET" && methodString != "HEAD" ) {
2015-11-02 22:23:53 +01:00
return "" , storagedriver . ErrUnsupportedMethod { }
2015-01-14 20:31:11 +01:00
}
}
2016-01-22 03:17:53 +01:00
expiresIn := 20 * time . Minute
2015-01-09 02:45:21 +01:00
expires , ok := options [ "expiry" ]
2015-01-09 02:10:32 +01:00
if ok {
et , ok := expires . ( time . Time )
if ok {
2016-01-22 03:17:53 +01:00
expiresIn = et . Sub ( time . Now ( ) )
2015-01-09 02:10:32 +01:00
}
}
2016-01-22 03:17:53 +01:00
var req * request . Request
switch methodString {
case "GET" :
req , _ = d . S3 . GetObjectRequest ( & s3 . GetObjectInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
} )
case "HEAD" :
req , _ = d . S3 . HeadObjectRequest ( & s3 . HeadObjectInput {
Bucket : aws . String ( d . Bucket ) ,
Key : aws . String ( d . s3Path ( path ) ) ,
} )
default :
panic ( "unreachable" )
}
return req . Presign ( expiresIn )
2015-01-07 17:31:38 +01:00
}
2015-02-04 01:54:52 +01:00
func ( d * driver ) s3Path ( path string ) string {
2015-01-24 01:46:43 +01:00
return strings . TrimLeft ( strings . TrimRight ( d . RootDirectory , "/" ) + path , "/" )
2014-12-19 18:16:51 +01:00
}
2014-10-25 01:37:25 +02:00
2015-04-07 01:23:31 +02:00
// S3BucketKey returns the s3 bucket key for the given storage driver path.
func ( d * Driver ) S3BucketKey ( path string ) string {
return d . StorageDriver . ( * driver ) . s3Path ( path )
}
2014-12-19 18:16:51 +01:00
func parseError ( path string , err error ) error {
2016-01-22 03:17:53 +01:00
if s3Err , ok := err . ( awserr . Error ) ; ok && s3Err . Code ( ) == "NoSuchKey" {
2014-12-19 18:16:51 +01:00
return storagedriver . PathNotFoundError { Path : path }
2014-10-25 01:37:25 +02:00
}
2014-12-19 18:16:51 +01:00
return err
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
func ( d * driver ) getEncryptionMode ( ) * string {
if d . Encrypt {
return aws . String ( "AES256" )
}
return nil
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
func ( d * driver ) getContentType ( ) * string {
return aws . String ( "application/octet-stream" )
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
func ( d * driver ) getACL ( ) * string {
return aws . String ( "private" )
2014-10-25 01:37:25 +02:00
}
2016-01-22 03:17:53 +01:00
func ( d * driver ) getStorageClass ( ) * string {
return aws . String ( d . StorageClass )
2014-10-25 01:37:25 +02:00
}
2015-04-23 00:07:18 +02:00
// getbuf returns a buffer from the driver's pool with length d.ChunkSize.
func ( d * driver ) getbuf ( ) [ ] byte {
return d . pool . Get ( ) . ( [ ] byte )
}
func ( d * driver ) putbuf ( p [ ] byte ) {
copy ( p , d . zeros )
d . pool . Put ( p )
}