mirror of
https://github.com/minio/minio.git
synced 2024-12-25 14:45:54 -05:00
multipart: Code cleanup
- More locking cleanup. Fix naming convention. - Simplify concatenation and blocking calls.
This commit is contained in:
parent
d79fcb1800
commit
a4c005ce30
@ -79,8 +79,8 @@ func setObjectHeaders(w http.ResponseWriter, metadata fs.ObjectMetadata, content
|
|||||||
lastModified := metadata.Created.Format(http.TimeFormat)
|
lastModified := metadata.Created.Format(http.TimeFormat)
|
||||||
// object related headers
|
// object related headers
|
||||||
w.Header().Set("Content-Type", metadata.ContentType)
|
w.Header().Set("Content-Type", metadata.ContentType)
|
||||||
if metadata.Md5 != "" {
|
if metadata.MD5 != "" {
|
||||||
w.Header().Set("ETag", "\""+metadata.Md5+"\"")
|
w.Header().Set("ETag", "\""+metadata.MD5+"\"")
|
||||||
}
|
}
|
||||||
w.Header().Set("Last-Modified", lastModified)
|
w.Header().Set("Last-Modified", lastModified)
|
||||||
|
|
||||||
|
@ -108,8 +108,8 @@ func generateListObjectsResponse(bucket, prefix, marker, delimiter string, maxKe
|
|||||||
}
|
}
|
||||||
content.Key = object.Object
|
content.Key = object.Object
|
||||||
content.LastModified = object.Created.Format(rfcFormat)
|
content.LastModified = object.Created.Format(rfcFormat)
|
||||||
if object.Md5 != "" {
|
if object.MD5 != "" {
|
||||||
content.ETag = "\"" + object.Md5 + "\""
|
content.ETag = "\"" + object.MD5 + "\""
|
||||||
}
|
}
|
||||||
content.Size = object.Size
|
content.Size = object.Size
|
||||||
content.StorageClass = "STANDARD"
|
content.StorageClass = "STANDARD"
|
||||||
|
@ -353,8 +353,8 @@ func (api CloudStorageAPI) PostPolicyBucketHandler(w http.ResponseWriter, req *h
|
|||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
if metadata.Md5 != "" {
|
if metadata.MD5 != "" {
|
||||||
w.Header().Set("ETag", "\""+metadata.Md5+"\"")
|
w.Header().Set("ETag", "\""+metadata.MD5+"\"")
|
||||||
}
|
}
|
||||||
writeSuccessResponse(w, nil)
|
writeSuccessResponse(w, nil)
|
||||||
}
|
}
|
||||||
|
@ -198,8 +198,8 @@ func (api CloudStorageAPI) PutObjectHandler(w http.ResponseWriter, req *http.Req
|
|||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
if metadata.Md5 != "" {
|
if metadata.MD5 != "" {
|
||||||
w.Header().Set("ETag", "\""+metadata.Md5+"\"")
|
w.Header().Set("ETag", "\""+metadata.MD5+"\"")
|
||||||
}
|
}
|
||||||
writeSuccessResponse(w, nil)
|
writeSuccessResponse(w, nil)
|
||||||
}
|
}
|
||||||
@ -501,7 +501,7 @@ func (api CloudStorageAPI) CompleteMultipartUploadHandler(w http.ResponseWriter,
|
|||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
response := generateCompleteMultpartUploadResponse(bucket, object, req.URL.String(), metadata.Md5)
|
response := generateCompleteMultpartUploadResponse(bucket, object, req.URL.String(), metadata.MD5)
|
||||||
encodedSuccessResponse := encodeSuccessResponse(response)
|
encodedSuccessResponse := encodeSuccessResponse(response)
|
||||||
// write headers
|
// write headers
|
||||||
setCommonHeaders(w)
|
setCommonHeaders(w)
|
||||||
|
@ -92,7 +92,7 @@ func testMultipartObjectCreation(c *check.C, create func() Filesystem) {
|
|||||||
c.Assert(e, check.IsNil)
|
c.Assert(e, check.IsNil)
|
||||||
objectMetadata, err := fs.CompleteMultipartUpload("bucket", "key", uploadID, bytes.NewReader(completedPartsBytes), nil)
|
objectMetadata, err := fs.CompleteMultipartUpload("bucket", "key", uploadID, bytes.NewReader(completedPartsBytes), nil)
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
c.Assert(objectMetadata.Md5, check.Equals, finalExpectedmd5SumHex)
|
c.Assert(objectMetadata.MD5, check.Equals, finalExpectedmd5SumHex)
|
||||||
}
|
}
|
||||||
|
|
||||||
func testMultipartObjectAbort(c *check.C, create func() Filesystem) {
|
func testMultipartObjectAbort(c *check.C, create func() Filesystem) {
|
||||||
@ -147,7 +147,7 @@ func testMultipleObjectCreation(c *check.C, create func() Filesystem) {
|
|||||||
objects[key] = []byte(randomString)
|
objects[key] = []byte(randomString)
|
||||||
objectMetadata, err := fs.CreateObject("bucket", key, expectedmd5Sum, int64(len(randomString)), bytes.NewBufferString(randomString), nil)
|
objectMetadata, err := fs.CreateObject("bucket", key, expectedmd5Sum, int64(len(randomString)), bytes.NewBufferString(randomString), nil)
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
c.Assert(objectMetadata.Md5, check.Equals, expectedmd5Sumhex)
|
c.Assert(objectMetadata.MD5, check.Equals, expectedmd5Sumhex)
|
||||||
}
|
}
|
||||||
|
|
||||||
for key, value := range objects {
|
for key, value := range objects {
|
||||||
@ -276,7 +276,7 @@ func testObjectOverwriteWorks(c *check.C, create func() Filesystem) {
|
|||||||
md5Sum1hex := hex.EncodeToString(hasher1.Sum(nil))
|
md5Sum1hex := hex.EncodeToString(hasher1.Sum(nil))
|
||||||
objectMetadata, err := fs.CreateObject("bucket", "object", md5Sum1, int64(len("one")), bytes.NewBufferString("one"), nil)
|
objectMetadata, err := fs.CreateObject("bucket", "object", md5Sum1, int64(len("one")), bytes.NewBufferString("one"), nil)
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
c.Assert(md5Sum1hex, check.Equals, objectMetadata.Md5)
|
c.Assert(md5Sum1hex, check.Equals, objectMetadata.MD5)
|
||||||
|
|
||||||
hasher2 := md5.New()
|
hasher2 := md5.New()
|
||||||
hasher2.Write([]byte("three"))
|
hasher2.Write([]byte("three"))
|
||||||
@ -326,7 +326,7 @@ func testPutObjectInSubdir(c *check.C, create func() Filesystem) {
|
|||||||
md5Sum1hex := hex.EncodeToString(hasher.Sum(nil))
|
md5Sum1hex := hex.EncodeToString(hasher.Sum(nil))
|
||||||
objectMetadata, err := fs.CreateObject("bucket", "dir1/dir2/object", md5Sum1, int64(len("hello world")), bytes.NewBufferString("hello world"), nil)
|
objectMetadata, err := fs.CreateObject("bucket", "dir1/dir2/object", md5Sum1, int64(len("hello world")), bytes.NewBufferString("hello world"), nil)
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
c.Assert(objectMetadata.Md5, check.Equals, md5Sum1hex)
|
c.Assert(objectMetadata.MD5, check.Equals, md5Sum1hex)
|
||||||
|
|
||||||
var bytesBuffer bytes.Buffer
|
var bytesBuffer bytes.Buffer
|
||||||
length, err := fs.GetObject(&bytesBuffer, "bucket", "dir1/dir2/object", 0, 0)
|
length, err := fs.GetObject(&bytesBuffer, "bucket", "dir1/dir2/object", 0, 0)
|
||||||
@ -458,7 +458,7 @@ func testDefaultContentType(c *check.C, create func() Filesystem) {
|
|||||||
c.Assert(metadata.ContentType, check.Equals, "application/octet-stream")
|
c.Assert(metadata.ContentType, check.Equals, "application/octet-stream")
|
||||||
}
|
}
|
||||||
|
|
||||||
func testContentMd5Set(c *check.C, create func() Filesystem) {
|
func testContentMD5Set(c *check.C, create func() Filesystem) {
|
||||||
fs := create()
|
fs := create()
|
||||||
err := fs.MakeBucket("bucket", "")
|
err := fs.MakeBucket("bucket", "")
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
|
@ -91,7 +91,7 @@ func testMultipartObjectCreation(c *check.C, create func() Filesystem) {
|
|||||||
c.Assert(e, check.IsNil)
|
c.Assert(e, check.IsNil)
|
||||||
objectMetadata, err := fs.CompleteMultipartUpload("bucket", "key", uploadID, bytes.NewReader(completedPartsBytes), nil)
|
objectMetadata, err := fs.CompleteMultipartUpload("bucket", "key", uploadID, bytes.NewReader(completedPartsBytes), nil)
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
c.Assert(objectMetadata.Md5, check.Equals, finalExpectedmd5SumHex)
|
c.Assert(objectMetadata.MD5, check.Equals, finalExpectedmd5SumHex)
|
||||||
}
|
}
|
||||||
|
|
||||||
func testMultipartObjectAbort(c *check.C, create func() Filesystem) {
|
func testMultipartObjectAbort(c *check.C, create func() Filesystem) {
|
||||||
@ -146,7 +146,7 @@ func testMultipleObjectCreation(c *check.C, create func() Filesystem) {
|
|||||||
objects[key] = []byte(randomString)
|
objects[key] = []byte(randomString)
|
||||||
objectMetadata, err := fs.CreateObject("bucket", key, expectedmd5Sum, int64(len(randomString)), bytes.NewBufferString(randomString), nil)
|
objectMetadata, err := fs.CreateObject("bucket", key, expectedmd5Sum, int64(len(randomString)), bytes.NewBufferString(randomString), nil)
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
c.Assert(objectMetadata.Md5, check.Equals, expectedmd5Sumhex)
|
c.Assert(objectMetadata.MD5, check.Equals, expectedmd5Sumhex)
|
||||||
}
|
}
|
||||||
|
|
||||||
for key, value := range objects {
|
for key, value := range objects {
|
||||||
@ -273,7 +273,7 @@ func testObjectOverwriteWorks(c *check.C, create func() Filesystem) {
|
|||||||
md5Sum1hex := hex.EncodeToString(hasher1.Sum(nil))
|
md5Sum1hex := hex.EncodeToString(hasher1.Sum(nil))
|
||||||
objectMetadata, err := fs.CreateObject("bucket", "object", md5Sum1, int64(len("one")), bytes.NewBufferString("one"), nil)
|
objectMetadata, err := fs.CreateObject("bucket", "object", md5Sum1, int64(len("one")), bytes.NewBufferString("one"), nil)
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
c.Assert(md5Sum1hex, check.Equals, objectMetadata.Md5)
|
c.Assert(md5Sum1hex, check.Equals, objectMetadata.MD5)
|
||||||
|
|
||||||
hasher2 := md5.New()
|
hasher2 := md5.New()
|
||||||
hasher2.Write([]byte("three"))
|
hasher2.Write([]byte("three"))
|
||||||
@ -323,7 +323,7 @@ func testPutObjectInSubdir(c *check.C, create func() Filesystem) {
|
|||||||
md5Sum1hex := hex.EncodeToString(hasher.Sum(nil))
|
md5Sum1hex := hex.EncodeToString(hasher.Sum(nil))
|
||||||
objectMetadata, err := fs.CreateObject("bucket", "dir1/dir2/object", md5Sum1, int64(len("hello world")), bytes.NewBufferString("hello world"), nil)
|
objectMetadata, err := fs.CreateObject("bucket", "dir1/dir2/object", md5Sum1, int64(len("hello world")), bytes.NewBufferString("hello world"), nil)
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
c.Assert(objectMetadata.Md5, check.Equals, md5Sum1hex)
|
c.Assert(objectMetadata.MD5, check.Equals, md5Sum1hex)
|
||||||
|
|
||||||
var bytesBuffer bytes.Buffer
|
var bytesBuffer bytes.Buffer
|
||||||
length, err := fs.GetObject(&bytesBuffer, "bucket", "dir1/dir2/object", 0, 0)
|
length, err := fs.GetObject(&bytesBuffer, "bucket", "dir1/dir2/object", 0, 0)
|
||||||
@ -459,7 +459,7 @@ func testDefaultContentType(c *check.C, create func() Filesystem) {
|
|||||||
c.Assert(metadata.ContentType, check.Equals, "application/octet-stream")
|
c.Assert(metadata.ContentType, check.Equals, "application/octet-stream")
|
||||||
}
|
}
|
||||||
|
|
||||||
func testContentMd5Set(c *check.C, create func() Filesystem) {
|
func testContentMD5Set(c *check.C, create func() Filesystem) {
|
||||||
fs := create()
|
fs := create()
|
||||||
err := fs.MakeBucket("bucket", "")
|
err := fs.MakeBucket("bucket", "")
|
||||||
c.Assert(err, check.IsNil)
|
c.Assert(err, check.IsNil)
|
||||||
|
@ -17,77 +17,42 @@
|
|||||||
package fs
|
package fs
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"path/filepath"
|
|
||||||
|
|
||||||
"github.com/minio/minio-xl/pkg/probe"
|
"github.com/minio/minio-xl/pkg/probe"
|
||||||
"github.com/minio/minio-xl/pkg/quick"
|
"github.com/minio/minio-xl/pkg/quick"
|
||||||
"github.com/minio/minio/pkg/user"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func getFSBucketsConfigPath() (string, *probe.Error) {
|
var multipartsMetadataPath, bucketsMetadataPath string
|
||||||
if customBucketsConfigPath != "" {
|
|
||||||
return customBucketsConfigPath, nil
|
// setFSBucketsMetadataPath - set fs buckets metadata path.
|
||||||
}
|
func setFSBucketsMetadataPath(metadataPath string) {
|
||||||
homeDir, e := user.HomeDir()
|
bucketsMetadataPath = metadataPath
|
||||||
if e != nil {
|
|
||||||
return "", probe.NewError(e)
|
|
||||||
}
|
|
||||||
fsBucketsConfigPath := filepath.Join(homeDir, ".minio", "$buckets.json")
|
|
||||||
return fsBucketsConfigPath, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func getFSMultipartsSessionConfigPath() (string, *probe.Error) {
|
// SetFSMultipartsMetadataPath - set custom multiparts session
|
||||||
if customMultipartsConfigPath != "" {
|
// metadata path.
|
||||||
return customMultipartsConfigPath, nil
|
func setFSMultipartsMetadataPath(metadataPath string) {
|
||||||
}
|
multipartsMetadataPath = metadataPath
|
||||||
homeDir, e := user.HomeDir()
|
|
||||||
if e != nil {
|
|
||||||
return "", probe.NewError(e)
|
|
||||||
}
|
|
||||||
fsMultipartsConfigPath := filepath.Join(homeDir, ".minio", "$multiparts-session.json")
|
|
||||||
return fsMultipartsConfigPath, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// internal variable only accessed via get/set methods
|
|
||||||
var customMultipartsConfigPath, customBucketsConfigPath string
|
|
||||||
|
|
||||||
// setFSBucketsConfigPath - set custom fs buckets config path
|
|
||||||
func setFSBucketsConfigPath(configPath string) {
|
|
||||||
customBucketsConfigPath = configPath
|
|
||||||
}
|
|
||||||
|
|
||||||
// SetFSMultipartsConfigPath - set custom multiparts session config path
|
|
||||||
func setFSMultipartsConfigPath(configPath string) {
|
|
||||||
customMultipartsConfigPath = configPath
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// saveMultipartsSession - save multiparts
|
// saveMultipartsSession - save multiparts
|
||||||
func saveMultipartsSession(multiparts *Multiparts) *probe.Error {
|
func saveMultipartsSession(multiparts Multiparts) *probe.Error {
|
||||||
fsMultipartsConfigPath, err := getFSMultipartsSessionConfigPath()
|
|
||||||
if err != nil {
|
|
||||||
return err.Trace()
|
|
||||||
}
|
|
||||||
qc, err := quick.New(multiparts)
|
qc, err := quick.New(multiparts)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err.Trace()
|
return err.Trace()
|
||||||
}
|
}
|
||||||
if err := qc.Save(fsMultipartsConfigPath); err != nil {
|
if err := qc.Save(multipartsMetadataPath); err != nil {
|
||||||
return err.Trace()
|
return err.Trace()
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// saveBucketsMetadata - save metadata of all buckets
|
// saveBucketsMetadata - save metadata of all buckets
|
||||||
func saveBucketsMetadata(buckets *Buckets) *probe.Error {
|
func saveBucketsMetadata(buckets Buckets) *probe.Error {
|
||||||
fsBucketsConfigPath, err := getFSBucketsConfigPath()
|
|
||||||
if err != nil {
|
|
||||||
return err.Trace()
|
|
||||||
}
|
|
||||||
qc, err := quick.New(buckets)
|
qc, err := quick.New(buckets)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err.Trace()
|
return err.Trace()
|
||||||
}
|
}
|
||||||
if err := qc.Save(fsBucketsConfigPath); err != nil {
|
if err := qc.Save(bucketsMetadataPath); err != nil {
|
||||||
return err.Trace()
|
return err.Trace()
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
@ -95,10 +60,6 @@ func saveBucketsMetadata(buckets *Buckets) *probe.Error {
|
|||||||
|
|
||||||
// loadMultipartsSession load multipart session file
|
// loadMultipartsSession load multipart session file
|
||||||
func loadMultipartsSession() (*Multiparts, *probe.Error) {
|
func loadMultipartsSession() (*Multiparts, *probe.Error) {
|
||||||
fsMultipartsConfigPath, err := getFSMultipartsSessionConfigPath()
|
|
||||||
if err != nil {
|
|
||||||
return nil, err.Trace()
|
|
||||||
}
|
|
||||||
multiparts := &Multiparts{}
|
multiparts := &Multiparts{}
|
||||||
multiparts.Version = "1"
|
multiparts.Version = "1"
|
||||||
multiparts.ActiveSession = make(map[string]*MultipartSession)
|
multiparts.ActiveSession = make(map[string]*MultipartSession)
|
||||||
@ -106,7 +67,7 @@ func loadMultipartsSession() (*Multiparts, *probe.Error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err.Trace()
|
return nil, err.Trace()
|
||||||
}
|
}
|
||||||
if err := qc.Load(fsMultipartsConfigPath); err != nil {
|
if err := qc.Load(multipartsMetadataPath); err != nil {
|
||||||
return nil, err.Trace()
|
return nil, err.Trace()
|
||||||
}
|
}
|
||||||
return qc.Data().(*Multiparts), nil
|
return qc.Data().(*Multiparts), nil
|
||||||
@ -114,10 +75,6 @@ func loadMultipartsSession() (*Multiparts, *probe.Error) {
|
|||||||
|
|
||||||
// loadBucketsMetadata load buckets metadata file
|
// loadBucketsMetadata load buckets metadata file
|
||||||
func loadBucketsMetadata() (*Buckets, *probe.Error) {
|
func loadBucketsMetadata() (*Buckets, *probe.Error) {
|
||||||
fsBucketsConfigPath, err := getFSBucketsConfigPath()
|
|
||||||
if err != nil {
|
|
||||||
return nil, err.Trace()
|
|
||||||
}
|
|
||||||
buckets := &Buckets{}
|
buckets := &Buckets{}
|
||||||
buckets.Version = "1"
|
buckets.Version = "1"
|
||||||
buckets.Metadata = make(map[string]*BucketMetadata)
|
buckets.Metadata = make(map[string]*BucketMetadata)
|
||||||
@ -125,7 +82,7 @@ func loadBucketsMetadata() (*Buckets, *probe.Error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err.Trace()
|
return nil, err.Trace()
|
||||||
}
|
}
|
||||||
if err := qc.Load(fsBucketsConfigPath); err != nil {
|
if err := qc.Load(bucketsMetadataPath); err != nil {
|
||||||
return nil, err.Trace()
|
return nil, err.Trace()
|
||||||
}
|
}
|
||||||
return qc.Data().(*Buckets), nil
|
return qc.Data().(*Buckets), nil
|
||||||
|
@ -68,7 +68,7 @@ type ObjectMetadata struct {
|
|||||||
ContentType string
|
ContentType string
|
||||||
Created time.Time
|
Created time.Time
|
||||||
Mode os.FileMode
|
Mode os.FileMode
|
||||||
Md5 string
|
MD5 string
|
||||||
Size int64
|
Size int64
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -157,7 +157,7 @@ func (e InvalidDisksArgument) Error() string {
|
|||||||
|
|
||||||
// BadDigest bad md5sum
|
// BadDigest bad md5sum
|
||||||
type BadDigest struct {
|
type BadDigest struct {
|
||||||
Md5 string
|
MD5 string
|
||||||
Bucket string
|
Bucket string
|
||||||
Object string
|
Object string
|
||||||
}
|
}
|
||||||
@ -222,11 +222,11 @@ type ImplementationError struct {
|
|||||||
Err error
|
Err error
|
||||||
}
|
}
|
||||||
|
|
||||||
// DigestError - Generic Md5 error
|
// DigestError - Generic MD5 error
|
||||||
type DigestError struct {
|
type DigestError struct {
|
||||||
Bucket string
|
Bucket string
|
||||||
Key string
|
Key string
|
||||||
Md5 string
|
MD5 string
|
||||||
}
|
}
|
||||||
|
|
||||||
/// ACL related errors
|
/// ACL related errors
|
||||||
@ -322,7 +322,7 @@ func (e BackendCorrupted) Error() string {
|
|||||||
|
|
||||||
// Return string an error formatted as the given text
|
// Return string an error formatted as the given text
|
||||||
func (e InvalidDigest) Error() string {
|
func (e InvalidDigest) Error() string {
|
||||||
return "Md5 provided " + e.Md5 + " is invalid"
|
return "MD5 provided " + e.MD5 + " is invalid"
|
||||||
}
|
}
|
||||||
|
|
||||||
// OperationNotPermitted - operation not permitted
|
// OperationNotPermitted - operation not permitted
|
||||||
|
@ -78,7 +78,13 @@ func (fs Filesystem) listObjects(bucket, prefix, marker, delimiter string, maxKe
|
|||||||
walkPath = prefixPath
|
walkPath = prefixPath
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
ioutils.FTW(walkPath, func(path string, info os.FileInfo, err error) error {
|
ioutils.FTW(walkPath, func(path string, info os.FileInfo, e error) error {
|
||||||
|
if e != nil {
|
||||||
|
return e
|
||||||
|
}
|
||||||
|
if strings.HasSuffix(path, "$multiparts") {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
// We don't need to list the walk path.
|
// We don't need to list the walk path.
|
||||||
if path == walkPath {
|
if path == walkPath {
|
||||||
return nil
|
return nil
|
||||||
@ -271,9 +277,6 @@ func (fs *Filesystem) listObjectsService() *probe.Error {
|
|||||||
// ListObjects - lists all objects for a given prefix, returns upto
|
// ListObjects - lists all objects for a given prefix, returns upto
|
||||||
// maxKeys number of objects per call.
|
// maxKeys number of objects per call.
|
||||||
func (fs Filesystem) ListObjects(bucket, prefix, marker, delimiter string, maxKeys int) (ListObjectsResult, *probe.Error) {
|
func (fs Filesystem) ListObjects(bucket, prefix, marker, delimiter string, maxKeys int) (ListObjectsResult, *probe.Error) {
|
||||||
fs.rwLock.RLock()
|
|
||||||
defer fs.rwLock.RUnlock()
|
|
||||||
|
|
||||||
// Input validation.
|
// Input validation.
|
||||||
if !IsValidBucketName(bucket) {
|
if !IsValidBucketName(bucket) {
|
||||||
return ListObjectsResult{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
return ListObjectsResult{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
||||||
|
@ -32,8 +32,6 @@ import (
|
|||||||
|
|
||||||
// DeleteBucket - delete bucket
|
// DeleteBucket - delete bucket
|
||||||
func (fs Filesystem) DeleteBucket(bucket string) *probe.Error {
|
func (fs Filesystem) DeleteBucket(bucket string) *probe.Error {
|
||||||
fs.rwLock.Lock()
|
|
||||||
defer fs.rwLock.Unlock()
|
|
||||||
// verify bucket path legal
|
// verify bucket path legal
|
||||||
if !IsValidBucketName(bucket) {
|
if !IsValidBucketName(bucket) {
|
||||||
return probe.NewError(BucketNameInvalid{Bucket: bucket})
|
return probe.NewError(BucketNameInvalid{Bucket: bucket})
|
||||||
@ -59,8 +57,10 @@ func (fs Filesystem) DeleteBucket(bucket string) *probe.Error {
|
|||||||
}
|
}
|
||||||
return probe.NewError(e)
|
return probe.NewError(e)
|
||||||
}
|
}
|
||||||
|
fs.rwLock.Lock()
|
||||||
delete(fs.buckets.Metadata, bucket)
|
delete(fs.buckets.Metadata, bucket)
|
||||||
if err := saveBucketsMetadata(fs.buckets); err != nil {
|
fs.rwLock.Unlock()
|
||||||
|
if err := saveBucketsMetadata(*fs.buckets); err != nil {
|
||||||
return err.Trace(bucket)
|
return err.Trace(bucket)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
@ -68,9 +68,6 @@ func (fs Filesystem) DeleteBucket(bucket string) *probe.Error {
|
|||||||
|
|
||||||
// ListBuckets - Get service.
|
// ListBuckets - Get service.
|
||||||
func (fs Filesystem) ListBuckets() ([]BucketMetadata, *probe.Error) {
|
func (fs Filesystem) ListBuckets() ([]BucketMetadata, *probe.Error) {
|
||||||
fs.rwLock.RLock()
|
|
||||||
defer fs.rwLock.RUnlock()
|
|
||||||
|
|
||||||
files, err := ioutils.ReadDirN(fs.path, fs.maxBuckets)
|
files, err := ioutils.ReadDirN(fs.path, fs.maxBuckets)
|
||||||
if err != nil && err != io.EOF {
|
if err != nil && err != io.EOF {
|
||||||
return []BucketMetadata{}, probe.NewError(err)
|
return []BucketMetadata{}, probe.NewError(err)
|
||||||
@ -118,9 +115,6 @@ func removeDuplicateBuckets(elements []BucketMetadata) (result []BucketMetadata)
|
|||||||
|
|
||||||
// MakeBucket - PUT Bucket.
|
// MakeBucket - PUT Bucket.
|
||||||
func (fs Filesystem) MakeBucket(bucket, acl string) *probe.Error {
|
func (fs Filesystem) MakeBucket(bucket, acl string) *probe.Error {
|
||||||
fs.rwLock.Lock()
|
|
||||||
defer fs.rwLock.Unlock()
|
|
||||||
|
|
||||||
di, err := disk.GetInfo(fs.path)
|
di, err := disk.GetInfo(fs.path)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return probe.NewError(err)
|
return probe.NewError(err)
|
||||||
@ -171,8 +165,10 @@ func (fs Filesystem) MakeBucket(bucket, acl string) *probe.Error {
|
|||||||
bucketMetadata.Name = fi.Name()
|
bucketMetadata.Name = fi.Name()
|
||||||
bucketMetadata.Created = fi.ModTime()
|
bucketMetadata.Created = fi.ModTime()
|
||||||
bucketMetadata.ACL = BucketACL(acl)
|
bucketMetadata.ACL = BucketACL(acl)
|
||||||
|
fs.rwLock.Lock()
|
||||||
fs.buckets.Metadata[bucket] = bucketMetadata
|
fs.buckets.Metadata[bucket] = bucketMetadata
|
||||||
if err := saveBucketsMetadata(fs.buckets); err != nil {
|
fs.rwLock.Unlock()
|
||||||
|
if err := saveBucketsMetadata(*fs.buckets); err != nil {
|
||||||
return err.Trace(bucket)
|
return err.Trace(bucket)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
@ -198,8 +194,6 @@ func (fs Filesystem) denormalizeBucket(bucket string) string {
|
|||||||
|
|
||||||
// GetBucketMetadata - get bucket metadata.
|
// GetBucketMetadata - get bucket metadata.
|
||||||
func (fs Filesystem) GetBucketMetadata(bucket string) (BucketMetadata, *probe.Error) {
|
func (fs Filesystem) GetBucketMetadata(bucket string) (BucketMetadata, *probe.Error) {
|
||||||
fs.rwLock.RLock()
|
|
||||||
defer fs.rwLock.RUnlock()
|
|
||||||
if !IsValidBucketName(bucket) {
|
if !IsValidBucketName(bucket) {
|
||||||
return BucketMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
return BucketMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
||||||
}
|
}
|
||||||
@ -215,7 +209,9 @@ func (fs Filesystem) GetBucketMetadata(bucket string) (BucketMetadata, *probe.Er
|
|||||||
}
|
}
|
||||||
return BucketMetadata{}, probe.NewError(e)
|
return BucketMetadata{}, probe.NewError(e)
|
||||||
}
|
}
|
||||||
|
fs.rwLock.RLock()
|
||||||
bucketMetadata, ok := fs.buckets.Metadata[bucket]
|
bucketMetadata, ok := fs.buckets.Metadata[bucket]
|
||||||
|
fs.rwLock.RUnlock()
|
||||||
if !ok {
|
if !ok {
|
||||||
bucketMetadata = &BucketMetadata{}
|
bucketMetadata = &BucketMetadata{}
|
||||||
bucketMetadata.Name = fi.Name()
|
bucketMetadata.Name = fi.Name()
|
||||||
@ -258,8 +254,10 @@ func (fs Filesystem) SetBucketMetadata(bucket string, metadata map[string]string
|
|||||||
bucketMetadata.Created = fi.ModTime()
|
bucketMetadata.Created = fi.ModTime()
|
||||||
}
|
}
|
||||||
bucketMetadata.ACL = BucketACL(acl)
|
bucketMetadata.ACL = BucketACL(acl)
|
||||||
|
fs.rwLock.Lock()
|
||||||
fs.buckets.Metadata[bucket] = bucketMetadata
|
fs.buckets.Metadata[bucket] = bucketMetadata
|
||||||
if err := saveBucketsMetadata(fs.buckets); err != nil {
|
fs.rwLock.Unlock()
|
||||||
|
if err := saveBucketsMetadata(*fs.buckets); err != nil {
|
||||||
return err.Trace(bucket)
|
return err.Trace(bucket)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
|
@ -27,7 +27,7 @@ import (
|
|||||||
|
|
||||||
// Metadata - carries metadata about object
|
// Metadata - carries metadata about object
|
||||||
type Metadata struct {
|
type Metadata struct {
|
||||||
Md5sum []byte
|
MD5sum []byte
|
||||||
ContentType string
|
ContentType string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -44,6 +44,8 @@ import (
|
|||||||
|
|
||||||
// isValidUploadID - is upload id.
|
// isValidUploadID - is upload id.
|
||||||
func (fs Filesystem) isValidUploadID(object, uploadID string) bool {
|
func (fs Filesystem) isValidUploadID(object, uploadID string) bool {
|
||||||
|
fs.rwLock.RLock()
|
||||||
|
defer fs.rwLock.RUnlock()
|
||||||
s, ok := fs.multiparts.ActiveSession[object]
|
s, ok := fs.multiparts.ActiveSession[object]
|
||||||
if !ok {
|
if !ok {
|
||||||
return false
|
return false
|
||||||
@ -56,9 +58,6 @@ func (fs Filesystem) isValidUploadID(object, uploadID string) bool {
|
|||||||
|
|
||||||
// ListMultipartUploads - list incomplete multipart sessions for a given BucketMultipartResourcesMetadata
|
// ListMultipartUploads - list incomplete multipart sessions for a given BucketMultipartResourcesMetadata
|
||||||
func (fs Filesystem) ListMultipartUploads(bucket string, resources BucketMultipartResourcesMetadata) (BucketMultipartResourcesMetadata, *probe.Error) {
|
func (fs Filesystem) ListMultipartUploads(bucket string, resources BucketMultipartResourcesMetadata) (BucketMultipartResourcesMetadata, *probe.Error) {
|
||||||
fs.rwLock.RLock()
|
|
||||||
defer fs.rwLock.RUnlock()
|
|
||||||
|
|
||||||
// Input validation.
|
// Input validation.
|
||||||
if !IsValidBucketName(bucket) {
|
if !IsValidBucketName(bucket) {
|
||||||
return BucketMultipartResourcesMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
return BucketMultipartResourcesMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
||||||
@ -73,6 +72,8 @@ func (fs Filesystem) ListMultipartUploads(bucket string, resources BucketMultipa
|
|||||||
return BucketMultipartResourcesMetadata{}, probe.NewError(e)
|
return BucketMultipartResourcesMetadata{}, probe.NewError(e)
|
||||||
}
|
}
|
||||||
var uploads []*UploadMetadata
|
var uploads []*UploadMetadata
|
||||||
|
fs.rwLock.RLock()
|
||||||
|
defer fs.rwLock.RUnlock()
|
||||||
for object, session := range fs.multiparts.ActiveSession {
|
for object, session := range fs.multiparts.ActiveSession {
|
||||||
if strings.HasPrefix(object, resources.Prefix) {
|
if strings.HasPrefix(object, resources.Prefix) {
|
||||||
if len(uploads) > resources.MaxUploads {
|
if len(uploads) > resources.MaxUploads {
|
||||||
@ -117,34 +118,81 @@ func (fs Filesystem) ListMultipartUploads(bucket string, resources BucketMultipa
|
|||||||
return resources, nil
|
return resources, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// concatenate parts.
|
// verify if parts sent over the network do really match with what we
|
||||||
func (fs Filesystem) concatParts(parts *CompleteMultipartUpload, objectPath string, mw io.Writer) *probe.Error {
|
// have for the session.
|
||||||
for _, part := range parts.Part {
|
func doPartsMatch(parts []CompletePart, savedParts []*PartMetadata) bool {
|
||||||
partFile, e := os.OpenFile(objectPath+fmt.Sprintf("$%d-$multiparts", part.PartNumber), os.O_RDONLY, 0600)
|
if parts == nil || savedParts == nil {
|
||||||
defer partFile.Close()
|
return false
|
||||||
|
}
|
||||||
|
// Range of incoming parts and compare them with saved parts.
|
||||||
|
for i, part := range parts {
|
||||||
|
if strings.Trim(part.ETag, "\"") != savedParts[i].ETag {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
type multiCloser struct {
|
||||||
|
Closers []io.Closer
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m multiCloser) Close() error {
|
||||||
|
for _, c := range m.Closers {
|
||||||
|
if e := c.Close(); e != nil {
|
||||||
|
return e
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// MultiCloser - returns a Closer that's the logical
|
||||||
|
// concatenation of the provided input closers. They're closed
|
||||||
|
// sequentially. If any of the closers return a non-nil error, Close
|
||||||
|
// will return that error.
|
||||||
|
func MultiCloser(closers ...io.Closer) io.Closer {
|
||||||
|
return multiCloser{closers}
|
||||||
|
}
|
||||||
|
|
||||||
|
// removeParts - remove all parts.
|
||||||
|
func removeParts(partPathPrefix string, parts []*PartMetadata) *probe.Error {
|
||||||
|
for _, part := range parts {
|
||||||
|
if e := os.Remove(partPathPrefix + fmt.Sprintf("$%d-$multiparts", part.PartNumber)); e != nil {
|
||||||
|
return probe.NewError(e)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if e := os.Remove(partPathPrefix + "$multiparts"); e != nil {
|
||||||
|
return probe.NewError(e)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// saveParts - concantenate and save all parts.
|
||||||
|
func saveParts(partPathPrefix string, mw io.Writer, parts []CompletePart) *probe.Error {
|
||||||
|
var partReaders []io.Reader
|
||||||
|
var partClosers []io.Closer
|
||||||
|
for _, part := range parts {
|
||||||
|
partFile, e := os.OpenFile(partPathPrefix+fmt.Sprintf("$%d-$multiparts", part.PartNumber), os.O_RDONLY, 0600)
|
||||||
if e != nil {
|
if e != nil {
|
||||||
return probe.NewError(e)
|
return probe.NewError(e)
|
||||||
}
|
}
|
||||||
|
partReaders = append(partReaders, partFile)
|
||||||
|
partClosers = append(partClosers, partFile)
|
||||||
|
}
|
||||||
|
// Concatenate a list of closers and close upon return.
|
||||||
|
closer := MultiCloser(partClosers...)
|
||||||
|
defer closer.Close()
|
||||||
|
|
||||||
recvMD5 := part.ETag
|
reader := io.MultiReader(partReaders...)
|
||||||
// Complete multipart request header md5sum per part is hex
|
readBuffer := make([]byte, 4*1024*1024)
|
||||||
// encoded trim it and decode if possible.
|
if _, e := io.CopyBuffer(mw, reader, readBuffer); e != nil {
|
||||||
if _, e = hex.DecodeString(strings.Trim(recvMD5, "\"")); e != nil {
|
return probe.NewError(e)
|
||||||
return probe.NewError(InvalidDigest{Md5: recvMD5})
|
|
||||||
}
|
|
||||||
|
|
||||||
if _, e = io.Copy(mw, partFile); e != nil {
|
|
||||||
return probe.NewError(e)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewMultipartUpload - initiate a new multipart session
|
// NewMultipartUpload - initiate a new multipart session
|
||||||
func (fs Filesystem) NewMultipartUpload(bucket, object string) (string, *probe.Error) {
|
func (fs Filesystem) NewMultipartUpload(bucket, object string) (string, *probe.Error) {
|
||||||
fs.rwLock.Lock()
|
|
||||||
defer fs.rwLock.Unlock()
|
|
||||||
|
|
||||||
di, e := disk.GetInfo(fs.path)
|
di, e := disk.GetInfo(fs.path)
|
||||||
if e != nil {
|
if e != nil {
|
||||||
return "", probe.NewError(e)
|
return "", probe.NewError(e)
|
||||||
@ -192,12 +240,6 @@ func (fs Filesystem) NewMultipartUpload(bucket, object string) (string, *probe.E
|
|||||||
uploadIDSum := sha512.Sum512(id)
|
uploadIDSum := sha512.Sum512(id)
|
||||||
uploadID := base64.URLEncoding.EncodeToString(uploadIDSum[:])[:47]
|
uploadID := base64.URLEncoding.EncodeToString(uploadIDSum[:])[:47]
|
||||||
|
|
||||||
multiPartfile, e := os.OpenFile(objectPath+"$multiparts", os.O_WRONLY|os.O_CREATE, 0600)
|
|
||||||
if e != nil {
|
|
||||||
return "", probe.NewError(e)
|
|
||||||
}
|
|
||||||
defer multiPartfile.Close()
|
|
||||||
|
|
||||||
// Initialize multipart session.
|
// Initialize multipart session.
|
||||||
mpartSession := &MultipartSession{}
|
mpartSession := &MultipartSession{}
|
||||||
mpartSession.TotalParts = 0
|
mpartSession.TotalParts = 0
|
||||||
@ -205,14 +247,23 @@ func (fs Filesystem) NewMultipartUpload(bucket, object string) (string, *probe.E
|
|||||||
mpartSession.Initiated = time.Now().UTC()
|
mpartSession.Initiated = time.Now().UTC()
|
||||||
var parts []*PartMetadata
|
var parts []*PartMetadata
|
||||||
mpartSession.Parts = parts
|
mpartSession.Parts = parts
|
||||||
fs.multiparts.ActiveSession[object] = mpartSession
|
|
||||||
|
|
||||||
encoder := json.NewEncoder(multiPartfile)
|
fs.rwLock.Lock()
|
||||||
if e = encoder.Encode(mpartSession); e != nil {
|
fs.multiparts.ActiveSession[object] = mpartSession
|
||||||
|
fs.rwLock.Unlock()
|
||||||
|
|
||||||
|
mpartSessionBytes, e := json.Marshal(mpartSession)
|
||||||
|
if e != nil {
|
||||||
return "", probe.NewError(e)
|
return "", probe.NewError(e)
|
||||||
}
|
}
|
||||||
if err := saveMultipartsSession(fs.multiparts); err != nil {
|
|
||||||
return "", err.Trace()
|
partPathPrefix := objectPath + uploadID
|
||||||
|
if e = ioutil.WriteFile(partPathPrefix+"$multiparts", mpartSessionBytes, 0600); e != nil {
|
||||||
|
return "", probe.NewError(e)
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := saveMultipartsSession(*fs.multiparts); err != nil {
|
||||||
|
return "", err.Trace(partPathPrefix)
|
||||||
}
|
}
|
||||||
return uploadID, nil
|
return uploadID, nil
|
||||||
}
|
}
|
||||||
@ -226,9 +277,6 @@ func (a partNumber) Less(i, j int) bool { return a[i].PartNumber < a[j].PartNumb
|
|||||||
|
|
||||||
// CreateObjectPart - create a part in a multipart session
|
// CreateObjectPart - create a part in a multipart session
|
||||||
func (fs Filesystem) CreateObjectPart(bucket, object, uploadID, expectedMD5Sum string, partID int, size int64, data io.Reader, signature *Signature) (string, *probe.Error) {
|
func (fs Filesystem) CreateObjectPart(bucket, object, uploadID, expectedMD5Sum string, partID int, size int64, data io.Reader, signature *Signature) (string, *probe.Error) {
|
||||||
fs.rwLock.Lock()
|
|
||||||
defer fs.rwLock.Unlock()
|
|
||||||
|
|
||||||
di, err := disk.GetInfo(fs.path)
|
di, err := disk.GetInfo(fs.path)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", probe.NewError(err)
|
return "", probe.NewError(err)
|
||||||
@ -266,7 +314,7 @@ func (fs Filesystem) CreateObjectPart(bucket, object, uploadID, expectedMD5Sum s
|
|||||||
expectedMD5SumBytes, err = base64.StdEncoding.DecodeString(strings.TrimSpace(expectedMD5Sum))
|
expectedMD5SumBytes, err = base64.StdEncoding.DecodeString(strings.TrimSpace(expectedMD5Sum))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// Pro-actively close the connection
|
// Pro-actively close the connection
|
||||||
return "", probe.NewError(InvalidDigest{Md5: expectedMD5Sum})
|
return "", probe.NewError(InvalidDigest{MD5: expectedMD5Sum})
|
||||||
}
|
}
|
||||||
expectedMD5Sum = hex.EncodeToString(expectedMD5SumBytes)
|
expectedMD5Sum = hex.EncodeToString(expectedMD5SumBytes)
|
||||||
}
|
}
|
||||||
@ -282,29 +330,32 @@ func (fs Filesystem) CreateObjectPart(bucket, object, uploadID, expectedMD5Sum s
|
|||||||
}
|
}
|
||||||
|
|
||||||
objectPath := filepath.Join(bucketPath, object)
|
objectPath := filepath.Join(bucketPath, object)
|
||||||
partPath := objectPath + fmt.Sprintf("$%d-$multiparts", partID)
|
partPathPrefix := objectPath + uploadID
|
||||||
|
partPath := partPathPrefix + fmt.Sprintf("$%d-$multiparts", partID)
|
||||||
partFile, e := atomic.FileCreateWithPrefix(partPath, "$multiparts")
|
partFile, e := atomic.FileCreateWithPrefix(partPath, "$multiparts")
|
||||||
if e != nil {
|
if e != nil {
|
||||||
return "", probe.NewError(e)
|
return "", probe.NewError(e)
|
||||||
}
|
}
|
||||||
h := md5.New()
|
|
||||||
sh := sha256.New()
|
md5Hasher := md5.New()
|
||||||
mw := io.MultiWriter(partFile, h, sh)
|
sha256Hasher := sha256.New()
|
||||||
if _, e = io.CopyN(mw, data, size); e != nil {
|
partWriter := io.MultiWriter(partFile, md5Hasher, sha256Hasher)
|
||||||
|
if _, e = io.CopyN(partWriter, data, size); e != nil {
|
||||||
partFile.CloseAndPurge()
|
partFile.CloseAndPurge()
|
||||||
return "", probe.NewError(e)
|
return "", probe.NewError(e)
|
||||||
}
|
}
|
||||||
md5sum := hex.EncodeToString(h.Sum(nil))
|
|
||||||
|
md5sum := hex.EncodeToString(md5Hasher.Sum(nil))
|
||||||
// Verify if the written object is equal to what is expected, only
|
// Verify if the written object is equal to what is expected, only
|
||||||
// if it is requested as such.
|
// if it is requested as such.
|
||||||
if strings.TrimSpace(expectedMD5Sum) != "" {
|
if strings.TrimSpace(expectedMD5Sum) != "" {
|
||||||
if !isMD5SumEqual(strings.TrimSpace(expectedMD5Sum), md5sum) {
|
if !isMD5SumEqual(strings.TrimSpace(expectedMD5Sum), md5sum) {
|
||||||
partFile.CloseAndPurge()
|
partFile.CloseAndPurge()
|
||||||
return "", probe.NewError(BadDigest{Md5: expectedMD5Sum, Bucket: bucket, Object: object})
|
return "", probe.NewError(BadDigest{MD5: expectedMD5Sum, Bucket: bucket, Object: object})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if signature != nil {
|
if signature != nil {
|
||||||
ok, err := signature.DoesSignatureMatch(hex.EncodeToString(sh.Sum(nil)))
|
ok, err := signature.DoesSignatureMatch(hex.EncodeToString(sha256Hasher.Sum(nil)))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
partFile.CloseAndPurge()
|
partFile.CloseAndPurge()
|
||||||
return "", err.Trace()
|
return "", err.Trace()
|
||||||
@ -326,24 +377,30 @@ func (fs Filesystem) CreateObjectPart(bucket, object, uploadID, expectedMD5Sum s
|
|||||||
partMetadata.Size = fi.Size()
|
partMetadata.Size = fi.Size()
|
||||||
partMetadata.LastModified = fi.ModTime()
|
partMetadata.LastModified = fi.ModTime()
|
||||||
|
|
||||||
multiPartfile, e := os.OpenFile(objectPath+"$multiparts", os.O_RDWR|os.O_APPEND, 0600)
|
multipartSessionBytes, e := ioutil.ReadFile(partPathPrefix + "$multiparts")
|
||||||
if e != nil {
|
if e != nil {
|
||||||
return "", probe.NewError(e)
|
return "", probe.NewError(e)
|
||||||
}
|
}
|
||||||
defer multiPartfile.Close()
|
|
||||||
|
|
||||||
var deserializedMultipartSession MultipartSession
|
var deserializedMultipartSession MultipartSession
|
||||||
decoder := json.NewDecoder(multiPartfile)
|
if e = json.Unmarshal(multipartSessionBytes, &deserializedMultipartSession); e != nil {
|
||||||
if e = decoder.Decode(&deserializedMultipartSession); e != nil {
|
|
||||||
return "", probe.NewError(e)
|
return "", probe.NewError(e)
|
||||||
}
|
}
|
||||||
deserializedMultipartSession.Parts = append(deserializedMultipartSession.Parts, &partMetadata)
|
deserializedMultipartSession.Parts = append(deserializedMultipartSession.Parts, &partMetadata)
|
||||||
deserializedMultipartSession.TotalParts++
|
deserializedMultipartSession.TotalParts++
|
||||||
fs.multiparts.ActiveSession[object] = &deserializedMultipartSession
|
|
||||||
|
|
||||||
|
fs.rwLock.Lock()
|
||||||
|
fs.multiparts.ActiveSession[object] = &deserializedMultipartSession
|
||||||
|
fs.rwLock.Unlock()
|
||||||
|
|
||||||
|
// Sort by part number before saving.
|
||||||
sort.Sort(partNumber(deserializedMultipartSession.Parts))
|
sort.Sort(partNumber(deserializedMultipartSession.Parts))
|
||||||
encoder := json.NewEncoder(multiPartfile)
|
|
||||||
if e = encoder.Encode(&deserializedMultipartSession); e != nil {
|
multipartSessionBytes, e = json.Marshal(deserializedMultipartSession)
|
||||||
|
if e != nil {
|
||||||
|
return "", probe.NewError(e)
|
||||||
|
}
|
||||||
|
if e = ioutil.WriteFile(partPathPrefix+"$multiparts", multipartSessionBytes, 0600); e != nil {
|
||||||
return "", probe.NewError(e)
|
return "", probe.NewError(e)
|
||||||
}
|
}
|
||||||
return partMetadata.ETag, nil
|
return partMetadata.ETag, nil
|
||||||
@ -351,9 +408,6 @@ func (fs Filesystem) CreateObjectPart(bucket, object, uploadID, expectedMD5Sum s
|
|||||||
|
|
||||||
// CompleteMultipartUpload - complete a multipart upload and persist the data
|
// CompleteMultipartUpload - complete a multipart upload and persist the data
|
||||||
func (fs Filesystem) CompleteMultipartUpload(bucket, object, uploadID string, data io.Reader, signature *Signature) (ObjectMetadata, *probe.Error) {
|
func (fs Filesystem) CompleteMultipartUpload(bucket, object, uploadID string, data io.Reader, signature *Signature) (ObjectMetadata, *probe.Error) {
|
||||||
fs.rwLock.Lock()
|
|
||||||
defer fs.rwLock.Unlock()
|
|
||||||
|
|
||||||
// Check bucket name is valid.
|
// Check bucket name is valid.
|
||||||
if !IsValidBucketName(bucket) {
|
if !IsValidBucketName(bucket) {
|
||||||
return ObjectMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
return ObjectMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
||||||
@ -384,8 +438,8 @@ func (fs Filesystem) CompleteMultipartUpload(bucket, object, uploadID string, da
|
|||||||
if e != nil {
|
if e != nil {
|
||||||
return ObjectMetadata{}, probe.NewError(e)
|
return ObjectMetadata{}, probe.NewError(e)
|
||||||
}
|
}
|
||||||
h := md5.New()
|
md5Hasher := md5.New()
|
||||||
mw := io.MultiWriter(file, h)
|
objectWriter := io.MultiWriter(file, md5Hasher)
|
||||||
|
|
||||||
partBytes, e := ioutil.ReadAll(data)
|
partBytes, e := ioutil.ReadAll(data)
|
||||||
if e != nil {
|
if e != nil {
|
||||||
@ -405,35 +459,45 @@ func (fs Filesystem) CompleteMultipartUpload(bucket, object, uploadID string, da
|
|||||||
return ObjectMetadata{}, probe.NewError(SignatureDoesNotMatch{})
|
return ObjectMetadata{}, probe.NewError(SignatureDoesNotMatch{})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
parts := &CompleteMultipartUpload{}
|
completeMultipartUpload := &CompleteMultipartUpload{}
|
||||||
if e := xml.Unmarshal(partBytes, parts); e != nil {
|
if e := xml.Unmarshal(partBytes, completeMultipartUpload); e != nil {
|
||||||
file.CloseAndPurge()
|
file.CloseAndPurge()
|
||||||
return ObjectMetadata{}, probe.NewError(MalformedXML{})
|
return ObjectMetadata{}, probe.NewError(MalformedXML{})
|
||||||
}
|
}
|
||||||
if !sort.IsSorted(completedParts(parts.Part)) {
|
if !sort.IsSorted(completedParts(completeMultipartUpload.Part)) {
|
||||||
file.CloseAndPurge()
|
file.CloseAndPurge()
|
||||||
return ObjectMetadata{}, probe.NewError(InvalidPartOrder{})
|
return ObjectMetadata{}, probe.NewError(InvalidPartOrder{})
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := fs.concatParts(parts, objectPath, mw); err != nil {
|
// Save parts for verification.
|
||||||
|
parts := completeMultipartUpload.Part
|
||||||
|
|
||||||
|
fs.rwLock.RLock()
|
||||||
|
savedParts := fs.multiparts.ActiveSession[object].Parts
|
||||||
|
fs.rwLock.RUnlock()
|
||||||
|
|
||||||
|
if !doPartsMatch(parts, savedParts) {
|
||||||
file.CloseAndPurge()
|
file.CloseAndPurge()
|
||||||
return ObjectMetadata{}, err.Trace()
|
return ObjectMetadata{}, probe.NewError(InvalidPart{})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
partPathPrefix := objectPath + uploadID
|
||||||
|
if err := saveParts(partPathPrefix, objectWriter, parts); err != nil {
|
||||||
|
file.CloseAndPurge()
|
||||||
|
return ObjectMetadata{}, err.Trace(partPathPrefix)
|
||||||
|
}
|
||||||
|
if err := removeParts(partPathPrefix, savedParts); err != nil {
|
||||||
|
file.CloseAndPurge()
|
||||||
|
return ObjectMetadata{}, err.Trace(partPathPrefix)
|
||||||
|
}
|
||||||
|
|
||||||
|
fs.rwLock.Lock()
|
||||||
delete(fs.multiparts.ActiveSession, object)
|
delete(fs.multiparts.ActiveSession, object)
|
||||||
for _, part := range parts.Part {
|
fs.rwLock.Unlock()
|
||||||
if e = os.Remove(objectPath + fmt.Sprintf("$%d-$multiparts", part.PartNumber)); e != nil {
|
|
||||||
file.CloseAndPurge()
|
if err := saveMultipartsSession(*fs.multiparts); err != nil {
|
||||||
return ObjectMetadata{}, probe.NewError(e)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if e := os.Remove(objectPath + "$multiparts"); e != nil {
|
|
||||||
file.CloseAndPurge()
|
file.CloseAndPurge()
|
||||||
return ObjectMetadata{}, probe.NewError(e)
|
return ObjectMetadata{}, err.Trace(partPathPrefix)
|
||||||
}
|
|
||||||
if e := saveMultipartsSession(fs.multiparts); e != nil {
|
|
||||||
file.CloseAndPurge()
|
|
||||||
return ObjectMetadata{}, e.Trace()
|
|
||||||
}
|
}
|
||||||
file.Close()
|
file.Close()
|
||||||
|
|
||||||
@ -451,16 +515,13 @@ func (fs Filesystem) CompleteMultipartUpload(bucket, object, uploadID string, da
|
|||||||
Created: st.ModTime(),
|
Created: st.ModTime(),
|
||||||
Size: st.Size(),
|
Size: st.Size(),
|
||||||
ContentType: contentType,
|
ContentType: contentType,
|
||||||
Md5: hex.EncodeToString(h.Sum(nil)),
|
MD5: hex.EncodeToString(md5Hasher.Sum(nil)),
|
||||||
}
|
}
|
||||||
return newObject, nil
|
return newObject, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// ListObjectParts - list parts from incomplete multipart session for a given ObjectResourcesMetadata
|
// ListObjectParts - list parts from incomplete multipart session for a given ObjectResourcesMetadata
|
||||||
func (fs Filesystem) ListObjectParts(bucket, object string, resources ObjectResourcesMetadata) (ObjectResourcesMetadata, *probe.Error) {
|
func (fs Filesystem) ListObjectParts(bucket, object string, resources ObjectResourcesMetadata) (ObjectResourcesMetadata, *probe.Error) {
|
||||||
fs.rwLock.Lock()
|
|
||||||
defer fs.rwLock.Unlock()
|
|
||||||
|
|
||||||
// Check bucket name is valid.
|
// Check bucket name is valid.
|
||||||
if !IsValidBucketName(bucket) {
|
if !IsValidBucketName(bucket) {
|
||||||
return ObjectResourcesMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
return ObjectResourcesMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
||||||
@ -498,17 +559,16 @@ func (fs Filesystem) ListObjectParts(bucket, object string, resources ObjectReso
|
|||||||
}
|
}
|
||||||
|
|
||||||
objectPath := filepath.Join(bucketPath, object)
|
objectPath := filepath.Join(bucketPath, object)
|
||||||
multiPartfile, e := os.OpenFile(objectPath+"$multiparts", os.O_RDONLY, 0600)
|
partPathPrefix := objectPath + resources.UploadID
|
||||||
|
multipartSessionBytes, e := ioutil.ReadFile(partPathPrefix + "$multiparts")
|
||||||
if e != nil {
|
if e != nil {
|
||||||
return ObjectResourcesMetadata{}, probe.NewError(e)
|
return ObjectResourcesMetadata{}, probe.NewError(e)
|
||||||
}
|
}
|
||||||
defer multiPartfile.Close()
|
|
||||||
|
|
||||||
var deserializedMultipartSession MultipartSession
|
var deserializedMultipartSession MultipartSession
|
||||||
decoder := json.NewDecoder(multiPartfile)
|
if e = json.Unmarshal(multipartSessionBytes, &deserializedMultipartSession); e != nil {
|
||||||
if e = decoder.Decode(&deserializedMultipartSession); e != nil {
|
|
||||||
return ObjectResourcesMetadata{}, probe.NewError(e)
|
return ObjectResourcesMetadata{}, probe.NewError(e)
|
||||||
}
|
}
|
||||||
|
|
||||||
var parts []*PartMetadata
|
var parts []*PartMetadata
|
||||||
for i := startPartNumber; i <= deserializedMultipartSession.TotalParts; i++ {
|
for i := startPartNumber; i <= deserializedMultipartSession.TotalParts; i++ {
|
||||||
if len(parts) > objectResourcesMetadata.MaxParts {
|
if len(parts) > objectResourcesMetadata.MaxParts {
|
||||||
@ -527,9 +587,6 @@ func (fs Filesystem) ListObjectParts(bucket, object string, resources ObjectReso
|
|||||||
|
|
||||||
// AbortMultipartUpload - abort an incomplete multipart session
|
// AbortMultipartUpload - abort an incomplete multipart session
|
||||||
func (fs Filesystem) AbortMultipartUpload(bucket, object, uploadID string) *probe.Error {
|
func (fs Filesystem) AbortMultipartUpload(bucket, object, uploadID string) *probe.Error {
|
||||||
fs.rwLock.Lock()
|
|
||||||
defer fs.rwLock.Unlock()
|
|
||||||
|
|
||||||
// Check bucket name valid.
|
// Check bucket name valid.
|
||||||
if !IsValidBucketName(bucket) {
|
if !IsValidBucketName(bucket) {
|
||||||
return probe.NewError(BucketNameInvalid{Bucket: bucket})
|
return probe.NewError(BucketNameInvalid{Bucket: bucket})
|
||||||
@ -555,15 +612,20 @@ func (fs Filesystem) AbortMultipartUpload(bucket, object, uploadID string) *prob
|
|||||||
}
|
}
|
||||||
|
|
||||||
objectPath := filepath.Join(bucketPath, object)
|
objectPath := filepath.Join(bucketPath, object)
|
||||||
for _, part := range fs.multiparts.ActiveSession[object].Parts {
|
partPathPrefix := objectPath + uploadID
|
||||||
e := os.RemoveAll(objectPath + fmt.Sprintf("$%d-$multiparts", part.PartNumber))
|
fs.rwLock.RLock()
|
||||||
if e != nil {
|
savedParts := fs.multiparts.ActiveSession[object].Parts
|
||||||
return probe.NewError(e)
|
fs.rwLock.RUnlock()
|
||||||
}
|
|
||||||
|
if err := removeParts(partPathPrefix, savedParts); err != nil {
|
||||||
|
return err.Trace(partPathPrefix)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fs.rwLock.Lock()
|
||||||
delete(fs.multiparts.ActiveSession, object)
|
delete(fs.multiparts.ActiveSession, object)
|
||||||
if e := os.RemoveAll(objectPath + "$multiparts"); e != nil {
|
fs.rwLock.Unlock()
|
||||||
return probe.NewError(e)
|
if err := saveMultipartsSession(*fs.multiparts); err != nil {
|
||||||
|
return err.Trace(partPathPrefix)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -102,9 +102,6 @@ func (fs Filesystem) GetObject(w io.Writer, bucket, object string, start, length
|
|||||||
|
|
||||||
// GetObjectMetadata - get object metadata.
|
// GetObjectMetadata - get object metadata.
|
||||||
func (fs Filesystem) GetObjectMetadata(bucket, object string) (ObjectMetadata, *probe.Error) {
|
func (fs Filesystem) GetObjectMetadata(bucket, object string) (ObjectMetadata, *probe.Error) {
|
||||||
fs.rwLock.RLock()
|
|
||||||
defer fs.rwLock.RUnlock()
|
|
||||||
|
|
||||||
// Input validation.
|
// Input validation.
|
||||||
if !IsValidBucketName(bucket) {
|
if !IsValidBucketName(bucket) {
|
||||||
return ObjectMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
return ObjectMetadata{}, probe.NewError(BucketNameInvalid{Bucket: bucket})
|
||||||
@ -193,9 +190,6 @@ func isMD5SumEqual(expectedMD5Sum, actualMD5Sum string) bool {
|
|||||||
|
|
||||||
// CreateObject - create an object.
|
// CreateObject - create an object.
|
||||||
func (fs Filesystem) CreateObject(bucket, object, expectedMD5Sum string, size int64, data io.Reader, signature *Signature) (ObjectMetadata, *probe.Error) {
|
func (fs Filesystem) CreateObject(bucket, object, expectedMD5Sum string, size int64, data io.Reader, signature *Signature) (ObjectMetadata, *probe.Error) {
|
||||||
fs.rwLock.Lock()
|
|
||||||
defer fs.rwLock.Unlock()
|
|
||||||
|
|
||||||
di, e := disk.GetInfo(fs.path)
|
di, e := disk.GetInfo(fs.path)
|
||||||
if e != nil {
|
if e != nil {
|
||||||
return ObjectMetadata{}, probe.NewError(e)
|
return ObjectMetadata{}, probe.NewError(e)
|
||||||
@ -233,7 +227,7 @@ func (fs Filesystem) CreateObject(bucket, object, expectedMD5Sum string, size in
|
|||||||
expectedMD5SumBytes, e = base64.StdEncoding.DecodeString(strings.TrimSpace(expectedMD5Sum))
|
expectedMD5SumBytes, e = base64.StdEncoding.DecodeString(strings.TrimSpace(expectedMD5Sum))
|
||||||
if e != nil {
|
if e != nil {
|
||||||
// Pro-actively close the connection.
|
// Pro-actively close the connection.
|
||||||
return ObjectMetadata{}, probe.NewError(InvalidDigest{Md5: expectedMD5Sum})
|
return ObjectMetadata{}, probe.NewError(InvalidDigest{MD5: expectedMD5Sum})
|
||||||
}
|
}
|
||||||
expectedMD5Sum = hex.EncodeToString(expectedMD5SumBytes)
|
expectedMD5Sum = hex.EncodeToString(expectedMD5SumBytes)
|
||||||
}
|
}
|
||||||
@ -276,7 +270,7 @@ func (fs Filesystem) CreateObject(bucket, object, expectedMD5Sum string, size in
|
|||||||
if strings.TrimSpace(expectedMD5Sum) != "" {
|
if strings.TrimSpace(expectedMD5Sum) != "" {
|
||||||
if !isMD5SumEqual(strings.TrimSpace(expectedMD5Sum), md5Sum) {
|
if !isMD5SumEqual(strings.TrimSpace(expectedMD5Sum), md5Sum) {
|
||||||
file.CloseAndPurge()
|
file.CloseAndPurge()
|
||||||
return ObjectMetadata{}, probe.NewError(BadDigest{Md5: expectedMD5Sum, Bucket: bucket, Object: object})
|
return ObjectMetadata{}, probe.NewError(BadDigest{MD5: expectedMD5Sum, Bucket: bucket, Object: object})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
sha256Sum := hex.EncodeToString(sh.Sum(nil))
|
sha256Sum := hex.EncodeToString(sh.Sum(nil))
|
||||||
@ -307,7 +301,7 @@ func (fs Filesystem) CreateObject(bucket, object, expectedMD5Sum string, size in
|
|||||||
Created: st.ModTime(),
|
Created: st.ModTime(),
|
||||||
Size: st.Size(),
|
Size: st.Size(),
|
||||||
ContentType: contentType,
|
ContentType: contentType,
|
||||||
Md5: md5Sum,
|
MD5: md5Sum,
|
||||||
}
|
}
|
||||||
return newObject, nil
|
return newObject, nil
|
||||||
}
|
}
|
||||||
@ -344,9 +338,6 @@ func deleteObjectPath(basePath, deletePath, bucket, object string) *probe.Error
|
|||||||
|
|
||||||
// DeleteObject - delete and object
|
// DeleteObject - delete and object
|
||||||
func (fs Filesystem) DeleteObject(bucket, object string) *probe.Error {
|
func (fs Filesystem) DeleteObject(bucket, object string) *probe.Error {
|
||||||
fs.rwLock.Lock()
|
|
||||||
defer fs.rwLock.Unlock()
|
|
||||||
|
|
||||||
// check bucket name valid
|
// check bucket name valid
|
||||||
if !IsValidBucketName(bucket) {
|
if !IsValidBucketName(bucket) {
|
||||||
return probe.NewError(BucketNameInvalid{Bucket: bucket})
|
return probe.NewError(BucketNameInvalid{Bucket: bucket})
|
||||||
|
@ -60,8 +60,8 @@ type Multiparts struct {
|
|||||||
|
|
||||||
// New instantiate a new donut
|
// New instantiate a new donut
|
||||||
func New(rootPath string, minFreeDisk int64, maxBuckets int) (Filesystem, *probe.Error) {
|
func New(rootPath string, minFreeDisk int64, maxBuckets int) (Filesystem, *probe.Error) {
|
||||||
setFSBucketsConfigPath(filepath.Join(rootPath, "$buckets.json"))
|
setFSBucketsMetadataPath(filepath.Join(rootPath, "$buckets.json"))
|
||||||
setFSMultipartsConfigPath(filepath.Join(rootPath, "$multiparts-session.json"))
|
setFSMultipartsMetadataPath(filepath.Join(rootPath, "$multiparts-session.json"))
|
||||||
|
|
||||||
var err *probe.Error
|
var err *probe.Error
|
||||||
// load multiparts session from disk
|
// load multiparts session from disk
|
||||||
@ -73,7 +73,7 @@ func New(rootPath string, minFreeDisk int64, maxBuckets int) (Filesystem, *probe
|
|||||||
Version: "1",
|
Version: "1",
|
||||||
ActiveSession: make(map[string]*MultipartSession),
|
ActiveSession: make(map[string]*MultipartSession),
|
||||||
}
|
}
|
||||||
if err := saveMultipartsSession(multiparts); err != nil {
|
if err := saveMultipartsSession(*multiparts); err != nil {
|
||||||
return Filesystem{}, err.Trace()
|
return Filesystem{}, err.Trace()
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
@ -94,7 +94,7 @@ func New(rootPath string, minFreeDisk int64, maxBuckets int) (Filesystem, *probe
|
|||||||
Version: "1",
|
Version: "1",
|
||||||
Metadata: make(map[string]*BucketMetadata),
|
Metadata: make(map[string]*BucketMetadata),
|
||||||
}
|
}
|
||||||
if err := saveBucketsMetadata(buckets); err != nil {
|
if err := saveBucketsMetadata(*buckets); err != nil {
|
||||||
return Filesystem{}, err.Trace()
|
return Filesystem{}, err.Trace()
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
|
Loading…
Reference in New Issue
Block a user