2016-08-16 22:24:23 -04:00
|
|
|
/*
|
|
|
|
* Minio Cloud Storage, (C) 2016 Minio, Inc.
|
|
|
|
*
|
|
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
* you may not use this file except in compliance with the License.
|
|
|
|
* You may obtain a copy of the License at
|
|
|
|
*
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
*
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
* See the License for the specific language governing permissions and
|
|
|
|
* limitations under the License.
|
|
|
|
*/
|
|
|
|
|
2016-08-18 19:23:42 -04:00
|
|
|
package cmd
|
2016-08-16 22:24:23 -04:00
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
2016-09-21 23:08:08 -04:00
|
|
|
"crypto/md5"
|
|
|
|
"encoding/hex"
|
|
|
|
"encoding/xml"
|
2016-08-16 22:24:23 -04:00
|
|
|
"io/ioutil"
|
|
|
|
"net/http"
|
|
|
|
"net/http/httptest"
|
|
|
|
"net/url"
|
2016-09-21 23:08:08 -04:00
|
|
|
"sync"
|
2016-08-16 22:24:23 -04:00
|
|
|
"testing"
|
|
|
|
)
|
|
|
|
|
|
|
|
// Wrapper for calling GetObject API handler tests for both XL multiple disks and FS single drive setup.
|
|
|
|
func TestAPIGetOjectHandler(t *testing.T) {
|
2016-09-10 17:47:27 -04:00
|
|
|
ExecObjectLayerAPITest(t, testAPIGetOjectHandler, []string{"GetObject"})
|
2016-08-16 22:24:23 -04:00
|
|
|
}
|
|
|
|
|
2016-09-10 17:47:27 -04:00
|
|
|
func testAPIGetOjectHandler(obj ObjectLayer, instanceType, bucketName string, apiRouter http.Handler,
|
|
|
|
credentials credential, t TestErrHandler) {
|
2016-08-16 22:24:23 -04:00
|
|
|
objectName := "test-object"
|
|
|
|
// set of byte data for PutObject.
|
|
|
|
// object has to be inserted before running tests for GetObject.
|
|
|
|
// this is required even to assert the GetObject data,
|
|
|
|
// since dataInserted === dataFetched back is a primary criteria for any object storage this assertion is critical.
|
|
|
|
bytesData := []struct {
|
|
|
|
byteData []byte
|
|
|
|
}{
|
|
|
|
{generateBytesData(6 * 1024 * 1024)},
|
|
|
|
}
|
|
|
|
// set of inputs for uploading the objects before tests for downloading is done.
|
|
|
|
putObjectInputs := []struct {
|
|
|
|
bucketName string
|
|
|
|
objectName string
|
|
|
|
contentLength int64
|
|
|
|
textData []byte
|
|
|
|
metaData map[string]string
|
|
|
|
}{
|
|
|
|
// case - 1.
|
|
|
|
{bucketName, objectName, int64(len(bytesData[0].byteData)), bytesData[0].byteData, make(map[string]string)},
|
|
|
|
}
|
|
|
|
// iterate through the above set of inputs and upload the object.
|
|
|
|
for i, input := range putObjectInputs {
|
|
|
|
// uploading the object.
|
2016-09-10 17:47:27 -04:00
|
|
|
_, err := obj.PutObject(input.bucketName, input.objectName, input.contentLength, bytes.NewBuffer(input.textData), input.metaData)
|
2016-08-16 22:24:23 -04:00
|
|
|
// if object upload fails stop the test.
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Put Object case %d: Error uploading object: <ERROR> %v", i+1, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// test cases with inputs and expected result for GetObject.
|
|
|
|
testCases := []struct {
|
|
|
|
bucketName string
|
|
|
|
objectName string
|
|
|
|
byteRange string // range of bytes to be fetched from GetObject.
|
|
|
|
// expected output.
|
|
|
|
expectedContent []byte // expected response body.
|
|
|
|
expectedRespStatus int // expected response status body.
|
|
|
|
}{
|
|
|
|
// Test case - 1.
|
|
|
|
// Fetching the entire object and validating its contents.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
byteRange: "",
|
|
|
|
expectedContent: bytesData[0].byteData,
|
|
|
|
expectedRespStatus: http.StatusOK,
|
|
|
|
},
|
|
|
|
// Test case - 2.
|
|
|
|
// Case with non-existent object name.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: "abcd",
|
|
|
|
byteRange: "",
|
|
|
|
expectedContent: encodeResponse(getAPIErrorResponse(getAPIError(ErrNoSuchKey), getGetObjectURL("", bucketName, "abcd"))),
|
|
|
|
expectedRespStatus: http.StatusNotFound,
|
|
|
|
},
|
|
|
|
// Test case - 3.
|
|
|
|
// Requesting from range 10-100.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
byteRange: "bytes=10-100",
|
|
|
|
expectedContent: bytesData[0].byteData[10:101],
|
|
|
|
expectedRespStatus: http.StatusPartialContent,
|
|
|
|
},
|
|
|
|
// Test case - 4.
|
|
|
|
// Test case with invalid range.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
byteRange: "bytes=-0",
|
|
|
|
expectedContent: encodeResponse(getAPIErrorResponse(getAPIError(ErrInvalidRange), getGetObjectURL("", bucketName, objectName))),
|
|
|
|
expectedRespStatus: http.StatusRequestedRangeNotSatisfiable,
|
|
|
|
},
|
|
|
|
// Test case - 5.
|
|
|
|
// Test case with byte range exceeding the object size.
|
|
|
|
// Expected to read till end of the object.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
byteRange: "bytes=10-1000000000000000",
|
|
|
|
expectedContent: bytesData[0].byteData[10:],
|
|
|
|
expectedRespStatus: http.StatusPartialContent,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
// Iterating over the cases, fetching the object validating the response.
|
|
|
|
for i, testCase := range testCases {
|
|
|
|
// initialize HTTP NewRecorder, this records any mutations to response writer inside the handler.
|
|
|
|
rec := httptest.NewRecorder()
|
|
|
|
// construct HTTP request for Get Object end point.
|
|
|
|
req, err := newTestSignedRequest("GET", getGetObjectURL("", testCase.bucketName, testCase.objectName),
|
|
|
|
0, nil, credentials.AccessKeyID, credentials.SecretAccessKey)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: Failed to create HTTP request for Get Object: <ERROR> %v", i+1, err)
|
|
|
|
}
|
|
|
|
if testCase.byteRange != "" {
|
|
|
|
req.Header.Add("Range", testCase.byteRange)
|
|
|
|
}
|
|
|
|
// Since `apiRouter` satisfies `http.Handler` it has a ServeHTTP to execute the logic of the handler.
|
|
|
|
// Call the ServeHTTP to execute the handler,`func (api objectAPIHandlers) GetObjectHandler` handles the request.
|
|
|
|
apiRouter.ServeHTTP(rec, req)
|
|
|
|
// Assert the response code with the expected status.
|
|
|
|
if rec.Code != testCase.expectedRespStatus {
|
|
|
|
t.Fatalf("Case %d: Expected the response status to be `%d`, but instead found `%d`", i+1, testCase.expectedRespStatus, rec.Code)
|
|
|
|
}
|
|
|
|
// read the response body.
|
|
|
|
actualContent, err := ioutil.ReadAll(rec.Body)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: %s: Failed parsing response body: <ERROR> %v", i+1, instanceType, err)
|
|
|
|
}
|
|
|
|
// Verify whether the bucket obtained object is same as the one inserted.
|
|
|
|
if !bytes.Equal(testCase.expectedContent, actualContent) {
|
|
|
|
t.Errorf("Test %d: %s: Object content differs from expected value.: %s", i+1, instanceType, string(actualContent))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-09-04 16:37:14 -04:00
|
|
|
// Wrapper for calling PutObject API handler tests using streaming signature v4 for both XL multiple disks and FS single drive setup.
|
|
|
|
func TestAPIPutObjectStreamSigV4Handler(t *testing.T) {
|
2016-09-10 17:47:27 -04:00
|
|
|
ExecObjectLayerAPITest(t, testAPIPutObjectStreamSigV4Handler, []string{"PutObject"})
|
2016-09-04 16:37:14 -04:00
|
|
|
}
|
|
|
|
|
2016-09-10 17:47:27 -04:00
|
|
|
func testAPIPutObjectStreamSigV4Handler(obj ObjectLayer, instanceType, bucketName string, apiRouter http.Handler,
|
|
|
|
credentials credential, t TestErrHandler) {
|
2016-09-04 16:37:14 -04:00
|
|
|
|
2016-09-10 17:47:27 -04:00
|
|
|
objectName := "test-object"
|
2016-09-04 16:37:14 -04:00
|
|
|
bytesDataLen := 65 * 1024
|
|
|
|
bytesData := bytes.Repeat([]byte{'a'}, bytesDataLen)
|
|
|
|
|
|
|
|
// byte data for PutObject.
|
|
|
|
// test cases with inputs and expected result for GetObject.
|
|
|
|
testCases := []struct {
|
|
|
|
bucketName string
|
|
|
|
objectName string
|
|
|
|
data []byte
|
|
|
|
dataLen int
|
2016-09-16 05:45:42 -04:00
|
|
|
chunkSize int64
|
2016-09-04 16:37:14 -04:00
|
|
|
// expected output.
|
|
|
|
expectedContent []byte // expected response body.
|
|
|
|
expectedRespStatus int // expected response status body.
|
2016-09-16 05:45:42 -04:00
|
|
|
// Access keys
|
|
|
|
accessKey string
|
|
|
|
secretKey string
|
|
|
|
shouldPass bool
|
|
|
|
removeAuthHeader bool
|
2016-09-04 16:37:14 -04:00
|
|
|
}{
|
|
|
|
// Test case - 1.
|
|
|
|
// Fetching the entire object and validating its contents.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
data: bytesData,
|
|
|
|
dataLen: len(bytesData),
|
2016-09-16 05:45:42 -04:00
|
|
|
chunkSize: 64 * 1024, // 64k
|
2016-09-04 16:37:14 -04:00
|
|
|
expectedContent: []byte{},
|
|
|
|
expectedRespStatus: http.StatusOK,
|
2016-09-16 05:45:42 -04:00
|
|
|
accessKey: credentials.AccessKeyID,
|
|
|
|
secretKey: credentials.SecretAccessKey,
|
|
|
|
shouldPass: true,
|
|
|
|
},
|
|
|
|
// Test case - 2
|
|
|
|
// Small chunk size.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
data: bytesData,
|
|
|
|
dataLen: len(bytesData),
|
|
|
|
chunkSize: 1 * 1024, // 1k
|
|
|
|
expectedContent: []byte{},
|
|
|
|
expectedRespStatus: http.StatusOK,
|
|
|
|
accessKey: credentials.AccessKeyID,
|
|
|
|
secretKey: credentials.SecretAccessKey,
|
|
|
|
shouldPass: true,
|
|
|
|
},
|
|
|
|
// Test case - 3
|
|
|
|
// Invalid access key id.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
data: bytesData,
|
|
|
|
dataLen: len(bytesData),
|
|
|
|
chunkSize: 64 * 1024, // 64k
|
|
|
|
expectedContent: []byte{},
|
|
|
|
expectedRespStatus: http.StatusForbidden,
|
|
|
|
accessKey: "",
|
|
|
|
secretKey: "",
|
|
|
|
shouldPass: false,
|
|
|
|
},
|
|
|
|
// Test case - 4
|
|
|
|
// Wrong auth header returns as bad request.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
data: bytesData,
|
|
|
|
dataLen: len(bytesData),
|
|
|
|
chunkSize: 64 * 1024, // 64k
|
|
|
|
expectedContent: []byte{},
|
|
|
|
expectedRespStatus: http.StatusBadRequest,
|
|
|
|
accessKey: credentials.AccessKeyID,
|
|
|
|
secretKey: credentials.SecretAccessKey,
|
|
|
|
shouldPass: false,
|
|
|
|
removeAuthHeader: true,
|
|
|
|
},
|
|
|
|
// Test case - 5
|
|
|
|
// Large chunk size.. also passes.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
data: bytesData,
|
|
|
|
dataLen: len(bytesData),
|
|
|
|
chunkSize: 100 * 1024, // 100k
|
|
|
|
expectedContent: []byte{},
|
|
|
|
expectedRespStatus: http.StatusOK,
|
|
|
|
accessKey: credentials.AccessKeyID,
|
|
|
|
secretKey: credentials.SecretAccessKey,
|
|
|
|
shouldPass: false,
|
2016-09-04 16:37:14 -04:00
|
|
|
},
|
|
|
|
}
|
|
|
|
// Iterating over the cases, fetching the object validating the response.
|
|
|
|
for i, testCase := range testCases {
|
|
|
|
// initialize HTTP NewRecorder, this records any mutations to response writer inside the handler.
|
|
|
|
rec := httptest.NewRecorder()
|
|
|
|
// construct HTTP request for Put Object end point.
|
|
|
|
req, err := newTestStreamingSignedRequest("PUT",
|
|
|
|
getPutObjectURL("", testCase.bucketName, testCase.objectName),
|
2016-09-16 05:45:42 -04:00
|
|
|
int64(testCase.dataLen), testCase.chunkSize, bytes.NewReader(testCase.data),
|
|
|
|
testCase.accessKey, testCase.secretKey)
|
2016-09-04 16:37:14 -04:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: Failed to create HTTP request for Put Object: <ERROR> %v", i+1, err)
|
|
|
|
}
|
2016-09-16 05:45:42 -04:00
|
|
|
// Removes auth header if test case requires it.
|
|
|
|
if testCase.removeAuthHeader {
|
|
|
|
req.Header.Del("Authorization")
|
|
|
|
}
|
2016-09-04 16:37:14 -04:00
|
|
|
// Since `apiRouter` satisfies `http.Handler` it has a ServeHTTP to execute the logic of the handler.
|
|
|
|
// Call the ServeHTTP to execute the handler,`func (api objectAPIHandlers) GetObjectHandler` handles the request.
|
|
|
|
apiRouter.ServeHTTP(rec, req)
|
|
|
|
// Assert the response code with the expected status.
|
|
|
|
if rec.Code != testCase.expectedRespStatus {
|
2016-09-16 05:45:42 -04:00
|
|
|
t.Errorf("Test %d: Expected the response status to be `%d`, but instead found `%d`", i+1, testCase.expectedRespStatus, rec.Code)
|
2016-09-04 16:37:14 -04:00
|
|
|
}
|
|
|
|
// read the response body.
|
|
|
|
actualContent, err := ioutil.ReadAll(rec.Body)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: %s: Failed parsing response body: <ERROR> %v", i+1, instanceType, err)
|
|
|
|
}
|
2016-09-16 05:45:42 -04:00
|
|
|
if testCase.shouldPass {
|
|
|
|
// Verify whether the bucket obtained object is same as the one inserted.
|
|
|
|
if !bytes.Equal(testCase.expectedContent, actualContent) {
|
|
|
|
t.Errorf("Test %d: %s: Object content differs from expected value.: %s", i+1, instanceType, string(actualContent))
|
|
|
|
}
|
2016-09-04 16:37:14 -04:00
|
|
|
|
2016-09-16 05:45:42 -04:00
|
|
|
buffer := new(bytes.Buffer)
|
|
|
|
err = obj.GetObject(testCase.bucketName, testCase.objectName, 0, int64(bytesDataLen), buffer)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: %s: Failed to fetch the copied object: <ERROR> %s", i+1, instanceType, err)
|
|
|
|
}
|
|
|
|
if !bytes.Equal(bytesData, buffer.Bytes()) {
|
|
|
|
t.Errorf("Test %d: %s: Data Mismatch: Data fetched back from the uploaded object doesn't match the original one.", i+1, instanceType)
|
|
|
|
}
|
|
|
|
buffer.Reset()
|
2016-09-04 16:37:14 -04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wrapper for calling PutObject API handler tests for both XL multiple disks and FS single drive setup.
|
|
|
|
func TestAPIPutObjectHandler(t *testing.T) {
|
2016-09-10 17:47:27 -04:00
|
|
|
ExecObjectLayerAPITest(t, testAPIPutObjectHandler, []string{"PutObject"})
|
2016-09-04 16:37:14 -04:00
|
|
|
}
|
|
|
|
|
2016-09-10 17:47:27 -04:00
|
|
|
func testAPIPutObjectHandler(obj ObjectLayer, instanceType, bucketName string, apiRouter http.Handler,
|
|
|
|
credentials credential, t TestErrHandler) {
|
2016-09-04 16:37:14 -04:00
|
|
|
|
2016-09-10 17:47:27 -04:00
|
|
|
objectName := "test-object"
|
2016-09-04 16:37:14 -04:00
|
|
|
// byte data for PutObject.
|
|
|
|
bytesData := generateBytesData(6 * 1024 * 1024)
|
|
|
|
|
|
|
|
// test cases with inputs and expected result for GetObject.
|
|
|
|
testCases := []struct {
|
|
|
|
bucketName string
|
|
|
|
objectName string
|
|
|
|
data []byte
|
|
|
|
dataLen int
|
|
|
|
// expected output.
|
|
|
|
expectedContent []byte // expected response body.
|
|
|
|
expectedRespStatus int // expected response status body.
|
|
|
|
}{
|
|
|
|
// Test case - 1.
|
|
|
|
// Fetching the entire object and validating its contents.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
data: bytesData,
|
|
|
|
dataLen: len(bytesData),
|
|
|
|
expectedContent: []byte{},
|
|
|
|
expectedRespStatus: http.StatusOK,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
// Iterating over the cases, fetching the object validating the response.
|
|
|
|
for i, testCase := range testCases {
|
|
|
|
// initialize HTTP NewRecorder, this records any mutations to response writer inside the handler.
|
|
|
|
rec := httptest.NewRecorder()
|
|
|
|
// construct HTTP request for Get Object end point.
|
|
|
|
req, err := newTestSignedRequest("PUT", getPutObjectURL("", testCase.bucketName, testCase.objectName),
|
|
|
|
int64(testCase.dataLen), bytes.NewReader(testCase.data), credentials.AccessKeyID, credentials.SecretAccessKey)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: Failed to create HTTP request for Put Object: <ERROR> %v", i+1, err)
|
|
|
|
}
|
|
|
|
// Since `apiRouter` satisfies `http.Handler` it has a ServeHTTP to execute the logic of the handler.
|
|
|
|
// Call the ServeHTTP to execute the handler,`func (api objectAPIHandlers) GetObjectHandler` handles the request.
|
|
|
|
apiRouter.ServeHTTP(rec, req)
|
|
|
|
// Assert the response code with the expected status.
|
|
|
|
if rec.Code != testCase.expectedRespStatus {
|
|
|
|
t.Fatalf("Case %d: Expected the response status to be `%d`, but instead found `%d`", i+1, testCase.expectedRespStatus, rec.Code)
|
|
|
|
}
|
|
|
|
// read the response body.
|
|
|
|
actualContent, err := ioutil.ReadAll(rec.Body)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: %s: Failed parsing response body: <ERROR> %v", i+1, instanceType, err)
|
|
|
|
}
|
|
|
|
// Verify whether the bucket obtained object is same as the one inserted.
|
|
|
|
if !bytes.Equal(testCase.expectedContent, actualContent) {
|
|
|
|
t.Errorf("Test %d: %s: Object content differs from expected value.: %s", i+1, instanceType, string(actualContent))
|
|
|
|
}
|
|
|
|
|
|
|
|
buffer := new(bytes.Buffer)
|
|
|
|
err = obj.GetObject(testCase.bucketName, testCase.objectName, 0, int64(len(bytesData)), buffer)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: %s: Failed to fetch the copied object: <ERROR> %s", i+1, instanceType, err)
|
|
|
|
}
|
|
|
|
if !bytes.Equal(bytesData, buffer.Bytes()) {
|
|
|
|
t.Errorf("Test %d: %s: Data Mismatch: Data fetched back from the uploaded object doesn't match the original one.", i+1, instanceType)
|
|
|
|
}
|
|
|
|
buffer.Reset()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-08-16 22:24:23 -04:00
|
|
|
// Wrapper for calling Copy Object API handler tests for both XL multiple disks and single node setup.
|
|
|
|
func TestAPICopyObjectHandler(t *testing.T) {
|
2016-09-10 17:47:27 -04:00
|
|
|
ExecObjectLayerAPITest(t, testAPICopyObjectHandler, []string{"CopyObject"})
|
2016-08-16 22:24:23 -04:00
|
|
|
}
|
|
|
|
|
2016-09-10 17:47:27 -04:00
|
|
|
func testAPICopyObjectHandler(obj ObjectLayer, instanceType, bucketName string, apiRouter http.Handler,
|
|
|
|
credentials credential, t TestErrHandler) {
|
2016-08-16 22:24:23 -04:00
|
|
|
|
2016-09-10 17:47:27 -04:00
|
|
|
objectName := "test-object"
|
|
|
|
// register event notifier.
|
|
|
|
err := initEventNotifier(obj)
|
2016-08-16 22:24:23 -04:00
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Initializing event notifiers failed")
|
|
|
|
}
|
|
|
|
|
|
|
|
// set of byte data for PutObject.
|
|
|
|
// object has to be inserted before running tests for Copy Object.
|
|
|
|
// this is required even to assert the copied object,
|
|
|
|
bytesData := []struct {
|
|
|
|
byteData []byte
|
|
|
|
}{
|
|
|
|
{generateBytesData(6 * 1024 * 1024)},
|
|
|
|
}
|
|
|
|
|
|
|
|
buffers := []*bytes.Buffer{
|
|
|
|
new(bytes.Buffer),
|
|
|
|
new(bytes.Buffer),
|
|
|
|
}
|
|
|
|
|
|
|
|
// set of inputs for uploading the objects before tests for downloading is done.
|
|
|
|
putObjectInputs := []struct {
|
|
|
|
bucketName string
|
|
|
|
objectName string
|
|
|
|
contentLength int64
|
|
|
|
textData []byte
|
|
|
|
metaData map[string]string
|
|
|
|
}{
|
|
|
|
// case - 1.
|
|
|
|
{bucketName, objectName, int64(len(bytesData[0].byteData)), bytesData[0].byteData, make(map[string]string)},
|
|
|
|
}
|
|
|
|
// iterate through the above set of inputs and upload the object.
|
|
|
|
for i, input := range putObjectInputs {
|
|
|
|
// uploading the object.
|
|
|
|
_, err = obj.PutObject(input.bucketName, input.objectName, input.contentLength, bytes.NewBuffer(input.textData), input.metaData)
|
|
|
|
// if object upload fails stop the test.
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Put Object case %d: Error uploading object: <ERROR> %v", i+1, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// test cases with inputs and expected result for Copy Object.
|
|
|
|
testCases := []struct {
|
|
|
|
bucketName string
|
|
|
|
newObjectName string // name of the newly copied object.
|
|
|
|
copySourceHeader string // data for "X-Amz-Copy-Source" header. Contains the object to be copied in the URL.
|
|
|
|
// expected output.
|
|
|
|
expectedRespStatus int
|
|
|
|
}{
|
|
|
|
// Test case - 1.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
newObjectName: "newObject1",
|
|
|
|
copySourceHeader: url.QueryEscape("/" + bucketName + "/" + objectName),
|
|
|
|
expectedRespStatus: http.StatusOK,
|
|
|
|
},
|
|
|
|
|
|
|
|
// Test case - 2.
|
|
|
|
// Test case with invalid source object.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
newObjectName: "newObject1",
|
|
|
|
copySourceHeader: url.QueryEscape("/"),
|
|
|
|
expectedRespStatus: http.StatusBadRequest,
|
|
|
|
},
|
|
|
|
// Test case - 3.
|
|
|
|
// Test case with new object name is same as object to be copied.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
newObjectName: objectName,
|
|
|
|
copySourceHeader: url.QueryEscape("/" + bucketName + "/" + objectName),
|
|
|
|
expectedRespStatus: http.StatusBadRequest,
|
|
|
|
},
|
|
|
|
// Test case - 4.
|
|
|
|
// Test case with non-existent source file.
|
|
|
|
// Case for the purpose of failing `api.ObjectAPI.GetObjectInfo`.
|
|
|
|
// Expecting the response status code to http.StatusNotFound (404).
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
newObjectName: objectName,
|
|
|
|
copySourceHeader: url.QueryEscape("/" + bucketName + "/" + "non-existent-object"),
|
|
|
|
expectedRespStatus: http.StatusNotFound,
|
|
|
|
},
|
|
|
|
// Test case - 5.
|
|
|
|
// Test case with non-existent source file.
|
|
|
|
// Case for the purpose of failing `api.ObjectAPI.PutObject`.
|
|
|
|
// Expecting the response status code to http.StatusNotFound (404).
|
|
|
|
{
|
|
|
|
bucketName: "non-existent-destination-bucket",
|
|
|
|
newObjectName: objectName,
|
|
|
|
copySourceHeader: url.QueryEscape("/" + bucketName + "/" + objectName),
|
|
|
|
expectedRespStatus: http.StatusNotFound,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
for i, testCase := range testCases {
|
|
|
|
// initialize HTTP NewRecorder, this records any mutations to response writer inside the handler.
|
|
|
|
rec := httptest.NewRecorder()
|
|
|
|
// construct HTTP request for copy object.
|
|
|
|
req, err := newTestSignedRequest("PUT", getCopyObjectURL("", testCase.bucketName, testCase.newObjectName),
|
|
|
|
0, nil, credentials.AccessKeyID, credentials.SecretAccessKey)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: Failed to create HTTP request for copy Object: <ERROR> %v", i+1, err)
|
|
|
|
}
|
|
|
|
// "X-Amz-Copy-Source" header contains the information about the source bucket and the object to copied.
|
|
|
|
if testCase.copySourceHeader != "" {
|
|
|
|
req.Header.Set("X-Amz-Copy-Source", testCase.copySourceHeader)
|
|
|
|
}
|
|
|
|
// Since `apiRouter` satisfies `http.Handler` it has a ServeHTTP to execute the logic of the handler.
|
|
|
|
// Call the ServeHTTP to execute the handler, `func (api objectAPIHandlers) CopyObjectHandler` handles the request.
|
|
|
|
apiRouter.ServeHTTP(rec, req)
|
|
|
|
// Assert the response code with the expected status.
|
|
|
|
if rec.Code != testCase.expectedRespStatus {
|
|
|
|
t.Fatalf("Test %d: %s: Expected the response status to be `%d`, but instead found `%d`", i+1, instanceType, testCase.expectedRespStatus, rec.Code)
|
|
|
|
}
|
|
|
|
if rec.Code == http.StatusOK {
|
|
|
|
// See if the new object is formed.
|
|
|
|
// testing whether the copy was successful.
|
|
|
|
err = obj.GetObject(testCase.bucketName, testCase.newObjectName, 0, int64(len(bytesData[0].byteData)), buffers[0])
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: %s: Failed to fetch the copied object: <ERROR> %s", i+1, instanceType, err)
|
|
|
|
}
|
|
|
|
if !bytes.Equal(bytesData[0].byteData, buffers[0].Bytes()) {
|
|
|
|
t.Errorf("Test %d: %s: Data Mismatch: Data fetched back from the copied object doesn't match the original one.", i+1, instanceType)
|
|
|
|
}
|
|
|
|
buffers[0].Reset()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2016-09-21 23:08:08 -04:00
|
|
|
|
|
|
|
// Wrapper for calling NewMultipartUpload tests for both XL multiple disks and single node setup.
|
|
|
|
// First register the HTTP handler for NewMutlipartUpload, then a HTTP request for NewMultipart upload is made.
|
|
|
|
// The UploadID from the response body is parsed and its existance is asserted with an attempt to ListParts using it.
|
|
|
|
func TestAPINewMultipartHandler(t *testing.T) {
|
|
|
|
ExecObjectLayerAPITest(t, testAPINewMultipartHandler, []string{"NewMultipart"})
|
|
|
|
}
|
|
|
|
|
|
|
|
func testAPINewMultipartHandler(obj ObjectLayer, instanceType, bucketName string, apiRouter http.Handler,
|
|
|
|
credentials credential, t TestErrHandler) {
|
|
|
|
|
|
|
|
objectName := "test-object-new-multipart"
|
|
|
|
rec := httptest.NewRecorder()
|
|
|
|
// construct HTTP request for copy object.
|
|
|
|
req, err := newTestSignedRequest("POST", getNewMultipartURL("", bucketName, objectName), 0, nil, credentials.AccessKeyID, credentials.SecretAccessKey)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Failed to create HTTP request for copy Object: <ERROR> %v", err)
|
|
|
|
}
|
|
|
|
// Since `apiRouter` satisfies `http.Handler` it has a ServeHTTP to execute the logic of the handler.
|
|
|
|
// Call the ServeHTTP to executes the registered handler.
|
|
|
|
apiRouter.ServeHTTP(rec, req)
|
|
|
|
// Assert the response code with the expected status.
|
|
|
|
if rec.Code != http.StatusOK {
|
|
|
|
t.Fatalf("%s: Expected the response status to be `%d`, but instead found `%d`", instanceType, http.StatusOK, rec.Code)
|
|
|
|
}
|
|
|
|
// decode the response body.
|
|
|
|
decoder := xml.NewDecoder(rec.Body)
|
|
|
|
multipartResponse := &InitiateMultipartUploadResponse{}
|
|
|
|
|
|
|
|
err = decoder.Decode(multipartResponse)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Error decoding the recorded response Body")
|
|
|
|
}
|
|
|
|
// verify the uploadID my making an attempt to list parts.
|
|
|
|
_, err = obj.ListObjectParts(bucketName, objectName, multipartResponse.UploadID, 0, 1)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Invalid UploadID: <ERROR> %s", err)
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wrapper for calling NewMultipartUploadParallel tests for both XL multiple disks and single node setup.
|
|
|
|
// The objective of the test is to initialte multipart upload on the same object 10 times concurrently,
|
|
|
|
// The UploadID from the response body is parsed and its existance is asserted with an attempt to ListParts using it.
|
|
|
|
func TestAPINewMultipartHandlerParallel(t *testing.T) {
|
|
|
|
ExecObjectLayerAPITest(t, testAPINewMultipartHandlerParallel, []string{"NewMultipart"})
|
|
|
|
}
|
|
|
|
|
|
|
|
func testAPINewMultipartHandlerParallel(obj ObjectLayer, instanceType, bucketName string, apiRouter http.Handler,
|
|
|
|
credentials credential, t TestErrHandler) {
|
|
|
|
// used for storing the uploadID's parsed on concurrent HTTP requests for NewMultipart upload on the same object.
|
|
|
|
testUploads := struct {
|
|
|
|
sync.Mutex
|
|
|
|
uploads []string
|
|
|
|
}{}
|
|
|
|
|
|
|
|
objectName := "test-object-new-multipart-parallel"
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
for i := 0; i < 10; i++ {
|
|
|
|
wg.Add(1)
|
|
|
|
// Initiate NewMultipart upload on the same object 10 times concurrrently.
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
rec := httptest.NewRecorder()
|
|
|
|
// construct HTTP request for copy object.
|
|
|
|
req, err := newTestSignedRequest("POST", getNewMultipartURL("", bucketName, objectName), 0, nil, credentials.AccessKeyID, credentials.SecretAccessKey)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Failed to create HTTP request for copy Object: <ERROR> %v", err)
|
|
|
|
}
|
|
|
|
// Since `apiRouter` satisfies `http.Handler` it has a ServeHTTP to execute the logic of the handler.
|
|
|
|
// Call the ServeHTTP to executes the registered handler.
|
|
|
|
apiRouter.ServeHTTP(rec, req)
|
|
|
|
// Assert the response code with the expected status.
|
|
|
|
if rec.Code != http.StatusOK {
|
|
|
|
t.Fatalf("Minio %s: Expected the response status to be `%d`, but instead found `%d`", instanceType, http.StatusOK, rec.Code)
|
|
|
|
}
|
|
|
|
// decode the response body.
|
|
|
|
decoder := xml.NewDecoder(rec.Body)
|
|
|
|
multipartResponse := &InitiateMultipartUploadResponse{}
|
|
|
|
|
|
|
|
err = decoder.Decode(multipartResponse)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Minio %s: Error decoding the recorded response Body", instanceType)
|
|
|
|
}
|
|
|
|
// push the obtained upload ID from the response into the array.
|
|
|
|
testUploads.Lock()
|
|
|
|
testUploads.uploads = append(testUploads.uploads, multipartResponse.UploadID)
|
|
|
|
testUploads.Unlock()
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
// Wait till all go routines finishes execution.
|
|
|
|
wg.Wait()
|
|
|
|
// Validate the upload ID by an attempt to list parts using it.
|
|
|
|
for _, uploadID := range testUploads.uploads {
|
|
|
|
_, err := obj.ListObjectParts(bucketName, objectName, uploadID, 0, 1)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Invalid UploadID: <ERROR> %s", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// The UploadID from the response body is parsed and its existance is asserted with an attempt to ListParts using it.
|
|
|
|
func TestAPICompleteMultipartHandler(t *testing.T) {
|
|
|
|
ExecObjectLayerAPITest(t, testAPICompleteMultipartHandler, []string{"CompleteMultipart"})
|
|
|
|
}
|
|
|
|
|
|
|
|
func testAPICompleteMultipartHandler(obj ObjectLayer, instanceType, bucketName string, apiRouter http.Handler,
|
|
|
|
credentials credential, t TestErrHandler) {
|
|
|
|
|
|
|
|
// Calculates MD5 sum of the given byte array.
|
|
|
|
findMD5 := func(toBeHashed []byte) string {
|
|
|
|
hasher := md5.New()
|
|
|
|
hasher.Write(toBeHashed)
|
|
|
|
return hex.EncodeToString(hasher.Sum(nil))
|
|
|
|
}
|
|
|
|
|
|
|
|
objectName := "test-object-new-multipart"
|
|
|
|
|
|
|
|
uploadID, err := obj.NewMultipartUpload(bucketName, objectName, nil)
|
|
|
|
if err != nil {
|
|
|
|
// Failed to create NewMultipartUpload, abort.
|
|
|
|
t.Fatalf("Minio %s : <ERROR> %s", instanceType, err)
|
|
|
|
}
|
|
|
|
var uploadIDs []string
|
|
|
|
uploadIDs = append(uploadIDs, uploadID)
|
|
|
|
// Parts with size greater than 5 MB.
|
|
|
|
// Generating a 6MB byte array.
|
|
|
|
validPart := bytes.Repeat([]byte("abcdef"), 1024*1024)
|
|
|
|
validPartMD5 := findMD5(validPart)
|
|
|
|
// Create multipart parts.
|
|
|
|
// Need parts to be uploaded before CompleteMultiPartUpload can be called tested.
|
|
|
|
parts := []struct {
|
|
|
|
bucketName string
|
|
|
|
objName string
|
|
|
|
uploadID string
|
|
|
|
PartID int
|
|
|
|
inputReaderData string
|
|
|
|
inputMd5 string
|
|
|
|
intputDataSize int64
|
|
|
|
}{
|
|
|
|
// Case 1-4.
|
|
|
|
// Creating sequence of parts for same uploadID.
|
|
|
|
{bucketName, objectName, uploadIDs[0], 1, "abcd", "e2fc714c4727ee9395f324cd2e7f331f", int64(len("abcd"))},
|
|
|
|
{bucketName, objectName, uploadIDs[0], 2, "efgh", "1f7690ebdd9b4caf8fab49ca1757bf27", int64(len("efgh"))},
|
|
|
|
{bucketName, objectName, uploadIDs[0], 3, "ijkl", "09a0877d04abf8759f99adec02baf579", int64(len("abcd"))},
|
|
|
|
{bucketName, objectName, uploadIDs[0], 4, "mnop", "e132e96a5ddad6da8b07bba6f6131fef", int64(len("abcd"))},
|
|
|
|
// Part with size larger than 5Mb.
|
|
|
|
{bucketName, objectName, uploadIDs[0], 5, string(validPart), validPartMD5, int64(len(string(validPart)))},
|
|
|
|
{bucketName, objectName, uploadIDs[0], 6, string(validPart), validPartMD5, int64(len(string(validPart)))},
|
|
|
|
}
|
|
|
|
// Iterating over creatPartCases to generate multipart chunks.
|
|
|
|
for _, part := range parts {
|
|
|
|
_, err = obj.PutObjectPart(part.bucketName, part.objName, part.uploadID, part.PartID, part.intputDataSize, bytes.NewBufferString(part.inputReaderData), part.inputMd5)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("%s : %s", instanceType, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Parts to be sent as input for CompleteMultipartUpload.
|
|
|
|
inputParts := []struct {
|
|
|
|
parts []completePart
|
|
|
|
}{
|
|
|
|
// inputParts - 0.
|
|
|
|
// Case for replicating ETag mismatch.
|
|
|
|
{
|
|
|
|
[]completePart{
|
|
|
|
{ETag: "abcd", PartNumber: 1},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// inputParts - 1.
|
|
|
|
// should error out with part too small.
|
|
|
|
{
|
|
|
|
[]completePart{
|
|
|
|
{ETag: "e2fc714c4727ee9395f324cd2e7f331f", PartNumber: 1},
|
|
|
|
{ETag: "1f7690ebdd9b4caf8fab49ca1757bf27", PartNumber: 2},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// inputParts - 2.
|
|
|
|
// Case with invalid Part number.
|
|
|
|
{
|
|
|
|
[]completePart{
|
|
|
|
{ETag: "e2fc714c4727ee9395f324cd2e7f331f", PartNumber: 10},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// inputParts - 3.
|
|
|
|
// Case with valid parts,but parts are unsorted.
|
|
|
|
// Part size greater than 5MB.
|
|
|
|
{
|
|
|
|
[]completePart{
|
|
|
|
{ETag: validPartMD5, PartNumber: 6},
|
|
|
|
{ETag: validPartMD5, PartNumber: 5},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
// inputParts - 4.
|
|
|
|
// Case with valid part.
|
|
|
|
// Part size greater than 5MB.
|
|
|
|
{
|
|
|
|
[]completePart{
|
|
|
|
{ETag: validPartMD5, PartNumber: 5},
|
|
|
|
{ETag: validPartMD5, PartNumber: 6},
|
|
|
|
},
|
|
|
|
},
|
|
|
|
}
|
|
|
|
// on succesfull complete multipart operation the s3MD5 for the parts uploaded iwll be returned.
|
|
|
|
s3MD5, err := completeMultipartMD5(inputParts[3].parts...)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Obtaining S3MD5 failed")
|
|
|
|
}
|
|
|
|
// generating the response body content for the success case.
|
|
|
|
successResponse := generateCompleteMultpartUploadResponse(bucketName, objectName, getGetObjectURL("", bucketName, objectName), s3MD5)
|
|
|
|
encodedSuccessResponse := encodeResponse(successResponse)
|
|
|
|
|
|
|
|
testCases := []struct {
|
|
|
|
bucket string
|
|
|
|
object string
|
|
|
|
uploadID string
|
|
|
|
parts []completePart
|
|
|
|
// Expected output of CompleteMultipartUpload.
|
|
|
|
expectedContent []byte
|
|
|
|
// Expected HTTP Response status.
|
|
|
|
expectedRespStatus int
|
|
|
|
}{
|
|
|
|
// Test case - 1.
|
|
|
|
// Upload and PartNumber exists, But a deliberate ETag mismatch is introduced.
|
|
|
|
{
|
|
|
|
bucket: bucketName,
|
|
|
|
object: objectName,
|
|
|
|
uploadID: uploadIDs[0],
|
|
|
|
parts: inputParts[0].parts,
|
|
|
|
expectedContent: encodeResponse(getAPIErrorResponse(getAPIError(toAPIErrorCode(BadDigest{})),
|
|
|
|
getGetObjectURL("", bucketName, objectName))),
|
|
|
|
expectedRespStatus: http.StatusBadRequest,
|
|
|
|
},
|
|
|
|
// Test case - 2.
|
|
|
|
// No parts specified in completePart{}.
|
|
|
|
// Should return ErrMalformedXML in the response body.
|
|
|
|
{
|
|
|
|
bucket: bucketName,
|
|
|
|
object: objectName,
|
|
|
|
uploadID: uploadIDs[0],
|
|
|
|
parts: []completePart{},
|
|
|
|
expectedContent: encodeResponse(getAPIErrorResponse(getAPIError(ErrMalformedXML), getGetObjectURL("", bucketName, objectName))),
|
|
|
|
expectedRespStatus: http.StatusBadRequest,
|
|
|
|
},
|
|
|
|
// Test case - 3.
|
|
|
|
// Non-Existant uploadID.
|
|
|
|
// 404 Not Found response status expected.
|
|
|
|
{
|
|
|
|
bucket: bucketName,
|
|
|
|
object: objectName,
|
|
|
|
uploadID: "abc",
|
|
|
|
parts: inputParts[0].parts,
|
|
|
|
expectedContent: encodeResponse(getAPIErrorResponse(getAPIError(toAPIErrorCode(InvalidUploadID{UploadID: "abc"})), getGetObjectURL("", bucketName, objectName))),
|
|
|
|
expectedRespStatus: http.StatusNotFound,
|
|
|
|
},
|
|
|
|
// Test case - 4.
|
|
|
|
// Case with part size being less than minimum allowed size.
|
|
|
|
{
|
|
|
|
bucket: bucketName,
|
|
|
|
object: objectName,
|
|
|
|
uploadID: uploadIDs[0],
|
|
|
|
parts: inputParts[1].parts,
|
|
|
|
expectedContent: encodeResponse(completeMultipartAPIError{int64(4), int64(5242880), 1, "e2fc714c4727ee9395f324cd2e7f331f",
|
|
|
|
getAPIErrorResponse(getAPIError(toAPIErrorCode(PartTooSmall{PartNumber: 1})),
|
|
|
|
getGetObjectURL("", bucketName, objectName))}),
|
|
|
|
expectedRespStatus: http.StatusBadRequest,
|
|
|
|
},
|
|
|
|
// Test case - 5.
|
|
|
|
// TestCase with invalid Part Number.
|
|
|
|
{
|
|
|
|
bucket: bucketName,
|
|
|
|
object: objectName,
|
|
|
|
uploadID: uploadIDs[0],
|
|
|
|
parts: inputParts[2].parts,
|
|
|
|
expectedContent: encodeResponse(getAPIErrorResponse(getAPIError(toAPIErrorCode(InvalidPart{})), getGetObjectURL("", bucketName, objectName))),
|
|
|
|
expectedRespStatus: http.StatusBadRequest,
|
|
|
|
},
|
|
|
|
// Test case - 6.
|
|
|
|
// Parts are not sorted according to the part number.
|
|
|
|
// This should return ErrInvalidPartOrder in the response body.
|
|
|
|
{
|
|
|
|
bucket: bucketName,
|
|
|
|
object: objectName,
|
|
|
|
uploadID: uploadIDs[0],
|
|
|
|
parts: inputParts[3].parts,
|
|
|
|
expectedContent: encodeResponse(getAPIErrorResponse(getAPIError(ErrInvalidPartOrder), getGetObjectURL("", bucketName, objectName))),
|
|
|
|
expectedRespStatus: http.StatusBadRequest,
|
|
|
|
},
|
|
|
|
// Test case - 7.
|
|
|
|
// Test case with proper parts.
|
|
|
|
// Should successed and the content in the response body is asserted.
|
|
|
|
{
|
|
|
|
bucket: bucketName,
|
|
|
|
object: objectName,
|
|
|
|
uploadID: uploadIDs[0],
|
|
|
|
parts: inputParts[4].parts,
|
|
|
|
expectedContent: encodedSuccessResponse,
|
|
|
|
expectedRespStatus: http.StatusOK,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
for i, testCase := range testCases {
|
|
|
|
var req *http.Request
|
|
|
|
// Complete multipart upload parts.
|
|
|
|
completeUploads := &completeMultipartUpload{
|
|
|
|
Parts: testCase.parts,
|
|
|
|
}
|
|
|
|
completeBytes, err := xml.Marshal(completeUploads)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Error XML encoding of parts: <ERROR> %s.", err)
|
|
|
|
}
|
|
|
|
// Indicating that all parts are uploaded and initiating completeMultipartUpload.
|
|
|
|
req, err = newTestSignedRequest("POST", getCompleteMultipartUploadURL("", bucketName, objectName, testCase.uploadID),
|
|
|
|
int64(len(completeBytes)), bytes.NewReader(completeBytes), credentials.AccessKeyID, credentials.SecretAccessKey)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Failed to create HTTP request for copy Object: <ERROR> %v", err)
|
|
|
|
}
|
|
|
|
rec := httptest.NewRecorder()
|
|
|
|
// construct HTTP request for copy object.
|
|
|
|
|
|
|
|
// Since `apiRouter` satisfies `http.Handler` it has a ServeHTTP to execute the logic of the handler.
|
|
|
|
// Call the ServeHTTP to executes the registered handler.
|
|
|
|
apiRouter.ServeHTTP(rec, req)
|
|
|
|
// Assert the response code with the expected status.
|
|
|
|
if rec.Code != testCase.expectedRespStatus {
|
|
|
|
t.Errorf("Case %d: Minio %s: Expected the response status to be `%d`, but instead found `%d`", i+1, instanceType, testCase.expectedRespStatus, rec.Code)
|
|
|
|
}
|
|
|
|
|
|
|
|
// read the response body.
|
|
|
|
actualContent, err := ioutil.ReadAll(rec.Body)
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d : Minio %s: Failed parsing response body: <ERROR> %v", i+1, instanceType, err)
|
|
|
|
}
|
|
|
|
// Verify whether the bucket obtained object is same as the one inserted.
|
|
|
|
if !bytes.Equal(testCase.expectedContent, actualContent) {
|
|
|
|
t.Errorf("Test %d : Minio %s: Object content differs from expected value.", i+1, instanceType)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wrapper for calling Delete Object API handler tests for both XL multiple disks and FS single drive setup.
|
|
|
|
func TestAPIDeleteOjectHandler(t *testing.T) {
|
|
|
|
ExecObjectLayerAPITest(t, testAPIDeleteOjectHandler, []string{"DeleteObject"})
|
|
|
|
}
|
|
|
|
|
|
|
|
func testAPIDeleteOjectHandler(obj ObjectLayer, instanceType, bucketName string, apiRouter http.Handler,
|
|
|
|
credentials credential, t TestErrHandler) {
|
|
|
|
|
|
|
|
switch obj.(type) {
|
|
|
|
case fsObjects:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
objectName := "test-object"
|
|
|
|
// set of byte data for PutObject.
|
|
|
|
// object has to be inserted before running tests for Deleting the object.
|
|
|
|
bytesData := []struct {
|
|
|
|
byteData []byte
|
|
|
|
}{
|
|
|
|
{generateBytesData(6 * 1024 * 1024)},
|
|
|
|
}
|
|
|
|
|
|
|
|
// set of inputs for uploading the objects before tests for deleting them is done.
|
|
|
|
putObjectInputs := []struct {
|
|
|
|
bucketName string
|
|
|
|
objectName string
|
|
|
|
contentLength int64
|
|
|
|
textData []byte
|
|
|
|
metaData map[string]string
|
|
|
|
}{
|
|
|
|
// case - 1.
|
|
|
|
{bucketName, objectName, int64(len(bytesData[0].byteData)), bytesData[0].byteData, make(map[string]string)},
|
|
|
|
}
|
|
|
|
// iterate through the above set of inputs and upload the object.
|
|
|
|
for i, input := range putObjectInputs {
|
|
|
|
// uploading the object.
|
|
|
|
_, err := obj.PutObject(input.bucketName, input.objectName, input.contentLength, bytes.NewBuffer(input.textData), input.metaData)
|
|
|
|
// if object upload fails stop the test.
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Put Object case %d: Error uploading object: <ERROR> %v", i+1, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// test cases with inputs and expected result for DeleteObject.
|
|
|
|
testCases := []struct {
|
|
|
|
bucketName string
|
|
|
|
objectName string
|
|
|
|
|
|
|
|
expectedRespStatus int // expected response status body.
|
|
|
|
}{
|
|
|
|
// Test case - 1.
|
|
|
|
// Deleting an existing object.
|
|
|
|
// Expected to return HTTP resposne status code 204.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
|
|
|
|
expectedRespStatus: http.StatusNoContent,
|
|
|
|
},
|
|
|
|
// Test case - 2.
|
|
|
|
// Attempt to delete an object which is already deleted.
|
|
|
|
// Still should return http response status 204.
|
|
|
|
{
|
|
|
|
bucketName: bucketName,
|
|
|
|
objectName: objectName,
|
|
|
|
|
|
|
|
expectedRespStatus: http.StatusNoContent,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
// Iterating over the cases, call DeleteObjectHandler and validate the HTTP response.
|
|
|
|
for i, testCase := range testCases {
|
|
|
|
// initialize HTTP NewRecorder, this records any mutations to response writer inside the handler.
|
|
|
|
rec := httptest.NewRecorder()
|
|
|
|
// construct HTTP request for Get Object end point.
|
|
|
|
req, err := newTestSignedRequest("DELETE", getDeleteObjectURL("", testCase.bucketName, testCase.objectName),
|
|
|
|
0, nil, credentials.AccessKeyID, credentials.SecretAccessKey)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
t.Fatalf("Test %d: Failed to create HTTP request for Get Object: <ERROR> %v", i+1, err)
|
|
|
|
}
|
|
|
|
// Since `apiRouter` satisfies `http.Handler` it has a ServeHTTP to execute the logic of the handler.
|
|
|
|
// Call the ServeHTTP to execute the handler,`func (api objectAPIHandlers) DeleteObjectHandler` handles the request.
|
|
|
|
apiRouter.ServeHTTP(rec, req)
|
|
|
|
// Assert the response code with the expected status.
|
|
|
|
if rec.Code != testCase.expectedRespStatus {
|
|
|
|
t.Fatalf("Minio %s: Case %d: Expected the response status to be `%d`, but instead found `%d`", instanceType, i+1, testCase.expectedRespStatus, rec.Code)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|