enhance getActualSize() to return valid values for most situations (#20228)

This commit is contained in:
Harshavardhana 2024-08-08 08:29:58 -07:00 committed by GitHub
parent 14876a4df1
commit 89c58ce87d
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
12 changed files with 64 additions and 49 deletions

View File

@ -1641,7 +1641,6 @@ func replicateObjectWithMultipart(ctx context.Context, c *minio.Core, bucket, ob
var ( var (
hr *hash.Reader hr *hash.Reader
pInfo minio.ObjectPart
isSSEC = crypto.SSEC.IsEncrypted(objInfo.UserDefined) isSSEC = crypto.SSEC.IsEncrypted(objInfo.UserDefined)
) )
@ -1668,18 +1667,19 @@ func replicateObjectWithMultipart(ctx context.Context, c *minio.Core, bucket, ob
CustomHeader: cHeader, CustomHeader: cHeader,
} }
var size int64
if isSSEC { if isSSEC {
objectSize += partInfo.Size size = partInfo.Size
pInfo, err = c.PutObjectPart(ctx, bucket, object, uploadID, partInfo.Number, hr, partInfo.Size, popts)
} else { } else {
objectSize += partInfo.ActualSize size = partInfo.ActualSize
pInfo, err = c.PutObjectPart(ctx, bucket, object, uploadID, partInfo.Number, hr, partInfo.ActualSize, popts)
} }
objectSize += size
pInfo, err := c.PutObjectPart(ctx, bucket, object, uploadID, partInfo.Number, hr, size, popts)
if err != nil { if err != nil {
return err return err
} }
if !isSSEC && pInfo.Size != partInfo.ActualSize { if pInfo.Size != size {
return fmt.Errorf("Part size mismatch: got %d, want %d", pInfo.Size, partInfo.ActualSize) return fmt.Errorf("ssec(%t): Part size mismatch: got %d, want %d", isSSEC, pInfo.Size, size)
} }
uploadedParts = append(uploadedParts, minio.CompletePart{ uploadedParts = append(uploadedParts, minio.CompletePart{
PartNumber: pInfo.PartNumber, PartNumber: pInfo.PartNumber,
@ -1691,7 +1691,7 @@ func replicateObjectWithMultipart(ctx context.Context, c *minio.Core, bucket, ob
}) })
} }
userMeta := map[string]string{ userMeta := map[string]string{
validSSEReplicationHeaders[ReservedMetadataPrefix+"Actual-Object-Size"]: objInfo.UserDefined[ReservedMetadataPrefix+"actual-size"], xhttp.MinIOReplicationActualObjectSize: objInfo.UserDefined[ReservedMetadataPrefix+"actual-size"],
} }
if isSSEC && objInfo.UserDefined[ReplicationSsecChecksumHeader] != "" { if isSSEC && objInfo.UserDefined[ReplicationSsecChecksumHeader] != "" {
userMeta[ReplicationSsecChecksumHeader] = objInfo.UserDefined[ReplicationSsecChecksumHeader] userMeta[ReplicationSsecChecksumHeader] = objInfo.UserDefined[ReplicationSsecChecksumHeader]

View File

@ -743,8 +743,9 @@ func (d *DecryptBlocksReader) Read(p []byte) (int, error) {
// but has an invalid size. // but has an invalid size.
func (o ObjectInfo) DecryptedSize() (int64, error) { func (o ObjectInfo) DecryptedSize() (int64, error) {
if _, ok := crypto.IsEncrypted(o.UserDefined); !ok { if _, ok := crypto.IsEncrypted(o.UserDefined); !ok {
return 0, errors.New("Cannot compute decrypted size of an unencrypted object") return -1, errors.New("Cannot compute decrypted size of an unencrypted object")
} }
if !o.isMultipart() { if !o.isMultipart() {
size, err := sio.DecryptedSize(uint64(o.Size)) size, err := sio.DecryptedSize(uint64(o.Size))
if err != nil { if err != nil {
@ -757,7 +758,7 @@ func (o ObjectInfo) DecryptedSize() (int64, error) {
for _, part := range o.Parts { for _, part := range o.Parts {
partSize, err := sio.DecryptedSize(uint64(part.Size)) partSize, err := sio.DecryptedSize(uint64(part.Size))
if err != nil { if err != nil {
return 0, errObjectTampered return -1, errObjectTampered
} }
size += int64(partSize) size += int64(partSize)
} }

View File

@ -1301,7 +1301,9 @@ func (er erasureObjects) CompleteMultipartUpload(ctx context.Context, bucket str
// Save the consolidated actual size. // Save the consolidated actual size.
if opts.ReplicationRequest { if opts.ReplicationRequest {
fi.Metadata[ReservedMetadataPrefix+"actual-size"] = opts.UserDefined["X-Minio-Internal-Actual-Object-Size"] if v := opts.UserDefined[ReservedMetadataPrefix+"Actual-Object-Size"]; v != "" {
fi.Metadata[ReservedMetadataPrefix+"actual-size"] = v
}
} else { } else {
fi.Metadata[ReservedMetadataPrefix+"actual-size"] = strconv.FormatInt(objectActualSize, 10) fi.Metadata[ReservedMetadataPrefix+"actual-size"] = strconv.FormatInt(objectActualSize, 10)
} }

View File

@ -97,7 +97,7 @@ func (er erasureObjects) CopyObject(ctx context.Context, srcBucket, srcObject, d
if srcOpts.VersionID != "" { if srcOpts.VersionID != "" {
metaArr, errs = readAllFileInfo(ctx, storageDisks, "", srcBucket, srcObject, srcOpts.VersionID, true, false) metaArr, errs = readAllFileInfo(ctx, storageDisks, "", srcBucket, srcObject, srcOpts.VersionID, true, false)
} else { } else {
metaArr, errs = readAllXL(ctx, storageDisks, srcBucket, srcObject, true, false, true) metaArr, errs = readAllXL(ctx, storageDisks, srcBucket, srcObject, true, false)
} }
readQuorum, writeQuorum, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount) readQuorum, writeQuorum, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount)
@ -575,11 +575,10 @@ func (er erasureObjects) deleteIfDangling(ctx context.Context, bucket, object st
return m, nil return m, nil
} }
func fileInfoFromRaw(ri RawFileInfo, bucket, object string, readData, inclFreeVers, allParts bool) (FileInfo, error) { func fileInfoFromRaw(ri RawFileInfo, bucket, object string, readData, inclFreeVers bool) (FileInfo, error) {
return getFileInfo(ri.Buf, bucket, object, "", fileInfoOpts{ return getFileInfo(ri.Buf, bucket, object, "", fileInfoOpts{
Data: readData, Data: readData,
InclFreeVersions: inclFreeVers, InclFreeVersions: inclFreeVers,
AllParts: allParts,
}) })
} }
@ -604,7 +603,7 @@ func readAllRawFileInfo(ctx context.Context, disks []StorageAPI, bucket, object
return rawFileInfos, g.Wait() return rawFileInfos, g.Wait()
} }
func pickLatestQuorumFilesInfo(ctx context.Context, rawFileInfos []RawFileInfo, errs []error, bucket, object string, readData, inclFreeVers, allParts bool) ([]FileInfo, []error) { func pickLatestQuorumFilesInfo(ctx context.Context, rawFileInfos []RawFileInfo, errs []error, bucket, object string, readData, inclFreeVers bool) ([]FileInfo, []error) {
metadataArray := make([]*xlMetaV2, len(rawFileInfos)) metadataArray := make([]*xlMetaV2, len(rawFileInfos))
metaFileInfos := make([]FileInfo, len(rawFileInfos)) metaFileInfos := make([]FileInfo, len(rawFileInfos))
metadataShallowVersions := make([][]xlMetaV2ShallowVersion, len(rawFileInfos)) metadataShallowVersions := make([][]xlMetaV2ShallowVersion, len(rawFileInfos))
@ -641,7 +640,7 @@ func pickLatestQuorumFilesInfo(ctx context.Context, rawFileInfos []RawFileInfo,
readQuorum := (len(rawFileInfos) + 1) / 2 readQuorum := (len(rawFileInfos) + 1) / 2
meta := &xlMetaV2{versions: mergeXLV2Versions(readQuorum, false, 1, metadataShallowVersions...)} meta := &xlMetaV2{versions: mergeXLV2Versions(readQuorum, false, 1, metadataShallowVersions...)}
lfi, err := meta.ToFileInfo(bucket, object, "", inclFreeVers, allParts) lfi, err := meta.ToFileInfo(bucket, object, "", inclFreeVers, true)
if err != nil { if err != nil {
for i := range errs { for i := range errs {
if errs[i] == nil { if errs[i] == nil {
@ -670,7 +669,7 @@ func pickLatestQuorumFilesInfo(ctx context.Context, rawFileInfos []RawFileInfo,
} }
// make sure to preserve this for diskmtime based healing bugfix. // make sure to preserve this for diskmtime based healing bugfix.
metaFileInfos[index], errs[index] = metadataArray[index].ToFileInfo(bucket, object, versionID, inclFreeVers, allParts) metaFileInfos[index], errs[index] = metadataArray[index].ToFileInfo(bucket, object, versionID, inclFreeVers, true)
if errs[index] != nil { if errs[index] != nil {
continue continue
} }
@ -715,9 +714,9 @@ func shouldCheckForDangling(err error, errs []error, bucket string) bool {
return false return false
} }
func readAllXL(ctx context.Context, disks []StorageAPI, bucket, object string, readData, inclFreeVers, allParts bool) ([]FileInfo, []error) { func readAllXL(ctx context.Context, disks []StorageAPI, bucket, object string, readData, inclFreeVers bool) ([]FileInfo, []error) {
rawFileInfos, errs := readAllRawFileInfo(ctx, disks, bucket, object, readData) rawFileInfos, errs := readAllRawFileInfo(ctx, disks, bucket, object, readData)
return pickLatestQuorumFilesInfo(ctx, rawFileInfos, errs, bucket, object, readData, inclFreeVers, allParts) return pickLatestQuorumFilesInfo(ctx, rawFileInfos, errs, bucket, object, readData, inclFreeVers)
} }
func (er erasureObjects) getObjectFileInfo(ctx context.Context, bucket, object string, opts ObjectOptions, readData bool) (FileInfo, []FileInfo, []StorageAPI, error) { func (er erasureObjects) getObjectFileInfo(ctx context.Context, bucket, object string, opts ObjectOptions, readData bool) (FileInfo, []FileInfo, []StorageAPI, error) {
@ -774,7 +773,7 @@ func (er erasureObjects) getObjectFileInfo(ctx context.Context, bucket, object s
// Read the latest version // Read the latest version
rfi, err = disk.ReadXL(ctx, bucket, object, readData) rfi, err = disk.ReadXL(ctx, bucket, object, readData)
if err == nil { if err == nil {
fi, err = fileInfoFromRaw(rfi, bucket, object, readData, opts.InclFreeVersions, true) fi, err = fileInfoFromRaw(rfi, bucket, object, readData, opts.InclFreeVersions)
} }
} }
@ -912,7 +911,7 @@ func (er erasureObjects) getObjectFileInfo(ctx context.Context, bucket, object s
// we must use rawFileInfo to resolve versions to figure out the latest version. // we must use rawFileInfo to resolve versions to figure out the latest version.
if opts.VersionID == "" && totalResp == er.setDriveCount { if opts.VersionID == "" && totalResp == er.setDriveCount {
fi, onlineMeta, onlineDisks, modTime, etag, err = calcQuorum(pickLatestQuorumFilesInfo(ctx, fi, onlineMeta, onlineDisks, modTime, etag, err = calcQuorum(pickLatestQuorumFilesInfo(ctx,
rawArr, errs, bucket, object, readData, opts.InclFreeVersions, true)) rawArr, errs, bucket, object, readData, opts.InclFreeVersions))
} else { } else {
fi, onlineMeta, onlineDisks, modTime, etag, err = calcQuorum(metaArr, errs) fi, onlineMeta, onlineDisks, modTime, etag, err = calcQuorum(metaArr, errs)
} }
@ -2142,7 +2141,7 @@ func (er erasureObjects) PutObjectMetadata(ctx context.Context, bucket, object s
if opts.VersionID != "" { if opts.VersionID != "" {
metaArr, errs = readAllFileInfo(ctx, disks, "", bucket, object, opts.VersionID, false, false) metaArr, errs = readAllFileInfo(ctx, disks, "", bucket, object, opts.VersionID, false, false)
} else { } else {
metaArr, errs = readAllXL(ctx, disks, bucket, object, false, false, true) metaArr, errs = readAllXL(ctx, disks, bucket, object, false, false)
} }
readQuorum, _, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount) readQuorum, _, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount)
@ -2219,7 +2218,7 @@ func (er erasureObjects) PutObjectTags(ctx context.Context, bucket, object strin
if opts.VersionID != "" { if opts.VersionID != "" {
metaArr, errs = readAllFileInfo(ctx, disks, "", bucket, object, opts.VersionID, false, false) metaArr, errs = readAllFileInfo(ctx, disks, "", bucket, object, opts.VersionID, false, false)
} else { } else {
metaArr, errs = readAllXL(ctx, disks, bucket, object, false, false, true) metaArr, errs = readAllXL(ctx, disks, bucket, object, false, false)
} }
readQuorum, _, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount) readQuorum, _, err := objectQuorumFromMeta(ctx, metaArr, errs, er.defaultParityCount)

View File

@ -257,7 +257,7 @@ func (e *metaCacheEntry) fileInfo(bucket string) (FileInfo, error) {
ModTime: timeSentinel1970, ModTime: timeSentinel1970,
}, nil }, nil
} }
return e.cached.ToFileInfo(bucket, e.name, "", false, false) return e.cached.ToFileInfo(bucket, e.name, "", false, true)
} }
return getFileInfo(e.metadata, bucket, e.name, "", fileInfoOpts{}) return getFileInfo(e.metadata, bucket, e.name, "", fileInfoOpts{})
} }
@ -300,7 +300,7 @@ func (e *metaCacheEntry) fileInfoVersions(bucket string) (FileInfoVersions, erro
}, nil }, nil
} }
// Too small gains to reuse cache here. // Too small gains to reuse cache here.
return getFileInfoVersions(e.metadata, bucket, e.name, false, true) return getFileInfoVersions(e.metadata, bucket, e.name, true)
} }
// metaCacheEntries is a slice of metacache entries. // metaCacheEntries is a slice of metacache entries.

View File

@ -482,7 +482,6 @@ func completeMultipartOpts(ctx context.Context, r *http.Request, bucket, object
} }
opts.MTime = mtime opts.MTime = mtime
opts.UserDefined = make(map[string]string) opts.UserDefined = make(map[string]string)
opts.UserDefined[ReservedMetadataPrefix+"Actual-Object-Size"] = r.Header.Get(xhttp.MinIOReplicationActualObjectSize)
// Transfer SSEC key in opts.EncryptFn // Transfer SSEC key in opts.EncryptFn
if crypto.SSEC.IsRequested(r.Header) { if crypto.SSEC.IsRequested(r.Header) {
key, err := ParseSSECustomerRequest(r) key, err := ParseSSECustomerRequest(r)
@ -495,6 +494,7 @@ func completeMultipartOpts(ctx context.Context, r *http.Request, bucket, object
} }
if _, ok := r.Header[xhttp.MinIOSourceReplicationRequest]; ok { if _, ok := r.Header[xhttp.MinIOSourceReplicationRequest]; ok {
opts.ReplicationRequest = true opts.ReplicationRequest = true
opts.UserDefined[ReservedMetadataPrefix+"Actual-Object-Size"] = r.Header.Get(xhttp.MinIOReplicationActualObjectSize)
} }
if r.Header.Get(ReplicationSsecChecksumHeader) != "" { if r.Header.Get(ReplicationSsecChecksumHeader) != "" {
opts.UserDefined[ReplicationSsecChecksumHeader] = r.Header.Get(ReplicationSsecChecksumHeader) opts.UserDefined[ReplicationSsecChecksumHeader] = r.Header.Get(ReplicationSsecChecksumHeader)

View File

@ -1,4 +1,4 @@
// Copyright (c) 2015-2021 MinIO, Inc. // Copyright (c) 2015-2024 MinIO, Inc.
// //
// This file is part of MinIO Object Storage stack // This file is part of MinIO Object Storage stack
// //
@ -554,28 +554,41 @@ func (o ObjectInfo) GetActualSize() (int64, error) {
return *o.ActualSize, nil return *o.ActualSize, nil
} }
if o.IsCompressed() { if o.IsCompressed() {
sizeStr, ok := o.UserDefined[ReservedMetadataPrefix+"actual-size"] sizeStr := o.UserDefined[ReservedMetadataPrefix+"actual-size"]
if !ok { if sizeStr != "" {
return -1, errInvalidDecompressedSize
}
size, err := strconv.ParseInt(sizeStr, 10, 64) size, err := strconv.ParseInt(sizeStr, 10, 64)
if err != nil { if err != nil {
return -1, errInvalidDecompressedSize return -1, errInvalidDecompressedSize
} }
return size, nil return size, nil
} }
var actualSize int64
for _, part := range o.Parts {
actualSize += part.ActualSize
}
if (actualSize == 0) && (actualSize != o.Size) {
return -1, errInvalidDecompressedSize
}
return actualSize, nil
}
if _, ok := crypto.IsEncrypted(o.UserDefined); ok { if _, ok := crypto.IsEncrypted(o.UserDefined); ok {
sizeStr, ok := o.UserDefined[ReservedMetadataPrefix+"actual-size"] sizeStr := o.UserDefined[ReservedMetadataPrefix+"actual-size"]
if ok { if sizeStr != "" {
size, err := strconv.ParseInt(sizeStr, 10, 64) size, err := strconv.ParseInt(sizeStr, 10, 64)
if err != nil { if err != nil {
return -1, errObjectTampered return -1, errObjectTampered
} }
return size, nil return size, nil
} }
return o.DecryptedSize() actualSize, err := o.DecryptedSize()
if err != nil {
return -1, err
}
if (actualSize == 0) && (actualSize != o.Size) {
return -1, errObjectTampered
}
return actualSize, nil
} }
return o.Size, nil return o.Size, nil
} }

View File

@ -609,7 +609,6 @@ func TestGetActualSize(t *testing.T) {
objInfo: ObjectInfo{ objInfo: ObjectInfo{
UserDefined: map[string]string{ UserDefined: map[string]string{
"X-Minio-Internal-compression": "klauspost/compress/s2", "X-Minio-Internal-compression": "klauspost/compress/s2",
"X-Minio-Internal-actual-size": "100000001",
"content-type": "application/octet-stream", "content-type": "application/octet-stream",
"etag": "b3ff3ef3789147152fbfbc50efba4bfd-2", "etag": "b3ff3ef3789147152fbfbc50efba4bfd-2",
}, },
@ -623,6 +622,7 @@ func TestGetActualSize(t *testing.T) {
ActualSize: 32891137, ActualSize: 32891137,
}, },
}, },
Size: 100000001,
}, },
result: 100000001, result: 100000001,
}, },
@ -635,6 +635,7 @@ func TestGetActualSize(t *testing.T) {
"etag": "b3ff3ef3789147152fbfbc50efba4bfd-2", "etag": "b3ff3ef3789147152fbfbc50efba4bfd-2",
}, },
Parts: []ObjectPartInfo{}, Parts: []ObjectPartInfo{},
Size: 841,
}, },
result: 841, result: 841,
}, },
@ -646,6 +647,7 @@ func TestGetActualSize(t *testing.T) {
"etag": "b3ff3ef3789147152fbfbc50efba4bfd-2", "etag": "b3ff3ef3789147152fbfbc50efba4bfd-2",
}, },
Parts: []ObjectPartInfo{}, Parts: []ObjectPartInfo{},
Size: 100,
}, },
result: -1, result: -1,
}, },

View File

@ -30,8 +30,8 @@ import (
// if inclFreeVersions is true all the versions are in fivs.Versions, free and non-free versions alike. // if inclFreeVersions is true all the versions are in fivs.Versions, free and non-free versions alike.
// //
// Note: Only the scanner requires fivs.Versions to have exclusively non-free versions. This is used while enforcing NewerNoncurrentVersions lifecycle element. // Note: Only the scanner requires fivs.Versions to have exclusively non-free versions. This is used while enforcing NewerNoncurrentVersions lifecycle element.
func getFileInfoVersions(xlMetaBuf []byte, volume, path string, allParts, inclFreeVersions bool) (FileInfoVersions, error) { func getFileInfoVersions(xlMetaBuf []byte, volume, path string, inclFreeVersions bool) (FileInfoVersions, error) {
fivs, err := getAllFileInfoVersions(xlMetaBuf, volume, path, allParts) fivs, err := getAllFileInfoVersions(xlMetaBuf, volume, path, true)
if err != nil { if err != nil {
return fivs, err return fivs, err
} }
@ -106,7 +106,6 @@ func getAllFileInfoVersions(xlMetaBuf []byte, volume, path string, allParts bool
type fileInfoOpts struct { type fileInfoOpts struct {
InclFreeVersions bool InclFreeVersions bool
Data bool Data bool
AllParts bool
} }
func getFileInfo(xlMetaBuf []byte, volume, path, versionID string, opts fileInfoOpts) (FileInfo, error) { func getFileInfo(xlMetaBuf []byte, volume, path, versionID string, opts fileInfoOpts) (FileInfo, error) {
@ -117,7 +116,7 @@ func getFileInfo(xlMetaBuf []byte, volume, path, versionID string, opts fileInfo
return FileInfo{}, e return FileInfo{}, e
} else if buf != nil { } else if buf != nil {
inData = data inData = data
fi, err = buf.ToFileInfo(volume, path, versionID, opts.AllParts) fi, err = buf.ToFileInfo(volume, path, versionID, true)
if len(buf) != 0 && errors.Is(err, errFileNotFound) { if len(buf) != 0 && errors.Is(err, errFileNotFound) {
// This special case is needed to handle len(xlMeta.versions) == 0 // This special case is needed to handle len(xlMeta.versions) == 0
return FileInfo{ return FileInfo{
@ -146,7 +145,7 @@ func getFileInfo(xlMetaBuf []byte, volume, path, versionID string, opts fileInfo
}, nil }, nil
} }
inData = xlMeta.data inData = xlMeta.data
fi, err = xlMeta.ToFileInfo(volume, path, versionID, opts.InclFreeVersions, opts.AllParts) fi, err = xlMeta.ToFileInfo(volume, path, versionID, opts.InclFreeVersions, true)
} }
if !opts.Data || err != nil { if !opts.Data || err != nil {
return fi, err return fi, err

View File

@ -178,7 +178,7 @@ func TestGetFileInfoVersions(t *testing.T) {
if err != nil { if err != nil {
t.Fatalf("Failed to serialize xlmeta %v", err) t.Fatalf("Failed to serialize xlmeta %v", err)
} }
fivs, err := getFileInfoVersions(buf, basefi.Volume, basefi.Name, true, false) fivs, err := getFileInfoVersions(buf, basefi.Volume, basefi.Name, false)
if err != nil { if err != nil {
t.Fatalf("getFileInfoVersions failed: %v", err) t.Fatalf("getFileInfoVersions failed: %v", err)
} }
@ -222,7 +222,7 @@ func TestGetFileInfoVersions(t *testing.T) {
// versions are stored in xl-meta sorted in descending order of their ModTime // versions are stored in xl-meta sorted in descending order of their ModTime
slices.Reverse(allVersionIDs) slices.Reverse(allVersionIDs)
fivs, err = getFileInfoVersions(buf, basefi.Volume, basefi.Name, true, true) fivs, err = getFileInfoVersions(buf, basefi.Volume, basefi.Name, true)
if err != nil { if err != nil {
t.Fatalf("getFileInfoVersions failed: %v", err) t.Fatalf("getFileInfoVersions failed: %v", err)
} }

View File

@ -590,7 +590,7 @@ func (s *xlStorage) NSScanner(ctx context.Context, cache dataUsageCache, updates
// Remove filename which is the meta file. // Remove filename which is the meta file.
item.transformMetaDir() item.transformMetaDir()
fivs, err := getFileInfoVersions(buf, item.bucket, item.objectPath(), false, false) fivs, err := getFileInfoVersions(buf, item.bucket, item.objectPath(), false)
metaDataPoolPut(buf) metaDataPoolPut(buf)
if err != nil { if err != nil {
res["err"] = err.Error() res["err"] = err.Error()
@ -1671,7 +1671,6 @@ func (s *xlStorage) ReadVersion(ctx context.Context, origvolume, volume, path, v
fi, err = getFileInfo(buf, volume, path, versionID, fileInfoOpts{ fi, err = getFileInfo(buf, volume, path, versionID, fileInfoOpts{
Data: opts.ReadData, Data: opts.ReadData,
InclFreeVersions: opts.InclFreeVersions, InclFreeVersions: opts.InclFreeVersions,
AllParts: true,
}) })
if err != nil { if err != nil {
return fi, err return fi, err

View File

@ -271,7 +271,7 @@ func TestXLStorageReadVersion(t *testing.T) {
} }
xlMeta, _ := os.ReadFile("testdata/xl.meta") xlMeta, _ := os.ReadFile("testdata/xl.meta")
fi, _ := getFileInfo(xlMeta, "exists", "as-file", "", fileInfoOpts{Data: false, AllParts: true}) fi, _ := getFileInfo(xlMeta, "exists", "as-file", "", fileInfoOpts{Data: false})
// Create files for the test cases. // Create files for the test cases.
if err = xlStorage.MakeVol(context.Background(), "exists"); err != nil { if err = xlStorage.MakeVol(context.Background(), "exists"); err != nil {