2021-04-18 15:41:13 -04:00
|
|
|
// Copyright (c) 2015-2021 MinIO, Inc.
|
|
|
|
//
|
|
|
|
// This file is part of MinIO Object Storage stack
|
|
|
|
//
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This program is distributed in the hope that it will be useful
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Affero General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
2016-10-17 05:10:23 -04:00
|
|
|
|
|
|
|
package cmd
|
|
|
|
|
2017-03-04 17:53:28 -05:00
|
|
|
import (
|
2021-03-29 20:00:55 -04:00
|
|
|
"bytes"
|
2018-04-05 18:04:40 -04:00
|
|
|
"context"
|
2024-06-10 11:51:27 -04:00
|
|
|
"slices"
|
2017-03-04 17:53:28 -05:00
|
|
|
"time"
|
2017-11-25 14:58:29 -05:00
|
|
|
|
2023-06-19 20:53:08 -04:00
|
|
|
"github.com/minio/madmin-go/v3"
|
2017-03-04 17:53:28 -05:00
|
|
|
)
|
2016-10-17 05:10:23 -04:00
|
|
|
|
2023-06-17 22:18:20 -04:00
|
|
|
func commonETags(etags []string) (etag string, maxima int) {
|
|
|
|
etagOccurrenceMap := make(map[string]int, len(etags))
|
|
|
|
|
|
|
|
// Ignore the uuid sentinel and count the rest.
|
|
|
|
for _, etag := range etags {
|
|
|
|
if etag == "" {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
etagOccurrenceMap[etag]++
|
|
|
|
}
|
|
|
|
|
|
|
|
maxima = 0 // Counter for remembering max occurrence of elements.
|
|
|
|
latest := ""
|
|
|
|
|
|
|
|
// Find the common cardinality from previously collected
|
|
|
|
// occurrences of elements.
|
|
|
|
for etag, count := range etagOccurrenceMap {
|
|
|
|
if count < maxima {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// We are at or above maxima
|
|
|
|
if count > maxima {
|
|
|
|
maxima = count
|
|
|
|
latest = etag
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return the collected common max time, with maxima
|
|
|
|
return latest, maxima
|
|
|
|
}
|
|
|
|
|
2016-10-17 05:10:23 -04:00
|
|
|
// commonTime returns a maximally occurring time from a list of time.
|
2024-02-22 01:26:06 -05:00
|
|
|
func commonTimeAndOccurrence(times []time.Time, group time.Duration) (maxTime time.Time, maxima int) {
|
|
|
|
timeOccurrenceMap := make(map[int64]int, len(times))
|
2021-12-21 13:08:26 -05:00
|
|
|
groupNano := group.Nanoseconds()
|
2016-10-17 05:10:23 -04:00
|
|
|
// Ignore the uuid sentinel and count the rest.
|
2021-12-21 13:08:26 -05:00
|
|
|
for _, t := range times {
|
2022-04-20 15:49:05 -04:00
|
|
|
if t.Equal(timeSentinel) || t.IsZero() {
|
2016-10-17 05:10:23 -04:00
|
|
|
continue
|
|
|
|
}
|
2021-12-21 13:08:26 -05:00
|
|
|
nano := t.UnixNano()
|
|
|
|
if group > 0 {
|
2024-02-22 01:26:06 -05:00
|
|
|
for k := range timeOccurrenceMap {
|
2021-12-21 13:08:26 -05:00
|
|
|
if k == nano {
|
|
|
|
// We add to ourself later
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
diff := k - nano
|
|
|
|
if diff < 0 {
|
|
|
|
diff = -diff
|
|
|
|
}
|
|
|
|
// We are within the limit
|
|
|
|
if diff < groupNano {
|
2024-02-22 01:26:06 -05:00
|
|
|
timeOccurrenceMap[k]++
|
2021-12-21 13:08:26 -05:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Add ourself...
|
2024-02-22 01:26:06 -05:00
|
|
|
timeOccurrenceMap[nano]++
|
2016-10-17 05:10:23 -04:00
|
|
|
}
|
2020-11-02 20:52:13 -05:00
|
|
|
|
2021-12-21 13:08:26 -05:00
|
|
|
maxima = 0 // Counter for remembering max occurrence of elements.
|
|
|
|
latest := int64(0)
|
2021-11-21 13:41:30 -05:00
|
|
|
|
2016-10-17 05:10:23 -04:00
|
|
|
// Find the common cardinality from previously collected
|
|
|
|
// occurrences of elements.
|
2024-02-22 01:26:06 -05:00
|
|
|
for nano, count := range timeOccurrenceMap {
|
2021-12-21 13:08:26 -05:00
|
|
|
if count < maxima {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// We are at or above maxima
|
|
|
|
if count > maxima || nano > latest {
|
2016-10-17 05:10:23 -04:00
|
|
|
maxima = count
|
2021-12-21 13:08:26 -05:00
|
|
|
latest = nano
|
2016-10-17 05:10:23 -04:00
|
|
|
}
|
|
|
|
}
|
2020-11-02 20:52:13 -05:00
|
|
|
|
2021-12-21 13:08:26 -05:00
|
|
|
// Return the collected common max time, with maxima
|
|
|
|
return time.Unix(0, latest).UTC(), maxima
|
|
|
|
}
|
|
|
|
|
2023-04-25 13:13:57 -04:00
|
|
|
// commonTime returns a maximally occurring time from a list of time if it
|
|
|
|
// occurs >= quorum, else return timeSentinel
|
|
|
|
func commonTime(modTimes []time.Time, quorum int) time.Time {
|
2024-02-22 01:26:06 -05:00
|
|
|
if modTime, count := commonTimeAndOccurrence(modTimes, 0); count >= quorum {
|
2023-04-25 13:13:57 -04:00
|
|
|
return modTime
|
|
|
|
}
|
|
|
|
|
|
|
|
return timeSentinel
|
2016-10-17 05:10:23 -04:00
|
|
|
}
|
|
|
|
|
2023-06-17 22:18:20 -04:00
|
|
|
func commonETag(etags []string, quorum int) string {
|
|
|
|
if etag, count := commonETags(etags); count >= quorum {
|
|
|
|
return etag
|
|
|
|
}
|
|
|
|
return ""
|
|
|
|
}
|
|
|
|
|
2016-10-17 05:10:23 -04:00
|
|
|
// Beginning of unix time is treated as sentinel value here.
|
2022-03-04 23:01:26 -05:00
|
|
|
var (
|
|
|
|
timeSentinel = time.Unix(0, 0).UTC()
|
|
|
|
timeSentinel1970 = time.Unix(0, 1).UTC() // 1970 used for special cases when xlmeta.version == 0
|
|
|
|
)
|
2016-10-17 05:10:23 -04:00
|
|
|
|
|
|
|
// Boot modTimes up to disk count, setting the value to time sentinel.
|
|
|
|
func bootModtimes(diskCount int) []time.Time {
|
|
|
|
modTimes := make([]time.Time, diskCount)
|
|
|
|
// Boots up all the modtimes.
|
|
|
|
for i := range modTimes {
|
|
|
|
modTimes[i] = timeSentinel
|
|
|
|
}
|
|
|
|
return modTimes
|
|
|
|
}
|
|
|
|
|
2023-06-17 22:18:20 -04:00
|
|
|
func listObjectETags(partsMetadata []FileInfo, errs []error, quorum int) (etags []string) {
|
|
|
|
etags = make([]string, len(partsMetadata))
|
|
|
|
vidMap := map[string]int{}
|
|
|
|
for index, metadata := range partsMetadata {
|
|
|
|
if errs[index] != nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
vid := metadata.VersionID
|
|
|
|
if metadata.VersionID == "" {
|
|
|
|
vid = nullVersionID
|
|
|
|
}
|
|
|
|
vidMap[vid]++
|
|
|
|
etags[index] = metadata.Metadata["etag"]
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, count := range vidMap {
|
|
|
|
// do we have enough common versions
|
|
|
|
// that have enough quorum to satisfy
|
|
|
|
// the etag.
|
|
|
|
if count >= quorum {
|
|
|
|
return etags
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return make([]string, len(partsMetadata))
|
|
|
|
}
|
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
// Extracts list of times from FileInfo slice and returns, skips
|
2017-03-04 17:53:28 -05:00
|
|
|
// slice elements which have errors.
|
2020-06-12 23:04:01 -04:00
|
|
|
func listObjectModtimes(partsMetadata []FileInfo, errs []error) (modTimes []time.Time) {
|
2016-10-17 05:10:23 -04:00
|
|
|
modTimes = bootModtimes(len(partsMetadata))
|
|
|
|
for index, metadata := range partsMetadata {
|
2017-03-04 17:53:28 -05:00
|
|
|
if errs[index] != nil {
|
|
|
|
continue
|
2016-10-17 05:10:23 -04:00
|
|
|
}
|
2017-03-04 17:53:28 -05:00
|
|
|
// Once the file is found, save the uuid saved on disk.
|
2020-06-12 23:04:01 -04:00
|
|
|
modTimes[index] = metadata.ModTime
|
2016-10-17 05:10:23 -04:00
|
|
|
}
|
|
|
|
return modTimes
|
|
|
|
}
|
|
|
|
|
2021-11-21 13:41:30 -05:00
|
|
|
func filterOnlineDisksInplace(fi FileInfo, partsMetadata []FileInfo, onlineDisks []StorageAPI) {
|
|
|
|
for i, meta := range partsMetadata {
|
|
|
|
if fi.XLV1 == meta.XLV1 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
onlineDisks[i] = nil
|
|
|
|
}
|
|
|
|
}
|
2021-10-04 11:53:38 -04:00
|
|
|
|
2017-03-04 17:53:28 -05:00
|
|
|
// Notes:
|
|
|
|
// There are 5 possible states a disk could be in,
|
2020-06-12 23:04:01 -04:00
|
|
|
// 1. __online__ - has the latest copy of xl.meta - returned by listOnlineDisks
|
2017-03-04 17:53:28 -05:00
|
|
|
//
|
|
|
|
// 2. __offline__ - err == errDiskNotFound
|
|
|
|
//
|
2020-06-12 23:04:01 -04:00
|
|
|
// 3. __availableWithParts__ - has the latest copy of xl.meta and has all
|
2017-03-04 17:53:28 -05:00
|
|
|
// parts with checksums matching; returned by disksWithAllParts
|
|
|
|
//
|
|
|
|
// 4. __outdated__ - returned by outDatedDisk, provided []StorageAPI
|
|
|
|
// returned by diskWithAllParts is passed for latestDisks.
|
2020-06-12 23:04:01 -04:00
|
|
|
// - has an old copy of xl.meta
|
|
|
|
// - doesn't have xl.meta (errFileNotFound)
|
|
|
|
// - has the latest xl.meta but one or more parts are corrupt
|
2017-03-04 17:53:28 -05:00
|
|
|
//
|
2020-06-12 23:04:01 -04:00
|
|
|
// 5. __missingParts__ - has the latest copy of xl.meta but has some parts
|
2017-03-04 17:53:28 -05:00
|
|
|
// missing. This is identified separately since this may need manual
|
|
|
|
// inspection to understand the root cause. E.g, this could be due to
|
|
|
|
// backend filesystem corruption.
|
|
|
|
|
|
|
|
// listOnlineDisks - returns
|
2020-06-12 23:04:01 -04:00
|
|
|
// - a slice of disks where disk having 'older' xl.meta (or nothing)
|
2017-03-04 17:53:28 -05:00
|
|
|
// are set to nil.
|
2023-04-25 13:13:57 -04:00
|
|
|
// - latest (in time) of the maximally occurring modTime(s), which has at least quorum occurrences.
|
2023-06-17 22:18:20 -04:00
|
|
|
func listOnlineDisks(disks []StorageAPI, partsMetadata []FileInfo, errs []error, quorum int) (onlineDisks []StorageAPI, modTime time.Time, etag string) {
|
2016-10-17 05:10:23 -04:00
|
|
|
onlineDisks = make([]StorageAPI, len(disks))
|
|
|
|
|
|
|
|
// List all the file commit ids from parts metadata.
|
|
|
|
modTimes := listObjectModtimes(partsMetadata, errs)
|
|
|
|
|
|
|
|
// Reduce list of UUIDs to a single common value.
|
2023-04-25 13:13:57 -04:00
|
|
|
modTime = commonTime(modTimes, quorum)
|
2016-10-17 05:10:23 -04:00
|
|
|
|
2023-06-17 22:18:20 -04:00
|
|
|
if modTime.IsZero() || modTime.Equal(timeSentinel) {
|
|
|
|
etags := listObjectETags(partsMetadata, errs, quorum)
|
|
|
|
|
|
|
|
etag = commonETag(etags, quorum)
|
|
|
|
|
|
|
|
if etag != "" { // allow this fallback only if a non-empty etag is found.
|
|
|
|
for index, e := range etags {
|
|
|
|
if partsMetadata[index].IsValid() && e == etag {
|
|
|
|
onlineDisks[index] = disks[index]
|
|
|
|
} else {
|
|
|
|
onlineDisks[index] = nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return onlineDisks, modTime, etag
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-10-17 05:10:23 -04:00
|
|
|
// Create a new online disks slice, which have common uuid.
|
|
|
|
for index, t := range modTimes {
|
2021-11-21 13:41:30 -05:00
|
|
|
if partsMetadata[index].IsValid() && t.Equal(modTime) {
|
2016-10-17 05:10:23 -04:00
|
|
|
onlineDisks[index] = disks[index]
|
|
|
|
} else {
|
|
|
|
onlineDisks[index] = nil
|
|
|
|
}
|
|
|
|
}
|
2021-04-21 22:06:08 -04:00
|
|
|
|
2023-06-17 22:18:20 -04:00
|
|
|
return onlineDisks, modTime, ""
|
2021-11-15 12:46:55 -05:00
|
|
|
}
|
|
|
|
|
2024-06-10 11:51:27 -04:00
|
|
|
// Convert verify or check parts returned error to integer representation
|
|
|
|
func convPartErrToInt(err error) int {
|
|
|
|
err = unwrapAll(err)
|
|
|
|
switch err {
|
|
|
|
case nil:
|
|
|
|
return checkPartSuccess
|
|
|
|
case errFileNotFound, errFileVersionNotFound:
|
|
|
|
return checkPartFileNotFound
|
|
|
|
case errFileCorrupt:
|
|
|
|
return checkPartFileCorrupt
|
|
|
|
case errVolumeNotFound:
|
|
|
|
return checkPartVolumeNotFound
|
|
|
|
case errDiskNotFound:
|
|
|
|
return checkPartDiskNotFound
|
|
|
|
default:
|
|
|
|
return checkPartUnknown
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func partNeedsHealing(partErrs []int) bool {
|
|
|
|
return slices.IndexFunc(partErrs, func(i int) bool { return i != checkPartSuccess && i != checkPartUnknown }) > -1
|
|
|
|
}
|
|
|
|
|
|
|
|
func hasPartErr(partErrs []int) bool {
|
|
|
|
return slices.IndexFunc(partErrs, func(i int) bool { return i != checkPartSuccess }) > -1
|
|
|
|
}
|
|
|
|
|
2017-03-04 17:53:28 -05:00
|
|
|
// disksWithAllParts - This function needs to be called with
|
|
|
|
// []StorageAPI returned by listOnlineDisks. Returns,
|
2017-09-28 18:57:19 -04:00
|
|
|
//
|
2020-06-12 23:04:01 -04:00
|
|
|
// - disks which have all parts specified in the latest xl.meta.
|
2017-09-28 18:57:19 -04:00
|
|
|
//
|
2022-08-26 15:52:29 -04:00
|
|
|
// - slice of errors about the state of data files on disk - can have
|
|
|
|
// a not-found error or a hash-mismatch error.
|
2021-07-26 14:48:09 -04:00
|
|
|
func disksWithAllParts(ctx context.Context, onlineDisks []StorageAPI, partsMetadata []FileInfo,
|
2024-10-01 11:19:10 -04:00
|
|
|
errs []error, latestMeta FileInfo, filterByETag bool, bucket, object string,
|
2024-06-10 11:51:27 -04:00
|
|
|
scanMode madmin.HealScanMode,
|
|
|
|
) (availableDisks []StorageAPI, dataErrsByDisk map[int][]int, dataErrsByPart map[int][]int) {
|
|
|
|
availableDisks = make([]StorageAPI, len(onlineDisks))
|
|
|
|
|
|
|
|
dataErrsByDisk = make(map[int][]int, len(onlineDisks))
|
|
|
|
for i := range onlineDisks {
|
|
|
|
dataErrsByDisk[i] = make([]int, len(latestMeta.Parts))
|
|
|
|
}
|
|
|
|
|
|
|
|
dataErrsByPart = make(map[int][]int, len(latestMeta.Parts))
|
|
|
|
for i := range latestMeta.Parts {
|
|
|
|
dataErrsByPart[i] = make([]int, len(onlineDisks))
|
|
|
|
}
|
2023-11-22 16:46:17 -05:00
|
|
|
|
2020-10-28 03:10:20 -04:00
|
|
|
inconsistent := 0
|
|
|
|
for i, meta := range partsMetadata {
|
|
|
|
if !meta.IsValid() {
|
|
|
|
// Since for majority of the cases erasure.Index matches with erasure.Distribution we can
|
|
|
|
// consider the offline disks as consistent.
|
|
|
|
continue
|
|
|
|
}
|
2021-07-26 14:48:09 -04:00
|
|
|
if !meta.Deleted {
|
|
|
|
if len(meta.Erasure.Distribution) != len(onlineDisks) {
|
|
|
|
// Erasure distribution seems to have lesser
|
|
|
|
// number of items than number of online disks.
|
|
|
|
inconsistent++
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if meta.Erasure.Distribution[i] != meta.Erasure.Index {
|
|
|
|
// Mismatch indexes with distribution order
|
|
|
|
inconsistent++
|
|
|
|
}
|
2020-10-28 03:10:20 -04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
erasureDistributionReliable := true
|
|
|
|
if inconsistent > len(partsMetadata)/2 {
|
|
|
|
// If there are too many inconsistent files, then we can't trust erasure.Distribution (most likely
|
|
|
|
// because of bugs found in CopyObject/PutObjectTags) https://github.com/minio/minio/pull/10772
|
|
|
|
erasureDistributionReliable = false
|
|
|
|
}
|
|
|
|
|
2024-06-10 11:51:27 -04:00
|
|
|
metaErrs := make([]error, len(errs))
|
|
|
|
|
2017-09-28 18:57:19 -04:00
|
|
|
for i, onlineDisk := range onlineDisks {
|
2020-06-12 23:04:01 -04:00
|
|
|
if errs[i] != nil {
|
2024-06-10 11:51:27 -04:00
|
|
|
metaErrs[i] = errs[i]
|
2017-03-04 17:53:28 -05:00
|
|
|
continue
|
|
|
|
}
|
2021-11-22 12:36:29 -05:00
|
|
|
if onlineDisk == OfflineDisk {
|
2024-06-10 11:51:27 -04:00
|
|
|
metaErrs[i] = errDiskNotFound
|
2020-06-12 23:04:01 -04:00
|
|
|
continue
|
|
|
|
}
|
2021-05-19 15:04:08 -04:00
|
|
|
|
2021-07-26 14:48:09 -04:00
|
|
|
meta := partsMetadata[i]
|
2024-10-01 11:19:10 -04:00
|
|
|
corrupted := false
|
|
|
|
if filterByETag {
|
|
|
|
corrupted = meta.Metadata["etag"] != latestMeta.Metadata["etag"]
|
|
|
|
} else {
|
|
|
|
corrupted = !meta.ModTime.Equal(latestMeta.ModTime) || meta.DataDir != latestMeta.DataDir
|
|
|
|
}
|
|
|
|
|
|
|
|
if corrupted {
|
2024-06-10 11:51:27 -04:00
|
|
|
metaErrs[i] = errFileCorrupt
|
2021-05-19 15:04:08 -04:00
|
|
|
partsMetadata[i] = FileInfo{}
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2020-10-28 03:10:20 -04:00
|
|
|
if erasureDistributionReliable {
|
|
|
|
if !meta.IsValid() {
|
2021-12-24 02:01:46 -05:00
|
|
|
partsMetadata[i] = FileInfo{}
|
2024-06-10 11:51:27 -04:00
|
|
|
metaErrs[i] = errFileCorrupt
|
2020-10-28 03:10:20 -04:00
|
|
|
continue
|
|
|
|
}
|
2020-10-28 22:24:01 -04:00
|
|
|
|
2021-07-26 14:48:09 -04:00
|
|
|
if !meta.Deleted {
|
|
|
|
if len(meta.Erasure.Distribution) != len(onlineDisks) {
|
|
|
|
// Erasure distribution is not the same as onlineDisks
|
|
|
|
// attempt a fix if possible, assuming other entries
|
|
|
|
// might have the right erasure distribution.
|
|
|
|
partsMetadata[i] = FileInfo{}
|
2024-06-10 11:51:27 -04:00
|
|
|
metaErrs[i] = errFileCorrupt
|
2021-07-26 14:48:09 -04:00
|
|
|
continue
|
|
|
|
}
|
2020-10-28 03:10:20 -04:00
|
|
|
}
|
|
|
|
}
|
2024-06-10 11:51:27 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
// Copy meta errors to part errors
|
|
|
|
for i, err := range metaErrs {
|
|
|
|
if err != nil {
|
|
|
|
partErr := convPartErrToInt(err)
|
|
|
|
for p := range latestMeta.Parts {
|
|
|
|
dataErrsByPart[p][i] = partErr
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for i, onlineDisk := range onlineDisks {
|
|
|
|
if metaErrs[i] != nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
meta := partsMetadata[i]
|
|
|
|
|
|
|
|
if meta.Deleted || meta.IsRemote() {
|
|
|
|
continue
|
|
|
|
}
|
2018-01-22 17:54:55 -05:00
|
|
|
|
2021-03-29 20:00:55 -04:00
|
|
|
// Always check data, if we got it.
|
2021-03-30 04:51:00 -04:00
|
|
|
if (len(meta.Data) > 0 || meta.Size == 0) && len(meta.Parts) > 0 {
|
2021-03-29 20:00:55 -04:00
|
|
|
checksumInfo := meta.Erasure.GetChecksumInfo(meta.Parts[0].Number)
|
2024-06-10 11:51:27 -04:00
|
|
|
verifyErr := bitrotVerify(bytes.NewReader(meta.Data),
|
2021-03-29 20:00:55 -04:00
|
|
|
int64(len(meta.Data)),
|
|
|
|
meta.Erasure.ShardFileSize(meta.Size),
|
|
|
|
checksumInfo.Algorithm,
|
|
|
|
checksumInfo.Hash, meta.Erasure.ShardSize())
|
2024-06-10 11:51:27 -04:00
|
|
|
dataErrsByPart[0][i] = convPartErrToInt(verifyErr)
|
2021-03-29 20:00:55 -04:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2024-06-10 11:51:27 -04:00
|
|
|
var (
|
|
|
|
verifyErr error
|
|
|
|
verifyResp *CheckPartsResp
|
|
|
|
)
|
|
|
|
|
2019-03-14 16:08:51 -04:00
|
|
|
switch scanMode {
|
|
|
|
case madmin.HealDeepScan:
|
2020-06-12 23:04:01 -04:00
|
|
|
// disk has a valid xl.meta but may not have all the
|
2019-03-14 16:08:51 -04:00
|
|
|
// parts. This is considered an outdated disk, since
|
|
|
|
// it needs healing too.
|
2024-06-10 11:51:27 -04:00
|
|
|
verifyResp, verifyErr = onlineDisk.VerifyFile(ctx, bucket, object, meta)
|
|
|
|
default:
|
|
|
|
verifyResp, verifyErr = onlineDisk.CheckParts(ctx, bucket, object, meta)
|
|
|
|
}
|
|
|
|
|
|
|
|
for p := range latestMeta.Parts {
|
|
|
|
if verifyErr != nil {
|
|
|
|
dataErrsByPart[p][i] = convPartErrToInt(verifyErr)
|
|
|
|
} else {
|
|
|
|
dataErrsByPart[p][i] = verifyResp.Results[p]
|
2021-04-19 13:30:42 -04:00
|
|
|
}
|
2017-09-28 18:57:19 -04:00
|
|
|
}
|
2024-06-10 11:51:27 -04:00
|
|
|
}
|
2017-09-28 18:57:19 -04:00
|
|
|
|
2024-06-10 11:51:27 -04:00
|
|
|
// Build dataErrs by disk from dataErrs by part
|
|
|
|
for part, disks := range dataErrsByPart {
|
|
|
|
for disk := range disks {
|
|
|
|
dataErrsByDisk[disk][part] = dataErrsByPart[part][disk]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for i, onlineDisk := range onlineDisks {
|
|
|
|
if metaErrs[i] == nil && !hasPartErr(dataErrsByDisk[i]) {
|
2017-09-28 18:57:19 -04:00
|
|
|
// All parts verified, mark it as all data available.
|
|
|
|
availableDisks[i] = onlineDisk
|
2021-05-14 19:50:47 -04:00
|
|
|
} else {
|
|
|
|
// upon errors just make that disk's fileinfo invalid
|
|
|
|
partsMetadata[i] = FileInfo{}
|
2017-03-04 17:53:28 -05:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-06-10 11:51:27 -04:00
|
|
|
return
|
2017-03-04 17:53:28 -05:00
|
|
|
}
|