2021-04-18 15:41:13 -04:00
|
|
|
// Copyright (c) 2015-2021 MinIO, Inc.
|
|
|
|
//
|
|
|
|
// This file is part of MinIO Object Storage stack
|
|
|
|
//
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This program is distributed in the hope that it will be useful
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Affero General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
2016-05-30 19:51:59 -04:00
|
|
|
|
2016-08-18 19:23:42 -04:00
|
|
|
package cmd
|
2016-05-30 19:51:59 -04:00
|
|
|
|
2017-09-25 17:47:58 -04:00
|
|
|
import (
|
2018-04-05 18:04:40 -04:00
|
|
|
"context"
|
2022-07-19 11:35:29 -04:00
|
|
|
"fmt"
|
2023-07-14 05:25:40 -04:00
|
|
|
"math/rand"
|
2020-09-29 12:54:41 -04:00
|
|
|
"sync"
|
2023-07-14 05:25:40 -04:00
|
|
|
"time"
|
2022-07-19 11:35:29 -04:00
|
|
|
|
|
|
|
"github.com/minio/minio/internal/logger"
|
2023-04-26 01:57:40 -04:00
|
|
|
"github.com/minio/pkg/sync/errgroup"
|
2017-09-25 17:47:58 -04:00
|
|
|
)
|
2016-05-30 19:51:59 -04:00
|
|
|
|
2022-01-10 12:07:49 -05:00
|
|
|
func (er erasureObjects) getOnlineDisks() (newDisks []StorageAPI) {
|
|
|
|
disks := er.getDisks()
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
var mu sync.Mutex
|
2023-07-14 05:25:40 -04:00
|
|
|
r := rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
|
|
for _, i := range r.Perm(len(disks)) {
|
2022-01-10 12:07:49 -05:00
|
|
|
i := i
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
2023-07-14 05:25:40 -04:00
|
|
|
if disks[i] == nil {
|
2022-01-10 12:07:49 -05:00
|
|
|
return
|
|
|
|
}
|
2023-07-14 05:25:40 -04:00
|
|
|
di, err := disks[i].DiskInfo(context.Background())
|
2022-01-10 12:07:49 -05:00
|
|
|
if err != nil || di.Healing {
|
|
|
|
// - Do not consume disks which are not reachable
|
|
|
|
// unformatted or simply not accessible for some reason.
|
|
|
|
//
|
|
|
|
// - Do not consume disks which are being healed
|
|
|
|
//
|
|
|
|
// - Future: skip busy disks
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
mu.Lock()
|
2023-07-14 05:25:40 -04:00
|
|
|
newDisks = append(newDisks, disks[i])
|
2022-01-10 12:07:49 -05:00
|
|
|
mu.Unlock()
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
return newDisks
|
|
|
|
}
|
|
|
|
|
2020-08-26 22:29:35 -04:00
|
|
|
func (er erasureObjects) getLoadBalancedLocalDisks() (newDisks []StorageAPI) {
|
|
|
|
disks := er.getDisks()
|
|
|
|
// Based on the random shuffling return back randomized disks.
|
2023-07-14 05:25:40 -04:00
|
|
|
r := rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
|
|
for _, i := range r.Perm(len(disks)) {
|
|
|
|
if disks[i] != nil && disks[i].IsLocal() {
|
|
|
|
newDisks = append(newDisks, disks[i])
|
2020-08-26 22:29:35 -04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return newDisks
|
|
|
|
}
|
|
|
|
|
2016-07-21 03:27:08 -04:00
|
|
|
// getLoadBalancedDisks - fetches load balanced (sufficiently randomized) disk slice.
|
2020-09-28 22:39:32 -04:00
|
|
|
// ensures to skip disks if they are not healing and online.
|
2020-10-09 18:40:46 -04:00
|
|
|
func (er erasureObjects) getLoadBalancedDisks(optimized bool) []StorageAPI {
|
2020-06-12 23:04:01 -04:00
|
|
|
disks := er.getDisks()
|
2020-09-28 22:39:32 -04:00
|
|
|
|
2023-07-14 05:25:40 -04:00
|
|
|
r := rand.New(rand.NewSource(time.Now().UnixNano()))
|
2020-10-09 18:40:46 -04:00
|
|
|
if !optimized {
|
|
|
|
var newDisks []StorageAPI
|
2023-07-14 05:25:40 -04:00
|
|
|
for _, i := range r.Perm(len(disks)) {
|
|
|
|
newDisks = append(newDisks, disks[i])
|
2020-10-09 18:40:46 -04:00
|
|
|
}
|
|
|
|
return newDisks
|
|
|
|
}
|
|
|
|
|
2020-09-29 12:54:41 -04:00
|
|
|
var wg sync.WaitGroup
|
|
|
|
var mu sync.Mutex
|
2022-01-02 12:15:06 -05:00
|
|
|
newDisks := map[uint64][]StorageAPI{}
|
2016-06-01 19:43:31 -04:00
|
|
|
// Based on the random shuffling return back randomized disks.
|
2023-07-14 05:25:40 -04:00
|
|
|
for _, i := range r.Perm(len(disks)) {
|
2020-09-29 12:54:41 -04:00
|
|
|
i := i
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
2023-07-14 05:25:40 -04:00
|
|
|
if disks[i] == nil {
|
2020-09-29 12:54:41 -04:00
|
|
|
return
|
|
|
|
}
|
2023-07-14 05:25:40 -04:00
|
|
|
di, err := disks[i].DiskInfo(context.Background())
|
2020-09-29 12:54:41 -04:00
|
|
|
if err != nil || di.Healing {
|
|
|
|
// - Do not consume disks which are not reachable
|
|
|
|
// unformatted or simply not accessible for some reason.
|
|
|
|
//
|
|
|
|
// - Do not consume disks which are being healed
|
|
|
|
//
|
|
|
|
// - Future: skip busy disks
|
|
|
|
return
|
|
|
|
}
|
2020-09-30 01:54:02 -04:00
|
|
|
|
2020-09-29 12:54:41 -04:00
|
|
|
mu.Lock()
|
2020-10-14 15:12:10 -04:00
|
|
|
// Capture disks usage wise upto resolution of MiB
|
2023-07-14 05:25:40 -04:00
|
|
|
newDisks[di.Used/1024/1024] = append(newDisks[di.Used/1024/1024], disks[i])
|
2020-09-29 12:54:41 -04:00
|
|
|
mu.Unlock()
|
|
|
|
}()
|
2016-06-01 19:43:31 -04:00
|
|
|
}
|
2020-09-29 12:54:41 -04:00
|
|
|
wg.Wait()
|
2020-09-30 01:54:02 -04:00
|
|
|
|
|
|
|
var max uint64
|
|
|
|
for k := range newDisks {
|
|
|
|
if k > max {
|
|
|
|
max = k
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return disks which have maximum disk usage common.
|
|
|
|
return newDisks[max]
|
2016-05-30 19:51:59 -04:00
|
|
|
}
|
2022-07-19 11:35:29 -04:00
|
|
|
|
|
|
|
// readMultipleFiles Reads raw data from all specified files from all disks.
|
|
|
|
func readMultipleFiles(ctx context.Context, disks []StorageAPI, req ReadMultipleReq, readQuorum int) ([]ReadMultipleResp, error) {
|
|
|
|
resps := make([]chan ReadMultipleResp, len(disks))
|
|
|
|
for i := range resps {
|
|
|
|
resps[i] = make(chan ReadMultipleResp, len(req.Files))
|
|
|
|
}
|
|
|
|
g := errgroup.WithNErrs(len(disks))
|
|
|
|
// Read files in parallel across disks.
|
|
|
|
for index := range disks {
|
|
|
|
index := index
|
|
|
|
g.Go(func() (err error) {
|
|
|
|
if disks[index] == nil {
|
|
|
|
return errDiskNotFound
|
|
|
|
}
|
|
|
|
return disks[index].ReadMultiple(ctx, req, resps[index])
|
|
|
|
}, index)
|
|
|
|
}
|
|
|
|
|
|
|
|
dataArray := make([]ReadMultipleResp, 0, len(req.Files))
|
|
|
|
// Merge results. They should come in order from each.
|
|
|
|
for _, wantFile := range req.Files {
|
|
|
|
quorum := 0
|
|
|
|
toAdd := ReadMultipleResp{
|
|
|
|
Bucket: req.Bucket,
|
|
|
|
Prefix: req.Prefix,
|
|
|
|
File: wantFile,
|
|
|
|
}
|
|
|
|
for i := range resps {
|
|
|
|
if disks[i] == nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
case gotFile, ok := <-resps[i]:
|
|
|
|
if !ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if gotFile.Error != "" || !gotFile.Exists {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if gotFile.File != wantFile || gotFile.Bucket != req.Bucket || gotFile.Prefix != req.Prefix {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
quorum++
|
|
|
|
if toAdd.Modtime.After(gotFile.Modtime) || len(gotFile.Data) < len(toAdd.Data) {
|
|
|
|
// Pick latest, or largest to avoid possible truncated entries.
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
toAdd = gotFile
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if quorum < readQuorum {
|
|
|
|
toAdd.Exists = false
|
|
|
|
toAdd.Error = errErasureReadQuorum.Error()
|
|
|
|
toAdd.Data = nil
|
|
|
|
}
|
|
|
|
dataArray = append(dataArray, toAdd)
|
|
|
|
}
|
|
|
|
|
|
|
|
errs := g.Wait()
|
|
|
|
for index, err := range errs {
|
|
|
|
if err == nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if !IsErr(err, []error{
|
|
|
|
errFileNotFound,
|
|
|
|
errVolumeNotFound,
|
|
|
|
errFileVersionNotFound,
|
|
|
|
errDiskNotFound,
|
|
|
|
errUnformattedDisk,
|
|
|
|
}...) {
|
|
|
|
logger.LogOnceIf(ctx, fmt.Errorf("Drive %s, path (%s/%s) returned an error (%w)",
|
|
|
|
disks[index], req.Bucket, req.Prefix, err),
|
|
|
|
disks[index].String())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return all the metadata.
|
|
|
|
return dataArray, nil
|
|
|
|
}
|