2021-04-18 15:41:13 -04:00
|
|
|
// Copyright (c) 2015-2021 MinIO, Inc.
|
|
|
|
//
|
|
|
|
// This file is part of MinIO Object Storage stack
|
|
|
|
//
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This program is distributed in the hope that it will be useful
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Affero General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
2017-08-14 21:08:42 -04:00
|
|
|
|
|
|
|
package cmd
|
|
|
|
|
|
|
|
import (
|
2018-04-05 18:04:40 -04:00
|
|
|
"context"
|
2020-08-03 21:17:48 -04:00
|
|
|
"errors"
|
2020-06-12 23:04:01 -04:00
|
|
|
"fmt"
|
2021-02-18 03:38:37 -05:00
|
|
|
"math/rand"
|
2021-02-26 12:52:27 -05:00
|
|
|
"os"
|
2022-08-08 19:16:44 -04:00
|
|
|
"runtime"
|
2020-06-12 23:04:01 -04:00
|
|
|
"sort"
|
2020-05-09 12:54:20 -04:00
|
|
|
"sync"
|
2020-06-12 23:04:01 -04:00
|
|
|
"time"
|
2017-08-14 21:08:42 -04:00
|
|
|
|
2021-05-06 11:52:02 -04:00
|
|
|
"github.com/minio/madmin-go"
|
2021-06-01 17:59:40 -04:00
|
|
|
"github.com/minio/minio/internal/bpool"
|
|
|
|
"github.com/minio/minio/internal/dsync"
|
|
|
|
"github.com/minio/minio/internal/logger"
|
|
|
|
"github.com/minio/minio/internal/sync/errgroup"
|
2021-05-28 18:17:01 -04:00
|
|
|
"github.com/minio/pkg/console"
|
2017-08-14 21:08:42 -04:00
|
|
|
)
|
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
// OfflineDisk represents an unavailable disk.
|
|
|
|
var OfflineDisk StorageAPI // zero value is nil
|
|
|
|
|
|
|
|
// erasureObjects - Implements ER object layer.
|
|
|
|
type erasureObjects struct {
|
2021-01-16 15:08:02 -05:00
|
|
|
setDriveCount int
|
|
|
|
defaultParityCount int
|
2020-12-07 13:04:07 -05:00
|
|
|
|
2021-03-04 17:36:23 -05:00
|
|
|
setIndex int
|
|
|
|
poolIndex int
|
2021-01-26 16:21:51 -05:00
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
// getDisks returns list of storageAPIs.
|
|
|
|
getDisks func() []StorageAPI
|
|
|
|
|
|
|
|
// getLockers returns list of remote and local lockers.
|
2020-09-25 22:21:52 -04:00
|
|
|
getLockers func() ([]dsync.NetLocker, string)
|
2020-06-12 23:04:01 -04:00
|
|
|
|
|
|
|
// getEndpoints returns list of endpoint strings belonging this set.
|
|
|
|
// some may be local and some remote.
|
2021-09-29 14:36:19 -04:00
|
|
|
getEndpoints func() []Endpoint
|
2020-06-12 23:04:01 -04:00
|
|
|
|
|
|
|
// Locker mutex map.
|
|
|
|
nsMutex *nsLockMap
|
|
|
|
|
|
|
|
// Byte pools used for temporary i/o buffers.
|
|
|
|
bp *bpool.BytePoolCap
|
|
|
|
|
2021-06-07 13:06:06 -04:00
|
|
|
// Byte pools used for temporary i/o buffers,
|
|
|
|
// legacy objects.
|
|
|
|
bpOld *bpool.BytePoolCap
|
|
|
|
|
2021-02-26 12:52:27 -05:00
|
|
|
deletedCleanupSleeper *dynamicSleeper
|
2017-08-14 21:08:42 -04:00
|
|
|
}
|
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
// NewNSLock - initialize a new namespace RWLocker instance.
|
2020-11-04 11:25:42 -05:00
|
|
|
func (er erasureObjects) NewNSLock(bucket string, objects ...string) RWLocker {
|
|
|
|
return er.nsMutex.NewNSLock(er.getLockers, bucket, objects...)
|
2020-06-12 23:04:01 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
// Shutdown function for object storage interface.
|
|
|
|
func (er erasureObjects) Shutdown(ctx context.Context) error {
|
|
|
|
// Add any object layer shutdown activities here.
|
2022-05-30 13:58:37 -04:00
|
|
|
closeStorageDisks(er.getDisks()...)
|
2020-06-12 23:04:01 -04:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2021-12-28 15:41:52 -05:00
|
|
|
// defaultWQuorum write quorum based on setDriveCount and defaultParityCount
|
|
|
|
func (er erasureObjects) defaultWQuorum() int {
|
|
|
|
dataCount := er.setDriveCount - er.defaultParityCount
|
|
|
|
if dataCount == er.defaultParityCount {
|
|
|
|
return dataCount + 1
|
|
|
|
}
|
|
|
|
return dataCount
|
|
|
|
}
|
|
|
|
|
|
|
|
func (er erasureObjects) defaultRQuorum() int {
|
|
|
|
return er.setDriveCount - er.defaultParityCount
|
|
|
|
}
|
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
// byDiskTotal is a collection satisfying sort.Interface.
|
2020-07-13 12:51:07 -04:00
|
|
|
type byDiskTotal []madmin.Disk
|
2020-06-12 23:04:01 -04:00
|
|
|
|
|
|
|
func (d byDiskTotal) Len() int { return len(d) }
|
|
|
|
func (d byDiskTotal) Swap(i, j int) { d[i], d[j] = d[j], d[i] }
|
|
|
|
func (d byDiskTotal) Less(i, j int) bool {
|
2020-07-13 12:51:07 -04:00
|
|
|
return d[i].TotalSpace < d[j].TotalSpace
|
|
|
|
}
|
|
|
|
|
|
|
|
func diskErrToDriveState(err error) (state string) {
|
2020-08-03 21:17:48 -04:00
|
|
|
switch {
|
2022-10-18 06:01:16 -04:00
|
|
|
case errors.Is(err, errDiskNotFound) || errors.Is(err, context.DeadlineExceeded):
|
2020-07-13 12:51:07 -04:00
|
|
|
state = madmin.DriveStateOffline
|
2020-08-03 21:17:48 -04:00
|
|
|
case errors.Is(err, errCorruptedFormat):
|
2020-07-13 12:51:07 -04:00
|
|
|
state = madmin.DriveStateCorrupt
|
2020-08-03 21:17:48 -04:00
|
|
|
case errors.Is(err, errUnformattedDisk):
|
2020-07-13 12:51:07 -04:00
|
|
|
state = madmin.DriveStateUnformatted
|
2020-08-03 21:17:48 -04:00
|
|
|
case errors.Is(err, errDiskAccessDenied):
|
2020-07-13 12:51:07 -04:00
|
|
|
state = madmin.DriveStatePermission
|
2020-08-03 21:17:48 -04:00
|
|
|
case errors.Is(err, errFaultyDisk):
|
2020-07-13 12:51:07 -04:00
|
|
|
state = madmin.DriveStateFaulty
|
2020-08-03 21:17:48 -04:00
|
|
|
case err == nil:
|
2020-07-13 12:51:07 -04:00
|
|
|
state = madmin.DriveStateOk
|
2022-10-13 19:41:44 -04:00
|
|
|
default:
|
|
|
|
state = fmt.Sprintf("%s (cause: %s)", madmin.DriveStateUnknown, err)
|
2020-07-13 12:51:07 -04:00
|
|
|
}
|
2022-10-13 19:41:44 -04:00
|
|
|
|
2020-07-13 12:51:07 -04:00
|
|
|
return
|
2020-06-12 23:04:01 -04:00
|
|
|
}
|
|
|
|
|
2020-12-21 12:35:19 -05:00
|
|
|
func getOnlineOfflineDisksStats(disksInfo []madmin.Disk) (onlineDisks, offlineDisks madmin.BackendDisks) {
|
2020-06-12 23:04:01 -04:00
|
|
|
onlineDisks = make(madmin.BackendDisks)
|
|
|
|
offlineDisks = make(madmin.BackendDisks)
|
|
|
|
|
2020-12-21 12:35:19 -05:00
|
|
|
for _, disk := range disksInfo {
|
|
|
|
ep := disk.Endpoint
|
2020-06-29 16:07:26 -04:00
|
|
|
if _, ok := offlineDisks[ep]; !ok {
|
|
|
|
offlineDisks[ep] = 0
|
2020-06-12 23:04:01 -04:00
|
|
|
}
|
2020-06-29 16:07:26 -04:00
|
|
|
if _, ok := onlineDisks[ep]; !ok {
|
|
|
|
onlineDisks[ep] = 0
|
2020-06-12 23:04:01 -04:00
|
|
|
}
|
2017-08-14 21:08:42 -04:00
|
|
|
}
|
2020-05-09 12:54:20 -04:00
|
|
|
|
2020-12-21 12:35:19 -05:00
|
|
|
// Wait for the routines.
|
|
|
|
for _, disk := range disksInfo {
|
|
|
|
ep := disk.Endpoint
|
|
|
|
state := disk.State
|
|
|
|
if state != madmin.DriveStateOk && state != madmin.DriveStateUnformatted {
|
|
|
|
offlineDisks[ep]++
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
onlineDisks[ep]++
|
|
|
|
}
|
|
|
|
|
|
|
|
rootDiskCount := 0
|
|
|
|
for _, di := range disksInfo {
|
|
|
|
if di.RootDisk {
|
|
|
|
rootDiskCount++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Count offline disks as well to ensure consistent
|
|
|
|
// reportability of offline drives on local setups.
|
|
|
|
if len(disksInfo) == (rootDiskCount + offlineDisks.Sum()) {
|
|
|
|
// Success.
|
|
|
|
return onlineDisks, offlineDisks
|
|
|
|
}
|
|
|
|
|
|
|
|
// Root disk should be considered offline
|
|
|
|
for i := range disksInfo {
|
|
|
|
ep := disksInfo[i].Endpoint
|
|
|
|
if disksInfo[i].RootDisk {
|
|
|
|
offlineDisks[ep]++
|
|
|
|
onlineDisks[ep]--
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return onlineDisks, offlineDisks
|
|
|
|
}
|
|
|
|
|
|
|
|
// getDisksInfo - fetch disks info across all other storage API.
|
2021-09-29 14:36:19 -04:00
|
|
|
func getDisksInfo(disks []StorageAPI, endpoints []Endpoint) (disksInfo []madmin.Disk, errs []error) {
|
2020-12-21 12:35:19 -05:00
|
|
|
disksInfo = make([]madmin.Disk, len(disks))
|
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
g := errgroup.WithNErrs(len(disks))
|
|
|
|
for index := range disks {
|
|
|
|
index := index
|
|
|
|
g.Go(func() error {
|
2022-10-13 19:41:44 -04:00
|
|
|
diskEndpoint := endpoints[index].String()
|
2020-06-12 23:04:01 -04:00
|
|
|
if disks[index] == OfflineDisk {
|
2020-09-17 00:14:35 -04:00
|
|
|
logger.LogIf(GlobalContext, fmt.Errorf("%s: %s", errDiskNotFound, endpoints[index]))
|
2020-07-13 12:51:07 -04:00
|
|
|
disksInfo[index] = madmin.Disk{
|
|
|
|
State: diskErrToDriveState(errDiskNotFound),
|
2022-10-13 19:41:44 -04:00
|
|
|
Endpoint: diskEndpoint,
|
2020-07-13 12:51:07 -04:00
|
|
|
}
|
2020-06-12 23:04:01 -04:00
|
|
|
// Storage disk is empty, perhaps ignored disk or not available.
|
|
|
|
return errDiskNotFound
|
|
|
|
}
|
2020-09-04 12:45:06 -04:00
|
|
|
info, err := disks[index].DiskInfo(context.TODO())
|
2020-07-13 12:51:07 -04:00
|
|
|
di := madmin.Disk{
|
2022-10-13 19:41:44 -04:00
|
|
|
Endpoint: diskEndpoint,
|
2020-07-16 17:43:48 -04:00
|
|
|
DrivePath: info.MountPath,
|
|
|
|
TotalSpace: info.Total,
|
|
|
|
UsedSpace: info.Used,
|
|
|
|
AvailableSpace: info.Free,
|
|
|
|
UUID: info.ID,
|
2022-08-16 10:13:49 -04:00
|
|
|
Major: info.Major,
|
|
|
|
Minor: info.Minor,
|
2020-08-18 17:37:26 -04:00
|
|
|
RootDisk: info.RootDisk,
|
2020-09-28 22:39:32 -04:00
|
|
|
Healing: info.Healing,
|
2022-08-16 10:13:49 -04:00
|
|
|
Scanning: info.Scanning,
|
2020-07-16 17:43:48 -04:00
|
|
|
State: diskErrToDriveState(err),
|
2021-05-06 02:03:40 -04:00
|
|
|
FreeInodes: info.FreeInodes,
|
2020-07-13 12:51:07 -04:00
|
|
|
}
|
2021-03-04 17:36:23 -05:00
|
|
|
di.PoolIndex, di.SetIndex, di.DiskIndex = disks[index].GetDiskLoc()
|
|
|
|
if info.Healing {
|
|
|
|
if hi := disks[index].Healing(); hi != nil {
|
|
|
|
hd := hi.toHealingDisk()
|
|
|
|
di.HealInfo = &hd
|
|
|
|
}
|
|
|
|
}
|
2021-03-16 23:06:57 -04:00
|
|
|
di.Metrics = &madmin.DiskMetrics{
|
2022-07-05 17:45:49 -04:00
|
|
|
LastMinute: make(map[string]madmin.TimedAction, len(info.Metrics.LastMinute)),
|
|
|
|
APICalls: make(map[string]uint64, len(info.Metrics.APICalls)),
|
2021-03-16 23:06:57 -04:00
|
|
|
}
|
2022-07-05 17:45:49 -04:00
|
|
|
for k, v := range info.Metrics.LastMinute {
|
|
|
|
if v.N > 0 {
|
|
|
|
di.Metrics.LastMinute[k] = v.asTimedAction()
|
|
|
|
}
|
2021-03-16 23:06:57 -04:00
|
|
|
}
|
|
|
|
for k, v := range info.Metrics.APICalls {
|
|
|
|
di.Metrics.APICalls[k] = v
|
|
|
|
}
|
2020-07-13 12:51:07 -04:00
|
|
|
if info.Total > 0 {
|
|
|
|
di.Utilization = float64(info.Used / info.Total * 100)
|
|
|
|
}
|
|
|
|
disksInfo[index] = di
|
|
|
|
return err
|
2020-06-12 23:04:01 -04:00
|
|
|
}, index)
|
2020-05-09 12:54:20 -04:00
|
|
|
}
|
|
|
|
|
2020-12-21 12:35:19 -05:00
|
|
|
return disksInfo, g.Wait()
|
2017-08-14 21:08:42 -04:00
|
|
|
}
|
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
// Get an aggregated storage info across all disks.
|
2021-09-29 14:36:19 -04:00
|
|
|
func getStorageInfo(disks []StorageAPI, endpoints []Endpoint) (StorageInfo, []error) {
|
2020-12-21 12:35:19 -05:00
|
|
|
disksInfo, errs := getDisksInfo(disks, endpoints)
|
2020-06-12 23:04:01 -04:00
|
|
|
|
|
|
|
// Sort so that the first element is the smallest.
|
|
|
|
sort.Sort(byDiskTotal(disksInfo))
|
|
|
|
|
|
|
|
storageInfo := StorageInfo{
|
2020-07-13 12:51:07 -04:00
|
|
|
Disks: disksInfo,
|
2017-08-14 21:08:42 -04:00
|
|
|
}
|
2020-06-12 23:04:01 -04:00
|
|
|
|
2021-03-04 17:36:23 -05:00
|
|
|
storageInfo.Backend.Type = madmin.Erasure
|
2020-06-12 23:04:01 -04:00
|
|
|
return storageInfo, errs
|
2017-08-14 21:08:42 -04:00
|
|
|
}
|
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
// StorageInfo - returns underlying storage statistics.
|
2021-01-04 12:42:09 -05:00
|
|
|
func (er erasureObjects) StorageInfo(ctx context.Context) (StorageInfo, []error) {
|
2020-06-12 23:04:01 -04:00
|
|
|
disks := er.getDisks()
|
2020-06-29 16:07:26 -04:00
|
|
|
endpoints := er.getEndpoints()
|
|
|
|
return getStorageInfo(disks, endpoints)
|
2020-06-12 23:04:01 -04:00
|
|
|
}
|
|
|
|
|
2021-03-02 20:28:04 -05:00
|
|
|
// LocalStorageInfo - returns underlying local storage statistics.
|
|
|
|
func (er erasureObjects) LocalStorageInfo(ctx context.Context) (StorageInfo, []error) {
|
2021-09-29 14:36:19 -04:00
|
|
|
disks := er.getDisks()
|
|
|
|
endpoints := er.getEndpoints()
|
|
|
|
|
|
|
|
var localDisks []StorageAPI
|
|
|
|
var localEndpoints []Endpoint
|
|
|
|
|
|
|
|
for i, endpoint := range endpoints {
|
|
|
|
if endpoint.IsLocal {
|
|
|
|
localDisks = append(localDisks, disks[i])
|
|
|
|
localEndpoints = append(localEndpoints, endpoint)
|
2021-03-02 20:28:04 -05:00
|
|
|
}
|
|
|
|
}
|
2021-09-29 14:36:19 -04:00
|
|
|
|
|
|
|
return getStorageInfo(localDisks, localEndpoints)
|
2021-03-02 20:28:04 -05:00
|
|
|
}
|
|
|
|
|
2021-01-19 05:40:52 -05:00
|
|
|
func (er erasureObjects) getOnlineDisksWithHealing() (newDisks []StorageAPI, healing bool) {
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
disks := er.getDisks()
|
|
|
|
infos := make([]DiskInfo, len(disks))
|
|
|
|
for _, i := range hashOrder(UTCNow().String(), len(disks)) {
|
|
|
|
i := i
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
|
|
|
|
|
|
|
disk := disks[i-1]
|
|
|
|
|
|
|
|
if disk == nil {
|
2022-08-04 19:10:08 -04:00
|
|
|
infos[i-1].Error = "nil drive"
|
2021-01-19 05:40:52 -05:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
di, err := disk.DiskInfo(context.Background())
|
|
|
|
if err != nil {
|
|
|
|
// - Do not consume disks which are not reachable
|
|
|
|
// unformatted or simply not accessible for some reason.
|
|
|
|
//
|
|
|
|
//
|
|
|
|
// - Future: skip busy disks
|
2021-02-02 20:04:37 -05:00
|
|
|
infos[i-1].Error = err.Error()
|
2021-01-19 05:40:52 -05:00
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
infos[i-1] = di
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
for i, info := range infos {
|
|
|
|
// Check if one of the drives in the set is being healed.
|
2021-02-17 15:04:11 -05:00
|
|
|
// this information is used by scanner to skip healing
|
2021-01-19 05:40:52 -05:00
|
|
|
// this erasure set while it calculates the usage.
|
2021-02-02 20:04:37 -05:00
|
|
|
if info.Healing || info.Error != "" {
|
2021-01-19 05:40:52 -05:00
|
|
|
healing = true
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
newDisks = append(newDisks, disks[i])
|
|
|
|
}
|
|
|
|
|
|
|
|
return newDisks, healing
|
|
|
|
}
|
|
|
|
|
2021-02-26 12:52:27 -05:00
|
|
|
// Clean-up previously deleted objects. from .minio.sys/tmp/.trash/
|
|
|
|
func (er erasureObjects) cleanupDeletedObjects(ctx context.Context) {
|
|
|
|
// run multiple cleanup's local to this server.
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
for _, disk := range er.getLoadBalancedLocalDisks() {
|
|
|
|
if disk != nil {
|
|
|
|
wg.Add(1)
|
|
|
|
go func(disk StorageAPI) {
|
|
|
|
defer wg.Done()
|
|
|
|
diskPath := disk.Endpoint().Path
|
|
|
|
readDirFn(pathJoin(diskPath, minioMetaTmpDeletedBucket), func(ddir string, typ os.FileMode) error {
|
|
|
|
wait := er.deletedCleanupSleeper.Timer(ctx)
|
|
|
|
removeAll(pathJoin(diskPath, minioMetaTmpDeletedBucket, ddir))
|
|
|
|
wait()
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
}(disk)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
}
|
|
|
|
|
2021-02-26 18:11:42 -05:00
|
|
|
// nsScanner will start scanning buckets and send updated totals as they are traversed.
|
2020-06-12 23:04:01 -04:00
|
|
|
// Updates are sent on a regular basis and the caller *must* consume them.
|
2022-04-07 11:10:40 -04:00
|
|
|
func (er erasureObjects) nsScanner(ctx context.Context, buckets []BucketInfo, bf *bloomFilter, wantCycle uint32, updates chan<- dataUsageCache, healScanMode madmin.HealScanMode) error {
|
2020-08-25 13:55:15 -04:00
|
|
|
if len(buckets) == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
2020-06-12 23:04:01 -04:00
|
|
|
|
2021-01-19 05:40:52 -05:00
|
|
|
// Collect disks we can use.
|
|
|
|
disks, healing := er.getOnlineDisksWithHealing()
|
2020-08-25 13:55:15 -04:00
|
|
|
if len(disks) == 0 {
|
2022-08-04 19:10:08 -04:00
|
|
|
logger.LogIf(ctx, errors.New("data-scanner: all drives are offline or being healed, skipping scanner cycle"))
|
2020-05-09 12:54:20 -04:00
|
|
|
return nil
|
|
|
|
}
|
2020-06-12 23:04:01 -04:00
|
|
|
|
|
|
|
// Load bucket totals
|
|
|
|
oldCache := dataUsageCache{}
|
2020-09-28 22:39:32 -04:00
|
|
|
if err := oldCache.load(ctx, er, dataUsageCacheName); err != nil {
|
2018-04-05 18:04:40 -04:00
|
|
|
return err
|
2017-08-14 21:08:42 -04:00
|
|
|
}
|
2019-01-17 07:58:18 -05:00
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
// New cache..
|
|
|
|
cache := dataUsageCache{
|
|
|
|
Info: dataUsageCacheInfo{
|
|
|
|
Name: dataUsageRoot,
|
|
|
|
NextCycle: oldCache.Info.NextCycle,
|
|
|
|
},
|
|
|
|
Cache: make(map[string]dataUsageEntry, len(oldCache.Cache)),
|
|
|
|
}
|
2020-06-17 11:54:41 -04:00
|
|
|
bloom := bf.bytes()
|
2019-01-17 07:58:18 -05:00
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
// Put all buckets into channel.
|
|
|
|
bucketCh := make(chan BucketInfo, len(buckets))
|
|
|
|
// Add new buckets first
|
|
|
|
for _, b := range buckets {
|
|
|
|
if oldCache.find(b.Name) == nil {
|
|
|
|
bucketCh <- b
|
|
|
|
}
|
2019-01-17 07:58:18 -05:00
|
|
|
}
|
2020-06-17 11:54:41 -04:00
|
|
|
|
2020-08-24 16:47:01 -04:00
|
|
|
// Add existing buckets.
|
2020-06-12 23:04:01 -04:00
|
|
|
for _, b := range buckets {
|
|
|
|
e := oldCache.find(b.Name)
|
|
|
|
if e != nil {
|
|
|
|
cache.replace(b.Name, dataUsageRoot, *e)
|
2020-08-24 16:47:01 -04:00
|
|
|
bucketCh <- b
|
2020-06-12 23:04:01 -04:00
|
|
|
}
|
2019-04-30 19:27:31 -04:00
|
|
|
}
|
2019-01-17 07:58:18 -05:00
|
|
|
|
2020-06-12 23:04:01 -04:00
|
|
|
close(bucketCh)
|
|
|
|
bucketResults := make(chan dataUsageEntryInfo, len(disks))
|
|
|
|
|
|
|
|
// Start async collector/saver.
|
|
|
|
// This goroutine owns the cache.
|
|
|
|
var saverWg sync.WaitGroup
|
|
|
|
saverWg.Add(1)
|
|
|
|
go func() {
|
2021-02-18 03:38:37 -05:00
|
|
|
// Add jitter to the update time so multiple sets don't sync up.
|
2022-01-02 12:15:06 -05:00
|
|
|
updateTime := 30*time.Second + time.Duration(float64(10*time.Second)*rand.Float64())
|
2020-06-12 23:04:01 -04:00
|
|
|
t := time.NewTicker(updateTime)
|
|
|
|
defer t.Stop()
|
|
|
|
defer saverWg.Done()
|
|
|
|
var lastSave time.Time
|
|
|
|
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
// Return without saving.
|
|
|
|
return
|
|
|
|
case <-t.C:
|
|
|
|
if cache.Info.LastUpdate.Equal(lastSave) {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
logger.LogIf(ctx, cache.save(ctx, er, dataUsageCacheName))
|
|
|
|
updates <- cache.clone()
|
|
|
|
lastSave = cache.Info.LastUpdate
|
|
|
|
case v, ok := <-bucketResults:
|
|
|
|
if !ok {
|
2020-12-27 01:58:06 -05:00
|
|
|
// Save final state...
|
2021-08-25 11:25:26 -04:00
|
|
|
cache.Info.NextCycle = wantCycle
|
2020-12-27 01:58:06 -05:00
|
|
|
cache.Info.LastUpdate = time.Now()
|
|
|
|
logger.LogIf(ctx, cache.save(ctx, er, dataUsageCacheName))
|
|
|
|
updates <- cache
|
|
|
|
return
|
2020-06-12 23:04:01 -04:00
|
|
|
}
|
|
|
|
cache.replace(v.Name, v.Parent, v.Entry)
|
|
|
|
cache.Info.LastUpdate = time.Now()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
2021-04-12 10:55:40 -04:00
|
|
|
// Shuffle disks to ensure a total randomness of bucket/disk association to ensure
|
|
|
|
// that objects that are not present in all disks are accounted and ILM applied.
|
|
|
|
r := rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
|
|
r.Shuffle(len(disks), func(i, j int) { disks[i], disks[j] = disks[j], disks[i] })
|
|
|
|
|
2022-08-08 19:16:44 -04:00
|
|
|
// Restrict parallelism for disk usage scanner
|
|
|
|
// upto GOMAXPROCS if GOMAXPROCS is < len(disks)
|
|
|
|
maxProcs := runtime.GOMAXPROCS(0)
|
|
|
|
if maxProcs < len(disks) {
|
|
|
|
disks = disks[:maxProcs]
|
|
|
|
}
|
|
|
|
|
2021-02-17 15:04:11 -05:00
|
|
|
// Start one scanner per disk
|
2020-06-12 23:04:01 -04:00
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(len(disks))
|
|
|
|
for i := range disks {
|
|
|
|
go func(i int) {
|
|
|
|
defer wg.Done()
|
|
|
|
disk := disks[i]
|
|
|
|
|
|
|
|
for bucket := range bucketCh {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load cache for bucket
|
|
|
|
cacheName := pathJoin(bucket.Name, dataUsageCacheName)
|
|
|
|
cache := dataUsageCache{}
|
|
|
|
logger.LogIf(ctx, cache.load(ctx, er, cacheName))
|
|
|
|
if cache.Info.Name == "" {
|
|
|
|
cache.Info.Name = bucket.Name
|
|
|
|
}
|
2020-06-17 11:54:41 -04:00
|
|
|
cache.Info.BloomFilter = bloom
|
2021-01-19 05:40:52 -05:00
|
|
|
cache.Info.SkipHealing = healing
|
2021-08-25 11:25:26 -04:00
|
|
|
cache.Info.NextCycle = wantCycle
|
2020-06-12 23:04:01 -04:00
|
|
|
if cache.Info.Name != bucket.Name {
|
|
|
|
logger.LogIf(ctx, fmt.Errorf("cache name mismatch: %s != %s", cache.Info.Name, bucket.Name))
|
|
|
|
cache.Info = dataUsageCacheInfo{
|
|
|
|
Name: bucket.Name,
|
|
|
|
LastUpdate: time.Time{},
|
2021-08-25 11:25:26 -04:00
|
|
|
NextCycle: wantCycle,
|
2020-06-12 23:04:01 -04:00
|
|
|
}
|
|
|
|
}
|
2021-05-19 17:38:30 -04:00
|
|
|
// Collect updates.
|
|
|
|
updates := make(chan dataUsageEntry, 1)
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(1)
|
2021-07-02 14:19:56 -04:00
|
|
|
go func(name string) {
|
2021-05-19 17:38:30 -04:00
|
|
|
defer wg.Done()
|
|
|
|
for update := range updates {
|
|
|
|
bucketResults <- dataUsageEntryInfo{
|
2021-07-02 14:19:56 -04:00
|
|
|
Name: name,
|
2021-05-19 17:38:30 -04:00
|
|
|
Parent: dataUsageRoot,
|
|
|
|
Entry: update,
|
|
|
|
}
|
|
|
|
if intDataUpdateTracker.debug {
|
2021-08-04 12:14:14 -04:00
|
|
|
console.Debugln("z:", er.poolIndex, "s:", er.setIndex, "bucket", name, "got update", update)
|
2021-05-19 17:38:30 -04:00
|
|
|
}
|
|
|
|
}
|
2021-07-02 14:19:56 -04:00
|
|
|
}(cache.Info.Name)
|
2020-06-12 23:04:01 -04:00
|
|
|
// Calc usage
|
|
|
|
before := cache.Info.LastUpdate
|
2020-09-28 22:39:32 -04:00
|
|
|
var err error
|
2022-04-07 11:10:40 -04:00
|
|
|
cache, err = disk.NSScanner(ctx, cache, updates, healScanMode)
|
2020-06-17 11:54:41 -04:00
|
|
|
cache.Info.BloomFilter = nil
|
2020-06-12 23:04:01 -04:00
|
|
|
if err != nil {
|
2021-02-22 13:04:32 -05:00
|
|
|
if !cache.Info.LastUpdate.IsZero() && cache.Info.LastUpdate.After(before) {
|
2020-06-12 23:04:01 -04:00
|
|
|
logger.LogIf(ctx, cache.save(ctx, er, cacheName))
|
2021-02-22 13:04:32 -05:00
|
|
|
} else {
|
|
|
|
logger.LogIf(ctx, err)
|
2020-06-12 23:04:01 -04:00
|
|
|
}
|
2022-01-19 03:46:43 -05:00
|
|
|
// This ensures that we don't close
|
|
|
|
// bucketResults channel while the
|
|
|
|
// updates-collector goroutine still
|
|
|
|
// holds a reference to this.
|
|
|
|
wg.Wait()
|
2020-06-12 23:04:01 -04:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2021-05-19 17:38:30 -04:00
|
|
|
wg.Wait()
|
2020-06-12 23:04:01 -04:00
|
|
|
var root dataUsageEntry
|
|
|
|
if r := cache.root(); r != nil {
|
|
|
|
root = cache.flatten(*r)
|
|
|
|
}
|
2021-02-18 03:38:37 -05:00
|
|
|
t := time.Now()
|
2020-06-12 23:04:01 -04:00
|
|
|
bucketResults <- dataUsageEntryInfo{
|
|
|
|
Name: cache.Info.Name,
|
|
|
|
Parent: dataUsageRoot,
|
|
|
|
Entry: root,
|
|
|
|
}
|
2021-02-18 03:38:37 -05:00
|
|
|
// We want to avoid synchronizing up all writes in case
|
|
|
|
// the results are piled up.
|
|
|
|
time.Sleep(time.Duration(float64(time.Since(t)) * rand.Float64()))
|
2020-06-12 23:04:01 -04:00
|
|
|
// Save cache
|
|
|
|
logger.LogIf(ctx, cache.save(ctx, er, cacheName))
|
|
|
|
}
|
|
|
|
}(i)
|
2019-01-17 07:58:18 -05:00
|
|
|
}
|
2020-06-12 23:04:01 -04:00
|
|
|
wg.Wait()
|
|
|
|
close(bucketResults)
|
|
|
|
saverWg.Wait()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|