mirror of
https://github.com/minio/minio.git
synced 2025-07-17 04:32:43 -04:00
heal multiple buckets in parallel
This commit is contained in:
parent
745a4b31ba
commit
fa1cd6dcce
@ -780,7 +780,8 @@ func (a adminAPIHandlers) HealSetsHandler(w http.ResponseWriter, r *http.Request
|
||||
go func(setNumber int) {
|
||||
defer wg.Done()
|
||||
lbDisks := z.serverSets[0].sets[setNumber].getOnlineDisks()
|
||||
if err := healErasureSet(ctx, setNumber, opts.sleepForIO, opts.sleepDuration, buckets, lbDisks); err != nil {
|
||||
setDriveCount := z.SetDriveCount()
|
||||
if err := healErasureSet(ctx, setNumber, setDriveCount, opts.sleepForIO, opts.sleepDuration, buckets, lbDisks); err != nil {
|
||||
logger.LogIf(ctx, err)
|
||||
}
|
||||
}(setNumber)
|
||||
|
@ -168,7 +168,8 @@ func monitorLocalDisksAndHeal(ctx context.Context, z *erasureServerSets, bgSeq *
|
||||
logger.Info("Healing disk '%s' on %s zone", disk, humanize.Ordinal(i+1))
|
||||
|
||||
lbDisks := z.serverSets[i].sets[setIndex].getOnlineDisks()
|
||||
if err := healErasureSet(ctx, setIndex, 10, time.Second, buckets, lbDisks); err != nil {
|
||||
setDriveCount := z.SetDriveCount()
|
||||
if err := healErasureSet(ctx, setIndex, setDriveCount, 10, time.Second, buckets, lbDisks); err != nil {
|
||||
logger.LogIf(ctx, err)
|
||||
continue
|
||||
}
|
||||
|
@ -98,7 +98,7 @@ func getLocalBackgroundHealStatus() (madmin.BgHealState, bool) {
|
||||
}
|
||||
|
||||
// healErasureSet lists and heals all objects in a specific erasure set
|
||||
func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.Duration, buckets []BucketInfo, disks []StorageAPI) error {
|
||||
func healErasureSet(ctx context.Context, setIndex int, setDriveCount int, maxIO int, maxSleep time.Duration, buckets []BucketInfo, disks []StorageAPI) error {
|
||||
// Get background heal sequence to send elements to heal
|
||||
var bgSeq *healSequence
|
||||
var ok bool
|
||||
@ -128,7 +128,12 @@ func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.
|
||||
}) // add metadata .minio.sys/ bucket prefixes to heal
|
||||
|
||||
// Heal all buckets with all objects
|
||||
var wg sync.WaitGroup
|
||||
for _, bucket := range buckets {
|
||||
wg.Add(1)
|
||||
go func(bucket BucketInfo, disks []StorageAPI) {
|
||||
defer wg.Done()
|
||||
|
||||
// Heal current bucket
|
||||
bgSeq.sourceCh <- healSource{
|
||||
bucket: bucket.Name,
|
||||
@ -136,12 +141,11 @@ func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.
|
||||
|
||||
var entryChs []FileInfoVersionsCh
|
||||
var mu sync.Mutex
|
||||
var wg sync.WaitGroup
|
||||
var wwg sync.WaitGroup
|
||||
for _, disk := range disks {
|
||||
disk := disk
|
||||
wg.Add(1)
|
||||
go func() {
|
||||
defer wg.Done()
|
||||
wwg.Add(1)
|
||||
go func(disk StorageAPI) {
|
||||
defer wwg.Done()
|
||||
entryCh, err := disk.WalkVersions(ctx, bucket.Name, "", "", true, ctx.Done())
|
||||
if err != nil {
|
||||
// Disk walk returned error, ignore it.
|
||||
@ -152,19 +156,23 @@ func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.
|
||||
Ch: entryCh,
|
||||
})
|
||||
mu.Unlock()
|
||||
}()
|
||||
}(disk)
|
||||
}
|
||||
wg.Wait()
|
||||
wwg.Wait()
|
||||
|
||||
entriesValid := make([]bool, len(entryChs))
|
||||
entries := make([]FileInfoVersions, len(entryChs))
|
||||
|
||||
for {
|
||||
entry, _, ok := lexicallySortedEntryVersions(entryChs, entries, entriesValid)
|
||||
entry, quorumCount, ok := lexicallySortedEntryVersions(entryChs, entries, entriesValid)
|
||||
if !ok {
|
||||
break
|
||||
}
|
||||
|
||||
if quorumCount == setDriveCount {
|
||||
continue
|
||||
}
|
||||
|
||||
for _, version := range entry.Versions {
|
||||
hsrc := healSource{
|
||||
bucket: bucket.Name,
|
||||
@ -180,7 +188,9 @@ func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.
|
||||
}
|
||||
}
|
||||
}
|
||||
}(bucket, disks)
|
||||
}
|
||||
wg.Wait()
|
||||
|
||||
return nil
|
||||
}
|
||||
|
Loading…
x
Reference in New Issue
Block a user