heal multiple buckets in parallel

This commit is contained in:
Harshavardhana 2021-02-01 20:20:12 -08:00
parent 745a4b31ba
commit fa1cd6dcce
3 changed files with 59 additions and 47 deletions

View File

@ -780,7 +780,8 @@ func (a adminAPIHandlers) HealSetsHandler(w http.ResponseWriter, r *http.Request
go func(setNumber int) {
defer wg.Done()
lbDisks := z.serverSets[0].sets[setNumber].getOnlineDisks()
if err := healErasureSet(ctx, setNumber, opts.sleepForIO, opts.sleepDuration, buckets, lbDisks); err != nil {
setDriveCount := z.SetDriveCount()
if err := healErasureSet(ctx, setNumber, setDriveCount, opts.sleepForIO, opts.sleepDuration, buckets, lbDisks); err != nil {
logger.LogIf(ctx, err)
}
}(setNumber)

View File

@ -168,7 +168,8 @@ func monitorLocalDisksAndHeal(ctx context.Context, z *erasureServerSets, bgSeq *
logger.Info("Healing disk '%s' on %s zone", disk, humanize.Ordinal(i+1))
lbDisks := z.serverSets[i].sets[setIndex].getOnlineDisks()
if err := healErasureSet(ctx, setIndex, 10, time.Second, buckets, lbDisks); err != nil {
setDriveCount := z.SetDriveCount()
if err := healErasureSet(ctx, setIndex, setDriveCount, 10, time.Second, buckets, lbDisks); err != nil {
logger.LogIf(ctx, err)
continue
}

View File

@ -98,7 +98,7 @@ func getLocalBackgroundHealStatus() (madmin.BgHealState, bool) {
}
// healErasureSet lists and heals all objects in a specific erasure set
func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.Duration, buckets []BucketInfo, disks []StorageAPI) error {
func healErasureSet(ctx context.Context, setIndex int, setDriveCount int, maxIO int, maxSleep time.Duration, buckets []BucketInfo, disks []StorageAPI) error {
// Get background heal sequence to send elements to heal
var bgSeq *healSequence
var ok bool
@ -128,7 +128,12 @@ func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.
}) // add metadata .minio.sys/ bucket prefixes to heal
// Heal all buckets with all objects
var wg sync.WaitGroup
for _, bucket := range buckets {
wg.Add(1)
go func(bucket BucketInfo, disks []StorageAPI) {
defer wg.Done()
// Heal current bucket
bgSeq.sourceCh <- healSource{
bucket: bucket.Name,
@ -136,12 +141,11 @@ func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.
var entryChs []FileInfoVersionsCh
var mu sync.Mutex
var wg sync.WaitGroup
var wwg sync.WaitGroup
for _, disk := range disks {
disk := disk
wg.Add(1)
go func() {
defer wg.Done()
wwg.Add(1)
go func(disk StorageAPI) {
defer wwg.Done()
entryCh, err := disk.WalkVersions(ctx, bucket.Name, "", "", true, ctx.Done())
if err != nil {
// Disk walk returned error, ignore it.
@ -152,19 +156,23 @@ func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.
Ch: entryCh,
})
mu.Unlock()
}()
}(disk)
}
wg.Wait()
wwg.Wait()
entriesValid := make([]bool, len(entryChs))
entries := make([]FileInfoVersions, len(entryChs))
for {
entry, _, ok := lexicallySortedEntryVersions(entryChs, entries, entriesValid)
entry, quorumCount, ok := lexicallySortedEntryVersions(entryChs, entries, entriesValid)
if !ok {
break
}
if quorumCount == setDriveCount {
continue
}
for _, version := range entry.Versions {
hsrc := healSource{
bucket: bucket.Name,
@ -180,7 +188,9 @@ func healErasureSet(ctx context.Context, setIndex int, maxIO int, maxSleep time.
}
}
}
}(bucket, disks)
}
wg.Wait()
return nil
}