// Copyright (c) 2015-2021 MinIO, Inc. // // This file is part of MinIO Object Storage stack // // This program is free software: you can redistribute it and/or modify // it under the terms of the GNU Affero General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // // This program is distributed in the hope that it will be useful // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU Affero General Public License for more details. // // You should have received a copy of the GNU Affero General Public License // along with this program. If not, see . package cmd import ( "context" "errors" "runtime/debug" "sort" "sync" "time" "github.com/minio/minio/internal/logger" "github.com/minio/pkg/console" ) // a bucketMetacache keeps track of all caches generated // for a bucket. type bucketMetacache struct { // Name of bucket bucket string // caches indexed by id. caches map[string]metacache // cache ids indexed by root paths cachesRoot map[string][]string `msg:"-"` // Internal state mu sync.RWMutex `msg:"-"` updated bool `msg:"-"` } // newBucketMetacache creates a new bucketMetacache. // Optionally remove all existing caches. func newBucketMetacache(bucket string, cleanup bool) *bucketMetacache { if cleanup { // Recursively delete all caches. objAPI := newObjectLayerFn() if objAPI != nil { ez, ok := objAPI.(renameAllStorager) if ok { ctx := context.Background() ez.renameAll(ctx, minioMetaBucket, metacachePrefixForID(bucket, slashSeparator)) } } } return &bucketMetacache{ bucket: bucket, caches: make(map[string]metacache, 10), cachesRoot: make(map[string][]string, 10), } } func (b *bucketMetacache) debugf(format string, data ...interface{}) { if serverDebugLog { console.Debugf(format+"\n", data...) } } // findCache will attempt to find a matching cache for the provided options. // If a cache with the same ID exists already it will be returned. // If none can be found a new is created with the provided ID. func (b *bucketMetacache) findCache(o listPathOptions) metacache { if b == nil { logger.Info("bucketMetacache.findCache: nil cache for bucket %s", o.Bucket) return metacache{} } if o.Bucket != b.bucket { logger.Info("bucketMetacache.findCache: bucket %s does not match this bucket %s", o.Bucket, b.bucket) debug.PrintStack() return metacache{} } // Grab a write lock, since we create one if we cannot find one. b.mu.Lock() defer b.mu.Unlock() // Check if exists already. if c, ok := b.caches[o.ID]; ok { c.lastHandout = time.Now() b.caches[o.ID] = c b.debugf("returning existing %v", o.ID) return c } if !o.Create { return metacache{ id: o.ID, bucket: o.Bucket, status: scanStateNone, } } // Create new and add. best := o.newMetacache() b.caches[o.ID] = best b.cachesRoot[best.root] = append(b.cachesRoot[best.root], best.id) b.updated = true b.debugf("returning new cache %s, bucket: %v", best.id, best.bucket) return best } // cleanup removes redundant and outdated entries. func (b *bucketMetacache) cleanup() { // Entries to remove. remove := make(map[string]struct{}) // Test on a copy // cleanup is the only one deleting caches. caches, _ := b.cloneCaches() for id, cache := range caches { if !cache.worthKeeping() { b.debugf("cache %s not worth keeping", id) remove[id] = struct{}{} continue } if cache.id != id { logger.Info("cache ID mismatch %s != %s", id, cache.id) remove[id] = struct{}{} continue } if cache.bucket != b.bucket { logger.Info("cache bucket mismatch %s != %s", b.bucket, cache.bucket) remove[id] = struct{}{} continue } } // If above limit, remove the caches with the oldest handout time. if len(caches)-len(remove) > metacacheMaxEntries { remainCaches := make([]metacache, 0, len(caches)-len(remove)) for id, cache := range caches { if _, ok := remove[id]; ok { continue } remainCaches = append(remainCaches, cache) } if len(remainCaches) > metacacheMaxEntries { // Sort oldest last... sort.Slice(remainCaches, func(i, j int) bool { return remainCaches[i].lastHandout.Before(remainCaches[j].lastHandout) }) // Keep first metacacheMaxEntries... for _, cache := range remainCaches[metacacheMaxEntries:] { if time.Since(cache.lastHandout) > metacacheMaxClientWait { remove[cache.id] = struct{}{} } } } } for id := range remove { b.deleteCache(id) } } // updateCacheEntry will update a cache. // Returns the updated status. func (b *bucketMetacache) updateCacheEntry(update metacache) (metacache, error) { b.mu.Lock() defer b.mu.Unlock() existing, ok := b.caches[update.id] if !ok { return update, errFileNotFound } existing.update(update) b.caches[update.id] = existing b.updated = true return existing, nil } // cloneCaches will return a clone of all current caches. func (b *bucketMetacache) cloneCaches() (map[string]metacache, map[string][]string) { b.mu.RLock() defer b.mu.RUnlock() dst := make(map[string]metacache, len(b.caches)) for k, v := range b.caches { dst[k] = v } // Copy indexes dst2 := make(map[string][]string, len(b.cachesRoot)) for k, v := range b.cachesRoot { tmp := make([]string, len(v)) copy(tmp, v) dst2[k] = tmp } return dst, dst2 } // deleteAll will delete all on disk data for ALL caches. // Deletes are performed concurrently. func (b *bucketMetacache) deleteAll() { ctx := context.Background() objAPI := newObjectLayerFn() if objAPI == nil { return } ez, ok := objAPI.(renameAllStorager) if !ok { logger.LogIf(ctx, errors.New("bucketMetacache: expected objAPI to be 'renameAllStorager'")) return } b.mu.Lock() defer b.mu.Unlock() b.updated = true // Delete all. ez.renameAll(ctx, minioMetaBucket, metacachePrefixForID(b.bucket, slashSeparator)) b.caches = make(map[string]metacache, 10) b.cachesRoot = make(map[string][]string, 10) } // deleteCache will delete a specific cache and all files related to it across the cluster. func (b *bucketMetacache) deleteCache(id string) { b.mu.Lock() c, ok := b.caches[id] if ok { // Delete from root map. list := b.cachesRoot[c.root] for i, lid := range list { if id == lid { list = append(list[:i], list[i+1:]...) break } } b.cachesRoot[c.root] = list delete(b.caches, id) b.updated = true } b.mu.Unlock() if ok { c.delete(context.Background()) } }