2021-04-18 15:41:13 -04:00
|
|
|
// Copyright (c) 2015-2021 MinIO, Inc.
|
|
|
|
//
|
|
|
|
// This file is part of MinIO Object Storage stack
|
|
|
|
//
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This program is distributed in the hope that it will be useful
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Affero General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
2020-10-28 12:18:35 -04:00
|
|
|
|
|
|
|
package cmd
|
|
|
|
|
|
|
|
import (
|
2021-02-25 01:24:38 -05:00
|
|
|
"bytes"
|
|
|
|
"context"
|
|
|
|
"errors"
|
|
|
|
"fmt"
|
|
|
|
"net/http"
|
2020-10-28 12:18:35 -04:00
|
|
|
"runtime/debug"
|
2021-02-08 14:36:16 -05:00
|
|
|
"sort"
|
2020-10-28 12:18:35 -04:00
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
|
2021-02-25 01:24:38 -05:00
|
|
|
"github.com/klauspost/compress/s2"
|
2021-06-01 17:59:40 -04:00
|
|
|
"github.com/minio/minio/internal/hash"
|
|
|
|
"github.com/minio/minio/internal/logger"
|
2021-05-28 18:17:01 -04:00
|
|
|
"github.com/minio/pkg/console"
|
2021-02-25 01:24:38 -05:00
|
|
|
"github.com/tinylib/msgp/msgp"
|
2020-10-28 12:18:35 -04:00
|
|
|
)
|
|
|
|
|
|
|
|
//go:generate msgp -file $GOFILE -unexported
|
|
|
|
|
|
|
|
// a bucketMetacache keeps track of all caches generated
|
|
|
|
// for a bucket.
|
|
|
|
type bucketMetacache struct {
|
|
|
|
// Name of bucket
|
|
|
|
bucket string
|
|
|
|
|
|
|
|
// caches indexed by id.
|
|
|
|
caches map[string]metacache
|
listcache: Add path index (#11063)
Add a root path index.
```
Before:
Benchmark_bucketMetacache_findCache-32 10000 730737 ns/op
With excluded prints:
Benchmark_bucketMetacache_findCache-32 10000 207100 ns/op
With the root path:
Benchmark_bucketMetacache_findCache-32 705765 1943 ns/op
```
Benchmark used (not linear):
```Go
func Benchmark_bucketMetacache_findCache(b *testing.B) {
bm := newBucketMetacache("", false)
for i := 0; i < b.N; i++ {
bm.findCache(listPathOptions{
ID: mustGetUUID(),
Bucket: "",
BaseDir: "prefix/" + mustGetUUID(),
Prefix: "",
FilterPrefix: "",
Marker: "",
Limit: 0,
AskDisks: 0,
Recursive: false,
Separator: slashSeparator,
Create: true,
CurrentCycle: 0,
OldestCycle: 0,
})
}
}
```
Replaces #11058
2020-12-09 11:37:43 -05:00
|
|
|
// cache ids indexed by root paths
|
|
|
|
cachesRoot map[string][]string `msg:"-"`
|
2020-10-28 12:18:35 -04:00
|
|
|
|
|
|
|
// Internal state
|
2021-07-05 18:34:41 -04:00
|
|
|
mu sync.RWMutex `msg:"-"`
|
|
|
|
updated bool `msg:"-"`
|
2020-10-28 12:18:35 -04:00
|
|
|
}
|
|
|
|
|
|
|
|
// newBucketMetacache creates a new bucketMetacache.
|
2020-10-31 13:26:16 -04:00
|
|
|
// Optionally remove all existing caches.
|
2021-02-25 01:24:38 -05:00
|
|
|
func newBucketMetacache(bucket string, cleanup bool) *bucketMetacache {
|
|
|
|
if cleanup {
|
|
|
|
// Recursively delete all caches.
|
|
|
|
objAPI := newObjectLayerFn()
|
|
|
|
ez, ok := objAPI.(*erasureServerPools)
|
|
|
|
if ok {
|
|
|
|
ctx := context.Background()
|
|
|
|
ez.renameAll(ctx, minioMetaBucket, metacachePrefixForID(bucket, slashSeparator))
|
|
|
|
}
|
|
|
|
}
|
2020-10-28 12:18:35 -04:00
|
|
|
return &bucketMetacache{
|
listcache: Add path index (#11063)
Add a root path index.
```
Before:
Benchmark_bucketMetacache_findCache-32 10000 730737 ns/op
With excluded prints:
Benchmark_bucketMetacache_findCache-32 10000 207100 ns/op
With the root path:
Benchmark_bucketMetacache_findCache-32 705765 1943 ns/op
```
Benchmark used (not linear):
```Go
func Benchmark_bucketMetacache_findCache(b *testing.B) {
bm := newBucketMetacache("", false)
for i := 0; i < b.N; i++ {
bm.findCache(listPathOptions{
ID: mustGetUUID(),
Bucket: "",
BaseDir: "prefix/" + mustGetUUID(),
Prefix: "",
FilterPrefix: "",
Marker: "",
Limit: 0,
AskDisks: 0,
Recursive: false,
Separator: slashSeparator,
Create: true,
CurrentCycle: 0,
OldestCycle: 0,
})
}
}
```
Replaces #11058
2020-12-09 11:37:43 -05:00
|
|
|
bucket: bucket,
|
|
|
|
caches: make(map[string]metacache, 10),
|
|
|
|
cachesRoot: make(map[string][]string, 10),
|
2020-10-28 12:18:35 -04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-12-14 16:13:33 -05:00
|
|
|
func (b *bucketMetacache) debugf(format string, data ...interface{}) {
|
2020-12-17 19:52:47 -05:00
|
|
|
if serverDebugLog {
|
|
|
|
console.Debugf(format+"\n", data...)
|
2020-12-14 16:13:33 -05:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-02-25 01:24:38 -05:00
|
|
|
// loadBucketMetaCache will load the cache from the object layer.
|
|
|
|
// If the cache cannot be found a new one is created.
|
|
|
|
func loadBucketMetaCache(ctx context.Context, bucket string) (*bucketMetacache, error) {
|
|
|
|
objAPI := newObjectLayerFn()
|
|
|
|
for objAPI == nil {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
return nil, ctx.Err()
|
|
|
|
default:
|
|
|
|
time.Sleep(250 * time.Millisecond)
|
|
|
|
}
|
|
|
|
objAPI = newObjectLayerFn()
|
|
|
|
if objAPI == nil {
|
|
|
|
logger.LogIf(ctx, fmt.Errorf("loadBucketMetaCache: object layer not ready. bucket: %q", bucket))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
var meta bucketMetacache
|
|
|
|
var decErr error
|
|
|
|
// Use global context for this.
|
|
|
|
r, err := objAPI.GetObjectNInfo(GlobalContext, minioMetaBucket, pathJoin("buckets", bucket, ".metacache", "index.s2"), nil, http.Header{}, readLock, ObjectOptions{})
|
|
|
|
if err == nil {
|
|
|
|
dec := s2DecPool.Get().(*s2.Reader)
|
|
|
|
dec.Reset(r)
|
|
|
|
decErr = meta.DecodeMsg(msgp.NewReader(dec))
|
|
|
|
dec.Reset(nil)
|
|
|
|
r.Close()
|
|
|
|
s2DecPool.Put(dec)
|
|
|
|
}
|
|
|
|
if err != nil {
|
|
|
|
switch err.(type) {
|
|
|
|
case ObjectNotFound:
|
|
|
|
err = nil
|
|
|
|
case InsufficientReadQuorum:
|
|
|
|
// Cache is likely lost. Clean up and return new.
|
|
|
|
return newBucketMetacache(bucket, true), nil
|
|
|
|
default:
|
|
|
|
logger.LogIf(ctx, err)
|
|
|
|
}
|
|
|
|
return newBucketMetacache(bucket, false), err
|
|
|
|
}
|
|
|
|
if decErr != nil {
|
|
|
|
if errors.Is(err, context.Canceled) {
|
|
|
|
return newBucketMetacache(bucket, false), err
|
|
|
|
}
|
|
|
|
// Log the error, but assume the data is lost and return a fresh bucket.
|
|
|
|
// Otherwise a broken cache will never recover.
|
|
|
|
logger.LogIf(ctx, decErr)
|
|
|
|
return newBucketMetacache(bucket, true), nil
|
|
|
|
}
|
|
|
|
// Sanity check...
|
|
|
|
if meta.bucket != bucket {
|
|
|
|
logger.Info("loadBucketMetaCache: loaded cache name mismatch, want %s, got %s. Discarding.", bucket, meta.bucket)
|
|
|
|
return newBucketMetacache(bucket, true), nil
|
|
|
|
}
|
|
|
|
meta.cachesRoot = make(map[string][]string, len(meta.caches)/10)
|
|
|
|
// Index roots
|
|
|
|
for id, cache := range meta.caches {
|
|
|
|
meta.cachesRoot[cache.root] = append(meta.cachesRoot[cache.root], id)
|
|
|
|
}
|
|
|
|
return &meta, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// save the bucket cache to the object storage.
|
|
|
|
func (b *bucketMetacache) save(ctx context.Context) error {
|
|
|
|
objAPI := newObjectLayerFn()
|
|
|
|
if objAPI == nil {
|
|
|
|
return errServerNotInitialized
|
|
|
|
}
|
|
|
|
|
|
|
|
// Keep lock while we marshal.
|
|
|
|
// We need a write lock since we update 'updated'
|
|
|
|
b.mu.Lock()
|
|
|
|
if !b.updated {
|
|
|
|
b.mu.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
// Save as s2 compressed msgpack
|
|
|
|
tmp := bytes.NewBuffer(make([]byte, 0, b.Msgsize()))
|
|
|
|
enc := s2.NewWriter(tmp)
|
|
|
|
err := msgp.Encode(enc, b)
|
|
|
|
if err != nil {
|
|
|
|
b.mu.Unlock()
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
err = enc.Close()
|
|
|
|
if err != nil {
|
|
|
|
b.mu.Unlock()
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
b.updated = false
|
|
|
|
b.mu.Unlock()
|
|
|
|
|
|
|
|
hr, err := hash.NewReader(tmp, int64(tmp.Len()), "", "", int64(tmp.Len()))
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
_, err = objAPI.PutObject(ctx, minioMetaBucket, pathJoin("buckets", b.bucket, ".metacache", "index.s2"), NewPutObjReader(hr), ObjectOptions{})
|
|
|
|
logger.LogIf(ctx, err)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2020-10-28 12:18:35 -04:00
|
|
|
// findCache will attempt to find a matching cache for the provided options.
|
|
|
|
// If a cache with the same ID exists already it will be returned.
|
|
|
|
// If none can be found a new is created with the provided ID.
|
|
|
|
func (b *bucketMetacache) findCache(o listPathOptions) metacache {
|
|
|
|
if b == nil {
|
|
|
|
logger.Info("bucketMetacache.findCache: nil cache for bucket %s", o.Bucket)
|
|
|
|
return metacache{}
|
|
|
|
}
|
|
|
|
|
2021-07-05 18:34:41 -04:00
|
|
|
if o.Bucket != b.bucket {
|
2020-10-28 12:18:35 -04:00
|
|
|
logger.Info("bucketMetacache.findCache: bucket %s does not match this bucket %s", o.Bucket, b.bucket)
|
|
|
|
debug.PrintStack()
|
|
|
|
return metacache{}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Grab a write lock, since we create one if we cannot find one.
|
2021-07-05 18:34:41 -04:00
|
|
|
b.mu.Lock()
|
|
|
|
defer b.mu.Unlock()
|
2020-10-28 12:18:35 -04:00
|
|
|
|
|
|
|
// Check if exists already.
|
|
|
|
if c, ok := b.caches[o.ID]; ok {
|
2021-07-05 18:34:41 -04:00
|
|
|
c.lastHandout = time.Now()
|
|
|
|
b.caches[o.ID] = c
|
2020-12-14 16:13:33 -05:00
|
|
|
b.debugf("returning existing %v", o.ID)
|
2020-10-28 12:18:35 -04:00
|
|
|
return c
|
|
|
|
}
|
2020-11-05 14:49:56 -05:00
|
|
|
|
2020-10-28 12:18:35 -04:00
|
|
|
if !o.Create {
|
|
|
|
return metacache{
|
|
|
|
id: o.ID,
|
|
|
|
bucket: o.Bucket,
|
|
|
|
status: scanStateNone,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create new and add.
|
2021-07-05 18:34:41 -04:00
|
|
|
best := o.newMetacache()
|
2020-10-28 12:18:35 -04:00
|
|
|
b.caches[o.ID] = best
|
listcache: Add path index (#11063)
Add a root path index.
```
Before:
Benchmark_bucketMetacache_findCache-32 10000 730737 ns/op
With excluded prints:
Benchmark_bucketMetacache_findCache-32 10000 207100 ns/op
With the root path:
Benchmark_bucketMetacache_findCache-32 705765 1943 ns/op
```
Benchmark used (not linear):
```Go
func Benchmark_bucketMetacache_findCache(b *testing.B) {
bm := newBucketMetacache("", false)
for i := 0; i < b.N; i++ {
bm.findCache(listPathOptions{
ID: mustGetUUID(),
Bucket: "",
BaseDir: "prefix/" + mustGetUUID(),
Prefix: "",
FilterPrefix: "",
Marker: "",
Limit: 0,
AskDisks: 0,
Recursive: false,
Separator: slashSeparator,
Create: true,
CurrentCycle: 0,
OldestCycle: 0,
})
}
}
```
Replaces #11058
2020-12-09 11:37:43 -05:00
|
|
|
b.cachesRoot[best.root] = append(b.cachesRoot[best.root], best.id)
|
2020-10-28 12:18:35 -04:00
|
|
|
b.updated = true
|
2020-12-14 16:13:33 -05:00
|
|
|
b.debugf("returning new cache %s, bucket: %v", best.id, best.bucket)
|
2020-10-28 12:18:35 -04:00
|
|
|
return best
|
|
|
|
}
|
|
|
|
|
|
|
|
// cleanup removes redundant and outdated entries.
|
|
|
|
func (b *bucketMetacache) cleanup() {
|
|
|
|
// Entries to remove.
|
|
|
|
remove := make(map[string]struct{})
|
|
|
|
|
2020-12-10 15:30:28 -05:00
|
|
|
// Test on a copy
|
|
|
|
// cleanup is the only one deleting caches.
|
2021-07-05 18:34:41 -04:00
|
|
|
caches, _ := b.cloneCaches()
|
2020-12-10 15:30:28 -05:00
|
|
|
|
|
|
|
for id, cache := range caches {
|
2021-07-05 18:34:41 -04:00
|
|
|
if !cache.worthKeeping() {
|
2020-12-14 16:13:33 -05:00
|
|
|
b.debugf("cache %s not worth keeping", id)
|
2020-10-28 12:18:35 -04:00
|
|
|
remove[id] = struct{}{}
|
2020-11-04 11:01:33 -05:00
|
|
|
continue
|
2020-10-28 12:18:35 -04:00
|
|
|
}
|
|
|
|
if cache.id != id {
|
|
|
|
logger.Info("cache ID mismatch %s != %s", id, cache.id)
|
|
|
|
remove[id] = struct{}{}
|
2020-11-04 11:01:33 -05:00
|
|
|
continue
|
2020-10-28 12:18:35 -04:00
|
|
|
}
|
2021-07-05 18:34:41 -04:00
|
|
|
if cache.bucket != b.bucket {
|
2020-10-28 12:18:35 -04:00
|
|
|
logger.Info("cache bucket mismatch %s != %s", b.bucket, cache.bucket)
|
|
|
|
remove[id] = struct{}{}
|
2020-11-04 11:01:33 -05:00
|
|
|
continue
|
2020-10-28 12:18:35 -04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-02-08 14:36:16 -05:00
|
|
|
// If above limit, remove the caches with the oldest handout time.
|
|
|
|
if len(caches)-len(remove) > metacacheMaxEntries {
|
|
|
|
remainCaches := make([]metacache, 0, len(caches)-len(remove))
|
|
|
|
for id, cache := range caches {
|
|
|
|
if _, ok := remove[id]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
remainCaches = append(remainCaches, cache)
|
|
|
|
}
|
|
|
|
if len(remainCaches) > metacacheMaxEntries {
|
|
|
|
// Sort oldest last...
|
|
|
|
sort.Slice(remainCaches, func(i, j int) bool {
|
|
|
|
return remainCaches[i].lastHandout.Before(remainCaches[j].lastHandout)
|
|
|
|
})
|
|
|
|
// Keep first metacacheMaxEntries...
|
|
|
|
for _, cache := range remainCaches[metacacheMaxEntries:] {
|
2021-02-11 13:22:03 -05:00
|
|
|
if time.Since(cache.lastHandout) > 30*time.Minute {
|
2021-02-08 14:36:16 -05:00
|
|
|
remove[cache.id] = struct{}{}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-10-28 12:18:35 -04:00
|
|
|
for id := range remove {
|
|
|
|
b.deleteCache(id)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// updateCacheEntry will update a cache.
|
|
|
|
// Returns the updated status.
|
|
|
|
func (b *bucketMetacache) updateCacheEntry(update metacache) (metacache, error) {
|
|
|
|
b.mu.Lock()
|
|
|
|
defer b.mu.Unlock()
|
|
|
|
existing, ok := b.caches[update.id]
|
|
|
|
if !ok {
|
|
|
|
return update, errFileNotFound
|
|
|
|
}
|
2020-11-05 10:34:08 -05:00
|
|
|
existing.update(update)
|
2020-10-28 12:18:35 -04:00
|
|
|
b.caches[update.id] = existing
|
|
|
|
b.updated = true
|
|
|
|
return existing, nil
|
|
|
|
}
|
|
|
|
|
2020-12-10 15:30:28 -05:00
|
|
|
// cloneCaches will return a clone of all current caches.
|
|
|
|
func (b *bucketMetacache) cloneCaches() (map[string]metacache, map[string][]string) {
|
|
|
|
b.mu.RLock()
|
|
|
|
defer b.mu.RUnlock()
|
|
|
|
dst := make(map[string]metacache, len(b.caches))
|
|
|
|
for k, v := range b.caches {
|
|
|
|
dst[k] = v
|
|
|
|
}
|
|
|
|
// Copy indexes
|
|
|
|
dst2 := make(map[string][]string, len(b.cachesRoot))
|
|
|
|
for k, v := range b.cachesRoot {
|
|
|
|
tmp := make([]string, len(v))
|
|
|
|
copy(tmp, v)
|
|
|
|
dst2[k] = tmp
|
|
|
|
}
|
|
|
|
|
|
|
|
return dst, dst2
|
|
|
|
}
|
|
|
|
|
2021-02-25 01:24:38 -05:00
|
|
|
// deleteAll will delete all on disk data for ALL caches.
|
|
|
|
// Deletes are performed concurrently.
|
|
|
|
func (b *bucketMetacache) deleteAll() {
|
|
|
|
ctx := context.Background()
|
|
|
|
ez, ok := newObjectLayerFn().(*erasureServerPools)
|
|
|
|
if !ok {
|
|
|
|
logger.LogIf(ctx, errors.New("bucketMetacache: expected objAPI to be *erasurePools"))
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
b.mu.Lock()
|
|
|
|
defer b.mu.Unlock()
|
|
|
|
|
|
|
|
b.updated = true
|
2021-07-05 18:34:41 -04:00
|
|
|
// Delete all.
|
|
|
|
ez.renameAll(ctx, minioMetaBucket, metacachePrefixForID(b.bucket, slashSeparator))
|
2021-02-25 01:24:38 -05:00
|
|
|
b.caches = make(map[string]metacache, 10)
|
2021-07-05 18:34:41 -04:00
|
|
|
b.cachesRoot = make(map[string][]string, 10)
|
2021-02-25 01:24:38 -05:00
|
|
|
}
|
|
|
|
|
2020-10-28 12:18:35 -04:00
|
|
|
// deleteCache will delete a specific cache and all files related to it across the cluster.
|
|
|
|
func (b *bucketMetacache) deleteCache(id string) {
|
|
|
|
b.mu.Lock()
|
|
|
|
c, ok := b.caches[id]
|
|
|
|
if ok {
|
listcache: Add path index (#11063)
Add a root path index.
```
Before:
Benchmark_bucketMetacache_findCache-32 10000 730737 ns/op
With excluded prints:
Benchmark_bucketMetacache_findCache-32 10000 207100 ns/op
With the root path:
Benchmark_bucketMetacache_findCache-32 705765 1943 ns/op
```
Benchmark used (not linear):
```Go
func Benchmark_bucketMetacache_findCache(b *testing.B) {
bm := newBucketMetacache("", false)
for i := 0; i < b.N; i++ {
bm.findCache(listPathOptions{
ID: mustGetUUID(),
Bucket: "",
BaseDir: "prefix/" + mustGetUUID(),
Prefix: "",
FilterPrefix: "",
Marker: "",
Limit: 0,
AskDisks: 0,
Recursive: false,
Separator: slashSeparator,
Create: true,
CurrentCycle: 0,
OldestCycle: 0,
})
}
}
```
Replaces #11058
2020-12-09 11:37:43 -05:00
|
|
|
// Delete from root map.
|
|
|
|
list := b.cachesRoot[c.root]
|
|
|
|
for i, lid := range list {
|
|
|
|
if id == lid {
|
|
|
|
list = append(list[:i], list[i+1:]...)
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
b.cachesRoot[c.root] = list
|
2020-10-28 12:18:35 -04:00
|
|
|
delete(b.caches, id)
|
|
|
|
b.updated = true
|
|
|
|
}
|
|
|
|
b.mu.Unlock()
|
2021-02-25 01:24:38 -05:00
|
|
|
if ok {
|
|
|
|
c.delete(context.Background())
|
|
|
|
}
|
2020-10-28 12:18:35 -04:00
|
|
|
}
|