mirror of
https://github.com/minio/minio.git
synced 2025-03-03 07:10:07 -05:00
fix: possible race in FS local lockMap (#9598)
This commit is contained in:
parent
56e0c6adf8
commit
b730bd1396
21
cmd/fs-v1.go
21
cmd/fs-v1.go
@ -320,12 +320,6 @@ func (fs *FSObjects) MakeBucketWithLocation(ctx context.Context, bucket, locatio
|
|||||||
return NotImplemented{}
|
return NotImplemented{}
|
||||||
}
|
}
|
||||||
|
|
||||||
bucketLock := fs.NewNSLock(ctx, bucket, "")
|
|
||||||
if err := bucketLock.GetLock(globalObjectTimeout); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
defer bucketLock.Unlock()
|
|
||||||
|
|
||||||
// Verify if bucket is valid.
|
// Verify if bucket is valid.
|
||||||
if s3utils.CheckValidBucketNameStrict(bucket) != nil {
|
if s3utils.CheckValidBucketNameStrict(bucket) != nil {
|
||||||
return BucketNameInvalid{Bucket: bucket}
|
return BucketNameInvalid{Bucket: bucket}
|
||||||
@ -356,12 +350,6 @@ func (fs *FSObjects) MakeBucketWithLocation(ctx context.Context, bucket, locatio
|
|||||||
|
|
||||||
// GetBucketInfo - fetch bucket metadata info.
|
// GetBucketInfo - fetch bucket metadata info.
|
||||||
func (fs *FSObjects) GetBucketInfo(ctx context.Context, bucket string) (bi BucketInfo, e error) {
|
func (fs *FSObjects) GetBucketInfo(ctx context.Context, bucket string) (bi BucketInfo, e error) {
|
||||||
bucketLock := fs.NewNSLock(ctx, bucket, "")
|
|
||||||
if e := bucketLock.GetRLock(globalObjectTimeout); e != nil {
|
|
||||||
return bi, e
|
|
||||||
}
|
|
||||||
defer bucketLock.RUnlock()
|
|
||||||
|
|
||||||
atomic.AddInt64(&fs.activeIOCount, 1)
|
atomic.AddInt64(&fs.activeIOCount, 1)
|
||||||
defer func() {
|
defer func() {
|
||||||
atomic.AddInt64(&fs.activeIOCount, -1)
|
atomic.AddInt64(&fs.activeIOCount, -1)
|
||||||
@ -438,15 +426,6 @@ func (fs *FSObjects) ListBuckets(ctx context.Context) ([]BucketInfo, error) {
|
|||||||
// DeleteBucket - delete a bucket and all the metadata associated
|
// DeleteBucket - delete a bucket and all the metadata associated
|
||||||
// with the bucket including pending multipart, object metadata.
|
// with the bucket including pending multipart, object metadata.
|
||||||
func (fs *FSObjects) DeleteBucket(ctx context.Context, bucket string, forceDelete bool) error {
|
func (fs *FSObjects) DeleteBucket(ctx context.Context, bucket string, forceDelete bool) error {
|
||||||
if !forceDelete {
|
|
||||||
bucketLock := fs.NewNSLock(ctx, bucket, "")
|
|
||||||
if err := bucketLock.GetLock(globalObjectTimeout); err != nil {
|
|
||||||
logger.LogIf(ctx, err)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
defer bucketLock.Unlock()
|
|
||||||
}
|
|
||||||
|
|
||||||
atomic.AddInt64(&fs.activeIOCount, 1)
|
atomic.AddInt64(&fs.activeIOCount, 1)
|
||||||
defer func() {
|
defer func() {
|
||||||
atomic.AddInt64(&fs.activeIOCount, -1)
|
atomic.AddInt64(&fs.activeIOCount, -1)
|
||||||
|
@ -18,18 +18,17 @@ package cmd
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
pathutil "path"
|
pathutil "path"
|
||||||
"runtime"
|
"runtime"
|
||||||
"sort"
|
"sort"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
"sync/atomic"
|
||||||
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/minio/lsync"
|
"github.com/minio/lsync"
|
||||||
"github.com/minio/minio/cmd/logger"
|
|
||||||
"github.com/minio/minio/pkg/dsync"
|
"github.com/minio/minio/pkg/dsync"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -58,8 +57,8 @@ func newNSLock(isDistXL bool) *nsLockMap {
|
|||||||
|
|
||||||
// nsLock - provides primitives for locking critical namespace regions.
|
// nsLock - provides primitives for locking critical namespace regions.
|
||||||
type nsLock struct {
|
type nsLock struct {
|
||||||
|
ref uint32
|
||||||
*lsync.LRWMutex
|
*lsync.LRWMutex
|
||||||
ref uint
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// nsLockMap - namespace lock map, provides primitives to Lock,
|
// nsLockMap - namespace lock map, provides primitives to Lock,
|
||||||
@ -68,7 +67,7 @@ type nsLockMap struct {
|
|||||||
// Indicates if namespace is part of a distributed setup.
|
// Indicates if namespace is part of a distributed setup.
|
||||||
isDistXL bool
|
isDistXL bool
|
||||||
lockMap map[string]*nsLock
|
lockMap map[string]*nsLock
|
||||||
lockMapMutex sync.RWMutex
|
lockMapMutex sync.Mutex
|
||||||
}
|
}
|
||||||
|
|
||||||
// Lock the namespace resource.
|
// Lock the namespace resource.
|
||||||
@ -78,17 +77,16 @@ func (n *nsLockMap) lock(ctx context.Context, volume string, path string, lockSo
|
|||||||
resource := pathJoin(volume, path)
|
resource := pathJoin(volume, path)
|
||||||
|
|
||||||
n.lockMapMutex.Lock()
|
n.lockMapMutex.Lock()
|
||||||
nsLk, found := n.lockMap[resource]
|
if _, found := n.lockMap[resource]; !found {
|
||||||
if !found {
|
n.lockMap[resource] = &nsLock{
|
||||||
nsLk = &nsLock{
|
|
||||||
LRWMutex: lsync.NewLRWMutex(ctx),
|
LRWMutex: lsync.NewLRWMutex(ctx),
|
||||||
ref: 1,
|
ref: 1,
|
||||||
}
|
}
|
||||||
n.lockMap[resource] = nsLk
|
|
||||||
} else {
|
} else {
|
||||||
// Update ref count here to avoid multiple races.
|
// Update ref count here to avoid multiple races.
|
||||||
nsLk.ref++
|
atomic.AddUint32(&n.lockMap[resource].ref, 1)
|
||||||
}
|
}
|
||||||
|
nsLk = n.lockMap[resource]
|
||||||
n.lockMapMutex.Unlock()
|
n.lockMapMutex.Unlock()
|
||||||
|
|
||||||
// Locking here will block (until timeout).
|
// Locking here will block (until timeout).
|
||||||
@ -101,13 +99,12 @@ func (n *nsLockMap) lock(ctx context.Context, volume string, path string, lockSo
|
|||||||
if !locked { // We failed to get the lock
|
if !locked { // We failed to get the lock
|
||||||
|
|
||||||
// Decrement ref count since we failed to get the lock
|
// Decrement ref count since we failed to get the lock
|
||||||
n.lockMapMutex.Lock()
|
if atomic.AddUint32(&nsLk.ref, ^uint32(0)) == 0 {
|
||||||
nsLk.ref--
|
|
||||||
if nsLk.ref == 0 {
|
|
||||||
// Remove from the map if there are no more references.
|
// Remove from the map if there are no more references.
|
||||||
|
n.lockMapMutex.Lock()
|
||||||
delete(n.lockMap, resource)
|
delete(n.lockMap, resource)
|
||||||
|
n.lockMapMutex.Unlock()
|
||||||
}
|
}
|
||||||
n.lockMapMutex.Unlock()
|
|
||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
@ -115,28 +112,21 @@ func (n *nsLockMap) lock(ctx context.Context, volume string, path string, lockSo
|
|||||||
// Unlock the namespace resource.
|
// Unlock the namespace resource.
|
||||||
func (n *nsLockMap) unlock(volume string, path string, readLock bool) {
|
func (n *nsLockMap) unlock(volume string, path string, readLock bool) {
|
||||||
resource := pathJoin(volume, path)
|
resource := pathJoin(volume, path)
|
||||||
n.lockMapMutex.RLock()
|
|
||||||
nsLk, found := n.lockMap[resource]
|
n.lockMapMutex.Lock()
|
||||||
n.lockMapMutex.RUnlock()
|
defer n.lockMapMutex.Unlock()
|
||||||
if !found {
|
if _, found := n.lockMap[resource]; !found {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
if readLock {
|
if readLock {
|
||||||
nsLk.RUnlock()
|
n.lockMap[resource].RUnlock()
|
||||||
} else {
|
} else {
|
||||||
nsLk.Unlock()
|
n.lockMap[resource].Unlock()
|
||||||
}
|
}
|
||||||
n.lockMapMutex.Lock()
|
if atomic.AddUint32(&n.lockMap[resource].ref, ^uint32(0)) == 0 {
|
||||||
if nsLk.ref == 0 {
|
// Remove from the map if there are no more references.
|
||||||
logger.LogIf(GlobalContext, errors.New("Namespace reference count cannot be 0"))
|
delete(n.lockMap, resource)
|
||||||
} else {
|
|
||||||
nsLk.ref--
|
|
||||||
if nsLk.ref == 0 {
|
|
||||||
// Remove from the map if there are no more references.
|
|
||||||
delete(n.lockMap, resource)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
n.lockMapMutex.Unlock()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// dsync's distributed lock instance.
|
// dsync's distributed lock instance.
|
||||||
@ -147,7 +137,7 @@ type distLockInstance struct {
|
|||||||
|
|
||||||
// Lock - block until write lock is taken or timeout has occurred.
|
// Lock - block until write lock is taken or timeout has occurred.
|
||||||
func (di *distLockInstance) GetLock(timeout *dynamicTimeout) (timedOutErr error) {
|
func (di *distLockInstance) GetLock(timeout *dynamicTimeout) (timedOutErr error) {
|
||||||
lockSource := getSource()
|
lockSource := getSource(2)
|
||||||
start := UTCNow()
|
start := UTCNow()
|
||||||
|
|
||||||
if !di.rwMutex.GetLock(di.opsID, lockSource, timeout.Timeout()) {
|
if !di.rwMutex.GetLock(di.opsID, lockSource, timeout.Timeout()) {
|
||||||
@ -165,7 +155,7 @@ func (di *distLockInstance) Unlock() {
|
|||||||
|
|
||||||
// RLock - block until read lock is taken or timeout has occurred.
|
// RLock - block until read lock is taken or timeout has occurred.
|
||||||
func (di *distLockInstance) GetRLock(timeout *dynamicTimeout) (timedOutErr error) {
|
func (di *distLockInstance) GetRLock(timeout *dynamicTimeout) (timedOutErr error) {
|
||||||
lockSource := getSource()
|
lockSource := getSource(2)
|
||||||
start := UTCNow()
|
start := UTCNow()
|
||||||
if !di.rwMutex.GetRLock(di.opsID, lockSource, timeout.Timeout()) {
|
if !di.rwMutex.GetRLock(di.opsID, lockSource, timeout.Timeout()) {
|
||||||
timeout.LogFailure()
|
timeout.LogFailure()
|
||||||
@ -206,7 +196,7 @@ func (n *nsLockMap) NewNSLock(ctx context.Context, lockersFn func() []dsync.NetL
|
|||||||
|
|
||||||
// Lock - block until write lock is taken or timeout has occurred.
|
// Lock - block until write lock is taken or timeout has occurred.
|
||||||
func (li *localLockInstance) GetLock(timeout *dynamicTimeout) (timedOutErr error) {
|
func (li *localLockInstance) GetLock(timeout *dynamicTimeout) (timedOutErr error) {
|
||||||
lockSource := getSource()
|
lockSource := getSource(2)
|
||||||
start := UTCNow()
|
start := UTCNow()
|
||||||
readLock := false
|
readLock := false
|
||||||
var success []int
|
var success []int
|
||||||
@ -234,7 +224,7 @@ func (li *localLockInstance) Unlock() {
|
|||||||
|
|
||||||
// RLock - block until read lock is taken or timeout has occurred.
|
// RLock - block until read lock is taken or timeout has occurred.
|
||||||
func (li *localLockInstance) GetRLock(timeout *dynamicTimeout) (timedOutErr error) {
|
func (li *localLockInstance) GetRLock(timeout *dynamicTimeout) (timedOutErr error) {
|
||||||
lockSource := getSource()
|
lockSource := getSource(2)
|
||||||
start := UTCNow()
|
start := UTCNow()
|
||||||
readLock := true
|
readLock := true
|
||||||
var success []int
|
var success []int
|
||||||
@ -260,9 +250,9 @@ func (li *localLockInstance) RUnlock() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func getSource() string {
|
func getSource(n int) string {
|
||||||
var funcName string
|
var funcName string
|
||||||
pc, filename, lineNum, ok := runtime.Caller(2)
|
pc, filename, lineNum, ok := runtime.Caller(n)
|
||||||
if ok {
|
if ok {
|
||||||
filename = pathutil.Base(filename)
|
filename = pathutil.Base(filename)
|
||||||
funcName = strings.TrimPrefix(runtime.FuncForPC(pc).Name(),
|
funcName = strings.TrimPrefix(runtime.FuncForPC(pc).Name(),
|
||||||
|
@ -27,7 +27,7 @@ import (
|
|||||||
// position will cause the line number to change and the test to FAIL
|
// position will cause the line number to change and the test to FAIL
|
||||||
// Tests getSource().
|
// Tests getSource().
|
||||||
func TestGetSource(t *testing.T) {
|
func TestGetSource(t *testing.T) {
|
||||||
currentSource := func() string { return getSource() }
|
currentSource := func() string { return getSource(2) }
|
||||||
gotSource := currentSource()
|
gotSource := currentSource()
|
||||||
// Hard coded line number, 31, in the "expectedSource" value
|
// Hard coded line number, 31, in the "expectedSource" value
|
||||||
expectedSource := "[namespace-lock_test.go:31:TestGetSource()]"
|
expectedSource := "[namespace-lock_test.go:31:TestGetSource()]"
|
||||||
|
@ -58,7 +58,7 @@ type check struct {
|
|||||||
// Assert - checks if gotValue is same as expectedValue, if not fails the test.
|
// Assert - checks if gotValue is same as expectedValue, if not fails the test.
|
||||||
func (c *check) Assert(gotValue interface{}, expectedValue interface{}) {
|
func (c *check) Assert(gotValue interface{}, expectedValue interface{}) {
|
||||||
if !reflect.DeepEqual(gotValue, expectedValue) {
|
if !reflect.DeepEqual(gotValue, expectedValue) {
|
||||||
c.Fatalf("Test %s:%s expected %v, got %v", getSource(), c.testType, expectedValue, gotValue)
|
c.Fatalf("Test %s:%s expected %v, got %v", getSource(2), c.testType, expectedValue, gotValue)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -57,6 +57,7 @@ import (
|
|||||||
"github.com/minio/minio-go/v6/pkg/s3utils"
|
"github.com/minio/minio-go/v6/pkg/s3utils"
|
||||||
"github.com/minio/minio-go/v6/pkg/signer"
|
"github.com/minio/minio-go/v6/pkg/signer"
|
||||||
"github.com/minio/minio/cmd/config"
|
"github.com/minio/minio/cmd/config"
|
||||||
|
"github.com/minio/minio/cmd/crypto"
|
||||||
"github.com/minio/minio/cmd/logger"
|
"github.com/minio/minio/cmd/logger"
|
||||||
"github.com/minio/minio/pkg/auth"
|
"github.com/minio/minio/pkg/auth"
|
||||||
"github.com/minio/minio/pkg/bucket/policy"
|
"github.com/minio/minio/pkg/bucket/policy"
|
||||||
@ -65,6 +66,18 @@ import (
|
|||||||
|
|
||||||
// Tests should initNSLock only once.
|
// Tests should initNSLock only once.
|
||||||
func init() {
|
func init() {
|
||||||
|
// disable ENVs which interfere with tests.
|
||||||
|
for _, env := range []string{
|
||||||
|
crypto.EnvAutoEncryptionLegacy,
|
||||||
|
crypto.EnvKMSAutoEncryption,
|
||||||
|
config.EnvAccessKey,
|
||||||
|
config.EnvAccessKeyOld,
|
||||||
|
config.EnvSecretKey,
|
||||||
|
config.EnvSecretKeyOld,
|
||||||
|
} {
|
||||||
|
os.Unsetenv(env)
|
||||||
|
}
|
||||||
|
|
||||||
// Set as non-distributed.
|
// Set as non-distributed.
|
||||||
globalIsDistXL = false
|
globalIsDistXL = false
|
||||||
|
|
||||||
@ -342,27 +355,9 @@ func UnstartedTestServer(t TestErrHandler, instanceType string) TestServer {
|
|||||||
globalMinioPort = port
|
globalMinioPort = port
|
||||||
globalMinioAddr = getEndpointsLocalAddr(testServer.Disks)
|
globalMinioAddr = getEndpointsLocalAddr(testServer.Disks)
|
||||||
|
|
||||||
globalConfigSys = NewConfigSys()
|
newAllSubsystems()
|
||||||
|
|
||||||
globalIAMSys = NewIAMSys()
|
initAllSubsystems(objLayer)
|
||||||
globalIAMSys.Init(ctx, objLayer)
|
|
||||||
|
|
||||||
buckets, err := objLayer.ListBuckets(ctx)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("Unable to list buckets on backend %s", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
globalPolicySys = NewPolicySys()
|
|
||||||
globalPolicySys.Init(buckets, objLayer)
|
|
||||||
|
|
||||||
globalNotificationSys = NewNotificationSys(testServer.Disks)
|
|
||||||
globalNotificationSys.Init(buckets, objLayer)
|
|
||||||
|
|
||||||
globalLifecycleSys = NewLifecycleSys()
|
|
||||||
globalLifecycleSys.Init(buckets, objLayer)
|
|
||||||
|
|
||||||
globalBucketSSEConfigSys = NewBucketSSEConfigSys()
|
|
||||||
globalBucketSSEConfigSys.Init(buckets, objLayer)
|
|
||||||
|
|
||||||
return testServer
|
return testServer
|
||||||
}
|
}
|
||||||
|
Loading…
x
Reference in New Issue
Block a user