mirror of
https://github.com/minio/minio.git
synced 2024-12-25 14:45:54 -05:00
200caab82b
This was a regression introduced in '14bb969782' this has the potential to cause corruption when there are concurrent overwrites attempting to update the content on the namespace. This PR adds a situation where PutObject(), CopyObject() compete properly for the same locks with NewMultipartUpload() however it ends up turning off competing locks for the actual object with GetObject() and DeleteObject() - since they do not compete due to concurrent I/O on a versioned bucket it can lead to loss of versions. This PR fixes this bug with multi-pool setup with replication that causes corruption of inlined data due to lack of competing locks in a multi-pool setup. Instead CompleteMultipartUpload holds the necessary locks when finishing the transaction, knowing the exact location of an object to schedule the multipart upload doesn't need to compete in this manner, a pool id location for existing object.
661 lines
18 KiB
Go
661 lines
18 KiB
Go
// Copyright (c) 2015-2021 MinIO, Inc.
|
|
//
|
|
// This file is part of MinIO Object Storage stack
|
|
//
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Affero General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package dsync
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"math/rand"
|
|
"os"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/minio/pkg/console"
|
|
)
|
|
|
|
// Indicator if logging is enabled.
|
|
var dsyncLog bool
|
|
|
|
func init() {
|
|
// Check for MINIO_DSYNC_TRACE env variable, if set logging will be enabled for failed REST operations.
|
|
dsyncLog = os.Getenv("MINIO_DSYNC_TRACE") == "1"
|
|
}
|
|
|
|
func log(format string, data ...interface{}) {
|
|
if dsyncLog {
|
|
console.Printf(format, data...)
|
|
}
|
|
}
|
|
|
|
// dRWMutexAcquireTimeout - tolerance limit to wait for lock acquisition before.
|
|
const drwMutexAcquireTimeout = 1 * time.Second // 1 second.
|
|
|
|
// dRWMutexRefreshTimeout - timeout for the refresh call
|
|
const drwMutexRefreshCallTimeout = 5 * time.Second
|
|
|
|
// dRWMutexUnlockTimeout - timeout for the unlock call
|
|
const drwMutexUnlockCallTimeout = 30 * time.Second
|
|
|
|
// dRWMutexForceUnlockTimeout - timeout for the unlock call
|
|
const drwMutexForceUnlockCallTimeout = 30 * time.Second
|
|
|
|
// dRWMutexRefreshInterval - the interval between two refresh calls
|
|
const drwMutexRefreshInterval = 10 * time.Second
|
|
|
|
const drwMutexInfinite = 1<<63 - 1
|
|
|
|
// A DRWMutex is a distributed mutual exclusion lock.
|
|
type DRWMutex struct {
|
|
Names []string
|
|
writeLocks []string // Array of nodes that granted a write lock
|
|
readersLocks [][]string // Array of array of nodes that granted reader locks
|
|
m sync.Mutex // Mutex to prevent multiple simultaneous locks from this node
|
|
clnt *Dsync
|
|
cancelRefresh context.CancelFunc
|
|
}
|
|
|
|
// Granted - represents a structure of a granted lock.
|
|
type Granted struct {
|
|
index int
|
|
lockUID string // Locked if set with UID string, unlocked if empty
|
|
}
|
|
|
|
func (g *Granted) isLocked() bool {
|
|
return isLocked(g.lockUID)
|
|
}
|
|
|
|
func isLocked(uid string) bool {
|
|
return len(uid) > 0
|
|
}
|
|
|
|
// NewDRWMutex - initializes a new dsync RW mutex.
|
|
func NewDRWMutex(clnt *Dsync, names ...string) *DRWMutex {
|
|
restClnts, _ := clnt.GetLockers()
|
|
return &DRWMutex{
|
|
writeLocks: make([]string, len(restClnts)),
|
|
Names: names,
|
|
clnt: clnt,
|
|
}
|
|
}
|
|
|
|
// Lock holds a write lock on dm.
|
|
//
|
|
// If the lock is already in use, the calling go routine
|
|
// blocks until the mutex is available.
|
|
func (dm *DRWMutex) Lock(id, source string) {
|
|
|
|
isReadLock := false
|
|
dm.lockBlocking(context.Background(), nil, id, source, isReadLock, Options{
|
|
Timeout: drwMutexInfinite,
|
|
})
|
|
}
|
|
|
|
// Options lock options.
|
|
type Options struct {
|
|
Timeout time.Duration
|
|
}
|
|
|
|
// GetLock tries to get a write lock on dm before the timeout elapses.
|
|
//
|
|
// If the lock is already in use, the calling go routine
|
|
// blocks until either the mutex becomes available and return success or
|
|
// more time has passed than the timeout value and return false.
|
|
func (dm *DRWMutex) GetLock(ctx context.Context, cancel context.CancelFunc, id, source string, opts Options) (locked bool) {
|
|
|
|
isReadLock := false
|
|
return dm.lockBlocking(ctx, cancel, id, source, isReadLock, opts)
|
|
}
|
|
|
|
// RLock holds a read lock on dm.
|
|
//
|
|
// If one or more read locks are already in use, it will grant another lock.
|
|
// Otherwise the calling go routine blocks until the mutex is available.
|
|
func (dm *DRWMutex) RLock(id, source string) {
|
|
|
|
isReadLock := true
|
|
dm.lockBlocking(context.Background(), nil, id, source, isReadLock, Options{
|
|
Timeout: drwMutexInfinite,
|
|
})
|
|
}
|
|
|
|
// GetRLock tries to get a read lock on dm before the timeout elapses.
|
|
//
|
|
// If one or more read locks are already in use, it will grant another lock.
|
|
// Otherwise the calling go routine blocks until either the mutex becomes
|
|
// available and return success or more time has passed than the timeout
|
|
// value and return false.
|
|
func (dm *DRWMutex) GetRLock(ctx context.Context, cancel context.CancelFunc, id, source string, opts Options) (locked bool) {
|
|
|
|
isReadLock := true
|
|
return dm.lockBlocking(ctx, cancel, id, source, isReadLock, opts)
|
|
}
|
|
|
|
const (
|
|
lockRetryInterval = 1 * time.Second
|
|
)
|
|
|
|
// lockBlocking will try to acquire either a read or a write lock
|
|
//
|
|
// The function will loop using a built-in timing randomized back-off
|
|
// algorithm until either the lock is acquired successfully or more
|
|
// time has elapsed than the timeout value.
|
|
func (dm *DRWMutex) lockBlocking(ctx context.Context, lockLossCallback func(), id, source string, isReadLock bool, opts Options) (locked bool) {
|
|
restClnts, _ := dm.clnt.GetLockers()
|
|
|
|
r := rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
|
|
// Create lock array to capture the successful lockers
|
|
locks := make([]string, len(restClnts))
|
|
|
|
// Add total timeout
|
|
ctx, cancel := context.WithTimeout(ctx, opts.Timeout)
|
|
defer cancel()
|
|
|
|
// Tolerance is not set, defaults to half of the locker clients.
|
|
tolerance := len(restClnts) / 2
|
|
|
|
// Quorum is effectively = total clients subtracted with tolerance limit
|
|
quorum := len(restClnts) - tolerance
|
|
if !isReadLock {
|
|
// In situations for write locks, as a special case
|
|
// to avoid split brains we make sure to acquire
|
|
// quorum + 1 when tolerance is exactly half of the
|
|
// total locker clients.
|
|
if quorum == tolerance {
|
|
quorum++
|
|
}
|
|
}
|
|
|
|
log("lockBlocking %s/%s for %#v: lockType readLock(%t), additional opts: %#v, quorum: %d, tolerance: %d, lockClients: %d\n", id, source, dm.Names, isReadLock, opts, quorum, tolerance, len(restClnts))
|
|
|
|
tolerance = len(restClnts) - quorum
|
|
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
return false
|
|
default:
|
|
// Try to acquire the lock.
|
|
if locked = lock(ctx, dm.clnt, &locks, id, source, isReadLock, tolerance, quorum, dm.Names...); locked {
|
|
dm.m.Lock()
|
|
|
|
// If success, copy array to object
|
|
if isReadLock {
|
|
// Append new array of strings at the end
|
|
dm.readersLocks = append(dm.readersLocks, make([]string, len(restClnts)))
|
|
// and copy stack array into last spot
|
|
copy(dm.readersLocks[len(dm.readersLocks)-1], locks[:])
|
|
} else {
|
|
copy(dm.writeLocks, locks[:])
|
|
}
|
|
|
|
dm.m.Unlock()
|
|
log("lockBlocking %s/%s for %#v: granted\n", id, source, dm.Names)
|
|
|
|
// Refresh lock continuously and cancel if there is no quorum in the lock anymore
|
|
dm.startContinousLockRefresh(lockLossCallback, id, source, quorum)
|
|
|
|
return locked
|
|
}
|
|
|
|
time.Sleep(time.Duration(r.Float64() * float64(lockRetryInterval)))
|
|
}
|
|
}
|
|
}
|
|
|
|
func (dm *DRWMutex) startContinousLockRefresh(lockLossCallback func(), id, source string, quorum int) {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
dm.m.Lock()
|
|
dm.cancelRefresh = cancel
|
|
dm.m.Unlock()
|
|
|
|
go func() {
|
|
defer cancel()
|
|
|
|
refreshTimer := time.NewTimer(drwMutexRefreshInterval)
|
|
defer refreshTimer.Stop()
|
|
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
return
|
|
case <-refreshTimer.C:
|
|
refreshTimer.Reset(drwMutexRefreshInterval)
|
|
|
|
refreshed, err := refresh(ctx, dm.clnt, id, source, quorum)
|
|
if err == nil && !refreshed {
|
|
// Clean the lock locally and in remote nodes
|
|
forceUnlock(ctx, dm.clnt, id)
|
|
// Execute the caller lock loss callback
|
|
if lockLossCallback != nil {
|
|
lockLossCallback()
|
|
}
|
|
return
|
|
}
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
|
|
func forceUnlock(ctx context.Context, ds *Dsync, id string) {
|
|
ctx, cancel := context.WithTimeout(ctx, drwMutexForceUnlockCallTimeout)
|
|
defer cancel()
|
|
|
|
restClnts, _ := ds.GetLockers()
|
|
|
|
args := LockArgs{
|
|
UID: id,
|
|
}
|
|
|
|
var wg sync.WaitGroup
|
|
for index, c := range restClnts {
|
|
wg.Add(1)
|
|
// Send refresh request to all nodes
|
|
go func(index int, c NetLocker) {
|
|
defer wg.Done()
|
|
c.ForceUnlock(ctx, args)
|
|
}(index, c)
|
|
}
|
|
wg.Wait()
|
|
}
|
|
|
|
type refreshResult struct {
|
|
offline bool
|
|
succeeded bool
|
|
}
|
|
|
|
func refresh(ctx context.Context, ds *Dsync, id, source string, quorum int) (bool, error) {
|
|
restClnts, _ := ds.GetLockers()
|
|
|
|
// Create buffered channel of size equal to total number of nodes.
|
|
ch := make(chan refreshResult, len(restClnts))
|
|
var wg sync.WaitGroup
|
|
|
|
args := LockArgs{
|
|
UID: id,
|
|
}
|
|
|
|
for index, c := range restClnts {
|
|
wg.Add(1)
|
|
// Send refresh request to all nodes
|
|
go func(index int, c NetLocker) {
|
|
defer wg.Done()
|
|
|
|
if c == nil {
|
|
ch <- refreshResult{offline: true}
|
|
return
|
|
}
|
|
|
|
ctx, cancel := context.WithTimeout(ctx, drwMutexRefreshCallTimeout)
|
|
defer cancel()
|
|
|
|
refreshed, err := c.Refresh(ctx, args)
|
|
if refreshed && err == nil {
|
|
ch <- refreshResult{succeeded: true}
|
|
} else {
|
|
if err != nil {
|
|
ch <- refreshResult{offline: true}
|
|
log("dsync: Unable to call Refresh failed with %s for %#v at %s\n", err, args, c)
|
|
} else {
|
|
ch <- refreshResult{succeeded: false}
|
|
log("dsync: Refresh returned false for %#v at %s\n", args, c)
|
|
}
|
|
}
|
|
|
|
}(index, c)
|
|
}
|
|
|
|
// Wait until we have either
|
|
//
|
|
// a) received all refresh responses
|
|
// b) received too many refreshed for quorum to be still possible
|
|
// c) timed out
|
|
//
|
|
i, refreshFailed, refreshSucceeded := 0, 0, 0
|
|
done := false
|
|
|
|
for ; i < len(restClnts); i++ {
|
|
select {
|
|
case refresh := <-ch:
|
|
if refresh.offline {
|
|
continue
|
|
}
|
|
if refresh.succeeded {
|
|
refreshSucceeded++
|
|
} else {
|
|
refreshFailed++
|
|
}
|
|
if refreshFailed > quorum {
|
|
// We know that we are not going to succeed with refresh
|
|
done = true
|
|
}
|
|
case <-ctx.Done():
|
|
// Refreshing is canceled
|
|
return false, ctx.Err()
|
|
}
|
|
|
|
if done {
|
|
break
|
|
}
|
|
}
|
|
|
|
refreshQuorum := refreshSucceeded >= quorum
|
|
if !refreshQuorum {
|
|
refreshQuorum = refreshFailed < quorum
|
|
}
|
|
|
|
// We may have some unused results in ch, release them async.
|
|
go func() {
|
|
wg.Wait()
|
|
close(ch)
|
|
for range ch {
|
|
}
|
|
}()
|
|
|
|
return refreshQuorum, nil
|
|
}
|
|
|
|
// lock tries to acquire the distributed lock, returning true or false.
|
|
func lock(ctx context.Context, ds *Dsync, locks *[]string, id, source string, isReadLock bool, tolerance, quorum int, lockNames ...string) bool {
|
|
for i := range *locks {
|
|
(*locks)[i] = ""
|
|
}
|
|
|
|
restClnts, owner := ds.GetLockers()
|
|
|
|
// Create buffered channel of size equal to total number of nodes.
|
|
ch := make(chan Granted, len(restClnts))
|
|
var wg sync.WaitGroup
|
|
|
|
args := LockArgs{
|
|
Owner: owner,
|
|
UID: id,
|
|
Resources: lockNames,
|
|
Source: source,
|
|
Quorum: quorum,
|
|
}
|
|
|
|
// Combined timeout for the lock attempt.
|
|
ctx, cancel := context.WithTimeout(ctx, drwMutexAcquireTimeout)
|
|
defer cancel()
|
|
for index, c := range restClnts {
|
|
wg.Add(1)
|
|
// broadcast lock request to all nodes
|
|
go func(index int, isReadLock bool, c NetLocker) {
|
|
defer wg.Done()
|
|
|
|
g := Granted{index: index}
|
|
if c == nil {
|
|
log("dsync: nil locker\n")
|
|
ch <- g
|
|
return
|
|
}
|
|
|
|
var locked bool
|
|
var err error
|
|
if isReadLock {
|
|
if locked, err = c.RLock(context.Background(), args); err != nil {
|
|
log("dsync: Unable to call RLock failed with %s for %#v at %s\n", err, args, c)
|
|
}
|
|
} else {
|
|
if locked, err = c.Lock(context.Background(), args); err != nil {
|
|
log("dsync: Unable to call Lock failed with %s for %#v at %s\n", err, args, c)
|
|
}
|
|
}
|
|
if locked {
|
|
g.lockUID = args.UID
|
|
}
|
|
ch <- g
|
|
|
|
}(index, isReadLock, c)
|
|
}
|
|
|
|
// Wait until we have either
|
|
//
|
|
// a) received all lock responses
|
|
// b) received too many 'non-'locks for quorum to be still possible
|
|
// c) timed out
|
|
//
|
|
i, locksFailed := 0, 0
|
|
done := false
|
|
|
|
for ; i < len(restClnts); i++ { // Loop until we acquired all locks
|
|
select {
|
|
case grant := <-ch:
|
|
if grant.isLocked() {
|
|
// Mark that this node has acquired the lock
|
|
(*locks)[grant.index] = grant.lockUID
|
|
} else {
|
|
locksFailed++
|
|
if locksFailed > tolerance {
|
|
// We know that we are not going to get the lock anymore,
|
|
// so exit out and release any locks that did get acquired
|
|
done = true
|
|
}
|
|
}
|
|
case <-ctx.Done():
|
|
// Capture timedout locks as failed or took too long
|
|
locksFailed++
|
|
if locksFailed > tolerance {
|
|
// We know that we are not going to get the lock anymore,
|
|
// so exit out and release any locks that did get acquired
|
|
done = true
|
|
}
|
|
}
|
|
|
|
if done {
|
|
break
|
|
}
|
|
}
|
|
|
|
quorumLocked := checkQuorumLocked(locks, quorum) && locksFailed <= tolerance
|
|
if !quorumLocked {
|
|
log("dsync: Unable to acquire lock in quorum %#v\n", args)
|
|
// Release all acquired locks without quorum.
|
|
if !releaseAll(ds, tolerance, owner, locks, isReadLock, restClnts, lockNames...) {
|
|
log("Unable to release acquired locks, these locks will expire automatically %#v\n", args)
|
|
}
|
|
}
|
|
|
|
// We may have some unused results in ch, release them async.
|
|
go func() {
|
|
wg.Wait()
|
|
close(ch)
|
|
for grantToBeReleased := range ch {
|
|
if grantToBeReleased.isLocked() {
|
|
// release abandoned lock
|
|
log("Releasing abandoned lock\n")
|
|
sendRelease(ds, restClnts[grantToBeReleased.index],
|
|
owner, grantToBeReleased.lockUID, isReadLock, lockNames...)
|
|
}
|
|
}
|
|
}()
|
|
|
|
return quorumLocked
|
|
}
|
|
|
|
// checkFailedUnlocks determines whether we have sufficiently unlocked all
|
|
// resources to ensure no deadlocks for future callers
|
|
func checkFailedUnlocks(locks []string, tolerance int) bool {
|
|
unlocksFailed := 0
|
|
for lockID := range locks {
|
|
if isLocked(locks[lockID]) {
|
|
unlocksFailed++
|
|
}
|
|
}
|
|
|
|
// Unlock failures are higher than tolerance limit
|
|
// for this instance of unlocker, we should let the
|
|
// caller know that lock is not successfully released
|
|
// yet.
|
|
if len(locks)-tolerance == tolerance {
|
|
// Incase of split brain scenarios where
|
|
// tolerance is exactly half of the len(*locks)
|
|
// then we need to make sure we have unlocked
|
|
// upto tolerance+1 - especially for RUnlock
|
|
// to ensure that we don't end up with active
|
|
// read locks on the resource after unlocking
|
|
// only half of the lockers.
|
|
return unlocksFailed >= tolerance
|
|
}
|
|
return unlocksFailed > tolerance
|
|
}
|
|
|
|
// checkQuorumLocked determines whether we have locked the required quorum of underlying locks or not
|
|
func checkQuorumLocked(locks *[]string, quorum int) bool {
|
|
count := 0
|
|
for _, uid := range *locks {
|
|
if isLocked(uid) {
|
|
count++
|
|
}
|
|
}
|
|
|
|
return count >= quorum
|
|
}
|
|
|
|
// releaseAll releases all locks that are marked as locked
|
|
func releaseAll(ds *Dsync, tolerance int, owner string, locks *[]string, isReadLock bool, restClnts []NetLocker, lockNames ...string) bool {
|
|
var wg sync.WaitGroup
|
|
for lockID := range restClnts {
|
|
wg.Add(1)
|
|
go func(lockID int) {
|
|
defer wg.Done()
|
|
if isLocked((*locks)[lockID]) {
|
|
if sendRelease(ds, restClnts[lockID], owner, (*locks)[lockID], isReadLock, lockNames...) {
|
|
(*locks)[lockID] = ""
|
|
}
|
|
}
|
|
}(lockID)
|
|
}
|
|
wg.Wait()
|
|
|
|
// Return true if releaseAll was successful, otherwise we return 'false'
|
|
// to indicate we haven't sufficiently unlocked lockers to avoid deadlocks.
|
|
//
|
|
// Caller may use this as an indication to call again.
|
|
return !checkFailedUnlocks(*locks, tolerance)
|
|
}
|
|
|
|
// Unlock unlocks the write lock.
|
|
//
|
|
// It is a run-time error if dm is not locked on entry to Unlock.
|
|
func (dm *DRWMutex) Unlock() {
|
|
dm.m.Lock()
|
|
dm.cancelRefresh()
|
|
dm.m.Unlock()
|
|
|
|
restClnts, owner := dm.clnt.GetLockers()
|
|
// create temp array on stack
|
|
locks := make([]string, len(restClnts))
|
|
|
|
{
|
|
dm.m.Lock()
|
|
defer dm.m.Unlock()
|
|
|
|
// Check if minimally a single bool is set in the writeLocks array
|
|
lockFound := false
|
|
for _, uid := range dm.writeLocks {
|
|
if isLocked(uid) {
|
|
lockFound = true
|
|
break
|
|
}
|
|
}
|
|
if !lockFound {
|
|
panic("Trying to Unlock() while no Lock() is active")
|
|
}
|
|
|
|
// Copy write locks to stack array
|
|
copy(locks, dm.writeLocks[:])
|
|
}
|
|
|
|
// Tolerance is not set, defaults to half of the locker clients.
|
|
tolerance := len(restClnts) / 2
|
|
|
|
isReadLock := false
|
|
r := rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
for !releaseAll(dm.clnt, tolerance, owner, &locks, isReadLock, restClnts, dm.Names...) {
|
|
time.Sleep(time.Duration(r.Float64() * float64(lockRetryInterval)))
|
|
}
|
|
}
|
|
|
|
// RUnlock releases a read lock held on dm.
|
|
//
|
|
// It is a run-time error if dm is not locked on entry to RUnlock.
|
|
func (dm *DRWMutex) RUnlock() {
|
|
dm.m.Lock()
|
|
dm.cancelRefresh()
|
|
dm.m.Unlock()
|
|
|
|
// create temp array on stack
|
|
restClnts, owner := dm.clnt.GetLockers()
|
|
|
|
locks := make([]string, len(restClnts))
|
|
{
|
|
dm.m.Lock()
|
|
defer dm.m.Unlock()
|
|
if len(dm.readersLocks) == 0 {
|
|
panic("Trying to RUnlock() while no RLock() is active")
|
|
}
|
|
// Copy out first element to release it first (FIFO)
|
|
copy(locks, dm.readersLocks[0][:])
|
|
// Drop first element from array
|
|
dm.readersLocks = dm.readersLocks[1:]
|
|
}
|
|
|
|
// Tolerance is not set, defaults to half of the locker clients.
|
|
tolerance := len(restClnts) / 2
|
|
|
|
isReadLock := true
|
|
r := rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
for !releaseAll(dm.clnt, tolerance, owner, &locks, isReadLock, restClnts, dm.Names...) {
|
|
time.Sleep(time.Duration(r.Float64() * float64(lockRetryInterval)))
|
|
}
|
|
}
|
|
|
|
// sendRelease sends a release message to a node that previously granted a lock
|
|
func sendRelease(ds *Dsync, c NetLocker, owner string, uid string, isReadLock bool, names ...string) bool {
|
|
if c == nil {
|
|
log("Unable to call RUnlock failed with %s\n", errors.New("netLocker is offline"))
|
|
return false
|
|
}
|
|
|
|
args := LockArgs{
|
|
Owner: owner,
|
|
UID: uid,
|
|
Resources: names,
|
|
}
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), drwMutexUnlockCallTimeout)
|
|
defer cancel()
|
|
|
|
if isReadLock {
|
|
if _, err := c.RUnlock(ctx, args); err != nil {
|
|
log("dsync: Unable to call RUnlock failed with %s for %#v at %s\n", err, args, c)
|
|
return false
|
|
}
|
|
} else {
|
|
if _, err := c.Unlock(ctx, args); err != nil {
|
|
log("dsync: Unable to call Unlock failed with %s for %#v at %s\n", err, args, c)
|
|
return false
|
|
}
|
|
}
|
|
|
|
return true
|
|
}
|