mirror of
https://github.com/minio/minio.git
synced 2024-12-26 07:05:55 -05:00
replication: limit larger uploads to a subset of workers (#17687)
Limit large uploads (> 128MiB) to a max of 10 workers, intent is to avoid larger uploads from using all replication bandwidth, giving room for smaller uploads to sync faster.
This commit is contained in:
parent
e7b60c4d65
commit
1a42693d68
@ -1645,6 +1645,7 @@ type ReplicationPool struct {
|
|||||||
|
|
||||||
// workers:
|
// workers:
|
||||||
workers []chan ReplicationWorkerOperation
|
workers []chan ReplicationWorkerOperation
|
||||||
|
lrgworkers []chan ReplicationWorkerOperation
|
||||||
existingWorkers chan ReplicationWorkerOperation
|
existingWorkers chan ReplicationWorkerOperation
|
||||||
|
|
||||||
// mrf:
|
// mrf:
|
||||||
@ -1679,6 +1680,9 @@ const (
|
|||||||
|
|
||||||
// MRFWorkerAutoDefault is default number of mrf workers for "auto" mode
|
// MRFWorkerAutoDefault is default number of mrf workers for "auto" mode
|
||||||
MRFWorkerAutoDefault = 4
|
MRFWorkerAutoDefault = 4
|
||||||
|
|
||||||
|
// LargeWorkerCount is default number of workers assigned to large uploads ( >= 128MiB)
|
||||||
|
LargeWorkerCount = 10
|
||||||
)
|
)
|
||||||
|
|
||||||
// NewReplicationPool creates a pool of replication workers of specified size
|
// NewReplicationPool creates a pool of replication workers of specified size
|
||||||
@ -1702,7 +1706,9 @@ func NewReplicationPool(ctx context.Context, o ObjectLayer, opts replicationPool
|
|||||||
|
|
||||||
pool := &ReplicationPool{
|
pool := &ReplicationPool{
|
||||||
workers: make([]chan ReplicationWorkerOperation, 0, workers),
|
workers: make([]chan ReplicationWorkerOperation, 0, workers),
|
||||||
|
lrgworkers: make([]chan ReplicationWorkerOperation, 0, LargeWorkerCount),
|
||||||
existingWorkers: make(chan ReplicationWorkerOperation, 100000),
|
existingWorkers: make(chan ReplicationWorkerOperation, 100000),
|
||||||
|
|
||||||
mrfReplicaCh: make(chan ReplicationWorkerOperation, 100000),
|
mrfReplicaCh: make(chan ReplicationWorkerOperation, 100000),
|
||||||
mrfWorkerKillCh: make(chan struct{}, failedWorkers),
|
mrfWorkerKillCh: make(chan struct{}, failedWorkers),
|
||||||
resyncer: newresyncer(),
|
resyncer: newresyncer(),
|
||||||
@ -1714,6 +1720,7 @@ func NewReplicationPool(ctx context.Context, o ObjectLayer, opts replicationPool
|
|||||||
priority: priority,
|
priority: priority,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pool.AddLargeWorkers()
|
||||||
pool.ResizeWorkers(workers, 0)
|
pool.ResizeWorkers(workers, 0)
|
||||||
pool.ResizeFailedWorkers(failedWorkers)
|
pool.ResizeFailedWorkers(failedWorkers)
|
||||||
go pool.AddWorker(pool.existingWorkers, nil)
|
go pool.AddWorker(pool.existingWorkers, nil)
|
||||||
@ -1785,6 +1792,19 @@ func (p *ReplicationPool) AddWorker(input <-chan ReplicationWorkerOperation, opT
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// AddLargeWorkers adds a static number of workers to handle large uploads
|
||||||
|
func (p *ReplicationPool) AddLargeWorkers() {
|
||||||
|
for i := 0; i < LargeWorkerCount; i++ {
|
||||||
|
p.lrgworkers = append(p.lrgworkers, make(chan ReplicationWorkerOperation, 100000))
|
||||||
|
}
|
||||||
|
go func() {
|
||||||
|
<-p.ctx.Done()
|
||||||
|
for i := 0; i < LargeWorkerCount; i++ {
|
||||||
|
close(p.lrgworkers[i])
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
// ActiveWorkers returns the number of active workers handling replication traffic.
|
// ActiveWorkers returns the number of active workers handling replication traffic.
|
||||||
func (p *ReplicationPool) ActiveWorkers() int {
|
func (p *ReplicationPool) ActiveWorkers() int {
|
||||||
return int(atomic.LoadInt32(&p.activeWorkers))
|
return int(atomic.LoadInt32(&p.activeWorkers))
|
||||||
@ -1861,9 +1881,14 @@ func (p *ReplicationPool) ResizeFailedWorkers(n int) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const (
|
||||||
|
minLargeObjSize = 128 * humanize.MiByte // 128MiB
|
||||||
|
)
|
||||||
|
|
||||||
// getWorkerCh gets a worker channel deterministically based on bucket and object names.
|
// getWorkerCh gets a worker channel deterministically based on bucket and object names.
|
||||||
// Must be able to grab read lock from p.
|
// Must be able to grab read lock from p.
|
||||||
func (p *ReplicationPool) getWorkerCh(bucket, object string) chan<- ReplicationWorkerOperation {
|
|
||||||
|
func (p *ReplicationPool) getWorkerCh(bucket, object string, sz int64) chan<- ReplicationWorkerOperation {
|
||||||
h := xxh3.HashString(bucket + object)
|
h := xxh3.HashString(bucket + object)
|
||||||
p.mu.RLock()
|
p.mu.RLock()
|
||||||
defer p.mu.RUnlock()
|
defer p.mu.RUnlock()
|
||||||
@ -1877,15 +1902,27 @@ func (p *ReplicationPool) queueReplicaTask(ri ReplicateObjectInfo) {
|
|||||||
if p == nil {
|
if p == nil {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
// if object is large, queue it to a static set of large workers
|
||||||
|
if ri.Size >= int64(minLargeObjSize) {
|
||||||
|
h := xxh3.HashString(ri.Bucket + ri.Name)
|
||||||
|
select {
|
||||||
|
case <-p.ctx.Done():
|
||||||
|
case p.lrgworkers[h%LargeWorkerCount] <- ri:
|
||||||
|
default:
|
||||||
|
globalReplicationPool.queueMRFSave(ri.ToMRFEntry())
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
var ch, healCh chan<- ReplicationWorkerOperation
|
var ch, healCh chan<- ReplicationWorkerOperation
|
||||||
switch ri.OpType {
|
switch ri.OpType {
|
||||||
case replication.ExistingObjectReplicationType:
|
case replication.ExistingObjectReplicationType:
|
||||||
ch = p.existingWorkers
|
ch = p.existingWorkers
|
||||||
case replication.HealReplicationType:
|
case replication.HealReplicationType:
|
||||||
ch = p.mrfReplicaCh
|
ch = p.mrfReplicaCh
|
||||||
healCh = p.getWorkerCh(ri.Name, ri.Bucket)
|
healCh = p.getWorkerCh(ri.Name, ri.Bucket, ri.Size)
|
||||||
default:
|
default:
|
||||||
ch = p.getWorkerCh(ri.Name, ri.Bucket)
|
ch = p.getWorkerCh(ri.Name, ri.Bucket, ri.Size)
|
||||||
}
|
}
|
||||||
if ch == nil && healCh == nil {
|
if ch == nil && healCh == nil {
|
||||||
return
|
return
|
||||||
@ -1945,7 +1982,7 @@ func (p *ReplicationPool) queueReplicaDeleteTask(doi DeletedObjectReplicationInf
|
|||||||
case replication.HealReplicationType:
|
case replication.HealReplicationType:
|
||||||
fallthrough
|
fallthrough
|
||||||
default:
|
default:
|
||||||
ch = p.getWorkerCh(doi.Bucket, doi.ObjectName)
|
ch = p.getWorkerCh(doi.Bucket, doi.ObjectName, 0)
|
||||||
}
|
}
|
||||||
|
|
||||||
select {
|
select {
|
||||||
|
Loading…
Reference in New Issue
Block a user