mirror of
https://github.com/minio/minio.git
synced 2024-12-27 15:45:55 -05:00
dd2542e96c
Original work here, #18474, refixed and updated.
197 lines
5.1 KiB
Go
197 lines
5.1 KiB
Go
// Copyright (c) 2015-2022 MinIO, Inc.
|
|
//
|
|
// This file is part of MinIO Object Storage stack
|
|
//
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Affero General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package cmd
|
|
|
|
import (
|
|
"bytes"
|
|
"compress/gzip"
|
|
"context"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"math/rand"
|
|
"net/url"
|
|
"time"
|
|
|
|
"github.com/minio/madmin-go/v3"
|
|
"github.com/minio/minio/internal/logger"
|
|
)
|
|
|
|
var callhomeLeaderLockTimeout = newDynamicTimeout(30*time.Second, 10*time.Second)
|
|
|
|
// initCallhome will start the callhome task in the background.
|
|
func initCallhome(ctx context.Context, objAPI ObjectLayer) {
|
|
if !globalCallhomeConfig.Enabled() {
|
|
return
|
|
}
|
|
|
|
go func() {
|
|
r := rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
// Leader node (that successfully acquires the lock inside runCallhome)
|
|
// will keep performing the callhome. If the leader goes down for some reason,
|
|
// the lock will be released and another node will acquire it and take over
|
|
// because of this loop.
|
|
for {
|
|
if !globalCallhomeConfig.Enabled() {
|
|
return
|
|
}
|
|
|
|
if !runCallhome(ctx, objAPI) {
|
|
// callhome was disabled or context was canceled
|
|
return
|
|
}
|
|
|
|
// callhome running on a different node.
|
|
// sleep for some time and try again.
|
|
duration := time.Duration(r.Float64() * float64(globalCallhomeConfig.FrequencyDur()))
|
|
if duration < time.Second {
|
|
// Make sure to sleep at least a second to avoid high CPU ticks.
|
|
duration = time.Second
|
|
}
|
|
time.Sleep(duration)
|
|
}
|
|
}()
|
|
}
|
|
|
|
func runCallhome(ctx context.Context, objAPI ObjectLayer) bool {
|
|
// Make sure only 1 callhome is running on the cluster.
|
|
locker := objAPI.NewNSLock(minioMetaBucket, "callhome/runCallhome.lock")
|
|
lkctx, err := locker.GetLock(ctx, callhomeLeaderLockTimeout)
|
|
if err != nil {
|
|
// lock timedout means some other node is the leader,
|
|
// cycle back return 'true'
|
|
return true
|
|
}
|
|
|
|
ctx = lkctx.Context()
|
|
defer locker.Unlock(lkctx)
|
|
|
|
callhomeTimer := time.NewTimer(globalCallhomeConfig.FrequencyDur())
|
|
defer callhomeTimer.Stop()
|
|
|
|
for {
|
|
if !globalCallhomeConfig.Enabled() {
|
|
// Stop the processing as callhome got disabled
|
|
return false
|
|
}
|
|
|
|
select {
|
|
case <-ctx.Done():
|
|
// indicates that we do not need to run callhome anymore
|
|
return false
|
|
case <-callhomeTimer.C:
|
|
if !globalCallhomeConfig.Enabled() {
|
|
// Stop the processing as callhome got disabled
|
|
return false
|
|
}
|
|
|
|
performCallhome(ctx)
|
|
|
|
// Reset the timer for next cycle.
|
|
callhomeTimer.Reset(globalCallhomeConfig.FrequencyDur())
|
|
}
|
|
}
|
|
}
|
|
|
|
func performCallhome(ctx context.Context) {
|
|
deadline := 10 * time.Second // Default deadline is 10secs for callhome
|
|
objectAPI := newObjectLayerFn()
|
|
if objectAPI == nil {
|
|
logger.LogIf(ctx, errors.New("Callhome: object layer not ready"))
|
|
return
|
|
}
|
|
|
|
healthCtx, healthCancel := context.WithTimeout(ctx, deadline)
|
|
defer healthCancel()
|
|
|
|
healthInfoCh := make(chan madmin.HealthInfo)
|
|
|
|
query := url.Values{}
|
|
for _, k := range madmin.HealthDataTypesList {
|
|
query.Set(string(k), "true")
|
|
}
|
|
|
|
healthInfo := madmin.HealthInfo{
|
|
TimeStamp: time.Now().UTC(),
|
|
Version: madmin.HealthInfoVersion,
|
|
Minio: madmin.MinioHealthInfo{
|
|
Info: madmin.MinioInfo{
|
|
DeploymentID: globalDeploymentID(),
|
|
},
|
|
},
|
|
}
|
|
|
|
go fetchHealthInfo(healthCtx, objectAPI, &query, healthInfoCh, healthInfo)
|
|
|
|
for {
|
|
select {
|
|
case hi, hasMore := <-healthInfoCh:
|
|
if !hasMore {
|
|
// Received all data. Send to SUBNET and return
|
|
err := sendHealthInfo(ctx, healthInfo)
|
|
if err != nil {
|
|
logger.LogIf(ctx, fmt.Errorf("Unable to perform callhome: %w", err))
|
|
}
|
|
return
|
|
}
|
|
healthInfo = hi
|
|
case <-healthCtx.Done():
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
const (
|
|
subnetHealthPath = "/api/health/upload"
|
|
)
|
|
|
|
func sendHealthInfo(ctx context.Context, healthInfo madmin.HealthInfo) error {
|
|
url := globalSubnetConfig.BaseURL + subnetHealthPath
|
|
|
|
filename := fmt.Sprintf("health_%s.json.gz", UTCNow().Format("20060102150405"))
|
|
url += "?filename=" + filename
|
|
|
|
_, err := globalSubnetConfig.Upload(url, filename, createHealthJSONGzip(ctx, healthInfo))
|
|
return err
|
|
}
|
|
|
|
func createHealthJSONGzip(ctx context.Context, healthInfo madmin.HealthInfo) []byte {
|
|
var b bytes.Buffer
|
|
gzWriter := gzip.NewWriter(&b)
|
|
|
|
header := struct {
|
|
Version string `json:"version"`
|
|
}{Version: healthInfo.Version}
|
|
|
|
enc := json.NewEncoder(gzWriter)
|
|
if e := enc.Encode(header); e != nil {
|
|
logger.LogIf(ctx, fmt.Errorf("Could not encode health info header: %w", e))
|
|
return nil
|
|
}
|
|
|
|
if e := enc.Encode(healthInfo); e != nil {
|
|
logger.LogIf(ctx, fmt.Errorf("Could not encode health info: %w", e))
|
|
return nil
|
|
}
|
|
|
|
gzWriter.Flush()
|
|
gzWriter.Close()
|
|
|
|
return b.Bytes()
|
|
}
|