mirror of
https://github.com/minio/minio.git
synced 2025-01-11 15:03:22 -05:00
Support persistent queue store for loggers (#17121)
This commit is contained in:
parent
42fb3cd95e
commit
57acacd5a7
@ -2511,13 +2511,13 @@ func fetchLoggerInfo() ([]madmin.Logger, []madmin.Audit) {
|
|||||||
var auditloggerInfo []madmin.Audit
|
var auditloggerInfo []madmin.Audit
|
||||||
for _, tgt := range logger.SystemTargets() {
|
for _, tgt := range logger.SystemTargets() {
|
||||||
if tgt.Endpoint() != "" {
|
if tgt.Endpoint() != "" {
|
||||||
loggerInfo = append(loggerInfo, madmin.Logger{tgt.String(): logger.TargetStatus(tgt)})
|
loggerInfo = append(loggerInfo, madmin.Logger{tgt.String(): logger.TargetStatus(GlobalContext, tgt)})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tgt := range logger.AuditTargets() {
|
for _, tgt := range logger.AuditTargets() {
|
||||||
if tgt.Endpoint() != "" {
|
if tgt.Endpoint() != "" {
|
||||||
auditloggerInfo = append(auditloggerInfo, madmin.Audit{tgt.String(): logger.TargetStatus(tgt)})
|
auditloggerInfo = append(auditloggerInfo, madmin.Audit{tgt.String(): logger.TargetStatus(GlobalContext, tgt)})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -588,7 +588,7 @@ func applyDynamicConfigForSubSys(ctx context.Context, objAPI ObjectLayer, s conf
|
|||||||
}
|
}
|
||||||
loggerCfg.HTTP[n] = l
|
loggerCfg.HTTP[n] = l
|
||||||
}
|
}
|
||||||
if errs := logger.UpdateSystemTargets(loggerCfg); len(errs) > 0 {
|
if errs := logger.UpdateSystemTargets(ctx, loggerCfg); len(errs) > 0 {
|
||||||
logger.LogIf(ctx, fmt.Errorf("Unable to update logger webhook config: %v", errs))
|
logger.LogIf(ctx, fmt.Errorf("Unable to update logger webhook config: %v", errs))
|
||||||
}
|
}
|
||||||
case config.AuditWebhookSubSys:
|
case config.AuditWebhookSubSys:
|
||||||
@ -606,7 +606,7 @@ func applyDynamicConfigForSubSys(ctx context.Context, objAPI ObjectLayer, s conf
|
|||||||
loggerCfg.AuditWebhook[n] = l
|
loggerCfg.AuditWebhook[n] = l
|
||||||
}
|
}
|
||||||
|
|
||||||
if errs := logger.UpdateAuditWebhookTargets(loggerCfg); len(errs) > 0 {
|
if errs := logger.UpdateAuditWebhookTargets(ctx, loggerCfg); len(errs) > 0 {
|
||||||
logger.LogIf(ctx, fmt.Errorf("Unable to update audit webhook targets: %v", errs))
|
logger.LogIf(ctx, fmt.Errorf("Unable to update audit webhook targets: %v", errs))
|
||||||
}
|
}
|
||||||
case config.AuditKafkaSubSys:
|
case config.AuditKafkaSubSys:
|
||||||
@ -620,7 +620,7 @@ func applyDynamicConfigForSubSys(ctx context.Context, objAPI ObjectLayer, s conf
|
|||||||
loggerCfg.AuditKafka[n] = l
|
loggerCfg.AuditKafka[n] = l
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if errs := logger.UpdateAuditKafkaTargets(loggerCfg); len(errs) > 0 {
|
if errs := logger.UpdateAuditKafkaTargets(ctx, loggerCfg); len(errs) > 0 {
|
||||||
logger.LogIf(ctx, fmt.Errorf("Unable to update audit kafka targets: %v", errs))
|
logger.LogIf(ctx, fmt.Errorf("Unable to update audit kafka targets: %v", errs))
|
||||||
}
|
}
|
||||||
case config.StorageClassSubSys:
|
case config.StorageClassSubSys:
|
||||||
|
@ -58,7 +58,7 @@ func NewConsoleLogger(ctx context.Context) *HTTPConsoleLoggerSys {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// IsOnline always true in case of console logger
|
// IsOnline always true in case of console logger
|
||||||
func (sys *HTTPConsoleLoggerSys) IsOnline() bool {
|
func (sys *HTTPConsoleLoggerSys) IsOnline(_ context.Context) bool {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -87,7 +87,7 @@ func (sys *HTTPConsoleLoggerSys) HasLogListeners() bool {
|
|||||||
func (sys *HTTPConsoleLoggerSys) Subscribe(subCh chan log.Info, doneCh <-chan struct{}, node string, last int, logKind madmin.LogMask, filter func(entry log.Info) bool) error {
|
func (sys *HTTPConsoleLoggerSys) Subscribe(subCh chan log.Info, doneCh <-chan struct{}, node string, last int, logKind madmin.LogMask, filter func(entry log.Info) bool) error {
|
||||||
// Enable console logging for remote client.
|
// Enable console logging for remote client.
|
||||||
if !sys.HasLogListeners() {
|
if !sys.HasLogListeners() {
|
||||||
logger.AddSystemTarget(sys)
|
logger.AddSystemTarget(GlobalContext, sys)
|
||||||
}
|
}
|
||||||
|
|
||||||
cnt := 0
|
cnt := 0
|
||||||
@ -128,7 +128,7 @@ func (sys *HTTPConsoleLoggerSys) Subscribe(subCh chan log.Info, doneCh <-chan st
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Init if HTTPConsoleLoggerSys is valid, always returns nil right now
|
// Init if HTTPConsoleLoggerSys is valid, always returns nil right now
|
||||||
func (sys *HTTPConsoleLoggerSys) Init() error {
|
func (sys *HTTPConsoleLoggerSys) Init(_ context.Context) error {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -180,7 +180,7 @@ func (sys *HTTPConsoleLoggerSys) Type() types.TargetType {
|
|||||||
|
|
||||||
// Send log message 'e' to console and publish to console
|
// Send log message 'e' to console and publish to console
|
||||||
// log pubsub system
|
// log pubsub system
|
||||||
func (sys *HTTPConsoleLoggerSys) Send(entry interface{}) error {
|
func (sys *HTTPConsoleLoggerSys) Send(ctx context.Context, entry interface{}) error {
|
||||||
var lg log.Info
|
var lg log.Info
|
||||||
switch e := entry.(type) {
|
switch e := entry.(type) {
|
||||||
case log.Entry:
|
case log.Entry:
|
||||||
|
@ -516,7 +516,7 @@ func serverMain(ctx *cli.Context) {
|
|||||||
|
|
||||||
// Initialize globalConsoleSys system
|
// Initialize globalConsoleSys system
|
||||||
globalConsoleSys = NewConsoleLogger(GlobalContext)
|
globalConsoleSys = NewConsoleLogger(GlobalContext)
|
||||||
logger.AddSystemTarget(globalConsoleSys)
|
logger.AddSystemTarget(GlobalContext, globalConsoleSys)
|
||||||
|
|
||||||
// Perform any self-tests
|
// Perform any self-tests
|
||||||
bitrotSelfTest()
|
bitrotSelfTest()
|
||||||
|
@ -42,7 +42,7 @@ func printStartupMessage(apiEndpoints []string, err error) {
|
|||||||
logger.Info(color.Bold("MinIO Object Storage Server"))
|
logger.Info(color.Bold("MinIO Object Storage Server"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if globalConsoleSys != nil {
|
if globalConsoleSys != nil {
|
||||||
globalConsoleSys.Send(fmt.Sprintf("Server startup failed with '%v', some features may be missing", err))
|
globalConsoleSys.Send(GlobalContext, fmt.Sprintf("Server startup failed with '%v', some features may be missing", err))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -30,6 +30,7 @@ import (
|
|||||||
|
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
"github.com/rabbitmq/amqp091-go"
|
"github.com/rabbitmq/amqp091-go"
|
||||||
@ -112,7 +113,7 @@ func (a *AMQPArgs) Validate() error {
|
|||||||
|
|
||||||
// AMQPTarget - AMQP target
|
// AMQPTarget - AMQP target
|
||||||
type AMQPTarget struct {
|
type AMQPTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args AMQPArgs
|
args AMQPArgs
|
||||||
@ -289,8 +290,8 @@ func (target *AMQPTarget) Save(eventData event.Event) error {
|
|||||||
return target.send(eventData, ch, confirms)
|
return target.send(eventData, ch, confirms)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - sends event to AMQP091.
|
// SendFromStore - reads an event from store and sends it to AMQP091.
|
||||||
func (target *AMQPTarget) Send(eventKey string) error {
|
func (target *AMQPTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -329,7 +330,7 @@ func (target *AMQPTarget) Close() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *AMQPTarget) init() error {
|
func (target *AMQPTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initAMQP)
|
return target.initOnce.Do(target.initAMQP)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (target *AMQPTarget) initAMQP() error {
|
func (target *AMQPTarget) initAMQP() error {
|
||||||
|
@ -36,6 +36,7 @@ import (
|
|||||||
"github.com/minio/highwayhash"
|
"github.com/minio/highwayhash"
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
"github.com/pkg/errors"
|
"github.com/pkg/errors"
|
||||||
@ -153,7 +154,7 @@ func (a ElasticsearchArgs) Validate() error {
|
|||||||
|
|
||||||
// ElasticsearchTarget - Elasticsearch target.
|
// ElasticsearchTarget - Elasticsearch target.
|
||||||
type ElasticsearchTarget struct {
|
type ElasticsearchTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args ElasticsearchArgs
|
args ElasticsearchArgs
|
||||||
@ -263,8 +264,8 @@ func (target *ElasticsearchTarget) send(eventData event.Event) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - reads an event from store and sends it to Elasticsearch.
|
// SendFromStore - reads an event from store and sends it to Elasticsearch.
|
||||||
func (target *ElasticsearchTarget) Send(eventKey string) error {
|
func (target *ElasticsearchTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -344,7 +345,7 @@ func (target *ElasticsearchTarget) checkAndInitClient(ctx context.Context) error
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *ElasticsearchTarget) init() error {
|
func (target *ElasticsearchTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initElasticsearch)
|
return target.initOnce.Do(target.initElasticsearch)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (target *ElasticsearchTarget) initElasticsearch() error {
|
func (target *ElasticsearchTarget) initElasticsearch() error {
|
||||||
|
@ -32,6 +32,7 @@ import (
|
|||||||
|
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
|
|
||||||
@ -125,7 +126,7 @@ func (k KafkaArgs) Validate() error {
|
|||||||
|
|
||||||
// KafkaTarget - Kafka target.
|
// KafkaTarget - Kafka target.
|
||||||
type KafkaTarget struct {
|
type KafkaTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args KafkaArgs
|
args KafkaArgs
|
||||||
@ -208,8 +209,8 @@ func (target *KafkaTarget) send(eventData event.Event) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - reads an event from store and sends it to Kafka.
|
// SendFromStore - reads an event from store and sends it to Kafka.
|
||||||
func (target *KafkaTarget) Send(eventKey string) error {
|
func (target *KafkaTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -278,7 +279,7 @@ func (k KafkaArgs) pingBrokers() bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *KafkaTarget) init() error {
|
func (target *KafkaTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initKafka)
|
return target.initOnce.Do(target.initKafka)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (target *KafkaTarget) initKafka() error {
|
func (target *KafkaTarget) initKafka() error {
|
||||||
|
@ -31,6 +31,7 @@ import (
|
|||||||
mqtt "github.com/eclipse/paho.mqtt.golang"
|
mqtt "github.com/eclipse/paho.mqtt.golang"
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
)
|
)
|
||||||
@ -107,7 +108,7 @@ func (m MQTTArgs) Validate() error {
|
|||||||
|
|
||||||
// MQTTTarget - MQTT target.
|
// MQTTTarget - MQTT target.
|
||||||
type MQTTTarget struct {
|
type MQTTTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args MQTTArgs
|
args MQTTArgs
|
||||||
@ -167,8 +168,8 @@ func (target *MQTTTarget) send(eventData event.Event) error {
|
|||||||
return token.Error()
|
return token.Error()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - reads an event from store and sends it to MQTT.
|
// SendFromStore - reads an event from store and sends it to MQTT.
|
||||||
func (target *MQTTTarget) Send(eventKey string) error {
|
func (target *MQTTTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -224,7 +225,7 @@ func (target *MQTTTarget) Close() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *MQTTTarget) init() error {
|
func (target *MQTTTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initMQTT)
|
return target.initOnce.Do(target.initMQTT)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (target *MQTTTarget) initMQTT() error {
|
func (target *MQTTTarget) initMQTT() error {
|
||||||
|
@ -33,6 +33,7 @@ import (
|
|||||||
"github.com/go-sql-driver/mysql"
|
"github.com/go-sql-driver/mysql"
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
)
|
)
|
||||||
@ -146,7 +147,7 @@ func (m MySQLArgs) Validate() error {
|
|||||||
|
|
||||||
// MySQLTarget - MySQL target.
|
// MySQLTarget - MySQL target.
|
||||||
type MySQLTarget struct {
|
type MySQLTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args MySQLArgs
|
args MySQLArgs
|
||||||
@ -252,8 +253,8 @@ func (target *MySQLTarget) send(eventData event.Event) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - reads an event from store and sends it to MySQL.
|
// SendFromStore - reads an event from store and sends it to MySQL.
|
||||||
func (target *MySQLTarget) Send(eventKey string) error {
|
func (target *MySQLTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -349,7 +350,7 @@ func (target *MySQLTarget) executeStmts() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *MySQLTarget) init() error {
|
func (target *MySQLTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initMySQL)
|
return target.initOnce.Do(target.initMySQL)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (target *MySQLTarget) initMySQL() error {
|
func (target *MySQLTarget) initMySQL() error {
|
||||||
|
@ -31,6 +31,7 @@ import (
|
|||||||
"github.com/google/uuid"
|
"github.com/google/uuid"
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
"github.com/nats-io/nats.go"
|
"github.com/nats-io/nats.go"
|
||||||
@ -216,7 +217,7 @@ func (n NATSArgs) connectStan() (stan.Conn, error) {
|
|||||||
|
|
||||||
// NATSTarget - NATS target.
|
// NATSTarget - NATS target.
|
||||||
type NATSTarget struct {
|
type NATSTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args NATSArgs
|
args NATSArgs
|
||||||
@ -333,8 +334,8 @@ func (target *NATSTarget) send(eventData event.Event) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - sends event to Nats.
|
// SendFromStore - reads an event from store and sends it to Nats.
|
||||||
func (target *NATSTarget) Send(eventKey string) error {
|
func (target *NATSTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -380,7 +381,7 @@ func (target *NATSTarget) Close() (err error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *NATSTarget) init() error {
|
func (target *NATSTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initNATS)
|
return target.initOnce.Do(target.initNATS)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (target *NATSTarget) initNATS() error {
|
func (target *NATSTarget) initNATS() error {
|
||||||
|
@ -31,6 +31,7 @@ import (
|
|||||||
|
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
)
|
)
|
||||||
@ -90,7 +91,7 @@ func (n NSQArgs) Validate() error {
|
|||||||
|
|
||||||
// NSQTarget - NSQ target.
|
// NSQTarget - NSQ target.
|
||||||
type NSQTarget struct {
|
type NSQTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args NSQArgs
|
args NSQArgs
|
||||||
@ -176,8 +177,8 @@ func (target *NSQTarget) send(eventData event.Event) error {
|
|||||||
return target.producer.Publish(target.args.Topic, data)
|
return target.producer.Publish(target.args.Topic, data)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - reads an event from store and sends it to NSQ.
|
// SendFromStore - reads an event from store and sends it to NSQ.
|
||||||
func (target *NSQTarget) Send(eventKey string) error {
|
func (target *NSQTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -216,7 +217,7 @@ func (target *NSQTarget) Close() (err error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *NSQTarget) init() error {
|
func (target *NSQTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initNSQ)
|
return target.initOnce.Do(target.initNSQ)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (target *NSQTarget) initNSQ() error {
|
func (target *NSQTarget) initNSQ() error {
|
||||||
|
@ -34,6 +34,7 @@ import (
|
|||||||
|
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
)
|
)
|
||||||
@ -138,7 +139,7 @@ func (p PostgreSQLArgs) Validate() error {
|
|||||||
|
|
||||||
// PostgreSQLTarget - PostgreSQL target.
|
// PostgreSQLTarget - PostgreSQL target.
|
||||||
type PostgreSQLTarget struct {
|
type PostgreSQLTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args PostgreSQLArgs
|
args PostgreSQLArgs
|
||||||
@ -249,8 +250,8 @@ func (target *PostgreSQLTarget) send(eventData event.Event) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - reads an event from store and sends it to PostgreSQL.
|
// SendFromStore - reads an event from store and sends it to PostgreSQL.
|
||||||
func (target *PostgreSQLTarget) Send(eventKey string) error {
|
func (target *PostgreSQLTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -345,7 +346,7 @@ func (target *PostgreSQLTarget) executeStmts() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *PostgreSQLTarget) init() error {
|
func (target *PostgreSQLTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initPostgreSQL)
|
return target.initOnce.Do(target.initPostgreSQL)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (target *PostgreSQLTarget) initPostgreSQL() error {
|
func (target *PostgreSQLTarget) initPostgreSQL() error {
|
||||||
|
@ -31,6 +31,7 @@ import (
|
|||||||
"github.com/gomodule/redigo/redis"
|
"github.com/gomodule/redigo/redis"
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
)
|
)
|
||||||
@ -118,7 +119,7 @@ func (r RedisArgs) validateFormat(c redis.Conn) error {
|
|||||||
|
|
||||||
// RedisTarget - Redis target.
|
// RedisTarget - Redis target.
|
||||||
type RedisTarget struct {
|
type RedisTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args RedisArgs
|
args RedisArgs
|
||||||
@ -221,8 +222,8 @@ func (target *RedisTarget) send(eventData event.Event) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - reads an event from store and sends it to redis.
|
// SendFromStore - reads an event from store and sends it to redis.
|
||||||
func (target *RedisTarget) Send(eventKey string) error {
|
func (target *RedisTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -276,7 +277,7 @@ func (target *RedisTarget) Close() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *RedisTarget) init() error {
|
func (target *RedisTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initRedis)
|
return target.initOnce.Do(target.initRedis)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (target *RedisTarget) initRedis() error {
|
func (target *RedisTarget) initRedis() error {
|
||||||
|
@ -35,6 +35,7 @@ import (
|
|||||||
|
|
||||||
"github.com/minio/minio/internal/event"
|
"github.com/minio/minio/internal/event"
|
||||||
"github.com/minio/minio/internal/logger"
|
"github.com/minio/minio/internal/logger"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
"github.com/minio/minio/internal/store"
|
"github.com/minio/minio/internal/store"
|
||||||
"github.com/minio/pkg/certs"
|
"github.com/minio/pkg/certs"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
@ -91,7 +92,7 @@ func (w WebhookArgs) Validate() error {
|
|||||||
|
|
||||||
// WebhookTarget - Webhook target.
|
// WebhookTarget - Webhook target.
|
||||||
type WebhookTarget struct {
|
type WebhookTarget struct {
|
||||||
lazyInit lazyInit
|
initOnce once.Init
|
||||||
|
|
||||||
id event.TargetID
|
id event.TargetID
|
||||||
args WebhookArgs
|
args WebhookArgs
|
||||||
@ -222,8 +223,8 @@ func (target *WebhookTarget) send(eventData event.Event) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - reads an event from store and sends it to webhook.
|
// SendFromStore - reads an event from store and sends it to webhook.
|
||||||
func (target *WebhookTarget) Send(eventKey string) error {
|
func (target *WebhookTarget) SendFromStore(eventKey string) error {
|
||||||
if err := target.init(); err != nil {
|
if err := target.init(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -256,7 +257,7 @@ func (target *WebhookTarget) Close() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (target *WebhookTarget) init() error {
|
func (target *WebhookTarget) init() error {
|
||||||
return target.lazyInit.Do(target.initWebhook)
|
return target.initOnce.Do(target.initWebhook)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Only called from init()
|
// Only called from init()
|
||||||
|
@ -34,7 +34,7 @@ type Target interface {
|
|||||||
ID() TargetID
|
ID() TargetID
|
||||||
IsActive() (bool, error)
|
IsActive() (bool, error)
|
||||||
Save(Event) error
|
Save(Event) error
|
||||||
Send(string) error
|
SendFromStore(string) error
|
||||||
Close() error
|
Close() error
|
||||||
Store() TargetStore
|
Store() TargetStore
|
||||||
}
|
}
|
||||||
|
@ -60,8 +60,8 @@ func (target ExampleTarget) send(eventData Event) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send - interface compatible method does no-op.
|
// SendFromStore - interface compatible method does no-op.
|
||||||
func (target ExampleTarget) Send(eventKey string) error {
|
func (target *ExampleTarget) SendFromStore(eventKey string) error {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -139,7 +139,7 @@ func AuditLog(ctx context.Context, w http.ResponseWriter, r *http.Request, reqCl
|
|||||||
|
|
||||||
// Send audit logs only to http targets.
|
// Send audit logs only to http targets.
|
||||||
for _, t := range auditTgts {
|
for _, t := range auditTgts {
|
||||||
if err := t.Send(entry); err != nil {
|
if err := t.Send(ctx, entry); err != nil {
|
||||||
LogAlwaysIf(context.Background(), fmt.Errorf("event(%v) was not sent to Audit target (%v): %v", entry, t, err), madmin.LogKindAll)
|
LogAlwaysIf(context.Background(), fmt.Errorf("event(%v) was not sent to Audit target (%v): %v", entry, t, err), madmin.LogKindAll)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -43,6 +43,7 @@ const (
|
|||||||
ClientCert = "client_cert"
|
ClientCert = "client_cert"
|
||||||
ClientKey = "client_key"
|
ClientKey = "client_key"
|
||||||
QueueSize = "queue_size"
|
QueueSize = "queue_size"
|
||||||
|
QueueDir = "queue_dir"
|
||||||
Proxy = "proxy"
|
Proxy = "proxy"
|
||||||
|
|
||||||
KafkaBrokers = "brokers"
|
KafkaBrokers = "brokers"
|
||||||
@ -57,6 +58,8 @@ const (
|
|||||||
KafkaClientTLSCert = "client_tls_cert"
|
KafkaClientTLSCert = "client_tls_cert"
|
||||||
KafkaClientTLSKey = "client_tls_key"
|
KafkaClientTLSKey = "client_tls_key"
|
||||||
KafkaVersion = "version"
|
KafkaVersion = "version"
|
||||||
|
KafkaQueueDir = "queue_dir"
|
||||||
|
KafkaQueueSize = "queue_size"
|
||||||
|
|
||||||
EnvLoggerWebhookEnable = "MINIO_LOGGER_WEBHOOK_ENABLE"
|
EnvLoggerWebhookEnable = "MINIO_LOGGER_WEBHOOK_ENABLE"
|
||||||
EnvLoggerWebhookEndpoint = "MINIO_LOGGER_WEBHOOK_ENDPOINT"
|
EnvLoggerWebhookEndpoint = "MINIO_LOGGER_WEBHOOK_ENDPOINT"
|
||||||
@ -65,6 +68,7 @@ const (
|
|||||||
EnvLoggerWebhookClientKey = "MINIO_LOGGER_WEBHOOK_CLIENT_KEY"
|
EnvLoggerWebhookClientKey = "MINIO_LOGGER_WEBHOOK_CLIENT_KEY"
|
||||||
EnvLoggerWebhookProxy = "MINIO_LOGGER_WEBHOOK_PROXY"
|
EnvLoggerWebhookProxy = "MINIO_LOGGER_WEBHOOK_PROXY"
|
||||||
EnvLoggerWebhookQueueSize = "MINIO_LOGGER_WEBHOOK_QUEUE_SIZE"
|
EnvLoggerWebhookQueueSize = "MINIO_LOGGER_WEBHOOK_QUEUE_SIZE"
|
||||||
|
EnvLoggerWebhookQueueDir = "MINIO_LOGGER_WEBHOOK_QUEUE_DIR"
|
||||||
|
|
||||||
EnvAuditWebhookEnable = "MINIO_AUDIT_WEBHOOK_ENABLE"
|
EnvAuditWebhookEnable = "MINIO_AUDIT_WEBHOOK_ENABLE"
|
||||||
EnvAuditWebhookEndpoint = "MINIO_AUDIT_WEBHOOK_ENDPOINT"
|
EnvAuditWebhookEndpoint = "MINIO_AUDIT_WEBHOOK_ENDPOINT"
|
||||||
@ -72,6 +76,7 @@ const (
|
|||||||
EnvAuditWebhookClientCert = "MINIO_AUDIT_WEBHOOK_CLIENT_CERT"
|
EnvAuditWebhookClientCert = "MINIO_AUDIT_WEBHOOK_CLIENT_CERT"
|
||||||
EnvAuditWebhookClientKey = "MINIO_AUDIT_WEBHOOK_CLIENT_KEY"
|
EnvAuditWebhookClientKey = "MINIO_AUDIT_WEBHOOK_CLIENT_KEY"
|
||||||
EnvAuditWebhookQueueSize = "MINIO_AUDIT_WEBHOOK_QUEUE_SIZE"
|
EnvAuditWebhookQueueSize = "MINIO_AUDIT_WEBHOOK_QUEUE_SIZE"
|
||||||
|
EnvAuditWebhookQueueDir = "MINIO_AUDIT_WEBHOOK_QUEUE_DIR"
|
||||||
|
|
||||||
EnvKafkaEnable = "MINIO_AUDIT_KAFKA_ENABLE"
|
EnvKafkaEnable = "MINIO_AUDIT_KAFKA_ENABLE"
|
||||||
EnvKafkaBrokers = "MINIO_AUDIT_KAFKA_BROKERS"
|
EnvKafkaBrokers = "MINIO_AUDIT_KAFKA_BROKERS"
|
||||||
@ -86,6 +91,11 @@ const (
|
|||||||
EnvKafkaClientTLSCert = "MINIO_AUDIT_KAFKA_CLIENT_TLS_CERT"
|
EnvKafkaClientTLSCert = "MINIO_AUDIT_KAFKA_CLIENT_TLS_CERT"
|
||||||
EnvKafkaClientTLSKey = "MINIO_AUDIT_KAFKA_CLIENT_TLS_KEY"
|
EnvKafkaClientTLSKey = "MINIO_AUDIT_KAFKA_CLIENT_TLS_KEY"
|
||||||
EnvKafkaVersion = "MINIO_AUDIT_KAFKA_VERSION"
|
EnvKafkaVersion = "MINIO_AUDIT_KAFKA_VERSION"
|
||||||
|
EnvKafkaQueueDir = "MINIO_AUDIT_KAFKA_QUEUE_DIR"
|
||||||
|
EnvKafkaQueueSize = "MINIO_AUDIT_KAFKA_QUEUE_SIZE"
|
||||||
|
|
||||||
|
loggerTargetNamePrefix = "logger-"
|
||||||
|
auditTargetNamePrefix = "audit-"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Default KVS for loggerHTTP and loggerAuditHTTP
|
// Default KVS for loggerHTTP and loggerAuditHTTP
|
||||||
@ -119,6 +129,10 @@ var (
|
|||||||
Key: QueueSize,
|
Key: QueueSize,
|
||||||
Value: "100000",
|
Value: "100000",
|
||||||
},
|
},
|
||||||
|
config.KV{
|
||||||
|
Key: QueueDir,
|
||||||
|
Value: "",
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
DefaultAuditWebhookKVS = config.KVS{
|
DefaultAuditWebhookKVS = config.KVS{
|
||||||
@ -146,6 +160,10 @@ var (
|
|||||||
Key: QueueSize,
|
Key: QueueSize,
|
||||||
Value: "100000",
|
Value: "100000",
|
||||||
},
|
},
|
||||||
|
config.KV{
|
||||||
|
Key: QueueDir,
|
||||||
|
Value: "",
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
DefaultAuditKafkaKVS = config.KVS{
|
DefaultAuditKafkaKVS = config.KVS{
|
||||||
@ -201,6 +219,14 @@ var (
|
|||||||
Key: KafkaVersion,
|
Key: KafkaVersion,
|
||||||
Value: "",
|
Value: "",
|
||||||
},
|
},
|
||||||
|
config.KV{
|
||||||
|
Key: QueueSize,
|
||||||
|
Value: "100000",
|
||||||
|
},
|
||||||
|
config.KV{
|
||||||
|
Key: QueueDir,
|
||||||
|
Value: "",
|
||||||
|
},
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -395,6 +421,25 @@ func lookupAuditKafkaConfig(scfg config.Config, cfg Config) (Config, error) {
|
|||||||
kafkaArgs.SASL.Password = env.Get(saslPasswordEnv, kv.Get(KafkaSASLPassword))
|
kafkaArgs.SASL.Password = env.Get(saslPasswordEnv, kv.Get(KafkaSASLPassword))
|
||||||
kafkaArgs.SASL.Mechanism = env.Get(saslMechanismEnv, kv.Get(KafkaSASLMechanism))
|
kafkaArgs.SASL.Mechanism = env.Get(saslMechanismEnv, kv.Get(KafkaSASLMechanism))
|
||||||
|
|
||||||
|
queueDirEnv := EnvKafkaQueueDir
|
||||||
|
if k != config.Default {
|
||||||
|
queueDirEnv = queueDirEnv + config.Default + k
|
||||||
|
}
|
||||||
|
kafkaArgs.QueueDir = env.Get(queueDirEnv, kv.Get(KafkaQueueDir))
|
||||||
|
|
||||||
|
queueSizeEnv := EnvKafkaQueueSize
|
||||||
|
if k != config.Default {
|
||||||
|
queueSizeEnv = queueSizeEnv + config.Default + k
|
||||||
|
}
|
||||||
|
queueSize, err := strconv.Atoi(env.Get(queueSizeEnv, kv.Get(KafkaQueueSize)))
|
||||||
|
if err != nil {
|
||||||
|
return cfg, err
|
||||||
|
}
|
||||||
|
if queueSize <= 0 {
|
||||||
|
return cfg, errors.New("invalid queue_size value")
|
||||||
|
}
|
||||||
|
kafkaArgs.QueueSize = queueSize
|
||||||
|
|
||||||
cfg.AuditKafka[k] = kafkaArgs
|
cfg.AuditKafka[k] = kafkaArgs
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -448,9 +493,9 @@ func lookupLoggerWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
|
|||||||
return cfg, err
|
return cfg, err
|
||||||
}
|
}
|
||||||
proxyEnv := EnvLoggerWebhookProxy
|
proxyEnv := EnvLoggerWebhookProxy
|
||||||
queueSizeEnv := EnvAuditWebhookQueueSize
|
queueSizeEnv := EnvLoggerWebhookQueueSize
|
||||||
if target != config.Default {
|
if target != config.Default {
|
||||||
queueSizeEnv = EnvAuditWebhookQueueSize + config.Default + target
|
queueSizeEnv = EnvLoggerWebhookQueueSize + config.Default + target
|
||||||
}
|
}
|
||||||
queueSize, err := strconv.Atoi(env.Get(queueSizeEnv, "100000"))
|
queueSize, err := strconv.Atoi(env.Get(queueSizeEnv, "100000"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -459,6 +504,10 @@ func lookupLoggerWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
|
|||||||
if queueSize <= 0 {
|
if queueSize <= 0 {
|
||||||
return cfg, errors.New("invalid queue_size value")
|
return cfg, errors.New("invalid queue_size value")
|
||||||
}
|
}
|
||||||
|
queueDirEnv := EnvLoggerWebhookQueueDir
|
||||||
|
if target != config.Default {
|
||||||
|
queueDirEnv = EnvLoggerWebhookQueueDir + config.Default + target
|
||||||
|
}
|
||||||
cfg.HTTP[target] = http.Config{
|
cfg.HTTP[target] = http.Config{
|
||||||
Enabled: true,
|
Enabled: true,
|
||||||
Endpoint: env.Get(endpointEnv, ""),
|
Endpoint: env.Get(endpointEnv, ""),
|
||||||
@ -467,7 +516,8 @@ func lookupLoggerWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
|
|||||||
ClientKey: env.Get(clientKeyEnv, ""),
|
ClientKey: env.Get(clientKeyEnv, ""),
|
||||||
Proxy: env.Get(proxyEnv, ""),
|
Proxy: env.Get(proxyEnv, ""),
|
||||||
QueueSize: queueSize,
|
QueueSize: queueSize,
|
||||||
Name: target,
|
QueueDir: env.Get(queueDirEnv, ""),
|
||||||
|
Name: loggerTargetNamePrefix + target,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -511,7 +561,8 @@ func lookupLoggerWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
|
|||||||
ClientKey: kv.Get(ClientKey),
|
ClientKey: kv.Get(ClientKey),
|
||||||
Proxy: kv.Get(Proxy),
|
Proxy: kv.Get(Proxy),
|
||||||
QueueSize: queueSize,
|
QueueSize: queueSize,
|
||||||
Name: starget,
|
QueueDir: kv.Get(QueueDir),
|
||||||
|
Name: loggerTargetNamePrefix + starget,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -574,6 +625,10 @@ func lookupAuditWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
|
|||||||
if queueSize <= 0 {
|
if queueSize <= 0 {
|
||||||
return cfg, errors.New("invalid queue_size value")
|
return cfg, errors.New("invalid queue_size value")
|
||||||
}
|
}
|
||||||
|
queueDirEnv := EnvAuditWebhookQueueDir
|
||||||
|
if target != config.Default {
|
||||||
|
queueDirEnv = EnvAuditWebhookQueueDir + config.Default + target
|
||||||
|
}
|
||||||
cfg.AuditWebhook[target] = http.Config{
|
cfg.AuditWebhook[target] = http.Config{
|
||||||
Enabled: true,
|
Enabled: true,
|
||||||
Endpoint: env.Get(endpointEnv, ""),
|
Endpoint: env.Get(endpointEnv, ""),
|
||||||
@ -581,7 +636,8 @@ func lookupAuditWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
|
|||||||
ClientCert: env.Get(clientCertEnv, ""),
|
ClientCert: env.Get(clientCertEnv, ""),
|
||||||
ClientKey: env.Get(clientKeyEnv, ""),
|
ClientKey: env.Get(clientKeyEnv, ""),
|
||||||
QueueSize: queueSize,
|
QueueSize: queueSize,
|
||||||
Name: target,
|
QueueDir: env.Get(queueDirEnv, ""),
|
||||||
|
Name: auditTargetNamePrefix + target,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -617,7 +673,6 @@ func lookupAuditWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
|
|||||||
if queueSize <= 0 {
|
if queueSize <= 0 {
|
||||||
return cfg, errors.New("invalid queue_size value")
|
return cfg, errors.New("invalid queue_size value")
|
||||||
}
|
}
|
||||||
|
|
||||||
cfg.AuditWebhook[starget] = http.Config{
|
cfg.AuditWebhook[starget] = http.Config{
|
||||||
Enabled: true,
|
Enabled: true,
|
||||||
Endpoint: kv.Get(Endpoint),
|
Endpoint: kv.Get(Endpoint),
|
||||||
@ -625,7 +680,8 @@ func lookupAuditWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
|
|||||||
ClientCert: kv.Get(ClientCert),
|
ClientCert: kv.Get(ClientCert),
|
||||||
ClientKey: kv.Get(ClientKey),
|
ClientKey: kv.Get(ClientKey),
|
||||||
QueueSize: queueSize,
|
QueueSize: queueSize,
|
||||||
Name: starget,
|
QueueDir: kv.Get(QueueDir),
|
||||||
|
Name: auditTargetNamePrefix + starget,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -347,7 +347,7 @@ func consoleLogIf(ctx context.Context, err error, errKind ...interface{}) {
|
|||||||
|
|
||||||
if consoleTgt != nil {
|
if consoleTgt != nil {
|
||||||
entry := errToEntry(ctx, err, errKind...)
|
entry := errToEntry(ctx, err, errKind...)
|
||||||
consoleTgt.Send(entry)
|
consoleTgt.Send(ctx, entry)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -366,10 +366,10 @@ func logIf(ctx context.Context, err error, errKind ...interface{}) {
|
|||||||
entry := errToEntry(ctx, err, errKind...)
|
entry := errToEntry(ctx, err, errKind...)
|
||||||
// Iterate over all logger targets to send the log entry
|
// Iterate over all logger targets to send the log entry
|
||||||
for _, t := range systemTgts {
|
for _, t := range systemTgts {
|
||||||
if err := t.Send(entry); err != nil {
|
if err := t.Send(ctx, entry); err != nil {
|
||||||
if consoleTgt != nil {
|
if consoleTgt != nil {
|
||||||
entry.Trace.Message = fmt.Sprintf("event(%#v) was not sent to Logger target (%#v): %#v", entry, t, err)
|
entry.Trace.Message = fmt.Sprintf("event(%#v) was not sent to Logger target (%#v): %#v", entry, t, err)
|
||||||
consoleTgt.Send(entry)
|
consoleTgt.Send(ctx, entry)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
// Copyright (c) 2015-2021 MinIO, Inc.
|
// Copyright (c) 2015-2023 MinIO, Inc.
|
||||||
//
|
//
|
||||||
// This file is part of MinIO Object Storage stack
|
// This file is part of MinIO Object Storage stack
|
||||||
//
|
//
|
||||||
@ -26,13 +26,17 @@ import (
|
|||||||
"math"
|
"math"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
"net/url"
|
||||||
"strings"
|
"os"
|
||||||
|
"path/filepath"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
xhttp "github.com/minio/minio/internal/http"
|
xhttp "github.com/minio/minio/internal/http"
|
||||||
"github.com/minio/minio/internal/logger/target/types"
|
"github.com/minio/minio/internal/logger/target/types"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
|
"github.com/minio/minio/internal/store"
|
||||||
|
xnet "github.com/minio/pkg/net"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -41,6 +45,9 @@ const (
|
|||||||
|
|
||||||
// maxWorkers is the maximum number of concurrent operations.
|
// maxWorkers is the maximum number of concurrent operations.
|
||||||
maxWorkers = 16
|
maxWorkers = 16
|
||||||
|
|
||||||
|
// the suffix for the configured queue dir where the logs will be persisted.
|
||||||
|
httpLoggerExtension = ".http.log"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@ -59,6 +66,7 @@ type Config struct {
|
|||||||
ClientCert string `json:"clientCert"`
|
ClientCert string `json:"clientCert"`
|
||||||
ClientKey string `json:"clientKey"`
|
ClientKey string `json:"clientKey"`
|
||||||
QueueSize int `json:"queueSize"`
|
QueueSize int `json:"queueSize"`
|
||||||
|
QueueDir string `json:"queueDir"`
|
||||||
Proxy string `json:"string"`
|
Proxy string `json:"string"`
|
||||||
Transport http.RoundTripper `json:"-"`
|
Transport http.RoundTripper `json:"-"`
|
||||||
|
|
||||||
@ -93,10 +101,21 @@ type Target struct {
|
|||||||
// will attempt to establish the connection.
|
// will attempt to establish the connection.
|
||||||
revive sync.Once
|
revive sync.Once
|
||||||
|
|
||||||
|
// store to persist and replay the logs to the target
|
||||||
|
// to avoid missing events when the target is down.
|
||||||
|
store store.Store[interface{}]
|
||||||
|
storeCtxCancel context.CancelFunc
|
||||||
|
initQueueStoreOnce once.Init
|
||||||
|
|
||||||
config Config
|
config Config
|
||||||
client *http.Client
|
client *http.Client
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Name returns the name of the target
|
||||||
|
func (h *Target) Name() string {
|
||||||
|
return "minio-http-" + h.config.Name
|
||||||
|
}
|
||||||
|
|
||||||
// Endpoint returns the backend endpoint
|
// Endpoint returns the backend endpoint
|
||||||
func (h *Target) Endpoint() string {
|
func (h *Target) Endpoint() string {
|
||||||
return h.config.Endpoint
|
return h.config.Endpoint
|
||||||
@ -106,9 +125,9 @@ func (h *Target) String() string {
|
|||||||
return h.config.Name
|
return h.config.Name
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsOnline returns true if the initialization was successful
|
// IsOnline returns true if the target is reachable.
|
||||||
func (h *Target) IsOnline() bool {
|
func (h *Target) IsOnline(ctx context.Context) bool {
|
||||||
return atomic.LoadInt32(&h.status) == statusOnline
|
return h.isAlive(ctx) == nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Stats returns the target statistics.
|
// Stats returns the target statistics.
|
||||||
@ -125,8 +144,34 @@ func (h *Target) Stats() types.TargetStats {
|
|||||||
return stats
|
return stats
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// This will check if we can reach the remote.
|
||||||
|
func (h *Target) isAlive(ctx context.Context) (err error) {
|
||||||
|
return h.send(ctx, []byte(`{}`), 2*webhookCallTimeout)
|
||||||
|
}
|
||||||
|
|
||||||
// Init validate and initialize the http target
|
// Init validate and initialize the http target
|
||||||
func (h *Target) Init() (err error) {
|
func (h *Target) Init(ctx context.Context) (err error) {
|
||||||
|
if h.config.QueueDir != "" {
|
||||||
|
return h.initQueueStoreOnce.DoWithContext(ctx, h.initQueueStore)
|
||||||
|
}
|
||||||
|
return h.initLogChannel(ctx)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *Target) initQueueStore(ctx context.Context) (err error) {
|
||||||
|
var queueStore store.Store[interface{}]
|
||||||
|
queueDir := filepath.Join(h.config.QueueDir, h.Name())
|
||||||
|
queueStore = store.NewQueueStore[interface{}](queueDir, uint64(h.config.QueueSize), httpLoggerExtension)
|
||||||
|
if err = queueStore.Open(); err != nil {
|
||||||
|
return fmt.Errorf("unable to initialize the queue store of %s webhook: %w", h.Name(), err)
|
||||||
|
}
|
||||||
|
ctx, cancel := context.WithCancel(ctx)
|
||||||
|
h.store = queueStore
|
||||||
|
h.storeCtxCancel = cancel
|
||||||
|
store.StreamItems(h.store, h, ctx.Done(), h.config.LogOnce)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *Target) initLogChannel(ctx context.Context) (err error) {
|
||||||
switch atomic.LoadInt32(&h.status) {
|
switch atomic.LoadInt32(&h.status) {
|
||||||
case statusOnline:
|
case statusOnline:
|
||||||
return nil
|
return nil
|
||||||
@ -134,45 +179,7 @@ func (h *Target) Init() (err error) {
|
|||||||
return errors.New("target is closed")
|
return errors.New("target is closed")
|
||||||
}
|
}
|
||||||
|
|
||||||
// This will check if we can reach the remote.
|
err = h.isAlive(ctx)
|
||||||
checkAlive := func() error {
|
|
||||||
ctx, cancel := context.WithTimeout(context.Background(), 2*webhookCallTimeout)
|
|
||||||
defer cancel()
|
|
||||||
|
|
||||||
req, err := http.NewRequestWithContext(ctx, http.MethodPost, h.config.Endpoint, strings.NewReader(`{}`))
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
req.Header.Set(xhttp.ContentType, "application/json")
|
|
||||||
|
|
||||||
// Set user-agent to indicate MinIO release
|
|
||||||
// version to the configured log endpoint
|
|
||||||
req.Header.Set("User-Agent", h.config.UserAgent)
|
|
||||||
|
|
||||||
if h.config.AuthToken != "" {
|
|
||||||
req.Header.Set("Authorization", h.config.AuthToken)
|
|
||||||
}
|
|
||||||
|
|
||||||
resp, err := h.client.Do(req)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// Drain any response.
|
|
||||||
xhttp.DrainBody(resp.Body)
|
|
||||||
|
|
||||||
if !acceptedResponseStatusCode(resp.StatusCode) {
|
|
||||||
if resp.StatusCode == http.StatusForbidden {
|
|
||||||
return fmt.Errorf("%s returned '%s', please check if your auth token is correctly set",
|
|
||||||
h.config.Endpoint, resp.Status)
|
|
||||||
}
|
|
||||||
return fmt.Errorf("%s returned '%s', please check your endpoint configuration",
|
|
||||||
h.config.Endpoint, resp.Status)
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
err = checkAlive()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// Start a goroutine that will continue to check if we can reach
|
// Start a goroutine that will continue to check if we can reach
|
||||||
h.revive.Do(func() {
|
h.revive.Do(func() {
|
||||||
@ -183,14 +190,14 @@ func (h *Target) Init() (err error) {
|
|||||||
if atomic.LoadInt32(&h.status) != statusOffline {
|
if atomic.LoadInt32(&h.status) != statusOffline {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
if err := checkAlive(); err == nil {
|
if err := h.isAlive(ctx); err == nil {
|
||||||
// We are online.
|
// We are online.
|
||||||
if atomic.CompareAndSwapInt32(&h.status, statusOffline, statusOnline) {
|
if atomic.CompareAndSwapInt32(&h.status, statusOffline, statusOnline) {
|
||||||
h.workerStartMu.Lock()
|
h.workerStartMu.Lock()
|
||||||
h.lastStarted = time.Now()
|
h.lastStarted = time.Now()
|
||||||
h.workerStartMu.Unlock()
|
h.workerStartMu.Unlock()
|
||||||
atomic.AddInt64(&h.workers, 1)
|
atomic.AddInt64(&h.workers, 1)
|
||||||
go h.startHTTPLogger()
|
go h.startHTTPLogger(ctx)
|
||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
@ -205,19 +212,51 @@ func (h *Target) Init() (err error) {
|
|||||||
h.lastStarted = time.Now()
|
h.lastStarted = time.Now()
|
||||||
h.workerStartMu.Unlock()
|
h.workerStartMu.Unlock()
|
||||||
atomic.AddInt64(&h.workers, 1)
|
atomic.AddInt64(&h.workers, 1)
|
||||||
go h.startHTTPLogger()
|
go h.startHTTPLogger(ctx)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Accepted HTTP Status Codes
|
func (h *Target) send(ctx context.Context, payload []byte, timeout time.Duration) (err error) {
|
||||||
var acceptedStatusCodeMap = map[int]bool{http.StatusOK: true, http.StatusCreated: true, http.StatusAccepted: true, http.StatusNoContent: true}
|
ctx, cancel := context.WithTimeout(ctx, timeout)
|
||||||
|
defer cancel()
|
||||||
|
req, err := http.NewRequestWithContext(ctx, http.MethodPost,
|
||||||
|
h.config.Endpoint, bytes.NewReader(payload))
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("%s returned '%w', please check your endpoint configuration", h.config.Endpoint, err)
|
||||||
|
}
|
||||||
|
req.Header.Set(xhttp.ContentType, "application/json")
|
||||||
|
req.Header.Set(xhttp.MinIOVersion, xhttp.GlobalMinIOVersion)
|
||||||
|
req.Header.Set(xhttp.MinioDeploymentID, xhttp.GlobalDeploymentID)
|
||||||
|
|
||||||
func acceptedResponseStatusCode(code int) bool {
|
// Set user-agent to indicate MinIO release
|
||||||
return acceptedStatusCodeMap[code]
|
// version to the configured log endpoint
|
||||||
|
req.Header.Set("User-Agent", h.config.UserAgent)
|
||||||
|
|
||||||
|
if h.config.AuthToken != "" {
|
||||||
|
req.Header.Set("Authorization", h.config.AuthToken)
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := h.client.Do(req)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("%s returned '%w', please check your endpoint configuration", h.config.Endpoint, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Drain any response.
|
||||||
|
xhttp.DrainBody(resp.Body)
|
||||||
|
|
||||||
|
switch resp.StatusCode {
|
||||||
|
case http.StatusOK, http.StatusCreated, http.StatusAccepted, http.StatusNoContent:
|
||||||
|
// accepted HTTP status codes.
|
||||||
|
return nil
|
||||||
|
case http.StatusForbidden:
|
||||||
|
return fmt.Errorf("%s returned '%s', please check if your auth token is correctly set", h.config.Endpoint, resp.Status)
|
||||||
|
default:
|
||||||
|
return fmt.Errorf("%s returned '%s', please check your endpoint configuration", h.config.Endpoint, resp.Status)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (h *Target) logEntry(entry interface{}) {
|
func (h *Target) logEntry(ctx context.Context, entry interface{}) {
|
||||||
logJSON, err := json.Marshal(&entry)
|
logJSON, err := json.Marshal(&entry)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
atomic.AddInt64(&h.failedMessages, 1)
|
atomic.AddInt64(&h.failedMessages, 1)
|
||||||
@ -239,52 +278,14 @@ func (h *Target) logEntry(entry interface{}) {
|
|||||||
time.Sleep(sleep)
|
time.Sleep(sleep)
|
||||||
}
|
}
|
||||||
tries++
|
tries++
|
||||||
ctx, cancel := context.WithTimeout(context.Background(), webhookCallTimeout)
|
if err := h.send(ctx, logJSON, webhookCallTimeout); err != nil {
|
||||||
defer cancel()
|
h.config.LogOnce(ctx, err, h.config.Endpoint)
|
||||||
req, err := http.NewRequestWithContext(ctx, http.MethodPost,
|
|
||||||
h.config.Endpoint, bytes.NewReader(logJSON))
|
|
||||||
if err != nil {
|
|
||||||
h.config.LogOnce(ctx, fmt.Errorf("%s returned '%w', please check your endpoint configuration", h.config.Endpoint, err), h.config.Endpoint)
|
|
||||||
atomic.AddInt64(&h.failedMessages, 1)
|
atomic.AddInt64(&h.failedMessages, 1)
|
||||||
continue
|
|
||||||
}
|
|
||||||
req.Header.Set(xhttp.ContentType, "application/json")
|
|
||||||
req.Header.Set(xhttp.MinIOVersion, xhttp.GlobalMinIOVersion)
|
|
||||||
req.Header.Set(xhttp.MinioDeploymentID, xhttp.GlobalDeploymentID)
|
|
||||||
|
|
||||||
// Set user-agent to indicate MinIO release
|
|
||||||
// version to the configured log endpoint
|
|
||||||
req.Header.Set("User-Agent", h.config.UserAgent)
|
|
||||||
|
|
||||||
if h.config.AuthToken != "" {
|
|
||||||
req.Header.Set("Authorization", h.config.AuthToken)
|
|
||||||
}
|
|
||||||
|
|
||||||
resp, err := h.client.Do(req)
|
|
||||||
if err != nil {
|
|
||||||
atomic.AddInt64(&h.failedMessages, 1)
|
|
||||||
h.config.LogOnce(ctx, fmt.Errorf("%s returned '%w', please check your endpoint configuration", h.config.Endpoint, err), h.config.Endpoint)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
// Drain any response.
|
|
||||||
xhttp.DrainBody(resp.Body)
|
|
||||||
|
|
||||||
if acceptedResponseStatusCode(resp.StatusCode) {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
// Log failure, retry
|
|
||||||
atomic.AddInt64(&h.failedMessages, 1)
|
|
||||||
switch resp.StatusCode {
|
|
||||||
case http.StatusForbidden:
|
|
||||||
h.config.LogOnce(ctx, fmt.Errorf("%s returned '%s', please check if your auth token is correctly set", h.config.Endpoint, resp.Status), h.config.Endpoint)
|
|
||||||
default:
|
|
||||||
h.config.LogOnce(ctx, fmt.Errorf("%s returned '%s', please check your endpoint configuration", h.config.Endpoint, resp.Status), h.config.Endpoint)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (h *Target) startHTTPLogger() {
|
func (h *Target) startHTTPLogger(ctx context.Context) {
|
||||||
h.logChMu.RLock()
|
h.logChMu.RLock()
|
||||||
logCh := h.logCh
|
logCh := h.logCh
|
||||||
if logCh != nil {
|
if logCh != nil {
|
||||||
@ -302,7 +303,7 @@ func (h *Target) startHTTPLogger() {
|
|||||||
// Send messages until channel is closed.
|
// Send messages until channel is closed.
|
||||||
for entry := range logCh {
|
for entry := range logCh {
|
||||||
atomic.AddInt64(&h.totalMessages, 1)
|
atomic.AddInt64(&h.totalMessages, 1)
|
||||||
h.logEntry(entry)
|
h.logEntry(ctx, entry)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -328,10 +329,41 @@ func New(config Config) *Target {
|
|||||||
return h
|
return h
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// SendFromStore - reads the log from store and sends it to webhook.
|
||||||
|
func (h *Target) SendFromStore(key string) (err error) {
|
||||||
|
var eventData interface{}
|
||||||
|
eventData, err = h.store.Get(key)
|
||||||
|
if err != nil {
|
||||||
|
if os.IsNotExist(err) {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
atomic.AddInt64(&h.totalMessages, 1)
|
||||||
|
logJSON, err := json.Marshal(&eventData)
|
||||||
|
if err != nil {
|
||||||
|
atomic.AddInt64(&h.failedMessages, 1)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if err := h.send(context.Background(), logJSON, webhookCallTimeout); err != nil {
|
||||||
|
atomic.AddInt64(&h.failedMessages, 1)
|
||||||
|
if xnet.IsNetworkOrHostDown(err, true) {
|
||||||
|
return store.ErrNotConnected
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
// Delete the event from store.
|
||||||
|
return h.store.Del(key)
|
||||||
|
}
|
||||||
|
|
||||||
// Send log message 'e' to http target.
|
// Send log message 'e' to http target.
|
||||||
// If servers are offline messages are queued until queue is full.
|
// If servers are offline messages are queued until queue is full.
|
||||||
// If Cancel has been called the message is ignored.
|
// If Cancel has been called the message is ignored.
|
||||||
func (h *Target) Send(entry interface{}) error {
|
func (h *Target) Send(ctx context.Context, entry interface{}) error {
|
||||||
|
if h.store != nil {
|
||||||
|
// save the entry to the queue store which will be replayed to the target.
|
||||||
|
return h.store.Put(entry)
|
||||||
|
}
|
||||||
if atomic.LoadInt32(&h.status) == statusClosed {
|
if atomic.LoadInt32(&h.status) == statusClosed {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@ -345,7 +377,7 @@ func (h *Target) Send(entry interface{}) error {
|
|||||||
case h.logCh <- entry:
|
case h.logCh <- entry:
|
||||||
default:
|
default:
|
||||||
// Drop messages until we are online.
|
// Drop messages until we are online.
|
||||||
if !h.IsOnline() {
|
if !h.IsOnline(ctx) {
|
||||||
return errors.New("log buffer full and remote offline")
|
return errors.New("log buffer full and remote offline")
|
||||||
}
|
}
|
||||||
nWorkers := atomic.LoadInt64(&h.workers)
|
nWorkers := atomic.LoadInt64(&h.workers)
|
||||||
@ -358,7 +390,7 @@ func (h *Target) Send(entry interface{}) error {
|
|||||||
if atomic.CompareAndSwapInt64(&h.workers, nWorkers, nWorkers+1) {
|
if atomic.CompareAndSwapInt64(&h.workers, nWorkers, nWorkers+1) {
|
||||||
// Start another logger.
|
// Start another logger.
|
||||||
h.lastStarted = time.Now()
|
h.lastStarted = time.Now()
|
||||||
go h.startHTTPLogger()
|
go h.startHTTPLogger(ctx)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
h.logCh <- entry
|
h.logCh <- entry
|
||||||
@ -380,6 +412,12 @@ func (h *Target) Send(entry interface{}) error {
|
|||||||
func (h *Target) Cancel() {
|
func (h *Target) Cancel() {
|
||||||
atomic.StoreInt32(&h.status, statusClosed)
|
atomic.StoreInt32(&h.status, statusClosed)
|
||||||
|
|
||||||
|
// If queuestore is configured, cancel it's context to
|
||||||
|
// stop the replay go-routine.
|
||||||
|
if h.store != nil {
|
||||||
|
h.storeCtxCancel()
|
||||||
|
}
|
||||||
|
|
||||||
// Set logch to nil and close it.
|
// Set logch to nil and close it.
|
||||||
// This will block all Send operations,
|
// This will block all Send operations,
|
||||||
// and finish the existing ones.
|
// and finish the existing ones.
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
// Copyright (c) 2015-2022 MinIO, Inc.
|
// Copyright (c) 2015-2023 MinIO, Inc.
|
||||||
//
|
//
|
||||||
// This file is part of MinIO Object Storage stack
|
// This file is part of MinIO Object Storage stack
|
||||||
//
|
//
|
||||||
@ -23,7 +23,10 @@ import (
|
|||||||
"crypto/x509"
|
"crypto/x509"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
|
"fmt"
|
||||||
"net"
|
"net"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
@ -34,94 +37,13 @@ import (
|
|||||||
saramatls "github.com/Shopify/sarama/tools/tls"
|
saramatls "github.com/Shopify/sarama/tools/tls"
|
||||||
|
|
||||||
"github.com/minio/minio/internal/logger/target/types"
|
"github.com/minio/minio/internal/logger/target/types"
|
||||||
|
"github.com/minio/minio/internal/once"
|
||||||
|
"github.com/minio/minio/internal/store"
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Target - Kafka target.
|
// the suffix for the configured queue dir where the logs will be persisted.
|
||||||
type Target struct {
|
const kafkaLoggerExtension = ".kafka.log"
|
||||||
totalMessages int64
|
|
||||||
failedMessages int64
|
|
||||||
|
|
||||||
wg sync.WaitGroup
|
|
||||||
doneCh chan struct{}
|
|
||||||
|
|
||||||
// Channel of log entries
|
|
||||||
logCh chan audit.Entry
|
|
||||||
|
|
||||||
// is the target online?
|
|
||||||
online bool
|
|
||||||
|
|
||||||
producer sarama.SyncProducer
|
|
||||||
kconfig Config
|
|
||||||
config *sarama.Config
|
|
||||||
}
|
|
||||||
|
|
||||||
// Send log message 'e' to kafka target.
|
|
||||||
func (h *Target) Send(entry interface{}) error {
|
|
||||||
if !h.online {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
select {
|
|
||||||
case <-h.doneCh:
|
|
||||||
return nil
|
|
||||||
default:
|
|
||||||
}
|
|
||||||
|
|
||||||
if e, ok := entry.(audit.Entry); ok {
|
|
||||||
select {
|
|
||||||
case <-h.doneCh:
|
|
||||||
case h.logCh <- e:
|
|
||||||
default:
|
|
||||||
// log channel is full, do not wait and return
|
|
||||||
// an error immediately to the caller
|
|
||||||
atomic.AddInt64(&h.totalMessages, 1)
|
|
||||||
atomic.AddInt64(&h.failedMessages, 1)
|
|
||||||
return errors.New("log buffer full")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (h *Target) logEntry(entry audit.Entry) {
|
|
||||||
atomic.AddInt64(&h.totalMessages, 1)
|
|
||||||
logJSON, err := json.Marshal(&entry)
|
|
||||||
if err != nil {
|
|
||||||
atomic.AddInt64(&h.failedMessages, 1)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
msg := sarama.ProducerMessage{
|
|
||||||
Topic: h.kconfig.Topic,
|
|
||||||
Key: sarama.StringEncoder(entry.RequestID),
|
|
||||||
Value: sarama.ByteEncoder(logJSON),
|
|
||||||
}
|
|
||||||
|
|
||||||
_, _, err = h.producer.SendMessage(&msg)
|
|
||||||
if err != nil {
|
|
||||||
atomic.AddInt64(&h.failedMessages, 1)
|
|
||||||
h.kconfig.LogOnce(context.Background(), err, h.kconfig.Topic)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (h *Target) startKakfaLogger() {
|
|
||||||
// Create a routine which sends json logs received
|
|
||||||
// from an internal channel.
|
|
||||||
h.wg.Add(1)
|
|
||||||
go func() {
|
|
||||||
defer h.wg.Done()
|
|
||||||
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case entry := <-h.logCh:
|
|
||||||
h.logEntry(entry)
|
|
||||||
case <-h.doneCh:
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
}
|
|
||||||
|
|
||||||
// Config - kafka target arguments.
|
// Config - kafka target arguments.
|
||||||
type Config struct {
|
type Config struct {
|
||||||
@ -143,6 +65,9 @@ type Config struct {
|
|||||||
Password string `json:"password"`
|
Password string `json:"password"`
|
||||||
Mechanism string `json:"mechanism"`
|
Mechanism string `json:"mechanism"`
|
||||||
} `json:"sasl"`
|
} `json:"sasl"`
|
||||||
|
// Queue store
|
||||||
|
QueueSize int `json:"queueSize"`
|
||||||
|
QueueDir string `json:"queueDir"`
|
||||||
|
|
||||||
// Custom logger
|
// Custom logger
|
||||||
LogOnce func(ctx context.Context, err error, id string, errKind ...interface{}) `json:"-"`
|
LogOnce func(ctx context.Context, err error, id string, errKind ...interface{}) `json:"-"`
|
||||||
@ -161,13 +86,44 @@ func (k Config) pingBrokers() (err error) {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Stats returns the target statistics.
|
// Target - Kafka target.
|
||||||
func (h *Target) Stats() types.TargetStats {
|
type Target struct {
|
||||||
return types.TargetStats{
|
totalMessages int64
|
||||||
TotalMessages: atomic.LoadInt64(&h.totalMessages),
|
failedMessages int64
|
||||||
FailedMessages: atomic.LoadInt64(&h.failedMessages),
|
|
||||||
QueueLength: len(h.logCh),
|
wg sync.WaitGroup
|
||||||
|
doneCh chan struct{}
|
||||||
|
|
||||||
|
// Channel of log entries
|
||||||
|
logCh chan audit.Entry
|
||||||
|
|
||||||
|
// store to persist and replay the logs to the target
|
||||||
|
// to avoid missing events when the target is down.
|
||||||
|
store store.Store[audit.Entry]
|
||||||
|
storeCtxCancel context.CancelFunc
|
||||||
|
initKafkaOnce once.Init
|
||||||
|
initQueueStoreOnce once.Init
|
||||||
|
|
||||||
|
producer sarama.SyncProducer
|
||||||
|
kconfig Config
|
||||||
|
config *sarama.Config
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *Target) validate() error {
|
||||||
|
if len(h.kconfig.Brokers) == 0 {
|
||||||
|
return errors.New("no broker address found")
|
||||||
}
|
}
|
||||||
|
for _, b := range h.kconfig.Brokers {
|
||||||
|
if _, err := xnet.ParseHost(b.String()); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Name returns the name of the target
|
||||||
|
func (h *Target) Name() string {
|
||||||
|
return "minio-kafka-audit"
|
||||||
}
|
}
|
||||||
|
|
||||||
// Endpoint - return kafka target
|
// Endpoint - return kafka target
|
||||||
@ -180,24 +136,95 @@ func (h *Target) String() string {
|
|||||||
return "kafka"
|
return "kafka"
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsOnline returns true if the initialization was successful
|
// Stats returns the target statistics.
|
||||||
func (h *Target) IsOnline() bool {
|
func (h *Target) Stats() types.TargetStats {
|
||||||
return h.online
|
return types.TargetStats{
|
||||||
|
TotalMessages: atomic.LoadInt64(&h.totalMessages),
|
||||||
|
FailedMessages: atomic.LoadInt64(&h.failedMessages),
|
||||||
|
QueueLength: len(h.logCh),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Init initialize kafka target
|
// Init initialize kafka target
|
||||||
func (h *Target) Init() error {
|
func (h *Target) Init(ctx context.Context) error {
|
||||||
if !h.kconfig.Enabled {
|
if !h.kconfig.Enabled {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if len(h.kconfig.Brokers) == 0 {
|
if err := h.validate(); err != nil {
|
||||||
return errors.New("no broker address found")
|
return err
|
||||||
}
|
}
|
||||||
for _, b := range h.kconfig.Brokers {
|
if h.kconfig.QueueDir != "" {
|
||||||
if _, err := xnet.ParseHost(b.String()); err != nil {
|
if err := h.initQueueStoreOnce.DoWithContext(ctx, h.initQueueStore); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
return h.initKafkaOnce.Do(h.init)
|
||||||
}
|
}
|
||||||
|
if err := h.init(); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
go h.startKakfaLogger()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *Target) initQueueStore(ctx context.Context) (err error) {
|
||||||
|
var queueStore store.Store[audit.Entry]
|
||||||
|
queueDir := filepath.Join(h.kconfig.QueueDir, h.Name())
|
||||||
|
queueStore = store.NewQueueStore[audit.Entry](queueDir, uint64(h.kconfig.QueueSize), kafkaLoggerExtension)
|
||||||
|
if err = queueStore.Open(); err != nil {
|
||||||
|
return fmt.Errorf("unable to initialize the queue store of %s webhook: %w", h.Name(), err)
|
||||||
|
}
|
||||||
|
ctx, cancel := context.WithCancel(ctx)
|
||||||
|
h.store = queueStore
|
||||||
|
h.storeCtxCancel = cancel
|
||||||
|
store.StreamItems(h.store, h, ctx.Done(), h.kconfig.LogOnce)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *Target) startKakfaLogger() {
|
||||||
|
// Create a routine which sends json logs received
|
||||||
|
// from an internal channel.
|
||||||
|
h.wg.Add(1)
|
||||||
|
go func() {
|
||||||
|
defer h.wg.Done()
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case entry := <-h.logCh:
|
||||||
|
h.logEntry(entry)
|
||||||
|
case <-h.doneCh:
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *Target) logEntry(entry audit.Entry) {
|
||||||
|
atomic.AddInt64(&h.totalMessages, 1)
|
||||||
|
if err := h.send(entry); err != nil {
|
||||||
|
atomic.AddInt64(&h.failedMessages, 1)
|
||||||
|
h.kconfig.LogOnce(context.Background(), err, h.kconfig.Topic)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *Target) send(entry audit.Entry) error {
|
||||||
|
if err := h.initKafkaOnce.Do(h.init); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
logJSON, err := json.Marshal(&entry)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
msg := sarama.ProducerMessage{
|
||||||
|
Topic: h.kconfig.Topic,
|
||||||
|
Key: sarama.StringEncoder(entry.RequestID),
|
||||||
|
Value: sarama.ByteEncoder(logJSON),
|
||||||
|
}
|
||||||
|
_, _, err = h.producer.SendMessage(&msg)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Init initialize kafka target
|
||||||
|
func (h *Target) init() error {
|
||||||
if err := h.kconfig.pingBrokers(); err != nil {
|
if err := h.kconfig.pingBrokers(); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -244,15 +271,73 @@ func (h *Target) Init() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
h.producer = producer
|
h.producer = producer
|
||||||
h.online = true
|
|
||||||
go h.startKakfaLogger()
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// IsOnline returns true if the target is online.
|
||||||
|
func (h *Target) IsOnline(_ context.Context) bool {
|
||||||
|
if err := h.initKafkaOnce.Do(h.init); err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return h.kconfig.pingBrokers() == nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Send log message 'e' to kafka target.
|
||||||
|
func (h *Target) Send(ctx context.Context, entry interface{}) error {
|
||||||
|
if auditEntry, ok := entry.(audit.Entry); ok {
|
||||||
|
if h.store != nil {
|
||||||
|
// save the entry to the queue store which will be replayed to the target.
|
||||||
|
return h.store.Put(auditEntry)
|
||||||
|
}
|
||||||
|
if err := h.initKafkaOnce.Do(h.init); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
select {
|
||||||
|
case <-h.doneCh:
|
||||||
|
case h.logCh <- auditEntry:
|
||||||
|
default:
|
||||||
|
// log channel is full, do not wait and return
|
||||||
|
// an error immediately to the caller
|
||||||
|
atomic.AddInt64(&h.totalMessages, 1)
|
||||||
|
atomic.AddInt64(&h.failedMessages, 1)
|
||||||
|
return errors.New("log buffer full")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// SendFromStore - reads the log from store and sends it to kafka.
|
||||||
|
func (h *Target) SendFromStore(key string) (err error) {
|
||||||
|
var auditEntry audit.Entry
|
||||||
|
auditEntry, err = h.store.Get(key)
|
||||||
|
if err != nil {
|
||||||
|
if os.IsNotExist(err) {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
atomic.AddInt64(&h.totalMessages, 1)
|
||||||
|
err = h.send(auditEntry)
|
||||||
|
if err != nil {
|
||||||
|
atomic.AddInt64(&h.failedMessages, 1)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// Delete the event from store.
|
||||||
|
return h.store.Del(key)
|
||||||
|
}
|
||||||
|
|
||||||
// Cancel - cancels the target
|
// Cancel - cancels the target
|
||||||
func (h *Target) Cancel() {
|
func (h *Target) Cancel() {
|
||||||
close(h.doneCh)
|
close(h.doneCh)
|
||||||
close(h.logCh)
|
close(h.logCh)
|
||||||
|
// If queuestore is configured, cancel it's context to
|
||||||
|
// stop the replay go-routine.
|
||||||
|
if h.store != nil {
|
||||||
|
h.storeCtxCancel()
|
||||||
|
}
|
||||||
|
if h.producer != nil {
|
||||||
|
h.producer.Close()
|
||||||
|
}
|
||||||
h.wg.Wait()
|
h.wg.Wait()
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -263,7 +348,6 @@ func New(config Config) *Target {
|
|||||||
logCh: make(chan audit.Entry, 10000),
|
logCh: make(chan audit.Entry, 10000),
|
||||||
doneCh: make(chan struct{}),
|
doneCh: make(chan struct{}),
|
||||||
kconfig: config,
|
kconfig: config,
|
||||||
online: false,
|
|
||||||
}
|
}
|
||||||
return target
|
return target
|
||||||
}
|
}
|
||||||
|
@ -18,6 +18,7 @@
|
|||||||
package logger
|
package logger
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
@ -36,10 +37,10 @@ type Target interface {
|
|||||||
String() string
|
String() string
|
||||||
Endpoint() string
|
Endpoint() string
|
||||||
Stats() types.TargetStats
|
Stats() types.TargetStats
|
||||||
Init() error
|
Init(ctx context.Context) error
|
||||||
IsOnline() bool
|
IsOnline(ctx context.Context) bool
|
||||||
Cancel()
|
Cancel()
|
||||||
Send(entry interface{}) error
|
Send(ctx context.Context, entry interface{}) error
|
||||||
Type() types.TargetType
|
Type() types.TargetType
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -57,12 +58,12 @@ var (
|
|||||||
)
|
)
|
||||||
|
|
||||||
// TargetStatus returns status of the target (online|offline)
|
// TargetStatus returns status of the target (online|offline)
|
||||||
func TargetStatus(h Target) madmin.Status {
|
func TargetStatus(ctx context.Context, h Target) madmin.Status {
|
||||||
if h.IsOnline() {
|
if h.IsOnline(ctx) {
|
||||||
return madmin.Status{Status: string(madmin.ItemOnline)}
|
return madmin.Status{Status: string(madmin.ItemOnline)}
|
||||||
}
|
}
|
||||||
// Previous initialization had failed. Try again.
|
// Previous initialization had failed. Try again.
|
||||||
if e := h.Init(); e == nil {
|
if e := h.Init(ctx); e == nil {
|
||||||
return madmin.Status{Status: string(madmin.ItemOnline)}
|
return madmin.Status{Status: string(madmin.ItemOnline)}
|
||||||
}
|
}
|
||||||
return madmin.Status{Status: string(madmin.ItemOffline)}
|
return madmin.Status{Status: string(madmin.ItemOffline)}
|
||||||
@ -124,8 +125,8 @@ var (
|
|||||||
|
|
||||||
// AddSystemTarget adds a new logger target to the
|
// AddSystemTarget adds a new logger target to the
|
||||||
// list of enabled loggers
|
// list of enabled loggers
|
||||||
func AddSystemTarget(t Target) error {
|
func AddSystemTarget(ctx context.Context, t Target) error {
|
||||||
if err := t.Init(); err != nil {
|
if err := t.Init(ctx); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -144,7 +145,7 @@ func AddSystemTarget(t Target) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func initSystemTargets(cfgMap map[string]http.Config) ([]Target, []error) {
|
func initSystemTargets(ctx context.Context, cfgMap map[string]http.Config) ([]Target, []error) {
|
||||||
tgts := []Target{}
|
tgts := []Target{}
|
||||||
errs := []error{}
|
errs := []error{}
|
||||||
for _, l := range cfgMap {
|
for _, l := range cfgMap {
|
||||||
@ -152,7 +153,7 @@ func initSystemTargets(cfgMap map[string]http.Config) ([]Target, []error) {
|
|||||||
t := http.New(l)
|
t := http.New(l)
|
||||||
tgts = append(tgts, t)
|
tgts = append(tgts, t)
|
||||||
|
|
||||||
e := t.Init()
|
e := t.Init(ctx)
|
||||||
if e != nil {
|
if e != nil {
|
||||||
errs = append(errs, e)
|
errs = append(errs, e)
|
||||||
}
|
}
|
||||||
@ -161,7 +162,7 @@ func initSystemTargets(cfgMap map[string]http.Config) ([]Target, []error) {
|
|||||||
return tgts, errs
|
return tgts, errs
|
||||||
}
|
}
|
||||||
|
|
||||||
func initKafkaTargets(cfgMap map[string]kafka.Config) ([]Target, []error) {
|
func initKafkaTargets(ctx context.Context, cfgMap map[string]kafka.Config) ([]Target, []error) {
|
||||||
tgts := []Target{}
|
tgts := []Target{}
|
||||||
errs := []error{}
|
errs := []error{}
|
||||||
for _, l := range cfgMap {
|
for _, l := range cfgMap {
|
||||||
@ -169,7 +170,7 @@ func initKafkaTargets(cfgMap map[string]kafka.Config) ([]Target, []error) {
|
|||||||
t := kafka.New(l)
|
t := kafka.New(l)
|
||||||
tgts = append(tgts, t)
|
tgts = append(tgts, t)
|
||||||
|
|
||||||
e := t.Init()
|
e := t.Init(ctx)
|
||||||
if e != nil {
|
if e != nil {
|
||||||
errs = append(errs, e)
|
errs = append(errs, e)
|
||||||
}
|
}
|
||||||
@ -200,8 +201,8 @@ func cancelTargets(targets []Target) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// UpdateSystemTargets swaps targets with newly loaded ones from the cfg
|
// UpdateSystemTargets swaps targets with newly loaded ones from the cfg
|
||||||
func UpdateSystemTargets(cfg Config) []error {
|
func UpdateSystemTargets(ctx context.Context, cfg Config) []error {
|
||||||
newTgts, errs := initSystemTargets(cfg.HTTP)
|
newTgts, errs := initSystemTargets(ctx, cfg.HTTP)
|
||||||
|
|
||||||
swapSystemMuRW.Lock()
|
swapSystemMuRW.Lock()
|
||||||
consoleTargets, otherTargets := splitTargets(systemTargets, types.TargetConsole)
|
consoleTargets, otherTargets := splitTargets(systemTargets, types.TargetConsole)
|
||||||
@ -214,8 +215,8 @@ func UpdateSystemTargets(cfg Config) []error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// UpdateAuditWebhookTargets swaps audit webhook targets with newly loaded ones from the cfg
|
// UpdateAuditWebhookTargets swaps audit webhook targets with newly loaded ones from the cfg
|
||||||
func UpdateAuditWebhookTargets(cfg Config) []error {
|
func UpdateAuditWebhookTargets(ctx context.Context, cfg Config) []error {
|
||||||
newWebhookTgts, errs := initSystemTargets(cfg.AuditWebhook)
|
newWebhookTgts, errs := initSystemTargets(ctx, cfg.AuditWebhook)
|
||||||
|
|
||||||
swapAuditMuRW.Lock()
|
swapAuditMuRW.Lock()
|
||||||
// Retain kafka targets
|
// Retain kafka targets
|
||||||
@ -229,8 +230,8 @@ func UpdateAuditWebhookTargets(cfg Config) []error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// UpdateAuditKafkaTargets swaps audit kafka targets with newly loaded ones from the cfg
|
// UpdateAuditKafkaTargets swaps audit kafka targets with newly loaded ones from the cfg
|
||||||
func UpdateAuditKafkaTargets(cfg Config) []error {
|
func UpdateAuditKafkaTargets(ctx context.Context, cfg Config) []error {
|
||||||
newKafkaTgts, errs := initKafkaTargets(cfg.AuditKafka)
|
newKafkaTgts, errs := initKafkaTargets(ctx, cfg.AuditKafka)
|
||||||
|
|
||||||
swapAuditMuRW.Lock()
|
swapAuditMuRW.Lock()
|
||||||
// Retain webhook targets
|
// Retain webhook targets
|
||||||
|
@ -15,9 +15,10 @@
|
|||||||
// You should have received a copy of the GNU Affero General Public License
|
// You should have received a copy of the GNU Affero General Public License
|
||||||
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
package target
|
package once
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
)
|
)
|
||||||
@ -25,19 +26,23 @@ import (
|
|||||||
// Inspired from Golang sync.Once but it is only marked
|
// Inspired from Golang sync.Once but it is only marked
|
||||||
// initialized when the provided function returns nil.
|
// initialized when the provided function returns nil.
|
||||||
|
|
||||||
type lazyInit struct {
|
// Init represents the structure.
|
||||||
|
type Init struct {
|
||||||
done uint32
|
done uint32
|
||||||
m sync.Mutex
|
m sync.Mutex
|
||||||
}
|
}
|
||||||
|
|
||||||
func (l *lazyInit) Do(f func() error) error {
|
// Do is similar to sync.Once.Do - makes one successful
|
||||||
|
// call to the function. ie, it invokes the function
|
||||||
|
// if it is not successful yet.
|
||||||
|
func (l *Init) Do(f func() error) error {
|
||||||
if atomic.LoadUint32(&l.done) == 0 {
|
if atomic.LoadUint32(&l.done) == 0 {
|
||||||
return l.doSlow(f)
|
return l.do(f)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (l *lazyInit) doSlow(f func() error) error {
|
func (l *Init) do(f func() error) error {
|
||||||
l.m.Lock()
|
l.m.Lock()
|
||||||
defer l.m.Unlock()
|
defer l.m.Unlock()
|
||||||
if atomic.LoadUint32(&l.done) == 0 {
|
if atomic.LoadUint32(&l.done) == 0 {
|
||||||
@ -49,3 +54,24 @@ func (l *lazyInit) doSlow(f func() error) error {
|
|||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// DoWithContext is similar to Do except that it accepts a context as an argument to be passed.
|
||||||
|
func (l *Init) DoWithContext(ctx context.Context, f func(context.Context) error) error {
|
||||||
|
if atomic.LoadUint32(&l.done) == 0 {
|
||||||
|
return l.doWithContext(ctx, f)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *Init) doWithContext(ctx context.Context, f func(context.Context) error) error {
|
||||||
|
l.m.Lock()
|
||||||
|
defer l.m.Unlock()
|
||||||
|
if atomic.LoadUint32(&l.done) == 0 {
|
||||||
|
if err := f(ctx); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
// Mark as done only when f() is successful
|
||||||
|
atomic.StoreUint32(&l.done, 1)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
@ -24,7 +24,6 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/minio/minio/internal/logger"
|
|
||||||
xnet "github.com/minio/pkg/net"
|
xnet "github.com/minio/pkg/net"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -32,13 +31,15 @@ const (
|
|||||||
retryInterval = 3 * time.Second
|
retryInterval = 3 * time.Second
|
||||||
)
|
)
|
||||||
|
|
||||||
|
type logger = func(ctx context.Context, err error, id string, errKind ...interface{})
|
||||||
|
|
||||||
// ErrNotConnected - indicates that the target connection is not active.
|
// ErrNotConnected - indicates that the target connection is not active.
|
||||||
var ErrNotConnected = errors.New("not connected to target server/service")
|
var ErrNotConnected = errors.New("not connected to target server/service")
|
||||||
|
|
||||||
// Target - store target interface
|
// Target - store target interface
|
||||||
type Target interface {
|
type Target interface {
|
||||||
Name() string
|
Name() string
|
||||||
Send(key string) error
|
SendFromStore(key string) error
|
||||||
}
|
}
|
||||||
|
|
||||||
// Store - Used to persist items.
|
// Store - Used to persist items.
|
||||||
@ -53,7 +54,7 @@ type Store[I any] interface {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// replayItems - Reads the items from the store and replays.
|
// replayItems - Reads the items from the store and replays.
|
||||||
func replayItems[I any](store Store[I], doneCh <-chan struct{}, loggerOnce logger.LogOnce, id string) <-chan string {
|
func replayItems[I any](store Store[I], doneCh <-chan struct{}, logger logger, id string) <-chan string {
|
||||||
itemKeyCh := make(chan string)
|
itemKeyCh := make(chan string)
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
@ -65,7 +66,7 @@ func replayItems[I any](store Store[I], doneCh <-chan struct{}, loggerOnce logge
|
|||||||
for {
|
for {
|
||||||
names, err := store.List()
|
names, err := store.List()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
loggerOnce(context.Background(), fmt.Errorf("store.List() failed with: %w", err), id)
|
logger(context.Background(), fmt.Errorf("store.List() failed with: %w", err), id)
|
||||||
} else {
|
} else {
|
||||||
for _, name := range names {
|
for _, name := range names {
|
||||||
select {
|
select {
|
||||||
@ -89,20 +90,20 @@ func replayItems[I any](store Store[I], doneCh <-chan struct{}, loggerOnce logge
|
|||||||
}
|
}
|
||||||
|
|
||||||
// sendItems - Reads items from the store and re-plays.
|
// sendItems - Reads items from the store and re-plays.
|
||||||
func sendItems(target Target, itemKeyCh <-chan string, doneCh <-chan struct{}, loggerOnce logger.LogOnce) {
|
func sendItems(target Target, itemKeyCh <-chan string, doneCh <-chan struct{}, logger logger) {
|
||||||
retryTicker := time.NewTicker(retryInterval)
|
retryTicker := time.NewTicker(retryInterval)
|
||||||
defer retryTicker.Stop()
|
defer retryTicker.Stop()
|
||||||
|
|
||||||
send := func(itemKey string) bool {
|
send := func(itemKey string) bool {
|
||||||
for {
|
for {
|
||||||
err := target.Send(itemKey)
|
err := target.SendFromStore(itemKey)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
if err != ErrNotConnected && !xnet.IsConnResetErr(err) {
|
if err != ErrNotConnected && !xnet.IsConnResetErr(err) {
|
||||||
loggerOnce(context.Background(),
|
logger(context.Background(),
|
||||||
fmt.Errorf("target.Send() failed with '%w'", err),
|
fmt.Errorf("target.SendFromStore() failed with '%w'", err),
|
||||||
target.Name())
|
target.Name())
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -135,11 +136,11 @@ func sendItems(target Target, itemKeyCh <-chan string, doneCh <-chan struct{}, l
|
|||||||
}
|
}
|
||||||
|
|
||||||
// StreamItems reads the keys from the store and replays the corresponding item to the target.
|
// StreamItems reads the keys from the store and replays the corresponding item to the target.
|
||||||
func StreamItems[I any](store Store[I], target Target, doneCh <-chan struct{}, loggerOnce logger.LogOnce) {
|
func StreamItems[I any](store Store[I], target Target, doneCh <-chan struct{}, logger logger) {
|
||||||
go func() {
|
go func() {
|
||||||
// Replays the items from the store.
|
// Replays the items from the store.
|
||||||
itemKeyCh := replayItems(store, doneCh, loggerOnce, target.Name())
|
itemKeyCh := replayItems(store, doneCh, logger, target.Name())
|
||||||
// Send items from the store.
|
// Send items from the store.
|
||||||
sendItems(target, itemKeyCh, doneCh, loggerOnce)
|
sendItems(target, itemKeyCh, doneCh, logger)
|
||||||
}()
|
}()
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user