Support persistent queue store for loggers (#17121)

This commit is contained in:
Praveen raj Mani 2023-05-09 09:50:31 +05:30 committed by GitHub
parent 42fb3cd95e
commit 57acacd5a7
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
25 changed files with 516 additions and 300 deletions

View File

@ -2511,13 +2511,13 @@ func fetchLoggerInfo() ([]madmin.Logger, []madmin.Audit) {
var auditloggerInfo []madmin.Audit var auditloggerInfo []madmin.Audit
for _, tgt := range logger.SystemTargets() { for _, tgt := range logger.SystemTargets() {
if tgt.Endpoint() != "" { if tgt.Endpoint() != "" {
loggerInfo = append(loggerInfo, madmin.Logger{tgt.String(): logger.TargetStatus(tgt)}) loggerInfo = append(loggerInfo, madmin.Logger{tgt.String(): logger.TargetStatus(GlobalContext, tgt)})
} }
} }
for _, tgt := range logger.AuditTargets() { for _, tgt := range logger.AuditTargets() {
if tgt.Endpoint() != "" { if tgt.Endpoint() != "" {
auditloggerInfo = append(auditloggerInfo, madmin.Audit{tgt.String(): logger.TargetStatus(tgt)}) auditloggerInfo = append(auditloggerInfo, madmin.Audit{tgt.String(): logger.TargetStatus(GlobalContext, tgt)})
} }
} }

View File

@ -588,7 +588,7 @@ func applyDynamicConfigForSubSys(ctx context.Context, objAPI ObjectLayer, s conf
} }
loggerCfg.HTTP[n] = l loggerCfg.HTTP[n] = l
} }
if errs := logger.UpdateSystemTargets(loggerCfg); len(errs) > 0 { if errs := logger.UpdateSystemTargets(ctx, loggerCfg); len(errs) > 0 {
logger.LogIf(ctx, fmt.Errorf("Unable to update logger webhook config: %v", errs)) logger.LogIf(ctx, fmt.Errorf("Unable to update logger webhook config: %v", errs))
} }
case config.AuditWebhookSubSys: case config.AuditWebhookSubSys:
@ -606,7 +606,7 @@ func applyDynamicConfigForSubSys(ctx context.Context, objAPI ObjectLayer, s conf
loggerCfg.AuditWebhook[n] = l loggerCfg.AuditWebhook[n] = l
} }
if errs := logger.UpdateAuditWebhookTargets(loggerCfg); len(errs) > 0 { if errs := logger.UpdateAuditWebhookTargets(ctx, loggerCfg); len(errs) > 0 {
logger.LogIf(ctx, fmt.Errorf("Unable to update audit webhook targets: %v", errs)) logger.LogIf(ctx, fmt.Errorf("Unable to update audit webhook targets: %v", errs))
} }
case config.AuditKafkaSubSys: case config.AuditKafkaSubSys:
@ -620,7 +620,7 @@ func applyDynamicConfigForSubSys(ctx context.Context, objAPI ObjectLayer, s conf
loggerCfg.AuditKafka[n] = l loggerCfg.AuditKafka[n] = l
} }
} }
if errs := logger.UpdateAuditKafkaTargets(loggerCfg); len(errs) > 0 { if errs := logger.UpdateAuditKafkaTargets(ctx, loggerCfg); len(errs) > 0 {
logger.LogIf(ctx, fmt.Errorf("Unable to update audit kafka targets: %v", errs)) logger.LogIf(ctx, fmt.Errorf("Unable to update audit kafka targets: %v", errs))
} }
case config.StorageClassSubSys: case config.StorageClassSubSys:

View File

@ -58,7 +58,7 @@ func NewConsoleLogger(ctx context.Context) *HTTPConsoleLoggerSys {
} }
// IsOnline always true in case of console logger // IsOnline always true in case of console logger
func (sys *HTTPConsoleLoggerSys) IsOnline() bool { func (sys *HTTPConsoleLoggerSys) IsOnline(_ context.Context) bool {
return true return true
} }
@ -87,7 +87,7 @@ func (sys *HTTPConsoleLoggerSys) HasLogListeners() bool {
func (sys *HTTPConsoleLoggerSys) Subscribe(subCh chan log.Info, doneCh <-chan struct{}, node string, last int, logKind madmin.LogMask, filter func(entry log.Info) bool) error { func (sys *HTTPConsoleLoggerSys) Subscribe(subCh chan log.Info, doneCh <-chan struct{}, node string, last int, logKind madmin.LogMask, filter func(entry log.Info) bool) error {
// Enable console logging for remote client. // Enable console logging for remote client.
if !sys.HasLogListeners() { if !sys.HasLogListeners() {
logger.AddSystemTarget(sys) logger.AddSystemTarget(GlobalContext, sys)
} }
cnt := 0 cnt := 0
@ -128,7 +128,7 @@ func (sys *HTTPConsoleLoggerSys) Subscribe(subCh chan log.Info, doneCh <-chan st
} }
// Init if HTTPConsoleLoggerSys is valid, always returns nil right now // Init if HTTPConsoleLoggerSys is valid, always returns nil right now
func (sys *HTTPConsoleLoggerSys) Init() error { func (sys *HTTPConsoleLoggerSys) Init(_ context.Context) error {
return nil return nil
} }
@ -180,7 +180,7 @@ func (sys *HTTPConsoleLoggerSys) Type() types.TargetType {
// Send log message 'e' to console and publish to console // Send log message 'e' to console and publish to console
// log pubsub system // log pubsub system
func (sys *HTTPConsoleLoggerSys) Send(entry interface{}) error { func (sys *HTTPConsoleLoggerSys) Send(ctx context.Context, entry interface{}) error {
var lg log.Info var lg log.Info
switch e := entry.(type) { switch e := entry.(type) {
case log.Entry: case log.Entry:

View File

@ -516,7 +516,7 @@ func serverMain(ctx *cli.Context) {
// Initialize globalConsoleSys system // Initialize globalConsoleSys system
globalConsoleSys = NewConsoleLogger(GlobalContext) globalConsoleSys = NewConsoleLogger(GlobalContext)
logger.AddSystemTarget(globalConsoleSys) logger.AddSystemTarget(GlobalContext, globalConsoleSys)
// Perform any self-tests // Perform any self-tests
bitrotSelfTest() bitrotSelfTest()

View File

@ -42,7 +42,7 @@ func printStartupMessage(apiEndpoints []string, err error) {
logger.Info(color.Bold("MinIO Object Storage Server")) logger.Info(color.Bold("MinIO Object Storage Server"))
if err != nil { if err != nil {
if globalConsoleSys != nil { if globalConsoleSys != nil {
globalConsoleSys.Send(fmt.Sprintf("Server startup failed with '%v', some features may be missing", err)) globalConsoleSys.Send(GlobalContext, fmt.Sprintf("Server startup failed with '%v', some features may be missing", err))
} }
} }

View File

@ -30,6 +30,7 @@ import (
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
"github.com/rabbitmq/amqp091-go" "github.com/rabbitmq/amqp091-go"
@ -112,7 +113,7 @@ func (a *AMQPArgs) Validate() error {
// AMQPTarget - AMQP target // AMQPTarget - AMQP target
type AMQPTarget struct { type AMQPTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args AMQPArgs args AMQPArgs
@ -289,8 +290,8 @@ func (target *AMQPTarget) Save(eventData event.Event) error {
return target.send(eventData, ch, confirms) return target.send(eventData, ch, confirms)
} }
// Send - sends event to AMQP091. // SendFromStore - reads an event from store and sends it to AMQP091.
func (target *AMQPTarget) Send(eventKey string) error { func (target *AMQPTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -329,7 +330,7 @@ func (target *AMQPTarget) Close() error {
} }
func (target *AMQPTarget) init() error { func (target *AMQPTarget) init() error {
return target.lazyInit.Do(target.initAMQP) return target.initOnce.Do(target.initAMQP)
} }
func (target *AMQPTarget) initAMQP() error { func (target *AMQPTarget) initAMQP() error {

View File

@ -36,6 +36,7 @@ import (
"github.com/minio/highwayhash" "github.com/minio/highwayhash"
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
"github.com/pkg/errors" "github.com/pkg/errors"
@ -153,7 +154,7 @@ func (a ElasticsearchArgs) Validate() error {
// ElasticsearchTarget - Elasticsearch target. // ElasticsearchTarget - Elasticsearch target.
type ElasticsearchTarget struct { type ElasticsearchTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args ElasticsearchArgs args ElasticsearchArgs
@ -263,8 +264,8 @@ func (target *ElasticsearchTarget) send(eventData event.Event) error {
return nil return nil
} }
// Send - reads an event from store and sends it to Elasticsearch. // SendFromStore - reads an event from store and sends it to Elasticsearch.
func (target *ElasticsearchTarget) Send(eventKey string) error { func (target *ElasticsearchTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -344,7 +345,7 @@ func (target *ElasticsearchTarget) checkAndInitClient(ctx context.Context) error
} }
func (target *ElasticsearchTarget) init() error { func (target *ElasticsearchTarget) init() error {
return target.lazyInit.Do(target.initElasticsearch) return target.initOnce.Do(target.initElasticsearch)
} }
func (target *ElasticsearchTarget) initElasticsearch() error { func (target *ElasticsearchTarget) initElasticsearch() error {

View File

@ -32,6 +32,7 @@ import (
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
@ -125,7 +126,7 @@ func (k KafkaArgs) Validate() error {
// KafkaTarget - Kafka target. // KafkaTarget - Kafka target.
type KafkaTarget struct { type KafkaTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args KafkaArgs args KafkaArgs
@ -208,8 +209,8 @@ func (target *KafkaTarget) send(eventData event.Event) error {
return err return err
} }
// Send - reads an event from store and sends it to Kafka. // SendFromStore - reads an event from store and sends it to Kafka.
func (target *KafkaTarget) Send(eventKey string) error { func (target *KafkaTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -278,7 +279,7 @@ func (k KafkaArgs) pingBrokers() bool {
} }
func (target *KafkaTarget) init() error { func (target *KafkaTarget) init() error {
return target.lazyInit.Do(target.initKafka) return target.initOnce.Do(target.initKafka)
} }
func (target *KafkaTarget) initKafka() error { func (target *KafkaTarget) initKafka() error {

View File

@ -31,6 +31,7 @@ import (
mqtt "github.com/eclipse/paho.mqtt.golang" mqtt "github.com/eclipse/paho.mqtt.golang"
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
) )
@ -107,7 +108,7 @@ func (m MQTTArgs) Validate() error {
// MQTTTarget - MQTT target. // MQTTTarget - MQTT target.
type MQTTTarget struct { type MQTTTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args MQTTArgs args MQTTArgs
@ -167,8 +168,8 @@ func (target *MQTTTarget) send(eventData event.Event) error {
return token.Error() return token.Error()
} }
// Send - reads an event from store and sends it to MQTT. // SendFromStore - reads an event from store and sends it to MQTT.
func (target *MQTTTarget) Send(eventKey string) error { func (target *MQTTTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -224,7 +225,7 @@ func (target *MQTTTarget) Close() error {
} }
func (target *MQTTTarget) init() error { func (target *MQTTTarget) init() error {
return target.lazyInit.Do(target.initMQTT) return target.initOnce.Do(target.initMQTT)
} }
func (target *MQTTTarget) initMQTT() error { func (target *MQTTTarget) initMQTT() error {

View File

@ -33,6 +33,7 @@ import (
"github.com/go-sql-driver/mysql" "github.com/go-sql-driver/mysql"
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
) )
@ -146,7 +147,7 @@ func (m MySQLArgs) Validate() error {
// MySQLTarget - MySQL target. // MySQLTarget - MySQL target.
type MySQLTarget struct { type MySQLTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args MySQLArgs args MySQLArgs
@ -252,8 +253,8 @@ func (target *MySQLTarget) send(eventData event.Event) error {
return nil return nil
} }
// Send - reads an event from store and sends it to MySQL. // SendFromStore - reads an event from store and sends it to MySQL.
func (target *MySQLTarget) Send(eventKey string) error { func (target *MySQLTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -349,7 +350,7 @@ func (target *MySQLTarget) executeStmts() error {
} }
func (target *MySQLTarget) init() error { func (target *MySQLTarget) init() error {
return target.lazyInit.Do(target.initMySQL) return target.initOnce.Do(target.initMySQL)
} }
func (target *MySQLTarget) initMySQL() error { func (target *MySQLTarget) initMySQL() error {

View File

@ -31,6 +31,7 @@ import (
"github.com/google/uuid" "github.com/google/uuid"
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
"github.com/nats-io/nats.go" "github.com/nats-io/nats.go"
@ -216,7 +217,7 @@ func (n NATSArgs) connectStan() (stan.Conn, error) {
// NATSTarget - NATS target. // NATSTarget - NATS target.
type NATSTarget struct { type NATSTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args NATSArgs args NATSArgs
@ -333,8 +334,8 @@ func (target *NATSTarget) send(eventData event.Event) error {
return err return err
} }
// Send - sends event to Nats. // SendFromStore - reads an event from store and sends it to Nats.
func (target *NATSTarget) Send(eventKey string) error { func (target *NATSTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -380,7 +381,7 @@ func (target *NATSTarget) Close() (err error) {
} }
func (target *NATSTarget) init() error { func (target *NATSTarget) init() error {
return target.lazyInit.Do(target.initNATS) return target.initOnce.Do(target.initNATS)
} }
func (target *NATSTarget) initNATS() error { func (target *NATSTarget) initNATS() error {

View File

@ -31,6 +31,7 @@ import (
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
) )
@ -90,7 +91,7 @@ func (n NSQArgs) Validate() error {
// NSQTarget - NSQ target. // NSQTarget - NSQ target.
type NSQTarget struct { type NSQTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args NSQArgs args NSQArgs
@ -176,8 +177,8 @@ func (target *NSQTarget) send(eventData event.Event) error {
return target.producer.Publish(target.args.Topic, data) return target.producer.Publish(target.args.Topic, data)
} }
// Send - reads an event from store and sends it to NSQ. // SendFromStore - reads an event from store and sends it to NSQ.
func (target *NSQTarget) Send(eventKey string) error { func (target *NSQTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -216,7 +217,7 @@ func (target *NSQTarget) Close() (err error) {
} }
func (target *NSQTarget) init() error { func (target *NSQTarget) init() error {
return target.lazyInit.Do(target.initNSQ) return target.initOnce.Do(target.initNSQ)
} }
func (target *NSQTarget) initNSQ() error { func (target *NSQTarget) initNSQ() error {

View File

@ -34,6 +34,7 @@ import (
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
) )
@ -138,7 +139,7 @@ func (p PostgreSQLArgs) Validate() error {
// PostgreSQLTarget - PostgreSQL target. // PostgreSQLTarget - PostgreSQL target.
type PostgreSQLTarget struct { type PostgreSQLTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args PostgreSQLArgs args PostgreSQLArgs
@ -249,8 +250,8 @@ func (target *PostgreSQLTarget) send(eventData event.Event) error {
return nil return nil
} }
// Send - reads an event from store and sends it to PostgreSQL. // SendFromStore - reads an event from store and sends it to PostgreSQL.
func (target *PostgreSQLTarget) Send(eventKey string) error { func (target *PostgreSQLTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -345,7 +346,7 @@ func (target *PostgreSQLTarget) executeStmts() error {
} }
func (target *PostgreSQLTarget) init() error { func (target *PostgreSQLTarget) init() error {
return target.lazyInit.Do(target.initPostgreSQL) return target.initOnce.Do(target.initPostgreSQL)
} }
func (target *PostgreSQLTarget) initPostgreSQL() error { func (target *PostgreSQLTarget) initPostgreSQL() error {

View File

@ -31,6 +31,7 @@ import (
"github.com/gomodule/redigo/redis" "github.com/gomodule/redigo/redis"
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
) )
@ -118,7 +119,7 @@ func (r RedisArgs) validateFormat(c redis.Conn) error {
// RedisTarget - Redis target. // RedisTarget - Redis target.
type RedisTarget struct { type RedisTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args RedisArgs args RedisArgs
@ -221,8 +222,8 @@ func (target *RedisTarget) send(eventData event.Event) error {
return nil return nil
} }
// Send - reads an event from store and sends it to redis. // SendFromStore - reads an event from store and sends it to redis.
func (target *RedisTarget) Send(eventKey string) error { func (target *RedisTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -276,7 +277,7 @@ func (target *RedisTarget) Close() error {
} }
func (target *RedisTarget) init() error { func (target *RedisTarget) init() error {
return target.lazyInit.Do(target.initRedis) return target.initOnce.Do(target.initRedis)
} }
func (target *RedisTarget) initRedis() error { func (target *RedisTarget) initRedis() error {

View File

@ -35,6 +35,7 @@ import (
"github.com/minio/minio/internal/event" "github.com/minio/minio/internal/event"
"github.com/minio/minio/internal/logger" "github.com/minio/minio/internal/logger"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store" "github.com/minio/minio/internal/store"
"github.com/minio/pkg/certs" "github.com/minio/pkg/certs"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
@ -91,7 +92,7 @@ func (w WebhookArgs) Validate() error {
// WebhookTarget - Webhook target. // WebhookTarget - Webhook target.
type WebhookTarget struct { type WebhookTarget struct {
lazyInit lazyInit initOnce once.Init
id event.TargetID id event.TargetID
args WebhookArgs args WebhookArgs
@ -222,8 +223,8 @@ func (target *WebhookTarget) send(eventData event.Event) error {
return nil return nil
} }
// Send - reads an event from store and sends it to webhook. // SendFromStore - reads an event from store and sends it to webhook.
func (target *WebhookTarget) Send(eventKey string) error { func (target *WebhookTarget) SendFromStore(eventKey string) error {
if err := target.init(); err != nil { if err := target.init(); err != nil {
return err return err
} }
@ -256,7 +257,7 @@ func (target *WebhookTarget) Close() error {
} }
func (target *WebhookTarget) init() error { func (target *WebhookTarget) init() error {
return target.lazyInit.Do(target.initWebhook) return target.initOnce.Do(target.initWebhook)
} }
// Only called from init() // Only called from init()

View File

@ -34,7 +34,7 @@ type Target interface {
ID() TargetID ID() TargetID
IsActive() (bool, error) IsActive() (bool, error)
Save(Event) error Save(Event) error
Send(string) error SendFromStore(string) error
Close() error Close() error
Store() TargetStore Store() TargetStore
} }

View File

@ -60,8 +60,8 @@ func (target ExampleTarget) send(eventData Event) error {
return nil return nil
} }
// Send - interface compatible method does no-op. // SendFromStore - interface compatible method does no-op.
func (target ExampleTarget) Send(eventKey string) error { func (target *ExampleTarget) SendFromStore(eventKey string) error {
return nil return nil
} }

View File

@ -139,7 +139,7 @@ func AuditLog(ctx context.Context, w http.ResponseWriter, r *http.Request, reqCl
// Send audit logs only to http targets. // Send audit logs only to http targets.
for _, t := range auditTgts { for _, t := range auditTgts {
if err := t.Send(entry); err != nil { if err := t.Send(ctx, entry); err != nil {
LogAlwaysIf(context.Background(), fmt.Errorf("event(%v) was not sent to Audit target (%v): %v", entry, t, err), madmin.LogKindAll) LogAlwaysIf(context.Background(), fmt.Errorf("event(%v) was not sent to Audit target (%v): %v", entry, t, err), madmin.LogKindAll)
} }
} }

View File

@ -43,6 +43,7 @@ const (
ClientCert = "client_cert" ClientCert = "client_cert"
ClientKey = "client_key" ClientKey = "client_key"
QueueSize = "queue_size" QueueSize = "queue_size"
QueueDir = "queue_dir"
Proxy = "proxy" Proxy = "proxy"
KafkaBrokers = "brokers" KafkaBrokers = "brokers"
@ -57,6 +58,8 @@ const (
KafkaClientTLSCert = "client_tls_cert" KafkaClientTLSCert = "client_tls_cert"
KafkaClientTLSKey = "client_tls_key" KafkaClientTLSKey = "client_tls_key"
KafkaVersion = "version" KafkaVersion = "version"
KafkaQueueDir = "queue_dir"
KafkaQueueSize = "queue_size"
EnvLoggerWebhookEnable = "MINIO_LOGGER_WEBHOOK_ENABLE" EnvLoggerWebhookEnable = "MINIO_LOGGER_WEBHOOK_ENABLE"
EnvLoggerWebhookEndpoint = "MINIO_LOGGER_WEBHOOK_ENDPOINT" EnvLoggerWebhookEndpoint = "MINIO_LOGGER_WEBHOOK_ENDPOINT"
@ -65,6 +68,7 @@ const (
EnvLoggerWebhookClientKey = "MINIO_LOGGER_WEBHOOK_CLIENT_KEY" EnvLoggerWebhookClientKey = "MINIO_LOGGER_WEBHOOK_CLIENT_KEY"
EnvLoggerWebhookProxy = "MINIO_LOGGER_WEBHOOK_PROXY" EnvLoggerWebhookProxy = "MINIO_LOGGER_WEBHOOK_PROXY"
EnvLoggerWebhookQueueSize = "MINIO_LOGGER_WEBHOOK_QUEUE_SIZE" EnvLoggerWebhookQueueSize = "MINIO_LOGGER_WEBHOOK_QUEUE_SIZE"
EnvLoggerWebhookQueueDir = "MINIO_LOGGER_WEBHOOK_QUEUE_DIR"
EnvAuditWebhookEnable = "MINIO_AUDIT_WEBHOOK_ENABLE" EnvAuditWebhookEnable = "MINIO_AUDIT_WEBHOOK_ENABLE"
EnvAuditWebhookEndpoint = "MINIO_AUDIT_WEBHOOK_ENDPOINT" EnvAuditWebhookEndpoint = "MINIO_AUDIT_WEBHOOK_ENDPOINT"
@ -72,6 +76,7 @@ const (
EnvAuditWebhookClientCert = "MINIO_AUDIT_WEBHOOK_CLIENT_CERT" EnvAuditWebhookClientCert = "MINIO_AUDIT_WEBHOOK_CLIENT_CERT"
EnvAuditWebhookClientKey = "MINIO_AUDIT_WEBHOOK_CLIENT_KEY" EnvAuditWebhookClientKey = "MINIO_AUDIT_WEBHOOK_CLIENT_KEY"
EnvAuditWebhookQueueSize = "MINIO_AUDIT_WEBHOOK_QUEUE_SIZE" EnvAuditWebhookQueueSize = "MINIO_AUDIT_WEBHOOK_QUEUE_SIZE"
EnvAuditWebhookQueueDir = "MINIO_AUDIT_WEBHOOK_QUEUE_DIR"
EnvKafkaEnable = "MINIO_AUDIT_KAFKA_ENABLE" EnvKafkaEnable = "MINIO_AUDIT_KAFKA_ENABLE"
EnvKafkaBrokers = "MINIO_AUDIT_KAFKA_BROKERS" EnvKafkaBrokers = "MINIO_AUDIT_KAFKA_BROKERS"
@ -86,6 +91,11 @@ const (
EnvKafkaClientTLSCert = "MINIO_AUDIT_KAFKA_CLIENT_TLS_CERT" EnvKafkaClientTLSCert = "MINIO_AUDIT_KAFKA_CLIENT_TLS_CERT"
EnvKafkaClientTLSKey = "MINIO_AUDIT_KAFKA_CLIENT_TLS_KEY" EnvKafkaClientTLSKey = "MINIO_AUDIT_KAFKA_CLIENT_TLS_KEY"
EnvKafkaVersion = "MINIO_AUDIT_KAFKA_VERSION" EnvKafkaVersion = "MINIO_AUDIT_KAFKA_VERSION"
EnvKafkaQueueDir = "MINIO_AUDIT_KAFKA_QUEUE_DIR"
EnvKafkaQueueSize = "MINIO_AUDIT_KAFKA_QUEUE_SIZE"
loggerTargetNamePrefix = "logger-"
auditTargetNamePrefix = "audit-"
) )
// Default KVS for loggerHTTP and loggerAuditHTTP // Default KVS for loggerHTTP and loggerAuditHTTP
@ -119,6 +129,10 @@ var (
Key: QueueSize, Key: QueueSize,
Value: "100000", Value: "100000",
}, },
config.KV{
Key: QueueDir,
Value: "",
},
} }
DefaultAuditWebhookKVS = config.KVS{ DefaultAuditWebhookKVS = config.KVS{
@ -146,6 +160,10 @@ var (
Key: QueueSize, Key: QueueSize,
Value: "100000", Value: "100000",
}, },
config.KV{
Key: QueueDir,
Value: "",
},
} }
DefaultAuditKafkaKVS = config.KVS{ DefaultAuditKafkaKVS = config.KVS{
@ -201,6 +219,14 @@ var (
Key: KafkaVersion, Key: KafkaVersion,
Value: "", Value: "",
}, },
config.KV{
Key: QueueSize,
Value: "100000",
},
config.KV{
Key: QueueDir,
Value: "",
},
} }
) )
@ -395,6 +421,25 @@ func lookupAuditKafkaConfig(scfg config.Config, cfg Config) (Config, error) {
kafkaArgs.SASL.Password = env.Get(saslPasswordEnv, kv.Get(KafkaSASLPassword)) kafkaArgs.SASL.Password = env.Get(saslPasswordEnv, kv.Get(KafkaSASLPassword))
kafkaArgs.SASL.Mechanism = env.Get(saslMechanismEnv, kv.Get(KafkaSASLMechanism)) kafkaArgs.SASL.Mechanism = env.Get(saslMechanismEnv, kv.Get(KafkaSASLMechanism))
queueDirEnv := EnvKafkaQueueDir
if k != config.Default {
queueDirEnv = queueDirEnv + config.Default + k
}
kafkaArgs.QueueDir = env.Get(queueDirEnv, kv.Get(KafkaQueueDir))
queueSizeEnv := EnvKafkaQueueSize
if k != config.Default {
queueSizeEnv = queueSizeEnv + config.Default + k
}
queueSize, err := strconv.Atoi(env.Get(queueSizeEnv, kv.Get(KafkaQueueSize)))
if err != nil {
return cfg, err
}
if queueSize <= 0 {
return cfg, errors.New("invalid queue_size value")
}
kafkaArgs.QueueSize = queueSize
cfg.AuditKafka[k] = kafkaArgs cfg.AuditKafka[k] = kafkaArgs
} }
@ -448,9 +493,9 @@ func lookupLoggerWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
return cfg, err return cfg, err
} }
proxyEnv := EnvLoggerWebhookProxy proxyEnv := EnvLoggerWebhookProxy
queueSizeEnv := EnvAuditWebhookQueueSize queueSizeEnv := EnvLoggerWebhookQueueSize
if target != config.Default { if target != config.Default {
queueSizeEnv = EnvAuditWebhookQueueSize + config.Default + target queueSizeEnv = EnvLoggerWebhookQueueSize + config.Default + target
} }
queueSize, err := strconv.Atoi(env.Get(queueSizeEnv, "100000")) queueSize, err := strconv.Atoi(env.Get(queueSizeEnv, "100000"))
if err != nil { if err != nil {
@ -459,6 +504,10 @@ func lookupLoggerWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
if queueSize <= 0 { if queueSize <= 0 {
return cfg, errors.New("invalid queue_size value") return cfg, errors.New("invalid queue_size value")
} }
queueDirEnv := EnvLoggerWebhookQueueDir
if target != config.Default {
queueDirEnv = EnvLoggerWebhookQueueDir + config.Default + target
}
cfg.HTTP[target] = http.Config{ cfg.HTTP[target] = http.Config{
Enabled: true, Enabled: true,
Endpoint: env.Get(endpointEnv, ""), Endpoint: env.Get(endpointEnv, ""),
@ -467,7 +516,8 @@ func lookupLoggerWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
ClientKey: env.Get(clientKeyEnv, ""), ClientKey: env.Get(clientKeyEnv, ""),
Proxy: env.Get(proxyEnv, ""), Proxy: env.Get(proxyEnv, ""),
QueueSize: queueSize, QueueSize: queueSize,
Name: target, QueueDir: env.Get(queueDirEnv, ""),
Name: loggerTargetNamePrefix + target,
} }
} }
@ -511,7 +561,8 @@ func lookupLoggerWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
ClientKey: kv.Get(ClientKey), ClientKey: kv.Get(ClientKey),
Proxy: kv.Get(Proxy), Proxy: kv.Get(Proxy),
QueueSize: queueSize, QueueSize: queueSize,
Name: starget, QueueDir: kv.Get(QueueDir),
Name: loggerTargetNamePrefix + starget,
} }
} }
@ -574,6 +625,10 @@ func lookupAuditWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
if queueSize <= 0 { if queueSize <= 0 {
return cfg, errors.New("invalid queue_size value") return cfg, errors.New("invalid queue_size value")
} }
queueDirEnv := EnvAuditWebhookQueueDir
if target != config.Default {
queueDirEnv = EnvAuditWebhookQueueDir + config.Default + target
}
cfg.AuditWebhook[target] = http.Config{ cfg.AuditWebhook[target] = http.Config{
Enabled: true, Enabled: true,
Endpoint: env.Get(endpointEnv, ""), Endpoint: env.Get(endpointEnv, ""),
@ -581,7 +636,8 @@ func lookupAuditWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
ClientCert: env.Get(clientCertEnv, ""), ClientCert: env.Get(clientCertEnv, ""),
ClientKey: env.Get(clientKeyEnv, ""), ClientKey: env.Get(clientKeyEnv, ""),
QueueSize: queueSize, QueueSize: queueSize,
Name: target, QueueDir: env.Get(queueDirEnv, ""),
Name: auditTargetNamePrefix + target,
} }
} }
@ -617,7 +673,6 @@ func lookupAuditWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
if queueSize <= 0 { if queueSize <= 0 {
return cfg, errors.New("invalid queue_size value") return cfg, errors.New("invalid queue_size value")
} }
cfg.AuditWebhook[starget] = http.Config{ cfg.AuditWebhook[starget] = http.Config{
Enabled: true, Enabled: true,
Endpoint: kv.Get(Endpoint), Endpoint: kv.Get(Endpoint),
@ -625,7 +680,8 @@ func lookupAuditWebhookConfig(scfg config.Config, cfg Config) (Config, error) {
ClientCert: kv.Get(ClientCert), ClientCert: kv.Get(ClientCert),
ClientKey: kv.Get(ClientKey), ClientKey: kv.Get(ClientKey),
QueueSize: queueSize, QueueSize: queueSize,
Name: starget, QueueDir: kv.Get(QueueDir),
Name: auditTargetNamePrefix + starget,
} }
} }

View File

@ -347,7 +347,7 @@ func consoleLogIf(ctx context.Context, err error, errKind ...interface{}) {
if consoleTgt != nil { if consoleTgt != nil {
entry := errToEntry(ctx, err, errKind...) entry := errToEntry(ctx, err, errKind...)
consoleTgt.Send(entry) consoleTgt.Send(ctx, entry)
} }
} }
@ -366,10 +366,10 @@ func logIf(ctx context.Context, err error, errKind ...interface{}) {
entry := errToEntry(ctx, err, errKind...) entry := errToEntry(ctx, err, errKind...)
// Iterate over all logger targets to send the log entry // Iterate over all logger targets to send the log entry
for _, t := range systemTgts { for _, t := range systemTgts {
if err := t.Send(entry); err != nil { if err := t.Send(ctx, entry); err != nil {
if consoleTgt != nil { if consoleTgt != nil {
entry.Trace.Message = fmt.Sprintf("event(%#v) was not sent to Logger target (%#v): %#v", entry, t, err) entry.Trace.Message = fmt.Sprintf("event(%#v) was not sent to Logger target (%#v): %#v", entry, t, err)
consoleTgt.Send(entry) consoleTgt.Send(ctx, entry)
} }
} }
} }

View File

@ -1,4 +1,4 @@
// Copyright (c) 2015-2021 MinIO, Inc. // Copyright (c) 2015-2023 MinIO, Inc.
// //
// This file is part of MinIO Object Storage stack // This file is part of MinIO Object Storage stack
// //
@ -26,13 +26,17 @@ import (
"math" "math"
"net/http" "net/http"
"net/url" "net/url"
"strings" "os"
"path/filepath"
"sync" "sync"
"sync/atomic" "sync/atomic"
"time" "time"
xhttp "github.com/minio/minio/internal/http" xhttp "github.com/minio/minio/internal/http"
"github.com/minio/minio/internal/logger/target/types" "github.com/minio/minio/internal/logger/target/types"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net"
) )
const ( const (
@ -41,6 +45,9 @@ const (
// maxWorkers is the maximum number of concurrent operations. // maxWorkers is the maximum number of concurrent operations.
maxWorkers = 16 maxWorkers = 16
// the suffix for the configured queue dir where the logs will be persisted.
httpLoggerExtension = ".http.log"
) )
const ( const (
@ -59,6 +66,7 @@ type Config struct {
ClientCert string `json:"clientCert"` ClientCert string `json:"clientCert"`
ClientKey string `json:"clientKey"` ClientKey string `json:"clientKey"`
QueueSize int `json:"queueSize"` QueueSize int `json:"queueSize"`
QueueDir string `json:"queueDir"`
Proxy string `json:"string"` Proxy string `json:"string"`
Transport http.RoundTripper `json:"-"` Transport http.RoundTripper `json:"-"`
@ -93,10 +101,21 @@ type Target struct {
// will attempt to establish the connection. // will attempt to establish the connection.
revive sync.Once revive sync.Once
// store to persist and replay the logs to the target
// to avoid missing events when the target is down.
store store.Store[interface{}]
storeCtxCancel context.CancelFunc
initQueueStoreOnce once.Init
config Config config Config
client *http.Client client *http.Client
} }
// Name returns the name of the target
func (h *Target) Name() string {
return "minio-http-" + h.config.Name
}
// Endpoint returns the backend endpoint // Endpoint returns the backend endpoint
func (h *Target) Endpoint() string { func (h *Target) Endpoint() string {
return h.config.Endpoint return h.config.Endpoint
@ -106,9 +125,9 @@ func (h *Target) String() string {
return h.config.Name return h.config.Name
} }
// IsOnline returns true if the initialization was successful // IsOnline returns true if the target is reachable.
func (h *Target) IsOnline() bool { func (h *Target) IsOnline(ctx context.Context) bool {
return atomic.LoadInt32(&h.status) == statusOnline return h.isAlive(ctx) == nil
} }
// Stats returns the target statistics. // Stats returns the target statistics.
@ -125,8 +144,34 @@ func (h *Target) Stats() types.TargetStats {
return stats return stats
} }
// This will check if we can reach the remote.
func (h *Target) isAlive(ctx context.Context) (err error) {
return h.send(ctx, []byte(`{}`), 2*webhookCallTimeout)
}
// Init validate and initialize the http target // Init validate and initialize the http target
func (h *Target) Init() (err error) { func (h *Target) Init(ctx context.Context) (err error) {
if h.config.QueueDir != "" {
return h.initQueueStoreOnce.DoWithContext(ctx, h.initQueueStore)
}
return h.initLogChannel(ctx)
}
func (h *Target) initQueueStore(ctx context.Context) (err error) {
var queueStore store.Store[interface{}]
queueDir := filepath.Join(h.config.QueueDir, h.Name())
queueStore = store.NewQueueStore[interface{}](queueDir, uint64(h.config.QueueSize), httpLoggerExtension)
if err = queueStore.Open(); err != nil {
return fmt.Errorf("unable to initialize the queue store of %s webhook: %w", h.Name(), err)
}
ctx, cancel := context.WithCancel(ctx)
h.store = queueStore
h.storeCtxCancel = cancel
store.StreamItems(h.store, h, ctx.Done(), h.config.LogOnce)
return
}
func (h *Target) initLogChannel(ctx context.Context) (err error) {
switch atomic.LoadInt32(&h.status) { switch atomic.LoadInt32(&h.status) {
case statusOnline: case statusOnline:
return nil return nil
@ -134,45 +179,7 @@ func (h *Target) Init() (err error) {
return errors.New("target is closed") return errors.New("target is closed")
} }
// This will check if we can reach the remote. err = h.isAlive(ctx)
checkAlive := func() error {
ctx, cancel := context.WithTimeout(context.Background(), 2*webhookCallTimeout)
defer cancel()
req, err := http.NewRequestWithContext(ctx, http.MethodPost, h.config.Endpoint, strings.NewReader(`{}`))
if err != nil {
return err
}
req.Header.Set(xhttp.ContentType, "application/json")
// Set user-agent to indicate MinIO release
// version to the configured log endpoint
req.Header.Set("User-Agent", h.config.UserAgent)
if h.config.AuthToken != "" {
req.Header.Set("Authorization", h.config.AuthToken)
}
resp, err := h.client.Do(req)
if err != nil {
return err
}
// Drain any response.
xhttp.DrainBody(resp.Body)
if !acceptedResponseStatusCode(resp.StatusCode) {
if resp.StatusCode == http.StatusForbidden {
return fmt.Errorf("%s returned '%s', please check if your auth token is correctly set",
h.config.Endpoint, resp.Status)
}
return fmt.Errorf("%s returned '%s', please check your endpoint configuration",
h.config.Endpoint, resp.Status)
}
return nil
}
err = checkAlive()
if err != nil { if err != nil {
// Start a goroutine that will continue to check if we can reach // Start a goroutine that will continue to check if we can reach
h.revive.Do(func() { h.revive.Do(func() {
@ -183,14 +190,14 @@ func (h *Target) Init() (err error) {
if atomic.LoadInt32(&h.status) != statusOffline { if atomic.LoadInt32(&h.status) != statusOffline {
return return
} }
if err := checkAlive(); err == nil { if err := h.isAlive(ctx); err == nil {
// We are online. // We are online.
if atomic.CompareAndSwapInt32(&h.status, statusOffline, statusOnline) { if atomic.CompareAndSwapInt32(&h.status, statusOffline, statusOnline) {
h.workerStartMu.Lock() h.workerStartMu.Lock()
h.lastStarted = time.Now() h.lastStarted = time.Now()
h.workerStartMu.Unlock() h.workerStartMu.Unlock()
atomic.AddInt64(&h.workers, 1) atomic.AddInt64(&h.workers, 1)
go h.startHTTPLogger() go h.startHTTPLogger(ctx)
} }
return return
} }
@ -205,19 +212,51 @@ func (h *Target) Init() (err error) {
h.lastStarted = time.Now() h.lastStarted = time.Now()
h.workerStartMu.Unlock() h.workerStartMu.Unlock()
atomic.AddInt64(&h.workers, 1) atomic.AddInt64(&h.workers, 1)
go h.startHTTPLogger() go h.startHTTPLogger(ctx)
} }
return nil return nil
} }
// Accepted HTTP Status Codes func (h *Target) send(ctx context.Context, payload []byte, timeout time.Duration) (err error) {
var acceptedStatusCodeMap = map[int]bool{http.StatusOK: true, http.StatusCreated: true, http.StatusAccepted: true, http.StatusNoContent: true} ctx, cancel := context.WithTimeout(ctx, timeout)
defer cancel()
req, err := http.NewRequestWithContext(ctx, http.MethodPost,
h.config.Endpoint, bytes.NewReader(payload))
if err != nil {
return fmt.Errorf("%s returned '%w', please check your endpoint configuration", h.config.Endpoint, err)
}
req.Header.Set(xhttp.ContentType, "application/json")
req.Header.Set(xhttp.MinIOVersion, xhttp.GlobalMinIOVersion)
req.Header.Set(xhttp.MinioDeploymentID, xhttp.GlobalDeploymentID)
func acceptedResponseStatusCode(code int) bool { // Set user-agent to indicate MinIO release
return acceptedStatusCodeMap[code] // version to the configured log endpoint
req.Header.Set("User-Agent", h.config.UserAgent)
if h.config.AuthToken != "" {
req.Header.Set("Authorization", h.config.AuthToken)
}
resp, err := h.client.Do(req)
if err != nil {
return fmt.Errorf("%s returned '%w', please check your endpoint configuration", h.config.Endpoint, err)
}
// Drain any response.
xhttp.DrainBody(resp.Body)
switch resp.StatusCode {
case http.StatusOK, http.StatusCreated, http.StatusAccepted, http.StatusNoContent:
// accepted HTTP status codes.
return nil
case http.StatusForbidden:
return fmt.Errorf("%s returned '%s', please check if your auth token is correctly set", h.config.Endpoint, resp.Status)
default:
return fmt.Errorf("%s returned '%s', please check your endpoint configuration", h.config.Endpoint, resp.Status)
}
} }
func (h *Target) logEntry(entry interface{}) { func (h *Target) logEntry(ctx context.Context, entry interface{}) {
logJSON, err := json.Marshal(&entry) logJSON, err := json.Marshal(&entry)
if err != nil { if err != nil {
atomic.AddInt64(&h.failedMessages, 1) atomic.AddInt64(&h.failedMessages, 1)
@ -239,52 +278,14 @@ func (h *Target) logEntry(entry interface{}) {
time.Sleep(sleep) time.Sleep(sleep)
} }
tries++ tries++
ctx, cancel := context.WithTimeout(context.Background(), webhookCallTimeout) if err := h.send(ctx, logJSON, webhookCallTimeout); err != nil {
defer cancel() h.config.LogOnce(ctx, err, h.config.Endpoint)
req, err := http.NewRequestWithContext(ctx, http.MethodPost,
h.config.Endpoint, bytes.NewReader(logJSON))
if err != nil {
h.config.LogOnce(ctx, fmt.Errorf("%s returned '%w', please check your endpoint configuration", h.config.Endpoint, err), h.config.Endpoint)
atomic.AddInt64(&h.failedMessages, 1) atomic.AddInt64(&h.failedMessages, 1)
continue
}
req.Header.Set(xhttp.ContentType, "application/json")
req.Header.Set(xhttp.MinIOVersion, xhttp.GlobalMinIOVersion)
req.Header.Set(xhttp.MinioDeploymentID, xhttp.GlobalDeploymentID)
// Set user-agent to indicate MinIO release
// version to the configured log endpoint
req.Header.Set("User-Agent", h.config.UserAgent)
if h.config.AuthToken != "" {
req.Header.Set("Authorization", h.config.AuthToken)
}
resp, err := h.client.Do(req)
if err != nil {
atomic.AddInt64(&h.failedMessages, 1)
h.config.LogOnce(ctx, fmt.Errorf("%s returned '%w', please check your endpoint configuration", h.config.Endpoint, err), h.config.Endpoint)
continue
}
// Drain any response.
xhttp.DrainBody(resp.Body)
if acceptedResponseStatusCode(resp.StatusCode) {
return
}
// Log failure, retry
atomic.AddInt64(&h.failedMessages, 1)
switch resp.StatusCode {
case http.StatusForbidden:
h.config.LogOnce(ctx, fmt.Errorf("%s returned '%s', please check if your auth token is correctly set", h.config.Endpoint, resp.Status), h.config.Endpoint)
default:
h.config.LogOnce(ctx, fmt.Errorf("%s returned '%s', please check your endpoint configuration", h.config.Endpoint, resp.Status), h.config.Endpoint)
} }
} }
} }
func (h *Target) startHTTPLogger() { func (h *Target) startHTTPLogger(ctx context.Context) {
h.logChMu.RLock() h.logChMu.RLock()
logCh := h.logCh logCh := h.logCh
if logCh != nil { if logCh != nil {
@ -302,7 +303,7 @@ func (h *Target) startHTTPLogger() {
// Send messages until channel is closed. // Send messages until channel is closed.
for entry := range logCh { for entry := range logCh {
atomic.AddInt64(&h.totalMessages, 1) atomic.AddInt64(&h.totalMessages, 1)
h.logEntry(entry) h.logEntry(ctx, entry)
} }
} }
@ -328,10 +329,41 @@ func New(config Config) *Target {
return h return h
} }
// SendFromStore - reads the log from store and sends it to webhook.
func (h *Target) SendFromStore(key string) (err error) {
var eventData interface{}
eventData, err = h.store.Get(key)
if err != nil {
if os.IsNotExist(err) {
return nil
}
return err
}
atomic.AddInt64(&h.totalMessages, 1)
logJSON, err := json.Marshal(&eventData)
if err != nil {
atomic.AddInt64(&h.failedMessages, 1)
return
}
if err := h.send(context.Background(), logJSON, webhookCallTimeout); err != nil {
atomic.AddInt64(&h.failedMessages, 1)
if xnet.IsNetworkOrHostDown(err, true) {
return store.ErrNotConnected
}
return err
}
// Delete the event from store.
return h.store.Del(key)
}
// Send log message 'e' to http target. // Send log message 'e' to http target.
// If servers are offline messages are queued until queue is full. // If servers are offline messages are queued until queue is full.
// If Cancel has been called the message is ignored. // If Cancel has been called the message is ignored.
func (h *Target) Send(entry interface{}) error { func (h *Target) Send(ctx context.Context, entry interface{}) error {
if h.store != nil {
// save the entry to the queue store which will be replayed to the target.
return h.store.Put(entry)
}
if atomic.LoadInt32(&h.status) == statusClosed { if atomic.LoadInt32(&h.status) == statusClosed {
return nil return nil
} }
@ -345,7 +377,7 @@ func (h *Target) Send(entry interface{}) error {
case h.logCh <- entry: case h.logCh <- entry:
default: default:
// Drop messages until we are online. // Drop messages until we are online.
if !h.IsOnline() { if !h.IsOnline(ctx) {
return errors.New("log buffer full and remote offline") return errors.New("log buffer full and remote offline")
} }
nWorkers := atomic.LoadInt64(&h.workers) nWorkers := atomic.LoadInt64(&h.workers)
@ -358,7 +390,7 @@ func (h *Target) Send(entry interface{}) error {
if atomic.CompareAndSwapInt64(&h.workers, nWorkers, nWorkers+1) { if atomic.CompareAndSwapInt64(&h.workers, nWorkers, nWorkers+1) {
// Start another logger. // Start another logger.
h.lastStarted = time.Now() h.lastStarted = time.Now()
go h.startHTTPLogger() go h.startHTTPLogger(ctx)
} }
} }
h.logCh <- entry h.logCh <- entry
@ -380,6 +412,12 @@ func (h *Target) Send(entry interface{}) error {
func (h *Target) Cancel() { func (h *Target) Cancel() {
atomic.StoreInt32(&h.status, statusClosed) atomic.StoreInt32(&h.status, statusClosed)
// If queuestore is configured, cancel it's context to
// stop the replay go-routine.
if h.store != nil {
h.storeCtxCancel()
}
// Set logch to nil and close it. // Set logch to nil and close it.
// This will block all Send operations, // This will block all Send operations,
// and finish the existing ones. // and finish the existing ones.

View File

@ -1,4 +1,4 @@
// Copyright (c) 2015-2022 MinIO, Inc. // Copyright (c) 2015-2023 MinIO, Inc.
// //
// This file is part of MinIO Object Storage stack // This file is part of MinIO Object Storage stack
// //
@ -23,7 +23,10 @@ import (
"crypto/x509" "crypto/x509"
"encoding/json" "encoding/json"
"errors" "errors"
"fmt"
"net" "net"
"os"
"path/filepath"
"sync" "sync"
"sync/atomic" "sync/atomic"
"time" "time"
@ -34,94 +37,13 @@ import (
saramatls "github.com/Shopify/sarama/tools/tls" saramatls "github.com/Shopify/sarama/tools/tls"
"github.com/minio/minio/internal/logger/target/types" "github.com/minio/minio/internal/logger/target/types"
"github.com/minio/minio/internal/once"
"github.com/minio/minio/internal/store"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
) )
// Target - Kafka target. // the suffix for the configured queue dir where the logs will be persisted.
type Target struct { const kafkaLoggerExtension = ".kafka.log"
totalMessages int64
failedMessages int64
wg sync.WaitGroup
doneCh chan struct{}
// Channel of log entries
logCh chan audit.Entry
// is the target online?
online bool
producer sarama.SyncProducer
kconfig Config
config *sarama.Config
}
// Send log message 'e' to kafka target.
func (h *Target) Send(entry interface{}) error {
if !h.online {
return nil
}
select {
case <-h.doneCh:
return nil
default:
}
if e, ok := entry.(audit.Entry); ok {
select {
case <-h.doneCh:
case h.logCh <- e:
default:
// log channel is full, do not wait and return
// an error immediately to the caller
atomic.AddInt64(&h.totalMessages, 1)
atomic.AddInt64(&h.failedMessages, 1)
return errors.New("log buffer full")
}
}
return nil
}
func (h *Target) logEntry(entry audit.Entry) {
atomic.AddInt64(&h.totalMessages, 1)
logJSON, err := json.Marshal(&entry)
if err != nil {
atomic.AddInt64(&h.failedMessages, 1)
return
}
msg := sarama.ProducerMessage{
Topic: h.kconfig.Topic,
Key: sarama.StringEncoder(entry.RequestID),
Value: sarama.ByteEncoder(logJSON),
}
_, _, err = h.producer.SendMessage(&msg)
if err != nil {
atomic.AddInt64(&h.failedMessages, 1)
h.kconfig.LogOnce(context.Background(), err, h.kconfig.Topic)
return
}
}
func (h *Target) startKakfaLogger() {
// Create a routine which sends json logs received
// from an internal channel.
h.wg.Add(1)
go func() {
defer h.wg.Done()
for {
select {
case entry := <-h.logCh:
h.logEntry(entry)
case <-h.doneCh:
return
}
}
}()
}
// Config - kafka target arguments. // Config - kafka target arguments.
type Config struct { type Config struct {
@ -143,6 +65,9 @@ type Config struct {
Password string `json:"password"` Password string `json:"password"`
Mechanism string `json:"mechanism"` Mechanism string `json:"mechanism"`
} `json:"sasl"` } `json:"sasl"`
// Queue store
QueueSize int `json:"queueSize"`
QueueDir string `json:"queueDir"`
// Custom logger // Custom logger
LogOnce func(ctx context.Context, err error, id string, errKind ...interface{}) `json:"-"` LogOnce func(ctx context.Context, err error, id string, errKind ...interface{}) `json:"-"`
@ -161,13 +86,44 @@ func (k Config) pingBrokers() (err error) {
return nil return nil
} }
// Stats returns the target statistics. // Target - Kafka target.
func (h *Target) Stats() types.TargetStats { type Target struct {
return types.TargetStats{ totalMessages int64
TotalMessages: atomic.LoadInt64(&h.totalMessages), failedMessages int64
FailedMessages: atomic.LoadInt64(&h.failedMessages),
QueueLength: len(h.logCh), wg sync.WaitGroup
doneCh chan struct{}
// Channel of log entries
logCh chan audit.Entry
// store to persist and replay the logs to the target
// to avoid missing events when the target is down.
store store.Store[audit.Entry]
storeCtxCancel context.CancelFunc
initKafkaOnce once.Init
initQueueStoreOnce once.Init
producer sarama.SyncProducer
kconfig Config
config *sarama.Config
}
func (h *Target) validate() error {
if len(h.kconfig.Brokers) == 0 {
return errors.New("no broker address found")
} }
for _, b := range h.kconfig.Brokers {
if _, err := xnet.ParseHost(b.String()); err != nil {
return err
}
}
return nil
}
// Name returns the name of the target
func (h *Target) Name() string {
return "minio-kafka-audit"
} }
// Endpoint - return kafka target // Endpoint - return kafka target
@ -180,24 +136,95 @@ func (h *Target) String() string {
return "kafka" return "kafka"
} }
// IsOnline returns true if the initialization was successful // Stats returns the target statistics.
func (h *Target) IsOnline() bool { func (h *Target) Stats() types.TargetStats {
return h.online return types.TargetStats{
TotalMessages: atomic.LoadInt64(&h.totalMessages),
FailedMessages: atomic.LoadInt64(&h.failedMessages),
QueueLength: len(h.logCh),
}
} }
// Init initialize kafka target // Init initialize kafka target
func (h *Target) Init() error { func (h *Target) Init(ctx context.Context) error {
if !h.kconfig.Enabled { if !h.kconfig.Enabled {
return nil return nil
} }
if len(h.kconfig.Brokers) == 0 { if err := h.validate(); err != nil {
return errors.New("no broker address found") return err
} }
for _, b := range h.kconfig.Brokers { if h.kconfig.QueueDir != "" {
if _, err := xnet.ParseHost(b.String()); err != nil { if err := h.initQueueStoreOnce.DoWithContext(ctx, h.initQueueStore); err != nil {
return err return err
} }
return h.initKafkaOnce.Do(h.init)
} }
if err := h.init(); err != nil {
return err
}
go h.startKakfaLogger()
return nil
}
func (h *Target) initQueueStore(ctx context.Context) (err error) {
var queueStore store.Store[audit.Entry]
queueDir := filepath.Join(h.kconfig.QueueDir, h.Name())
queueStore = store.NewQueueStore[audit.Entry](queueDir, uint64(h.kconfig.QueueSize), kafkaLoggerExtension)
if err = queueStore.Open(); err != nil {
return fmt.Errorf("unable to initialize the queue store of %s webhook: %w", h.Name(), err)
}
ctx, cancel := context.WithCancel(ctx)
h.store = queueStore
h.storeCtxCancel = cancel
store.StreamItems(h.store, h, ctx.Done(), h.kconfig.LogOnce)
return
}
func (h *Target) startKakfaLogger() {
// Create a routine which sends json logs received
// from an internal channel.
h.wg.Add(1)
go func() {
defer h.wg.Done()
for {
select {
case entry := <-h.logCh:
h.logEntry(entry)
case <-h.doneCh:
return
}
}
}()
}
func (h *Target) logEntry(entry audit.Entry) {
atomic.AddInt64(&h.totalMessages, 1)
if err := h.send(entry); err != nil {
atomic.AddInt64(&h.failedMessages, 1)
h.kconfig.LogOnce(context.Background(), err, h.kconfig.Topic)
}
}
func (h *Target) send(entry audit.Entry) error {
if err := h.initKafkaOnce.Do(h.init); err != nil {
return err
}
logJSON, err := json.Marshal(&entry)
if err != nil {
return err
}
msg := sarama.ProducerMessage{
Topic: h.kconfig.Topic,
Key: sarama.StringEncoder(entry.RequestID),
Value: sarama.ByteEncoder(logJSON),
}
_, _, err = h.producer.SendMessage(&msg)
return err
}
// Init initialize kafka target
func (h *Target) init() error {
if err := h.kconfig.pingBrokers(); err != nil { if err := h.kconfig.pingBrokers(); err != nil {
return err return err
} }
@ -244,15 +271,73 @@ func (h *Target) Init() error {
} }
h.producer = producer h.producer = producer
h.online = true
go h.startKakfaLogger()
return nil return nil
} }
// IsOnline returns true if the target is online.
func (h *Target) IsOnline(_ context.Context) bool {
if err := h.initKafkaOnce.Do(h.init); err != nil {
return false
}
return h.kconfig.pingBrokers() == nil
}
// Send log message 'e' to kafka target.
func (h *Target) Send(ctx context.Context, entry interface{}) error {
if auditEntry, ok := entry.(audit.Entry); ok {
if h.store != nil {
// save the entry to the queue store which will be replayed to the target.
return h.store.Put(auditEntry)
}
if err := h.initKafkaOnce.Do(h.init); err != nil {
return err
}
select {
case <-h.doneCh:
case h.logCh <- auditEntry:
default:
// log channel is full, do not wait and return
// an error immediately to the caller
atomic.AddInt64(&h.totalMessages, 1)
atomic.AddInt64(&h.failedMessages, 1)
return errors.New("log buffer full")
}
}
return nil
}
// SendFromStore - reads the log from store and sends it to kafka.
func (h *Target) SendFromStore(key string) (err error) {
var auditEntry audit.Entry
auditEntry, err = h.store.Get(key)
if err != nil {
if os.IsNotExist(err) {
return nil
}
return err
}
atomic.AddInt64(&h.totalMessages, 1)
err = h.send(auditEntry)
if err != nil {
atomic.AddInt64(&h.failedMessages, 1)
return
}
// Delete the event from store.
return h.store.Del(key)
}
// Cancel - cancels the target // Cancel - cancels the target
func (h *Target) Cancel() { func (h *Target) Cancel() {
close(h.doneCh) close(h.doneCh)
close(h.logCh) close(h.logCh)
// If queuestore is configured, cancel it's context to
// stop the replay go-routine.
if h.store != nil {
h.storeCtxCancel()
}
if h.producer != nil {
h.producer.Close()
}
h.wg.Wait() h.wg.Wait()
} }
@ -263,7 +348,6 @@ func New(config Config) *Target {
logCh: make(chan audit.Entry, 10000), logCh: make(chan audit.Entry, 10000),
doneCh: make(chan struct{}), doneCh: make(chan struct{}),
kconfig: config, kconfig: config,
online: false,
} }
return target return target
} }

View File

@ -18,6 +18,7 @@
package logger package logger
import ( import (
"context"
"fmt" "fmt"
"strings" "strings"
"sync" "sync"
@ -36,10 +37,10 @@ type Target interface {
String() string String() string
Endpoint() string Endpoint() string
Stats() types.TargetStats Stats() types.TargetStats
Init() error Init(ctx context.Context) error
IsOnline() bool IsOnline(ctx context.Context) bool
Cancel() Cancel()
Send(entry interface{}) error Send(ctx context.Context, entry interface{}) error
Type() types.TargetType Type() types.TargetType
} }
@ -57,12 +58,12 @@ var (
) )
// TargetStatus returns status of the target (online|offline) // TargetStatus returns status of the target (online|offline)
func TargetStatus(h Target) madmin.Status { func TargetStatus(ctx context.Context, h Target) madmin.Status {
if h.IsOnline() { if h.IsOnline(ctx) {
return madmin.Status{Status: string(madmin.ItemOnline)} return madmin.Status{Status: string(madmin.ItemOnline)}
} }
// Previous initialization had failed. Try again. // Previous initialization had failed. Try again.
if e := h.Init(); e == nil { if e := h.Init(ctx); e == nil {
return madmin.Status{Status: string(madmin.ItemOnline)} return madmin.Status{Status: string(madmin.ItemOnline)}
} }
return madmin.Status{Status: string(madmin.ItemOffline)} return madmin.Status{Status: string(madmin.ItemOffline)}
@ -124,8 +125,8 @@ var (
// AddSystemTarget adds a new logger target to the // AddSystemTarget adds a new logger target to the
// list of enabled loggers // list of enabled loggers
func AddSystemTarget(t Target) error { func AddSystemTarget(ctx context.Context, t Target) error {
if err := t.Init(); err != nil { if err := t.Init(ctx); err != nil {
return err return err
} }
@ -144,7 +145,7 @@ func AddSystemTarget(t Target) error {
return nil return nil
} }
func initSystemTargets(cfgMap map[string]http.Config) ([]Target, []error) { func initSystemTargets(ctx context.Context, cfgMap map[string]http.Config) ([]Target, []error) {
tgts := []Target{} tgts := []Target{}
errs := []error{} errs := []error{}
for _, l := range cfgMap { for _, l := range cfgMap {
@ -152,7 +153,7 @@ func initSystemTargets(cfgMap map[string]http.Config) ([]Target, []error) {
t := http.New(l) t := http.New(l)
tgts = append(tgts, t) tgts = append(tgts, t)
e := t.Init() e := t.Init(ctx)
if e != nil { if e != nil {
errs = append(errs, e) errs = append(errs, e)
} }
@ -161,7 +162,7 @@ func initSystemTargets(cfgMap map[string]http.Config) ([]Target, []error) {
return tgts, errs return tgts, errs
} }
func initKafkaTargets(cfgMap map[string]kafka.Config) ([]Target, []error) { func initKafkaTargets(ctx context.Context, cfgMap map[string]kafka.Config) ([]Target, []error) {
tgts := []Target{} tgts := []Target{}
errs := []error{} errs := []error{}
for _, l := range cfgMap { for _, l := range cfgMap {
@ -169,7 +170,7 @@ func initKafkaTargets(cfgMap map[string]kafka.Config) ([]Target, []error) {
t := kafka.New(l) t := kafka.New(l)
tgts = append(tgts, t) tgts = append(tgts, t)
e := t.Init() e := t.Init(ctx)
if e != nil { if e != nil {
errs = append(errs, e) errs = append(errs, e)
} }
@ -200,8 +201,8 @@ func cancelTargets(targets []Target) {
} }
// UpdateSystemTargets swaps targets with newly loaded ones from the cfg // UpdateSystemTargets swaps targets with newly loaded ones from the cfg
func UpdateSystemTargets(cfg Config) []error { func UpdateSystemTargets(ctx context.Context, cfg Config) []error {
newTgts, errs := initSystemTargets(cfg.HTTP) newTgts, errs := initSystemTargets(ctx, cfg.HTTP)
swapSystemMuRW.Lock() swapSystemMuRW.Lock()
consoleTargets, otherTargets := splitTargets(systemTargets, types.TargetConsole) consoleTargets, otherTargets := splitTargets(systemTargets, types.TargetConsole)
@ -214,8 +215,8 @@ func UpdateSystemTargets(cfg Config) []error {
} }
// UpdateAuditWebhookTargets swaps audit webhook targets with newly loaded ones from the cfg // UpdateAuditWebhookTargets swaps audit webhook targets with newly loaded ones from the cfg
func UpdateAuditWebhookTargets(cfg Config) []error { func UpdateAuditWebhookTargets(ctx context.Context, cfg Config) []error {
newWebhookTgts, errs := initSystemTargets(cfg.AuditWebhook) newWebhookTgts, errs := initSystemTargets(ctx, cfg.AuditWebhook)
swapAuditMuRW.Lock() swapAuditMuRW.Lock()
// Retain kafka targets // Retain kafka targets
@ -229,8 +230,8 @@ func UpdateAuditWebhookTargets(cfg Config) []error {
} }
// UpdateAuditKafkaTargets swaps audit kafka targets with newly loaded ones from the cfg // UpdateAuditKafkaTargets swaps audit kafka targets with newly loaded ones from the cfg
func UpdateAuditKafkaTargets(cfg Config) []error { func UpdateAuditKafkaTargets(ctx context.Context, cfg Config) []error {
newKafkaTgts, errs := initKafkaTargets(cfg.AuditKafka) newKafkaTgts, errs := initKafkaTargets(ctx, cfg.AuditKafka)
swapAuditMuRW.Lock() swapAuditMuRW.Lock()
// Retain webhook targets // Retain webhook targets

View File

@ -15,9 +15,10 @@
// You should have received a copy of the GNU Affero General Public License // You should have received a copy of the GNU Affero General Public License
// along with this program. If not, see <http://www.gnu.org/licenses/>. // along with this program. If not, see <http://www.gnu.org/licenses/>.
package target package once
import ( import (
"context"
"sync" "sync"
"sync/atomic" "sync/atomic"
) )
@ -25,19 +26,23 @@ import (
// Inspired from Golang sync.Once but it is only marked // Inspired from Golang sync.Once but it is only marked
// initialized when the provided function returns nil. // initialized when the provided function returns nil.
type lazyInit struct { // Init represents the structure.
type Init struct {
done uint32 done uint32
m sync.Mutex m sync.Mutex
} }
func (l *lazyInit) Do(f func() error) error { // Do is similar to sync.Once.Do - makes one successful
// call to the function. ie, it invokes the function
// if it is not successful yet.
func (l *Init) Do(f func() error) error {
if atomic.LoadUint32(&l.done) == 0 { if atomic.LoadUint32(&l.done) == 0 {
return l.doSlow(f) return l.do(f)
} }
return nil return nil
} }
func (l *lazyInit) doSlow(f func() error) error { func (l *Init) do(f func() error) error {
l.m.Lock() l.m.Lock()
defer l.m.Unlock() defer l.m.Unlock()
if atomic.LoadUint32(&l.done) == 0 { if atomic.LoadUint32(&l.done) == 0 {
@ -49,3 +54,24 @@ func (l *lazyInit) doSlow(f func() error) error {
} }
return nil return nil
} }
// DoWithContext is similar to Do except that it accepts a context as an argument to be passed.
func (l *Init) DoWithContext(ctx context.Context, f func(context.Context) error) error {
if atomic.LoadUint32(&l.done) == 0 {
return l.doWithContext(ctx, f)
}
return nil
}
func (l *Init) doWithContext(ctx context.Context, f func(context.Context) error) error {
l.m.Lock()
defer l.m.Unlock()
if atomic.LoadUint32(&l.done) == 0 {
if err := f(ctx); err != nil {
return err
}
// Mark as done only when f() is successful
atomic.StoreUint32(&l.done, 1)
}
return nil
}

View File

@ -24,7 +24,6 @@ import (
"strings" "strings"
"time" "time"
"github.com/minio/minio/internal/logger"
xnet "github.com/minio/pkg/net" xnet "github.com/minio/pkg/net"
) )
@ -32,13 +31,15 @@ const (
retryInterval = 3 * time.Second retryInterval = 3 * time.Second
) )
type logger = func(ctx context.Context, err error, id string, errKind ...interface{})
// ErrNotConnected - indicates that the target connection is not active. // ErrNotConnected - indicates that the target connection is not active.
var ErrNotConnected = errors.New("not connected to target server/service") var ErrNotConnected = errors.New("not connected to target server/service")
// Target - store target interface // Target - store target interface
type Target interface { type Target interface {
Name() string Name() string
Send(key string) error SendFromStore(key string) error
} }
// Store - Used to persist items. // Store - Used to persist items.
@ -53,7 +54,7 @@ type Store[I any] interface {
} }
// replayItems - Reads the items from the store and replays. // replayItems - Reads the items from the store and replays.
func replayItems[I any](store Store[I], doneCh <-chan struct{}, loggerOnce logger.LogOnce, id string) <-chan string { func replayItems[I any](store Store[I], doneCh <-chan struct{}, logger logger, id string) <-chan string {
itemKeyCh := make(chan string) itemKeyCh := make(chan string)
go func() { go func() {
@ -65,7 +66,7 @@ func replayItems[I any](store Store[I], doneCh <-chan struct{}, loggerOnce logge
for { for {
names, err := store.List() names, err := store.List()
if err != nil { if err != nil {
loggerOnce(context.Background(), fmt.Errorf("store.List() failed with: %w", err), id) logger(context.Background(), fmt.Errorf("store.List() failed with: %w", err), id)
} else { } else {
for _, name := range names { for _, name := range names {
select { select {
@ -89,20 +90,20 @@ func replayItems[I any](store Store[I], doneCh <-chan struct{}, loggerOnce logge
} }
// sendItems - Reads items from the store and re-plays. // sendItems - Reads items from the store and re-plays.
func sendItems(target Target, itemKeyCh <-chan string, doneCh <-chan struct{}, loggerOnce logger.LogOnce) { func sendItems(target Target, itemKeyCh <-chan string, doneCh <-chan struct{}, logger logger) {
retryTicker := time.NewTicker(retryInterval) retryTicker := time.NewTicker(retryInterval)
defer retryTicker.Stop() defer retryTicker.Stop()
send := func(itemKey string) bool { send := func(itemKey string) bool {
for { for {
err := target.Send(itemKey) err := target.SendFromStore(itemKey)
if err == nil { if err == nil {
break break
} }
if err != ErrNotConnected && !xnet.IsConnResetErr(err) { if err != ErrNotConnected && !xnet.IsConnResetErr(err) {
loggerOnce(context.Background(), logger(context.Background(),
fmt.Errorf("target.Send() failed with '%w'", err), fmt.Errorf("target.SendFromStore() failed with '%w'", err),
target.Name()) target.Name())
} }
@ -135,11 +136,11 @@ func sendItems(target Target, itemKeyCh <-chan string, doneCh <-chan struct{}, l
} }
// StreamItems reads the keys from the store and replays the corresponding item to the target. // StreamItems reads the keys from the store and replays the corresponding item to the target.
func StreamItems[I any](store Store[I], target Target, doneCh <-chan struct{}, loggerOnce logger.LogOnce) { func StreamItems[I any](store Store[I], target Target, doneCh <-chan struct{}, logger logger) {
go func() { go func() {
// Replays the items from the store. // Replays the items from the store.
itemKeyCh := replayItems(store, doneCh, loggerOnce, target.Name()) itemKeyCh := replayItems(store, doneCh, logger, target.Name())
// Send items from the store. // Send items from the store.
sendItems(target, itemKeyCh, doneCh, loggerOnce) sendItems(target, itemKeyCh, doneCh, logger)
}() }()
} }