Files
akvorado/inlet/kafka/root.go
Vincent Bernat c6a9319b57 common/schema: turns into a component
This is a first step to make it accept configuration. Most of the
changes are quite trivial, but I also ran into some difficulties with
query columns and filters. They need the schema for parsing, but parsing
happens before dependencies are instantiated (and even if it was not the
case, parsing is stateless). Therefore, I have added a `Validate()`
method that must be called after instantiation. Various bits `panic()`
if not validated to ensure we catch all cases.

The alternative to make the component manages a global state would have
been simpler but it would break once we add the ability to add or
disable columns.
2023-01-18 12:22:10 +01:00

142 lines
4.0 KiB
Go

// SPDX-FileCopyrightText: 2022 Free Mobile
// SPDX-License-Identifier: AGPL-3.0-only
// Package kafka handles flow exports to Kafka.
package kafka
import (
"encoding/binary"
"fmt"
"math/rand"
"strings"
"time"
"github.com/Shopify/sarama"
"gopkg.in/tomb.v2"
"akvorado/common/daemon"
"akvorado/common/kafka"
"akvorado/common/reporter"
"akvorado/common/schema"
)
// Component represents the Kafka exporter.
type Component struct {
r *reporter.Reporter
d *Dependencies
t tomb.Tomb
config Configuration
kafkaTopic string
kafkaConfig *sarama.Config
kafkaProducer sarama.AsyncProducer
createKafkaProducer func() (sarama.AsyncProducer, error)
metrics metrics
}
// Dependencies define the dependencies of the Kafka exporter.
type Dependencies struct {
Daemon daemon.Component
Schema *schema.Component
}
// New creates a new HTTP component.
func New(reporter *reporter.Reporter, configuration Configuration, dependencies Dependencies) (*Component, error) {
// Build Kafka configuration
kafkaConfig, err := kafka.NewConfig(configuration.Configuration)
if err != nil {
return nil, err
}
kafkaConfig.Metadata.AllowAutoTopicCreation = true
kafkaConfig.Producer.MaxMessageBytes = configuration.MaxMessageBytes
kafkaConfig.Producer.Compression = sarama.CompressionCodec(configuration.CompressionCodec)
kafkaConfig.Producer.Return.Successes = false
kafkaConfig.Producer.Return.Errors = true
kafkaConfig.Producer.Flush.Bytes = configuration.FlushBytes
kafkaConfig.Producer.Flush.Frequency = configuration.FlushInterval
kafkaConfig.Producer.Partitioner = sarama.NewHashPartitioner
kafkaConfig.ChannelBufferSize = configuration.QueueSize / 2
if err := kafkaConfig.Validate(); err != nil {
return nil, fmt.Errorf("cannot validate Kafka configuration: %w", err)
}
c := Component{
r: reporter,
d: &dependencies,
config: configuration,
kafkaConfig: kafkaConfig,
kafkaTopic: fmt.Sprintf("%s-%s", configuration.Topic, dependencies.Schema.ProtobufMessageHash()),
}
c.initMetrics()
c.createKafkaProducer = func() (sarama.AsyncProducer, error) {
return sarama.NewAsyncProducer(c.config.Brokers, c.kafkaConfig)
}
c.d.Daemon.Track(&c.t, "inlet/kafka")
return &c, nil
}
// Start starts the Kafka component.
func (c *Component) Start() error {
c.r.Info().Msg("starting Kafka component")
kafka.GlobalKafkaLogger.Register(c.r)
// Create producer
kafkaProducer, err := c.createKafkaProducer()
if err != nil {
c.r.Err(err).
Str("brokers", strings.Join(c.config.Brokers, ",")).
Msg("unable to create async producer")
return fmt.Errorf("unable to create Kafka async producer: %w", err)
}
c.kafkaProducer = kafkaProducer
// Main loop
c.t.Go(func() error {
defer kafkaProducer.Close()
defer c.kafkaConfig.MetricRegistry.UnregisterAll()
errLogger := c.r.Sample(reporter.BurstSampler(10*time.Second, 3))
for {
select {
case <-c.t.Dying():
c.r.Debug().Msg("stop error logger")
return nil
case msg := <-kafkaProducer.Errors():
if msg != nil {
c.metrics.errors.WithLabelValues(msg.Error()).Inc()
errLogger.Err(msg.Err).
Str("topic", msg.Msg.Topic).
Int64("offset", msg.Msg.Offset).
Int32("partition", msg.Msg.Partition).
Msg("Kafka producer error")
}
}
}
})
return nil
}
// Stop stops the Kafka component
func (c *Component) Stop() error {
defer func() {
kafka.GlobalKafkaLogger.Unregister()
c.r.Info().Msg("Kafka component stopped")
}()
c.r.Info().Msg("stopping Kafka component")
c.t.Kill(nil)
return c.t.Wait()
}
// Send a message to Kafka.
func (c *Component) Send(exporter string, payload []byte) {
c.metrics.bytesSent.WithLabelValues(exporter).Add(float64(len(payload)))
c.metrics.messagesSent.WithLabelValues(exporter).Inc()
key := make([]byte, 4)
binary.BigEndian.PutUint32(key, rand.Uint32())
c.kafkaProducer.Input() <- &sarama.ProducerMessage{
Topic: c.kafkaTopic,
Key: sarama.ByteEncoder(key),
Value: sarama.ByteEncoder(payload),
}
}