mirror of
https://github.com/akvorado/akvorado.git
synced 2025-12-12 06:24:10 +01:00
Using the exporter IP address is not helpful at all as we may not have many of them and it would make big exporters difficult to scale if one thread is not enough to ingest on the ClickHouse side. Fix #75
137 lines
3.9 KiB
Go
137 lines
3.9 KiB
Go
// SPDX-FileCopyrightText: 2022 Free Mobile
|
|
// SPDX-License-Identifier: AGPL-3.0-only
|
|
|
|
// Package kafka handles flow exports to Kafka.
|
|
package kafka
|
|
|
|
import (
|
|
"encoding/binary"
|
|
"fmt"
|
|
"math/rand"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/Shopify/sarama"
|
|
"gopkg.in/tomb.v2"
|
|
|
|
"akvorado/common/daemon"
|
|
"akvorado/common/kafka"
|
|
"akvorado/common/reporter"
|
|
"akvorado/inlet/flow"
|
|
)
|
|
|
|
// Component represents the Kafka exporter.
|
|
type Component struct {
|
|
r *reporter.Reporter
|
|
d *Dependencies
|
|
t tomb.Tomb
|
|
config Configuration
|
|
|
|
kafkaTopic string
|
|
kafkaConfig *sarama.Config
|
|
kafkaProducer sarama.AsyncProducer
|
|
createKafkaProducer func() (sarama.AsyncProducer, error)
|
|
metrics metrics
|
|
}
|
|
|
|
// Dependencies define the dependencies of the Kafka exporter.
|
|
type Dependencies struct {
|
|
Daemon daemon.Component
|
|
}
|
|
|
|
// New creates a new HTTP component.
|
|
func New(reporter *reporter.Reporter, configuration Configuration, dependencies Dependencies) (*Component, error) {
|
|
// Build Kafka configuration
|
|
kafkaConfig := sarama.NewConfig()
|
|
kafkaConfig.Version = sarama.KafkaVersion(configuration.Version)
|
|
kafkaConfig.Metadata.AllowAutoTopicCreation = true
|
|
kafkaConfig.Producer.MaxMessageBytes = configuration.MaxMessageBytes
|
|
kafkaConfig.Producer.Compression = sarama.CompressionCodec(configuration.CompressionCodec)
|
|
kafkaConfig.Producer.Return.Successes = false
|
|
kafkaConfig.Producer.Return.Errors = true
|
|
kafkaConfig.Producer.Flush.Bytes = configuration.FlushBytes
|
|
kafkaConfig.Producer.Flush.Frequency = configuration.FlushInterval
|
|
kafkaConfig.Producer.Partitioner = sarama.NewHashPartitioner
|
|
kafkaConfig.ChannelBufferSize = configuration.QueueSize / 2
|
|
if err := kafkaConfig.Validate(); err != nil {
|
|
return nil, fmt.Errorf("cannot validate Kafka configuration: %w", err)
|
|
}
|
|
|
|
c := Component{
|
|
r: reporter,
|
|
d: &dependencies,
|
|
config: configuration,
|
|
|
|
kafkaConfig: kafkaConfig,
|
|
kafkaTopic: fmt.Sprintf("%s-v%d", configuration.Topic, flow.CurrentSchemaVersion),
|
|
}
|
|
c.initMetrics()
|
|
c.createKafkaProducer = func() (sarama.AsyncProducer, error) {
|
|
return sarama.NewAsyncProducer(c.config.Brokers, c.kafkaConfig)
|
|
}
|
|
c.d.Daemon.Track(&c.t, "inlet/kafka")
|
|
return &c, nil
|
|
}
|
|
|
|
// Start starts the Kafka component.
|
|
func (c *Component) Start() error {
|
|
c.r.Info().Msg("starting Kafka component")
|
|
kafka.GlobalKafkaLogger.Register(c.r)
|
|
|
|
// Create producer
|
|
kafkaProducer, err := c.createKafkaProducer()
|
|
if err != nil {
|
|
c.r.Err(err).
|
|
Str("brokers", strings.Join(c.config.Brokers, ",")).
|
|
Msg("unable to create async producer")
|
|
return fmt.Errorf("unable to create Kafka async producer: %w", err)
|
|
}
|
|
c.kafkaProducer = kafkaProducer
|
|
|
|
// Main loop
|
|
c.t.Go(func() error {
|
|
defer kafkaProducer.Close()
|
|
defer c.kafkaConfig.MetricRegistry.UnregisterAll()
|
|
errLogger := c.r.Sample(reporter.BurstSampler(10*time.Second, 3))
|
|
for {
|
|
select {
|
|
case <-c.t.Dying():
|
|
c.r.Debug().Msg("stop error logger")
|
|
return nil
|
|
case msg := <-kafkaProducer.Errors():
|
|
c.metrics.errors.WithLabelValues(msg.Error()).Inc()
|
|
errLogger.Err(msg.Err).
|
|
Str("topic", msg.Msg.Topic).
|
|
Int64("offset", msg.Msg.Offset).
|
|
Int32("partition", msg.Msg.Partition).
|
|
Msg("Kafka producer error")
|
|
}
|
|
}
|
|
})
|
|
return nil
|
|
}
|
|
|
|
// Stop stops the Kafka component
|
|
func (c *Component) Stop() error {
|
|
defer func() {
|
|
kafka.GlobalKafkaLogger.Unregister()
|
|
c.r.Info().Msg("Kafka component stopped")
|
|
}()
|
|
c.r.Info().Msg("stopping Kafka component")
|
|
c.t.Kill(nil)
|
|
return c.t.Wait()
|
|
}
|
|
|
|
// Send a message to Kafka.
|
|
func (c *Component) Send(exporter string, payload []byte) {
|
|
c.metrics.bytesSent.WithLabelValues(exporter).Add(float64(len(payload)))
|
|
c.metrics.messagesSent.WithLabelValues(exporter).Inc()
|
|
key := make([]byte, 4)
|
|
binary.BigEndian.PutUint32(key, rand.Uint32())
|
|
c.kafkaProducer.Input() <- &sarama.ProducerMessage{
|
|
Topic: c.kafkaTopic,
|
|
Key: sarama.ByteEncoder(key),
|
|
Value: sarama.ByteEncoder(payload),
|
|
}
|
|
}
|